David Molina Concha, Kyeonghyeon Park, Hyun-Rok Lee, Taesik Lee, Chi-Guhn Lee
{"title":"Algorithmic Contract Design with Reinforcement Learning Agents","authors":"David Molina Concha, Kyeonghyeon Park, Hyun-Rok Lee, Taesik Lee, Chi-Guhn Lee","doi":"arxiv-2408.09686","DOIUrl":null,"url":null,"abstract":"We introduce a novel problem setting for algorithmic contract design, named\nthe principal-MARL contract design problem. This setting extends traditional\ncontract design to account for dynamic and stochastic environments using Markov\nGames and Multi-Agent Reinforcement Learning. To tackle this problem, we\npropose a Multi-Objective Bayesian Optimization (MOBO) framework named\nConstrained Pareto Maximum Entropy Search (cPMES). Our approach integrates MOBO\nand MARL to explore the highly constrained contract design space, identifying\npromising incentive and recruitment decisions. cPMES transforms the\nprincipal-MARL contract design problem into an unconstrained multi-objective\nproblem, leveraging the probability of feasibility as part of the objectives\nand ensuring promising designs predicted on the feasibility border are included\nin the Pareto front. By focusing the entropy prediction on designs within the\nPareto set, cPMES mitigates the risk of the search strategy being overwhelmed\nby entropy from constraints. We demonstrate the effectiveness of cPMES through\nextensive benchmark studies in synthetic and simulated environments, showing\nits ability to find feasible contract designs that maximize the principal's\nobjectives. Additionally, we provide theoretical support with a sub-linear\nregret bound concerning the number of iterations.","PeriodicalId":501315,"journal":{"name":"arXiv - CS - Multiagent Systems","volume":"63-65 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Multiagent Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.09686","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
We introduce a novel problem setting for algorithmic contract design, named
the principal-MARL contract design problem. This setting extends traditional
contract design to account for dynamic and stochastic environments using Markov
Games and Multi-Agent Reinforcement Learning. To tackle this problem, we
propose a Multi-Objective Bayesian Optimization (MOBO) framework named
Constrained Pareto Maximum Entropy Search (cPMES). Our approach integrates MOBO
and MARL to explore the highly constrained contract design space, identifying
promising incentive and recruitment decisions. cPMES transforms the
principal-MARL contract design problem into an unconstrained multi-objective
problem, leveraging the probability of feasibility as part of the objectives
and ensuring promising designs predicted on the feasibility border are included
in the Pareto front. By focusing the entropy prediction on designs within the
Pareto set, cPMES mitigates the risk of the search strategy being overwhelmed
by entropy from constraints. We demonstrate the effectiveness of cPMES through
extensive benchmark studies in synthetic and simulated environments, showing
its ability to find feasible contract designs that maximize the principal's
objectives. Additionally, we provide theoretical support with a sub-linear
regret bound concerning the number of iterations.