Siyuan Li;Yafei Chen;Hui Niu;Jiahao Zheng;Zhouchi Lin;Jian Li;Jian Guo;Zhen Wang
{"title":"迈向自动做市:一种带有预测表示学习的模仿强化学习方法","authors":"Siyuan Li;Yafei Chen;Hui Niu;Jiahao Zheng;Zhouchi Lin;Jian Li;Jian Guo;Zhen Wang","doi":"10.1109/TETCI.2024.3451476","DOIUrl":null,"url":null,"abstract":"Market making (MM) is a crucial trading problem, where a market maker stands ready to buy and sell the asset at a publicly quoted price to provide market liquidity continuously. The primary challenges in market making include position risk, liquidity risk, and adverse selection. Emerging research works investigate applying reinforcement learning (RL) techniques to derive automatic MM strategies. However, existing methods mainly focus on addressing inventory risk using only single-level quotes, which restricts the trading flexibility. In this paper, we shed light on the optimization of market makers' returns under a smaller risk while ensuring market liquidity and depth. This paper proposes a novel RL-based market-making strategy <italic>Predictive and Imitative Market Making Agent</i> (PIMMA). First, to ensure adequate liquidity, we design an action space to enable stably allocating orders of multi-level volumes and prices. Beyond that, we apply queue position information from these multi-price levels to encode them in the state representations. Second, aiming at alleviating adverse selection, we draw auxiliary signals into state representation and design a representation learning network structure to catch implicit information from the price-volume fluctuations. Finally, we develop a novel reward function to earn a fortune while avoiding holding a large inventory. With a provided expert demonstration, our method augments the RL objective with imitation learning and learns an effective MM policy. Experiments are conducted to evaluate the proposed method based on realistic historical data, and the results demonstrate PIMMA outperforms RL-based strategy in the perspectives of earning decent revenue and information by adopting the multi-risk aversion strategy.","PeriodicalId":13135,"journal":{"name":"IEEE Transactions on Emerging Topics in Computational Intelligence","volume":"9 3","pages":"2427-2439"},"PeriodicalIF":5.3000,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Toward Automatic Market Making: An Imitative Reinforcement Learning Approach With Predictive Representation Learning\",\"authors\":\"Siyuan Li;Yafei Chen;Hui Niu;Jiahao Zheng;Zhouchi Lin;Jian Li;Jian Guo;Zhen Wang\",\"doi\":\"10.1109/TETCI.2024.3451476\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Market making (MM) is a crucial trading problem, where a market maker stands ready to buy and sell the asset at a publicly quoted price to provide market liquidity continuously. The primary challenges in market making include position risk, liquidity risk, and adverse selection. Emerging research works investigate applying reinforcement learning (RL) techniques to derive automatic MM strategies. However, existing methods mainly focus on addressing inventory risk using only single-level quotes, which restricts the trading flexibility. In this paper, we shed light on the optimization of market makers' returns under a smaller risk while ensuring market liquidity and depth. This paper proposes a novel RL-based market-making strategy <italic>Predictive and Imitative Market Making Agent</i> (PIMMA). First, to ensure adequate liquidity, we design an action space to enable stably allocating orders of multi-level volumes and prices. Beyond that, we apply queue position information from these multi-price levels to encode them in the state representations. Second, aiming at alleviating adverse selection, we draw auxiliary signals into state representation and design a representation learning network structure to catch implicit information from the price-volume fluctuations. Finally, we develop a novel reward function to earn a fortune while avoiding holding a large inventory. With a provided expert demonstration, our method augments the RL objective with imitation learning and learns an effective MM policy. Experiments are conducted to evaluate the proposed method based on realistic historical data, and the results demonstrate PIMMA outperforms RL-based strategy in the perspectives of earning decent revenue and information by adopting the multi-risk aversion strategy.\",\"PeriodicalId\":13135,\"journal\":{\"name\":\"IEEE Transactions on Emerging Topics in Computational Intelligence\",\"volume\":\"9 3\",\"pages\":\"2427-2439\"},\"PeriodicalIF\":5.3000,\"publicationDate\":\"2024-09-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Emerging Topics in Computational Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10688395/\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Emerging Topics in Computational Intelligence","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10688395/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Toward Automatic Market Making: An Imitative Reinforcement Learning Approach With Predictive Representation Learning
Market making (MM) is a crucial trading problem, where a market maker stands ready to buy and sell the asset at a publicly quoted price to provide market liquidity continuously. The primary challenges in market making include position risk, liquidity risk, and adverse selection. Emerging research works investigate applying reinforcement learning (RL) techniques to derive automatic MM strategies. However, existing methods mainly focus on addressing inventory risk using only single-level quotes, which restricts the trading flexibility. In this paper, we shed light on the optimization of market makers' returns under a smaller risk while ensuring market liquidity and depth. This paper proposes a novel RL-based market-making strategy Predictive and Imitative Market Making Agent (PIMMA). First, to ensure adequate liquidity, we design an action space to enable stably allocating orders of multi-level volumes and prices. Beyond that, we apply queue position information from these multi-price levels to encode them in the state representations. Second, aiming at alleviating adverse selection, we draw auxiliary signals into state representation and design a representation learning network structure to catch implicit information from the price-volume fluctuations. Finally, we develop a novel reward function to earn a fortune while avoiding holding a large inventory. With a provided expert demonstration, our method augments the RL objective with imitation learning and learns an effective MM policy. Experiments are conducted to evaluate the proposed method based on realistic historical data, and the results demonstrate PIMMA outperforms RL-based strategy in the perspectives of earning decent revenue and information by adopting the multi-risk aversion strategy.
期刊介绍:
The IEEE Transactions on Emerging Topics in Computational Intelligence (TETCI) publishes original articles on emerging aspects of computational intelligence, including theory, applications, and surveys.
TETCI is an electronics only publication. TETCI publishes six issues per year.
Authors are encouraged to submit manuscripts in any emerging topic in computational intelligence, especially nature-inspired computing topics not covered by other IEEE Computational Intelligence Society journals. A few such illustrative examples are glial cell networks, computational neuroscience, Brain Computer Interface, ambient intelligence, non-fuzzy computing with words, artificial life, cultural learning, artificial endocrine networks, social reasoning, artificial hormone networks, computational intelligence for the IoT and Smart-X technologies.