{"title":"金融时间序列无模型控制的课程学习与模仿学习","authors":"Woosung Koh, Insu Choi, Yuntae Jang, Gimin Kang, Woo Chang Kim","doi":"arxiv-2311.13326","DOIUrl":null,"url":null,"abstract":"Curriculum learning and imitation learning have been leveraged extensively in\nthe robotics domain. However, minimal research has been done on leveraging\nthese ideas on control tasks over highly stochastic time-series data. Here, we\ntheoretically and empirically explore these approaches in a representative\ncontrol task over complex time-series data. We implement the fundamental ideas\nof curriculum learning via data augmentation, while imitation learning is\nimplemented via policy distillation from an oracle. Our findings reveal that\ncurriculum learning should be considered a novel direction in improving\ncontrol-task performance over complex time-series. Our ample random-seed\nout-sample empirics and ablation studies are highly encouraging for curriculum\nlearning for time-series control. These findings are especially encouraging as\nwe tune all overlapping hyperparameters on the baseline -- giving an advantage\nto the baseline. On the other hand, we find that imitation learning should be\nused with caution.","PeriodicalId":501045,"journal":{"name":"arXiv - QuantFin - Portfolio Management","volume":"7 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2023-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Curriculum Learning and Imitation Learning for Model-free Control on Financial Time-series\",\"authors\":\"Woosung Koh, Insu Choi, Yuntae Jang, Gimin Kang, Woo Chang Kim\",\"doi\":\"arxiv-2311.13326\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Curriculum learning and imitation learning have been leveraged extensively in\\nthe robotics domain. However, minimal research has been done on leveraging\\nthese ideas on control tasks over highly stochastic time-series data. Here, we\\ntheoretically and empirically explore these approaches in a representative\\ncontrol task over complex time-series data. We implement the fundamental ideas\\nof curriculum learning via data augmentation, while imitation learning is\\nimplemented via policy distillation from an oracle. Our findings reveal that\\ncurriculum learning should be considered a novel direction in improving\\ncontrol-task performance over complex time-series. Our ample random-seed\\nout-sample empirics and ablation studies are highly encouraging for curriculum\\nlearning for time-series control. These findings are especially encouraging as\\nwe tune all overlapping hyperparameters on the baseline -- giving an advantage\\nto the baseline. On the other hand, we find that imitation learning should be\\nused with caution.\",\"PeriodicalId\":501045,\"journal\":{\"name\":\"arXiv - QuantFin - Portfolio Management\",\"volume\":\"7 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-11-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - QuantFin - Portfolio Management\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2311.13326\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuantFin - Portfolio Management","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2311.13326","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Curriculum Learning and Imitation Learning for Model-free Control on Financial Time-series
Curriculum learning and imitation learning have been leveraged extensively in
the robotics domain. However, minimal research has been done on leveraging
these ideas on control tasks over highly stochastic time-series data. Here, we
theoretically and empirically explore these approaches in a representative
control task over complex time-series data. We implement the fundamental ideas
of curriculum learning via data augmentation, while imitation learning is
implemented via policy distillation from an oracle. Our findings reveal that
curriculum learning should be considered a novel direction in improving
control-task performance over complex time-series. Our ample random-seed
out-sample empirics and ablation studies are highly encouraging for curriculum
learning for time-series control. These findings are especially encouraging as
we tune all overlapping hyperparameters on the baseline -- giving an advantage
to the baseline. On the other hand, we find that imitation learning should be
used with caution.