{"title":"Sample-Efficient Reinforcement Learning From Human Feedback via Information-Directed Sampling","authors":"Han Qi;Haochen Yang;Qiaosheng Zhang;Zhuoran Yang","doi":"10.1109/TIT.2025.3598296","DOIUrl":null,"url":null,"abstract":"We study the problem of reinforcement learning from human feedback (RLHF), a critical problem in training large language models, from a theoretical perspective. Our main contribution is the design of novel sample-efficient RLHF algorithms based on information-directed sampling (IDS), an online decision-making principle inspired by information theory. Our algorithms maximize the sum of the value function and a mutual information term that encourages exploration of the unknown environment (which quantifies the information gained about the environment through observed human feedback data). To tackle the challenge of large state spaces and improve sample efficiency, we construct a simplified <italic>surrogate environment</i> and introduce a novel distance measure (named the <inline-formula> <tex-math>$\\ell _{g}$ </tex-math></inline-formula><italic>-distance</i>), enabling our IDS-based algorithm to achieve a Bayesian regret upper bound of order <inline-formula> <tex-math>$O(H^{3/2}\\sqrt {\\log (K(\\epsilon)) T})$ </tex-math></inline-formula>, where <italic>H</i> is the episode length, <italic>T</i> is the number of episode and <inline-formula> <tex-math>$K(\\epsilon)$ </tex-math></inline-formula> is related to the covering number of the environment. Specializing to the tabular settings, this regret bound is of order <inline-formula> <tex-math>$\\tilde {O}(H^{2}\\sqrt {SAT})$ </tex-math></inline-formula>, where <italic>S</i> and <italic>A</i> are the numbers of states and actions. Finally, we propose an Approximate-IDS algorithm that is computationally more efficient while maintaining nearly the same sample efficiency. The design principle of this approximate algorithm is not only effective in RLHF settings but also applicable to the standard RL framework. Moreover, our work showcases the value of information theory in reinforcement learning and in the training of large language models.","PeriodicalId":13494,"journal":{"name":"IEEE Transactions on Information Theory","volume":"71 10","pages":"7942-7958"},"PeriodicalIF":2.9000,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Information Theory","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/11123904/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
We study the problem of reinforcement learning from human feedback (RLHF), a critical problem in training large language models, from a theoretical perspective. Our main contribution is the design of novel sample-efficient RLHF algorithms based on information-directed sampling (IDS), an online decision-making principle inspired by information theory. Our algorithms maximize the sum of the value function and a mutual information term that encourages exploration of the unknown environment (which quantifies the information gained about the environment through observed human feedback data). To tackle the challenge of large state spaces and improve sample efficiency, we construct a simplified surrogate environment and introduce a novel distance measure (named the $\ell _{g}$ -distance), enabling our IDS-based algorithm to achieve a Bayesian regret upper bound of order $O(H^{3/2}\sqrt {\log (K(\epsilon)) T})$ , where H is the episode length, T is the number of episode and $K(\epsilon)$ is related to the covering number of the environment. Specializing to the tabular settings, this regret bound is of order $\tilde {O}(H^{2}\sqrt {SAT})$ , where S and A are the numbers of states and actions. Finally, we propose an Approximate-IDS algorithm that is computationally more efficient while maintaining nearly the same sample efficiency. The design principle of this approximate algorithm is not only effective in RLHF settings but also applicable to the standard RL framework. Moreover, our work showcases the value of information theory in reinforcement learning and in the training of large language models.
期刊介绍:
The IEEE Transactions on Information Theory is a journal that publishes theoretical and experimental papers concerned with the transmission, processing, and utilization of information. The boundaries of acceptable subject matter are intentionally not sharply delimited. Rather, it is hoped that as the focus of research activity changes, a flexible policy will permit this Transactions to follow suit. Current appropriate topics are best reflected by recent Tables of Contents; they are summarized in the titles of editorial areas that appear on the inside front cover.