Anas Krichel, Nikolay Malkin, Salem Lahlou, Yoshua Bengio
{"title":"On Generalization for Generative Flow Networks","authors":"Anas Krichel, Nikolay Malkin, Salem Lahlou, Yoshua Bengio","doi":"arxiv-2407.03105","DOIUrl":null,"url":null,"abstract":"Generative Flow Networks (GFlowNets) have emerged as an innovative learning\nparadigm designed to address the challenge of sampling from an unnormalized\nprobability distribution, called the reward function. This framework learns a\npolicy on a constructed graph, which enables sampling from an approximation of\nthe target probability distribution through successive steps of sampling from\nthe learned policy. To achieve this, GFlowNets can be trained with various\nobjectives, each of which can lead to the model s ultimate goal. The\naspirational strength of GFlowNets lies in their potential to discern intricate\npatterns within the reward function and their capacity to generalize\neffectively to novel, unseen parts of the reward function. This paper attempts\nto formalize generalization in the context of GFlowNets, to link generalization\nwith stability, and also to design experiments that assess the capacity of\nthese models to uncover unseen parts of the reward function. The experiments\nwill focus on length generalization meaning generalization to states that can\nbe constructed only by longer trajectories than those seen in training.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":"6 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Machine Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.03105","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Generative Flow Networks (GFlowNets) have emerged as an innovative learning
paradigm designed to address the challenge of sampling from an unnormalized
probability distribution, called the reward function. This framework learns a
policy on a constructed graph, which enables sampling from an approximation of
the target probability distribution through successive steps of sampling from
the learned policy. To achieve this, GFlowNets can be trained with various
objectives, each of which can lead to the model s ultimate goal. The
aspirational strength of GFlowNets lies in their potential to discern intricate
patterns within the reward function and their capacity to generalize
effectively to novel, unseen parts of the reward function. This paper attempts
to formalize generalization in the context of GFlowNets, to link generalization
with stability, and also to design experiments that assess the capacity of
these models to uncover unseen parts of the reward function. The experiments
will focus on length generalization meaning generalization to states that can
be constructed only by longer trajectories than those seen in training.