Ibrahim Musaddequr Rahman, Stanford White, Katelyn Crockett, Yu Gu, D. A. Dutra, Guilherme A. S. Pereira
{"title":"Imitating Swarm Behaviors by Learning Agent-Level Controllers","authors":"Ibrahim Musaddequr Rahman, Stanford White, Katelyn Crockett, Yu Gu, D. A. Dutra, Guilherme A. S. Pereira","doi":"10.23919/ACC55779.2023.10156561","DOIUrl":null,"url":null,"abstract":"A main challenge in swarm robotics is the unknown mapping between simple agent-level behavior rules and emergent global behaviors. Currently, there is no known swarm control algorithm that maps global behaviors to local control policies. This paper proposes a novel method to circumvent this problem by learning the agent-level controllers of an observed swarm to imitate its emergent behavior. Agent-level controllers are treated as a set of policies that are combined to dictate the agent’s change in velocity. The trajectory data of known swarms is used with linear regression and nonlinear optimization methods to learn the relative weight of each policy. To show our approach’s ability for imitating swarm behavior, we apply this methodology to both simulated and physical swarms (i.e., a school of fish) exhibiting a multitude of distinct emergent behaviors. We found that our pipeline was effective at imitating the simulated behaviors using both accurate and inaccurate assumptions, being able to closely identify not only the policy gains, but also the agent’s radius of communication and their maximum velocity constraint.","PeriodicalId":397401,"journal":{"name":"2023 American Control Conference (ACC)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 American Control Conference (ACC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.23919/ACC55779.2023.10156561","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
A main challenge in swarm robotics is the unknown mapping between simple agent-level behavior rules and emergent global behaviors. Currently, there is no known swarm control algorithm that maps global behaviors to local control policies. This paper proposes a novel method to circumvent this problem by learning the agent-level controllers of an observed swarm to imitate its emergent behavior. Agent-level controllers are treated as a set of policies that are combined to dictate the agent’s change in velocity. The trajectory data of known swarms is used with linear regression and nonlinear optimization methods to learn the relative weight of each policy. To show our approach’s ability for imitating swarm behavior, we apply this methodology to both simulated and physical swarms (i.e., a school of fish) exhibiting a multitude of distinct emergent behaviors. We found that our pipeline was effective at imitating the simulated behaviors using both accurate and inaccurate assumptions, being able to closely identify not only the policy gains, but also the agent’s radius of communication and their maximum velocity constraint.