A motivational-based learning model for mobile robots

IF 4.6 Q2 MATERIALS SCIENCE, BIOMATERIALS
Letícia Berto , Paula Costa , Alexandre Simões , Ricardo Gudwin , Esther Colombini
{"title":"A motivational-based learning model for mobile robots","authors":"Letícia Berto ,&nbsp;Paula Costa ,&nbsp;Alexandre Simões ,&nbsp;Ricardo Gudwin ,&nbsp;Esther Colombini","doi":"10.1016/j.cogsys.2024.101278","DOIUrl":null,"url":null,"abstract":"<div><p>Humans have needs motivating their behavior according to intensity and context. However, we also create preferences associated with each action’s perceived pleasure, which is susceptible to changes over time. This makes decision-making more complex, requiring learning to balance <em>needs</em> and preferences according to the context. To understand how this process works and enable the development of robots with a motivational-based learning model, we computationally model a motivation theory proposed by Hull. In this model, the agent (an abstraction of a mobile robot) is motivated to keep itself in a state of homeostasis. We introduced hedonic dimensions to explore the impact of preferences on decision-making and employed reinforcement learning to train our motivated-based agents. In our experiments, we deploy three agents with distinct energy decay rates, simulating different metabolic rates, within two diverse environments. We investigate the influence of these conditions on their strategies, movement patterns, and overall behavior. The findings reveal that agents excel at learning more effective strategies when the environment allows for choices that align with their metabolic requirements. Furthermore, we observe that incorporating pleasure as a component of the motivational mechanism affects behavior learning, particularly for agents with regular metabolisms depending on the environment. Our study also unveils that, when confronted with survival challenges, agents prioritize immediate <em>needs</em> over pleasure and equilibrium. These insights shed light on how robotic agents can adapt and make informed decisions in demanding scenarios, demonstrating the intricate interplay between motivation, pleasure, and environmental context in autonomous systems.</p></div>","PeriodicalId":2,"journal":{"name":"ACS Applied Bio Materials","volume":null,"pages":null},"PeriodicalIF":4.6000,"publicationDate":"2024-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACS Applied Bio Materials","FirstCategoryId":"102","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S138904172400072X","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MATERIALS SCIENCE, BIOMATERIALS","Score":null,"Total":0}
引用次数: 0

Abstract

Humans have needs motivating their behavior according to intensity and context. However, we also create preferences associated with each action’s perceived pleasure, which is susceptible to changes over time. This makes decision-making more complex, requiring learning to balance needs and preferences according to the context. To understand how this process works and enable the development of robots with a motivational-based learning model, we computationally model a motivation theory proposed by Hull. In this model, the agent (an abstraction of a mobile robot) is motivated to keep itself in a state of homeostasis. We introduced hedonic dimensions to explore the impact of preferences on decision-making and employed reinforcement learning to train our motivated-based agents. In our experiments, we deploy three agents with distinct energy decay rates, simulating different metabolic rates, within two diverse environments. We investigate the influence of these conditions on their strategies, movement patterns, and overall behavior. The findings reveal that agents excel at learning more effective strategies when the environment allows for choices that align with their metabolic requirements. Furthermore, we observe that incorporating pleasure as a component of the motivational mechanism affects behavior learning, particularly for agents with regular metabolisms depending on the environment. Our study also unveils that, when confronted with survival challenges, agents prioritize immediate needs over pleasure and equilibrium. These insights shed light on how robotic agents can adapt and make informed decisions in demanding scenarios, demonstrating the intricate interplay between motivation, pleasure, and environmental context in autonomous systems.

基于动机的移动机器人学习模型
人类会根据不同的强度和情境产生不同的需求,从而激发他们的行为。然而,我们也会产生与每个行为的感知快感相关的偏好,而这种快感很容易随着时间的推移而发生变化。这就使得决策变得更加复杂,需要学会根据情境平衡需求和偏好。为了了解这一过程是如何进行的,并开发出具有基于动机的学习模型的机器人,我们对赫尔提出的动机理论进行了计算建模。在该模型中,代理(移动机器人的抽象)的动机是使自身保持平衡状态。我们引入了享乐维度来探索偏好对决策的影响,并采用强化学习来训练基于动机的代理。在实验中,我们在两种不同的环境中部署了三个具有不同能量衰减率的代理,模拟不同的新陈代谢率。我们研究了这些条件对它们的策略、运动模式和整体行为的影响。研究结果表明,当环境允许代理人选择符合其新陈代谢需求的策略时,代理人就能学习到更有效的策略。此外,我们还观察到,将快乐作为动机机制的一个组成部分会影响行为学习,尤其是对于新陈代谢随环境而有规律的特工来说。我们的研究还发现,在面临生存挑战时,机器人会优先考虑眼前的需要,而不是快乐和平衡。这些见解揭示了机器人如何在苛刻的环境中适应并做出明智的决定,展示了自主系统中动机、快乐和环境背景之间错综复杂的相互作用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
ACS Applied Bio Materials
ACS Applied Bio Materials Chemistry-Chemistry (all)
CiteScore
9.40
自引率
2.10%
发文量
464
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信