{"title":"装配场景下机器人自然语言指令中注视与手势检测的集成","authors":"Jianwei Zhang, T. Baier, M. Hueser","doi":"10.1109/ROMAN.2002.1045629","DOIUrl":null,"url":null,"abstract":"We present the development of and experiments with a robot system showing multimodal interaction capabilities. We focus on the understanding of human instructions in natural language by integrating gaze and pointing hand gestures. A typical application of such a system is interactive assembly. A human communicator sharing a view of the assembly scenario with the robot instructs the latter by speaking to it in the same way that he would communicate with a child. His instructions can be under-specified, incomplete and/or context-dependent. By detecting gaze and gesture information of the instructor, which are described in two sections of the paper, the ambiguities in the interactions can be resolved in a natural manner. Finally, we outline a list of future topics for extending our research.","PeriodicalId":222409,"journal":{"name":"Proceedings. 11th IEEE International Workshop on Robot and Human Interactive Communication","volume":"40 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2002-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Integration of gaze and gesture detection in nature language instructing of robot in an assembly scenario\",\"authors\":\"Jianwei Zhang, T. Baier, M. Hueser\",\"doi\":\"10.1109/ROMAN.2002.1045629\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We present the development of and experiments with a robot system showing multimodal interaction capabilities. We focus on the understanding of human instructions in natural language by integrating gaze and pointing hand gestures. A typical application of such a system is interactive assembly. A human communicator sharing a view of the assembly scenario with the robot instructs the latter by speaking to it in the same way that he would communicate with a child. His instructions can be under-specified, incomplete and/or context-dependent. By detecting gaze and gesture information of the instructor, which are described in two sections of the paper, the ambiguities in the interactions can be resolved in a natural manner. Finally, we outline a list of future topics for extending our research.\",\"PeriodicalId\":222409,\"journal\":{\"name\":\"Proceedings. 11th IEEE International Workshop on Robot and Human Interactive Communication\",\"volume\":\"40 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2002-12-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings. 11th IEEE International Workshop on Robot and Human Interactive Communication\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ROMAN.2002.1045629\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings. 11th IEEE International Workshop on Robot and Human Interactive Communication","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ROMAN.2002.1045629","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Integration of gaze and gesture detection in nature language instructing of robot in an assembly scenario
We present the development of and experiments with a robot system showing multimodal interaction capabilities. We focus on the understanding of human instructions in natural language by integrating gaze and pointing hand gestures. A typical application of such a system is interactive assembly. A human communicator sharing a view of the assembly scenario with the robot instructs the latter by speaking to it in the same way that he would communicate with a child. His instructions can be under-specified, incomplete and/or context-dependent. By detecting gaze and gesture information of the instructor, which are described in two sections of the paper, the ambiguities in the interactions can be resolved in a natural manner. Finally, we outline a list of future topics for extending our research.