Martin Rothbucher, Tim Habigt, Julian Habigt, Thomas Riedmaier, K. Diepold
{"title":"测量HRTF个性化的人体测量数据","authors":"Martin Rothbucher, Tim Habigt, Julian Habigt, Thomas Riedmaier, K. Diepold","doi":"10.1109/SITIS.2010.27","DOIUrl":null,"url":null,"abstract":"Nowadays, multimodal human-like sensing, e.g. vision, hap tics and audition seeks to improve interaction between an operator (human) and a teleoperator (robot) in human centered robotic systems. Head Related Transfer Function (HRTF) based sound rendering techniques, which seek to create a realistic virtual auditory space for listeners, have become a prominent concept in human robot interaction. Applications that demand high quality 3D sound synthesis are usually based on measured HRTFs of listeners. Recently, researchers propose to construct a set of personalized HRTFs using multiple linear regression models between anthropometric data and measured HRTFs, which implies the existence of a training HRTF dataset together with the corresponding anthropometric data. This paper focuses on the measurement of Head-Related transfer Functions (HRTFs) and the corresponding anthropometric data of a listener. Several state-of-the-art techniques of measuring the HRTFs are described. For measuring the anthropometric data, we develop a low budget approach, which enables us to measure the anthropometry of a person within short time at a high accuracy, whereas the hardware costs for the scanning system are significantly reduced.","PeriodicalId":128396,"journal":{"name":"2010 Sixth International Conference on Signal-Image Technology and Internet Based Systems","volume":"92 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"Measuring Anthropometric Data for HRTF Personalization\",\"authors\":\"Martin Rothbucher, Tim Habigt, Julian Habigt, Thomas Riedmaier, K. Diepold\",\"doi\":\"10.1109/SITIS.2010.27\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Nowadays, multimodal human-like sensing, e.g. vision, hap tics and audition seeks to improve interaction between an operator (human) and a teleoperator (robot) in human centered robotic systems. Head Related Transfer Function (HRTF) based sound rendering techniques, which seek to create a realistic virtual auditory space for listeners, have become a prominent concept in human robot interaction. Applications that demand high quality 3D sound synthesis are usually based on measured HRTFs of listeners. Recently, researchers propose to construct a set of personalized HRTFs using multiple linear regression models between anthropometric data and measured HRTFs, which implies the existence of a training HRTF dataset together with the corresponding anthropometric data. This paper focuses on the measurement of Head-Related transfer Functions (HRTFs) and the corresponding anthropometric data of a listener. Several state-of-the-art techniques of measuring the HRTFs are described. For measuring the anthropometric data, we develop a low budget approach, which enables us to measure the anthropometry of a person within short time at a high accuracy, whereas the hardware costs for the scanning system are significantly reduced.\",\"PeriodicalId\":128396,\"journal\":{\"name\":\"2010 Sixth International Conference on Signal-Image Technology and Internet Based Systems\",\"volume\":\"92 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2010-12-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2010 Sixth International Conference on Signal-Image Technology and Internet Based Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SITIS.2010.27\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 Sixth International Conference on Signal-Image Technology and Internet Based Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SITIS.2010.27","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Measuring Anthropometric Data for HRTF Personalization
Nowadays, multimodal human-like sensing, e.g. vision, hap tics and audition seeks to improve interaction between an operator (human) and a teleoperator (robot) in human centered robotic systems. Head Related Transfer Function (HRTF) based sound rendering techniques, which seek to create a realistic virtual auditory space for listeners, have become a prominent concept in human robot interaction. Applications that demand high quality 3D sound synthesis are usually based on measured HRTFs of listeners. Recently, researchers propose to construct a set of personalized HRTFs using multiple linear regression models between anthropometric data and measured HRTFs, which implies the existence of a training HRTF dataset together with the corresponding anthropometric data. This paper focuses on the measurement of Head-Related transfer Functions (HRTFs) and the corresponding anthropometric data of a listener. Several state-of-the-art techniques of measuring the HRTFs are described. For measuring the anthropometric data, we develop a low budget approach, which enables us to measure the anthropometry of a person within short time at a high accuracy, whereas the hardware costs for the scanning system are significantly reduced.