Martin Rothbucher, Tim Habigt, Julian Habigt, Thomas Riedmaier, K. Diepold
{"title":"Measuring Anthropometric Data for HRTF Personalization","authors":"Martin Rothbucher, Tim Habigt, Julian Habigt, Thomas Riedmaier, K. Diepold","doi":"10.1109/SITIS.2010.27","DOIUrl":null,"url":null,"abstract":"Nowadays, multimodal human-like sensing, e.g. vision, hap tics and audition seeks to improve interaction between an operator (human) and a teleoperator (robot) in human centered robotic systems. Head Related Transfer Function (HRTF) based sound rendering techniques, which seek to create a realistic virtual auditory space for listeners, have become a prominent concept in human robot interaction. Applications that demand high quality 3D sound synthesis are usually based on measured HRTFs of listeners. Recently, researchers propose to construct a set of personalized HRTFs using multiple linear regression models between anthropometric data and measured HRTFs, which implies the existence of a training HRTF dataset together with the corresponding anthropometric data. This paper focuses on the measurement of Head-Related transfer Functions (HRTFs) and the corresponding anthropometric data of a listener. Several state-of-the-art techniques of measuring the HRTFs are described. For measuring the anthropometric data, we develop a low budget approach, which enables us to measure the anthropometry of a person within short time at a high accuracy, whereas the hardware costs for the scanning system are significantly reduced.","PeriodicalId":128396,"journal":{"name":"2010 Sixth International Conference on Signal-Image Technology and Internet Based Systems","volume":"92 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 Sixth International Conference on Signal-Image Technology and Internet Based Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SITIS.2010.27","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7
Abstract
Nowadays, multimodal human-like sensing, e.g. vision, hap tics and audition seeks to improve interaction between an operator (human) and a teleoperator (robot) in human centered robotic systems. Head Related Transfer Function (HRTF) based sound rendering techniques, which seek to create a realistic virtual auditory space for listeners, have become a prominent concept in human robot interaction. Applications that demand high quality 3D sound synthesis are usually based on measured HRTFs of listeners. Recently, researchers propose to construct a set of personalized HRTFs using multiple linear regression models between anthropometric data and measured HRTFs, which implies the existence of a training HRTF dataset together with the corresponding anthropometric data. This paper focuses on the measurement of Head-Related transfer Functions (HRTFs) and the corresponding anthropometric data of a listener. Several state-of-the-art techniques of measuring the HRTFs are described. For measuring the anthropometric data, we develop a low budget approach, which enables us to measure the anthropometry of a person within short time at a high accuracy, whereas the hardware costs for the scanning system are significantly reduced.