Sensing Super-position: Human Sensing Beyond the Visual Spectrum

D. Maluf, P. Tran
{"title":"Sensing Super-position: Human Sensing Beyond the Visual Spectrum","authors":"D. Maluf, P. Tran","doi":"10.1109/IRI.2006.252481","DOIUrl":null,"url":null,"abstract":"The coming decade of fast, cheap and miniaturized electronics and sensory devices opens new pathways for the development of sophisticated equipment to overcome limitations of the human senses. This paper addresses the technical feasibility of augmenting human vision through sensing super-position by mixing natural human sensing. The current implementation of the device translates visual and other passive or active sensory instruments into sounds, which become relevant when the visual resolution is insufficient for very difficult and particular sensing tasks. A successful sensing super-position meets many human and pilot vehicle system requirements. The system can be further developed into cheap, portable, and low power taking into account the limited capabilities of the human user as well as the typical characteristics of his dynamic environment. The system operates in real time, giving the desired information for the particular augmented sensing tasks. The sensing super-position device increases the image resolution perception and is obtained via an auditory representation as well as the visual representation. Auditory mapping is performed to distribute an image in time. The three-dimensional spatial brightness and multi-spectral maps of a sensed image are processed using real-time image processing techniques (e.g. histogram normalization) and transformed into a two-dimensional map of an audio signal as a function of frequency and time. This paper details the approach of developing sensing super-position systems as a way to augment the human vision system by exploiting the capabilities of the human hearing system as an additional neural input. The human hearing system is capable of learning to process and interpret extremely complicated and rapidly changing auditory patterns. The known capabilities of the human hearing system to learn and understand complicated auditory patterns provided the basic motivation for developing an image-to-sound mapping system. The human brain is superior to most existing computer systems in rapidly extracting relevant information from blurred, noisy, and redundant images. From a theoretical viewpoint, this means that the available bandwidth is not exploited in an optimal way. While image-processing techniques can manipulate, condense and focus the information (e.g., Fourier transforms), keeping the mapping as direct and simple as possible might also reduce the risk of accidentally filtering out important clues. After all, especially a perfect non-redundant sound representation is prone to loss of relevant information in the non-perfect human hearing system. Also, a complicated non-redundant image-to-sound mapping may well be far more difficult to learn and comprehend than a straightforward mapping, while the mapping system would increase in complexity and cost. This work demonstrates some basic information processing for optimal information capture for head-mounted systems","PeriodicalId":402255,"journal":{"name":"2006 IEEE International Conference on Information Reuse & Integration","volume":"100 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2006-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2006 IEEE International Conference on Information Reuse & Integration","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IRI.2006.252481","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

The coming decade of fast, cheap and miniaturized electronics and sensory devices opens new pathways for the development of sophisticated equipment to overcome limitations of the human senses. This paper addresses the technical feasibility of augmenting human vision through sensing super-position by mixing natural human sensing. The current implementation of the device translates visual and other passive or active sensory instruments into sounds, which become relevant when the visual resolution is insufficient for very difficult and particular sensing tasks. A successful sensing super-position meets many human and pilot vehicle system requirements. The system can be further developed into cheap, portable, and low power taking into account the limited capabilities of the human user as well as the typical characteristics of his dynamic environment. The system operates in real time, giving the desired information for the particular augmented sensing tasks. The sensing super-position device increases the image resolution perception and is obtained via an auditory representation as well as the visual representation. Auditory mapping is performed to distribute an image in time. The three-dimensional spatial brightness and multi-spectral maps of a sensed image are processed using real-time image processing techniques (e.g. histogram normalization) and transformed into a two-dimensional map of an audio signal as a function of frequency and time. This paper details the approach of developing sensing super-position systems as a way to augment the human vision system by exploiting the capabilities of the human hearing system as an additional neural input. The human hearing system is capable of learning to process and interpret extremely complicated and rapidly changing auditory patterns. The known capabilities of the human hearing system to learn and understand complicated auditory patterns provided the basic motivation for developing an image-to-sound mapping system. The human brain is superior to most existing computer systems in rapidly extracting relevant information from blurred, noisy, and redundant images. From a theoretical viewpoint, this means that the available bandwidth is not exploited in an optimal way. While image-processing techniques can manipulate, condense and focus the information (e.g., Fourier transforms), keeping the mapping as direct and simple as possible might also reduce the risk of accidentally filtering out important clues. After all, especially a perfect non-redundant sound representation is prone to loss of relevant information in the non-perfect human hearing system. Also, a complicated non-redundant image-to-sound mapping may well be far more difficult to learn and comprehend than a straightforward mapping, while the mapping system would increase in complexity and cost. This work demonstrates some basic information processing for optimal information capture for head-mounted systems
感知叠加:超越视觉光谱的人类感知
在未来的十年里,快速、廉价和小型化的电子和传感设备为开发复杂的设备开辟了新的途径,以克服人类感官的局限性。本文探讨了混合人类自然感知,通过感知叠加增强人类视觉的技术可行性。该设备目前的实现将视觉和其他被动或主动感知仪器转换为声音,当视觉分辨率不足以完成非常困难和特殊的感知任务时,声音就变得相关了。一个成功的传感叠加可以满足许多人类和驾驶车辆系统的要求。考虑到人类用户有限的能力以及其动态环境的典型特征,该系统可以进一步发展为廉价、便携和低功耗。该系统实时运行,为特定的增强传感任务提供所需的信息。感测叠加装置增加图像分辨率感知,并通过听觉表示和视觉表示获得。执行听觉映射以及时分配图像。利用实时图像处理技术(如直方图归一化)处理感测图像的三维空间亮度和多光谱图,并将其转换为音频信号的二维图,作为频率和时间的函数。本文详细介绍了开发传感叠加系统的方法,通过利用人类听觉系统作为额外神经输入的能力来增强人类视觉系统。人类的听觉系统能够学习处理和解释极其复杂和快速变化的听觉模式。人类听觉系统学习和理解复杂听觉模式的已知能力为开发图像到声音映射系统提供了基本动机。人脑在从模糊、嘈杂和冗余的图像中快速提取相关信息方面优于大多数现有的计算机系统。从理论上讲,这意味着可用带宽没有以最优的方式被利用。虽然图像处理技术可以操纵、浓缩和聚焦信息(例如,傅里叶变换),但保持映射尽可能直接和简单也可能降低意外过滤掉重要线索的风险。毕竟,在不完美的人类听觉系统中,尤其是一个完美的、无冗余的声音表示,容易丢失相关信息。此外,复杂的非冗余图像到声音映射可能比直接映射更难学习和理解,而映射系统的复杂性和成本也会增加。这项工作演示了一些基本的信息处理,以实现头戴式系统的最佳信息捕获
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信