D. Looney, N. Rehman, D. Mandic, Tomasz M. Rutkowski, A. Heidenreich, Dagmar Beyer
{"title":"为智能环境调节多模态信息","authors":"D. Looney, N. Rehman, D. Mandic, Tomasz M. Rutkowski, A. Heidenreich, Dagmar Beyer","doi":"10.1109/ICDSC.2009.5289373","DOIUrl":null,"url":null,"abstract":"This study aims at providing signal processing solutions for the conditioning of multimodal information in audio-aided smart camera environments. A novel approach is introduced for processing audio and video within a unified ‘data fusion via fission’ framework. This is achieved using empirical mode decomposition (EMD), a fully data-driven algorithm which facilitates analysis at multiple time-frequency scales. Its adaptive nature makes it suitable for processing real-world data and allows, for example, signal conditioning (denoising, illumination invariant video) and robust feature extraction. Furthermore, complex extension of the EMD algorithm are used to quantify shared dynamics between the conditioned modalities facilitating multimodal fusion. The proposed collaborative approach is used to model human-human interaction.","PeriodicalId":324810,"journal":{"name":"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2009-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Conditioning multimodal information for smart environments\",\"authors\":\"D. Looney, N. Rehman, D. Mandic, Tomasz M. Rutkowski, A. Heidenreich, Dagmar Beyer\",\"doi\":\"10.1109/ICDSC.2009.5289373\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This study aims at providing signal processing solutions for the conditioning of multimodal information in audio-aided smart camera environments. A novel approach is introduced for processing audio and video within a unified ‘data fusion via fission’ framework. This is achieved using empirical mode decomposition (EMD), a fully data-driven algorithm which facilitates analysis at multiple time-frequency scales. Its adaptive nature makes it suitable for processing real-world data and allows, for example, signal conditioning (denoising, illumination invariant video) and robust feature extraction. Furthermore, complex extension of the EMD algorithm are used to quantify shared dynamics between the conditioned modalities facilitating multimodal fusion. The proposed collaborative approach is used to model human-human interaction.\",\"PeriodicalId\":324810,\"journal\":{\"name\":\"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)\",\"volume\":\"16 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2009-10-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICDSC.2009.5289373\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDSC.2009.5289373","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Conditioning multimodal information for smart environments
This study aims at providing signal processing solutions for the conditioning of multimodal information in audio-aided smart camera environments. A novel approach is introduced for processing audio and video within a unified ‘data fusion via fission’ framework. This is achieved using empirical mode decomposition (EMD), a fully data-driven algorithm which facilitates analysis at multiple time-frequency scales. Its adaptive nature makes it suitable for processing real-world data and allows, for example, signal conditioning (denoising, illumination invariant video) and robust feature extraction. Furthermore, complex extension of the EMD algorithm are used to quantify shared dynamics between the conditioned modalities facilitating multimodal fusion. The proposed collaborative approach is used to model human-human interaction.