UnTran:使用迁移学习识别未标记数据的不可见活动

Md Abdullah Al Hafiz Khan, Nirmalya Roy
{"title":"UnTran:使用迁移学习识别未标记数据的不可见活动","authors":"Md Abdullah Al Hafiz Khan, Nirmalya Roy","doi":"10.1109/IoTDI.2018.00014","DOIUrl":null,"url":null,"abstract":"The success and impact of activity recognition algorithms largely depends on the availability of the labeled training samples and adaptability of activity recognition models across various domains. In a new environment, the pre-trained activity recognition models face challenges in presence of sensing bias- ness, device heterogeneities, and inherent variabilities in human behaviors and activities. Activity Recognition (AR) system built in one environment does not scale well in another environment, if it has to learn new activities and the annotated activity samples are scarce. Indeed building a new activity recognition model and training the model with large annotated samples often help overcome this challenging problem. However, collecting annotated samples is cost-sensitive and learning activity model at wild is computationally expensive. In this work, we propose an activity recognition framework, UnTran that utilizes source domains' pre-trained autoencoder enabled activity model that transfers two layers of this network to generate a common feature space for both source and target domain activities. We postulate a hybrid AR framework that helps fuse the decisions from a trained model in source domain and two activity models (raw and deep-feature based activity model) in target domain reducing the demand of annotated activity samples to help recognize unseen activities. We evaluated our framework with three real-world data traces consisting of 41 users and 26 activities in total. Our proposed UnTran AR framework achieves ≈ 75% F1 score in recognizing unseen new activities using only 10% labeled activity data in the target domain. UnTran attains ≈ 98% F1 score while recognizing seen activities in presence of only 2-3% of labeled activity samples.","PeriodicalId":149725,"journal":{"name":"2018 IEEE/ACM Third International Conference on Internet-of-Things Design and Implementation (IoTDI)","volume":"51 2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"18","resultStr":"{\"title\":\"UnTran: Recognizing Unseen Activities with Unlabeled Data Using Transfer Learning\",\"authors\":\"Md Abdullah Al Hafiz Khan, Nirmalya Roy\",\"doi\":\"10.1109/IoTDI.2018.00014\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The success and impact of activity recognition algorithms largely depends on the availability of the labeled training samples and adaptability of activity recognition models across various domains. In a new environment, the pre-trained activity recognition models face challenges in presence of sensing bias- ness, device heterogeneities, and inherent variabilities in human behaviors and activities. Activity Recognition (AR) system built in one environment does not scale well in another environment, if it has to learn new activities and the annotated activity samples are scarce. Indeed building a new activity recognition model and training the model with large annotated samples often help overcome this challenging problem. However, collecting annotated samples is cost-sensitive and learning activity model at wild is computationally expensive. In this work, we propose an activity recognition framework, UnTran that utilizes source domains' pre-trained autoencoder enabled activity model that transfers two layers of this network to generate a common feature space for both source and target domain activities. We postulate a hybrid AR framework that helps fuse the decisions from a trained model in source domain and two activity models (raw and deep-feature based activity model) in target domain reducing the demand of annotated activity samples to help recognize unseen activities. We evaluated our framework with three real-world data traces consisting of 41 users and 26 activities in total. Our proposed UnTran AR framework achieves ≈ 75% F1 score in recognizing unseen new activities using only 10% labeled activity data in the target domain. UnTran attains ≈ 98% F1 score while recognizing seen activities in presence of only 2-3% of labeled activity samples.\",\"PeriodicalId\":149725,\"journal\":{\"name\":\"2018 IEEE/ACM Third International Conference on Internet-of-Things Design and Implementation (IoTDI)\",\"volume\":\"51 2 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-04-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"18\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 IEEE/ACM Third International Conference on Internet-of-Things Design and Implementation (IoTDI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IoTDI.2018.00014\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE/ACM Third International Conference on Internet-of-Things Design and Implementation (IoTDI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IoTDI.2018.00014","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 18

摘要

活动识别算法的成功和影响在很大程度上取决于标记训练样本的可用性和活动识别模型在各个领域的适应性。在新的环境下,预训练的活动识别模型面临着感知偏差、设备异质性以及人类行为和活动的内在变异性的挑战。在一种环境中构建的活动识别(AR)系统,如果需要学习新的活动,并且标注的活动样本很少,则无法很好地扩展到另一种环境中。实际上,构建一个新的活动识别模型并使用大量带注释的样本训练模型通常有助于克服这个具有挑战性的问题。然而,收集带注释的样本是成本敏感的,并且在野外学习活动模型的计算成本很高。在这项工作中,我们提出了一个活动识别框架,UnTran,它利用源域的预训练自动编码器支持的活动模型,该模型传输该网络的两层,以生成源域和目标域活动的公共特征空间。我们假设了一个混合AR框架,它有助于融合源域的训练模型和目标域的两个活动模型(原始和基于深度特征的活动模型)的决策,减少了对注释活动样本的需求,以帮助识别未见过的活动。我们用总共41个用户和26个活动组成的三个真实数据跟踪来评估我们的框架。我们提出的UnTran AR框架在使用目标域中仅10%的标记活动数据识别未见的新活动方面达到了≈75%的F1分数。UnTran在仅识别2-3%标记活性样本的情况下,获得了≈98%的F1分数。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
UnTran: Recognizing Unseen Activities with Unlabeled Data Using Transfer Learning
The success and impact of activity recognition algorithms largely depends on the availability of the labeled training samples and adaptability of activity recognition models across various domains. In a new environment, the pre-trained activity recognition models face challenges in presence of sensing bias- ness, device heterogeneities, and inherent variabilities in human behaviors and activities. Activity Recognition (AR) system built in one environment does not scale well in another environment, if it has to learn new activities and the annotated activity samples are scarce. Indeed building a new activity recognition model and training the model with large annotated samples often help overcome this challenging problem. However, collecting annotated samples is cost-sensitive and learning activity model at wild is computationally expensive. In this work, we propose an activity recognition framework, UnTran that utilizes source domains' pre-trained autoencoder enabled activity model that transfers two layers of this network to generate a common feature space for both source and target domain activities. We postulate a hybrid AR framework that helps fuse the decisions from a trained model in source domain and two activity models (raw and deep-feature based activity model) in target domain reducing the demand of annotated activity samples to help recognize unseen activities. We evaluated our framework with three real-world data traces consisting of 41 users and 26 activities in total. Our proposed UnTran AR framework achieves ≈ 75% F1 score in recognizing unseen new activities using only 10% labeled activity data in the target domain. UnTran attains ≈ 98% F1 score while recognizing seen activities in presence of only 2-3% of labeled activity samples.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信