{"title":"EventAug:基于事件学习的多方面时空数据增强方法","authors":"Yukun Tian, Hao Chen, Yongjian Deng, Feihong Shen, Kepan Liu, Wei You, Ziyang Zhang","doi":"arxiv-2409.11813","DOIUrl":null,"url":null,"abstract":"The event camera has demonstrated significant success across a wide range of\nareas due to its low time latency and high dynamic range. However, the\ncommunity faces challenges such as data deficiency and limited diversity, often\nresulting in over-fitting and inadequate feature learning. Notably, the\nexploration of data augmentation techniques in the event community remains\nscarce. This work aims to address this gap by introducing a systematic\naugmentation scheme named EventAug to enrich spatial-temporal diversity. In\nparticular, we first propose Multi-scale Temporal Integration (MSTI) to\ndiversify the motion speed of objects, then introduce Spatial-salient Event\nMask (SSEM) and Temporal-salient Event Mask (TSEM) to enrich object variants.\nOur EventAug can facilitate models learning with richer motion patterns, object\nvariants and local spatio-temporal relations, thus improving model robustness\nto varied moving speeds, occlusions, and action disruptions. Experiment results\nshow that our augmentation method consistently yields significant improvements\nacross different tasks and backbones (e.g., a 4.87% accuracy gain on DVS128\nGesture). Our code will be publicly available for this community.","PeriodicalId":501130,"journal":{"name":"arXiv - CS - Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"EventAug: Multifaceted Spatio-Temporal Data Augmentation Methods for Event-based Learning\",\"authors\":\"Yukun Tian, Hao Chen, Yongjian Deng, Feihong Shen, Kepan Liu, Wei You, Ziyang Zhang\",\"doi\":\"arxiv-2409.11813\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The event camera has demonstrated significant success across a wide range of\\nareas due to its low time latency and high dynamic range. However, the\\ncommunity faces challenges such as data deficiency and limited diversity, often\\nresulting in over-fitting and inadequate feature learning. Notably, the\\nexploration of data augmentation techniques in the event community remains\\nscarce. This work aims to address this gap by introducing a systematic\\naugmentation scheme named EventAug to enrich spatial-temporal diversity. In\\nparticular, we first propose Multi-scale Temporal Integration (MSTI) to\\ndiversify the motion speed of objects, then introduce Spatial-salient Event\\nMask (SSEM) and Temporal-salient Event Mask (TSEM) to enrich object variants.\\nOur EventAug can facilitate models learning with richer motion patterns, object\\nvariants and local spatio-temporal relations, thus improving model robustness\\nto varied moving speeds, occlusions, and action disruptions. Experiment results\\nshow that our augmentation method consistently yields significant improvements\\nacross different tasks and backbones (e.g., a 4.87% accuracy gain on DVS128\\nGesture). Our code will be publicly available for this community.\",\"PeriodicalId\":501130,\"journal\":{\"name\":\"arXiv - CS - Computer Vision and Pattern Recognition\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Computer Vision and Pattern Recognition\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.11813\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computer Vision and Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11813","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
EventAug: Multifaceted Spatio-Temporal Data Augmentation Methods for Event-based Learning
The event camera has demonstrated significant success across a wide range of
areas due to its low time latency and high dynamic range. However, the
community faces challenges such as data deficiency and limited diversity, often
resulting in over-fitting and inadequate feature learning. Notably, the
exploration of data augmentation techniques in the event community remains
scarce. This work aims to address this gap by introducing a systematic
augmentation scheme named EventAug to enrich spatial-temporal diversity. In
particular, we first propose Multi-scale Temporal Integration (MSTI) to
diversify the motion speed of objects, then introduce Spatial-salient Event
Mask (SSEM) and Temporal-salient Event Mask (TSEM) to enrich object variants.
Our EventAug can facilitate models learning with richer motion patterns, object
variants and local spatio-temporal relations, thus improving model robustness
to varied moving speeds, occlusions, and action disruptions. Experiment results
show that our augmentation method consistently yields significant improvements
across different tasks and backbones (e.g., a 4.87% accuracy gain on DVS128
Gesture). Our code will be publicly available for this community.