{"title":"基于传感器融合的自编码器特征提取用于三维目标检测","authors":"Junmin Lee, Wonjun Hwang","doi":"10.1049/ell2.70295","DOIUrl":null,"url":null,"abstract":"<p>Knowledge distillation is a widely adopted model compression method aimed at narrowing the performance gap between a high-capacity teacher network and a lightweight student network. However, in the context of sensor fusion-based 3D object detection, existing distillation methods predominantly emphasize accuracy enhancement through the introduction of multiple loss functions, which often leads to overly complex training procedures. To address this limitation, we propose a sensor fusion-based feature distillation framework tailored for camera and radar modalities. Our proposed method utilizes an autoencoder to facilitate efficient knowledge transfer from the teacher to the student model. Additionally, we introduce image-context and radar-context knowledge distillation strategies to capture and transfer modality-specific features effectively. We demonstrate the effectiveness of the proposed method on the nuScenes dataset using a ResNet-based architecture.</p>","PeriodicalId":11556,"journal":{"name":"Electronics Letters","volume":"61 1","pages":""},"PeriodicalIF":0.7000,"publicationDate":"2025-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/ell2.70295","citationCount":"0","resultStr":"{\"title\":\"Sensor Fusion-Based Autoencoder Feature Distillation for 3D Object Detection\",\"authors\":\"Junmin Lee, Wonjun Hwang\",\"doi\":\"10.1049/ell2.70295\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Knowledge distillation is a widely adopted model compression method aimed at narrowing the performance gap between a high-capacity teacher network and a lightweight student network. However, in the context of sensor fusion-based 3D object detection, existing distillation methods predominantly emphasize accuracy enhancement through the introduction of multiple loss functions, which often leads to overly complex training procedures. To address this limitation, we propose a sensor fusion-based feature distillation framework tailored for camera and radar modalities. Our proposed method utilizes an autoencoder to facilitate efficient knowledge transfer from the teacher to the student model. Additionally, we introduce image-context and radar-context knowledge distillation strategies to capture and transfer modality-specific features effectively. We demonstrate the effectiveness of the proposed method on the nuScenes dataset using a ResNet-based architecture.</p>\",\"PeriodicalId\":11556,\"journal\":{\"name\":\"Electronics Letters\",\"volume\":\"61 1\",\"pages\":\"\"},\"PeriodicalIF\":0.7000,\"publicationDate\":\"2025-05-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1049/ell2.70295\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Electronics Letters\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1049/ell2.70295\",\"RegionNum\":4,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Electronics Letters","FirstCategoryId":"5","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1049/ell2.70295","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Sensor Fusion-Based Autoencoder Feature Distillation for 3D Object Detection
Knowledge distillation is a widely adopted model compression method aimed at narrowing the performance gap between a high-capacity teacher network and a lightweight student network. However, in the context of sensor fusion-based 3D object detection, existing distillation methods predominantly emphasize accuracy enhancement through the introduction of multiple loss functions, which often leads to overly complex training procedures. To address this limitation, we propose a sensor fusion-based feature distillation framework tailored for camera and radar modalities. Our proposed method utilizes an autoencoder to facilitate efficient knowledge transfer from the teacher to the student model. Additionally, we introduce image-context and radar-context knowledge distillation strategies to capture and transfer modality-specific features effectively. We demonstrate the effectiveness of the proposed method on the nuScenes dataset using a ResNet-based architecture.
期刊介绍:
Electronics Letters is an internationally renowned peer-reviewed rapid-communication journal that publishes short original research papers every two weeks. Its broad and interdisciplinary scope covers the latest developments in all electronic engineering related fields including communication, biomedical, optical and device technologies. Electronics Letters also provides further insight into some of the latest developments through special features and interviews.
Scope
As a journal at the forefront of its field, Electronics Letters publishes papers covering all themes of electronic and electrical engineering. The major themes of the journal are listed below.
Antennas and Propagation
Biomedical and Bioinspired Technologies, Signal Processing and Applications
Control Engineering
Electromagnetism: Theory, Materials and Devices
Electronic Circuits and Systems
Image, Video and Vision Processing and Applications
Information, Computing and Communications
Instrumentation and Measurement
Microwave Technology
Optical Communications
Photonics and Opto-Electronics
Power Electronics, Energy and Sustainability
Radar, Sonar and Navigation
Semiconductor Technology
Signal Processing
MIMO