{"title":"SwiftDepth++:用于精确深度估计的高效轻量级模型","authors":"Y. Dayoub, I. Makarov","doi":"10.1134/S1064562424602038","DOIUrl":null,"url":null,"abstract":"<p>Depth estimation is a crucial task across various domains, but the high cost of collecting labeled depth data has led to growing interest in self-supervised monocular depth estimation methods. In this paper, we introduce SwiftDepth++, a lightweight depth estimation model that delivers competitive results while maintaining a low computational budget. The core innovation of SwiftDepth++ lies in its novel depth decoder, which enhances efficiency by rapidly compressing features while preserving essential information. Additionally, we incorporate a teacher-student knowledge distillation framework that guides the student model in refining its predictions. We evaluate SwiftDepth++ on the KITTI and NYU datasets, where it achieves an absolute relative error (Abs_rel) of 10.2% on the KITTI dataset and 22% on the NYU dataset without fine-tuning, all with approximately 6 million parameters. These results demonstrate that SwiftDepth++ not only meets the demands of modern depth estimation tasks but also significantly reduces computational complexity, making it a practical choice for real-world applications.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S162 - S171"},"PeriodicalIF":0.5000,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602038.pdf","citationCount":"0","resultStr":"{\"title\":\"SwiftDepth++: An Efficient and Lightweight Model for Accurate Depth Estimation\",\"authors\":\"Y. Dayoub, I. Makarov\",\"doi\":\"10.1134/S1064562424602038\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Depth estimation is a crucial task across various domains, but the high cost of collecting labeled depth data has led to growing interest in self-supervised monocular depth estimation methods. In this paper, we introduce SwiftDepth++, a lightweight depth estimation model that delivers competitive results while maintaining a low computational budget. The core innovation of SwiftDepth++ lies in its novel depth decoder, which enhances efficiency by rapidly compressing features while preserving essential information. Additionally, we incorporate a teacher-student knowledge distillation framework that guides the student model in refining its predictions. We evaluate SwiftDepth++ on the KITTI and NYU datasets, where it achieves an absolute relative error (Abs_rel) of 10.2% on the KITTI dataset and 22% on the NYU dataset without fine-tuning, all with approximately 6 million parameters. These results demonstrate that SwiftDepth++ not only meets the demands of modern depth estimation tasks but also significantly reduces computational complexity, making it a practical choice for real-world applications.</p>\",\"PeriodicalId\":531,\"journal\":{\"name\":\"Doklady Mathematics\",\"volume\":\"110 1 supplement\",\"pages\":\"S162 - S171\"},\"PeriodicalIF\":0.5000,\"publicationDate\":\"2025-03-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://link.springer.com/content/pdf/10.1134/S1064562424602038.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Doklady Mathematics\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://link.springer.com/article/10.1134/S1064562424602038\",\"RegionNum\":4,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"MATHEMATICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Doklady Mathematics","FirstCategoryId":"100","ListUrlMain":"https://link.springer.com/article/10.1134/S1064562424602038","RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"MATHEMATICS","Score":null,"Total":0}
SwiftDepth++: An Efficient and Lightweight Model for Accurate Depth Estimation
Depth estimation is a crucial task across various domains, but the high cost of collecting labeled depth data has led to growing interest in self-supervised monocular depth estimation methods. In this paper, we introduce SwiftDepth++, a lightweight depth estimation model that delivers competitive results while maintaining a low computational budget. The core innovation of SwiftDepth++ lies in its novel depth decoder, which enhances efficiency by rapidly compressing features while preserving essential information. Additionally, we incorporate a teacher-student knowledge distillation framework that guides the student model in refining its predictions. We evaluate SwiftDepth++ on the KITTI and NYU datasets, where it achieves an absolute relative error (Abs_rel) of 10.2% on the KITTI dataset and 22% on the NYU dataset without fine-tuning, all with approximately 6 million parameters. These results demonstrate that SwiftDepth++ not only meets the demands of modern depth estimation tasks but also significantly reduces computational complexity, making it a practical choice for real-world applications.
期刊介绍:
Doklady Mathematics is a journal of the Presidium of the Russian Academy of Sciences. It contains English translations of papers published in Doklady Akademii Nauk (Proceedings of the Russian Academy of Sciences), which was founded in 1933 and is published 36 times a year. Doklady Mathematics includes the materials from the following areas: mathematics, mathematical physics, computer science, control theory, and computers. It publishes brief scientific reports on previously unpublished significant new research in mathematics and its applications. The main contributors to the journal are Members of the RAS, Corresponding Members of the RAS, and scientists from the former Soviet Union and other foreign countries. Among the contributors are the outstanding Russian mathematicians.