Franziska Jurosch, Janik Zeller, Lars Wagner, Ege Özsoy, Alissa Jell, Sven Kolb, Dirk Wilhelm
{"title":"基于视频的多目标多摄像机术后相位识别。","authors":"Franziska Jurosch, Janik Zeller, Lars Wagner, Ege Özsoy, Alissa Jell, Sven Kolb, Dirk Wilhelm","doi":"10.1007/s11548-025-03344-x","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>Deep learning methods are commonly used to generate context understanding to support surgeons and medical professionals. By expanding the current focus beyond the operating room (OR) to postoperative workflows, new forms of assistance are possible. In this article, we propose a novel multi-target multi-camera tracking (MTMCT) architecture for postoperative phase recognition, location tracking, and automatic timestamp generation.</p><p><strong>Methods: </strong>Three RGB cameras were used to create a multi-camera data set containing 19 reenacted postoperative patient flows. Patients and beds were annotated and used to train the custom MTMCT architecture. It includes bed and patient tracking for each camera and a postoperative patient state module to provide the postoperative phase, current location of the patient, and automatically generated timestamps.</p><p><strong>Results: </strong>The architecture demonstrates robust performance for single- and multi-patient scenarios by embedding medical domain-specific knowledge. In multi-patient scenarios, the state machine representing the postoperative phases has a traversal accuracy of <math><mrow><mn>84.9</mn> <mo>±</mo> <mn>6.0</mn> <mo>%</mo></mrow> </math> , <math><mrow><mn>91.4</mn> <mo>±</mo> <mn>1.5</mn> <mo>%</mo></mrow> </math> of timestamps are generated correctly, and the patient tracking IDF1 reaches <math><mrow><mn>92.0</mn> <mo>±</mo> <mn>3.6</mn> <mo>%</mo></mrow> </math> . Comparative experiments show the effectiveness of using AFLink for matching partial trajectories in postoperative settings.</p><p><strong>Conclusion: </strong>As our approach shows promising results, it lays the foundation for real-time surgeon support, enhancing clinical documentation and ultimately improving patient care.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3000,"publicationDate":"2025-04-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Video-based multi-target multi-camera tracking for postoperative phase recognition.\",\"authors\":\"Franziska Jurosch, Janik Zeller, Lars Wagner, Ege Özsoy, Alissa Jell, Sven Kolb, Dirk Wilhelm\",\"doi\":\"10.1007/s11548-025-03344-x\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Purpose: </strong>Deep learning methods are commonly used to generate context understanding to support surgeons and medical professionals. By expanding the current focus beyond the operating room (OR) to postoperative workflows, new forms of assistance are possible. In this article, we propose a novel multi-target multi-camera tracking (MTMCT) architecture for postoperative phase recognition, location tracking, and automatic timestamp generation.</p><p><strong>Methods: </strong>Three RGB cameras were used to create a multi-camera data set containing 19 reenacted postoperative patient flows. Patients and beds were annotated and used to train the custom MTMCT architecture. It includes bed and patient tracking for each camera and a postoperative patient state module to provide the postoperative phase, current location of the patient, and automatically generated timestamps.</p><p><strong>Results: </strong>The architecture demonstrates robust performance for single- and multi-patient scenarios by embedding medical domain-specific knowledge. In multi-patient scenarios, the state machine representing the postoperative phases has a traversal accuracy of <math><mrow><mn>84.9</mn> <mo>±</mo> <mn>6.0</mn> <mo>%</mo></mrow> </math> , <math><mrow><mn>91.4</mn> <mo>±</mo> <mn>1.5</mn> <mo>%</mo></mrow> </math> of timestamps are generated correctly, and the patient tracking IDF1 reaches <math><mrow><mn>92.0</mn> <mo>±</mo> <mn>3.6</mn> <mo>%</mo></mrow> </math> . Comparative experiments show the effectiveness of using AFLink for matching partial trajectories in postoperative settings.</p><p><strong>Conclusion: </strong>As our approach shows promising results, it lays the foundation for real-time surgeon support, enhancing clinical documentation and ultimately improving patient care.</p>\",\"PeriodicalId\":51251,\"journal\":{\"name\":\"International Journal of Computer Assisted Radiology and Surgery\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":2.3000,\"publicationDate\":\"2025-04-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Computer Assisted Radiology and Surgery\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://doi.org/10.1007/s11548-025-03344-x\",\"RegionNum\":3,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"ENGINEERING, BIOMEDICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Computer Assisted Radiology and Surgery","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1007/s11548-025-03344-x","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
Video-based multi-target multi-camera tracking for postoperative phase recognition.
Purpose: Deep learning methods are commonly used to generate context understanding to support surgeons and medical professionals. By expanding the current focus beyond the operating room (OR) to postoperative workflows, new forms of assistance are possible. In this article, we propose a novel multi-target multi-camera tracking (MTMCT) architecture for postoperative phase recognition, location tracking, and automatic timestamp generation.
Methods: Three RGB cameras were used to create a multi-camera data set containing 19 reenacted postoperative patient flows. Patients and beds were annotated and used to train the custom MTMCT architecture. It includes bed and patient tracking for each camera and a postoperative patient state module to provide the postoperative phase, current location of the patient, and automatically generated timestamps.
Results: The architecture demonstrates robust performance for single- and multi-patient scenarios by embedding medical domain-specific knowledge. In multi-patient scenarios, the state machine representing the postoperative phases has a traversal accuracy of , of timestamps are generated correctly, and the patient tracking IDF1 reaches . Comparative experiments show the effectiveness of using AFLink for matching partial trajectories in postoperative settings.
Conclusion: As our approach shows promising results, it lays the foundation for real-time surgeon support, enhancing clinical documentation and ultimately improving patient care.
期刊介绍:
The International Journal for Computer Assisted Radiology and Surgery (IJCARS) is a peer-reviewed journal that provides a platform for closing the gap between medical and technical disciplines, and encourages interdisciplinary research and development activities in an international environment.