TU Darmstadt / ULB / TUprints

Simulation-to-real domain adaptation with teacher–student learning for endoscopic instrument segmentation

Sahu, Manish ; Mukhopadhyay, Anirban ; Zachow, Stefan (2024)
Simulation-to-real domain adaptation with teacher–student learning for endoscopic instrument segmentation.
In: International Journal of Computer Assisted Radiology and Surgery, 2021, 16 (5)
doi: 10.26083/tuprints-00023530
Article, Secondary publication, Publisher's Version

[img] Text
s11548-021-02383-4.pdf
Copyright Information: CC BY 4.0 International - Creative Commons, Attribution.

Download (2MB)
Item Type: Article
Type of entry: Secondary publication
Title: Simulation-to-real domain adaptation with teacher–student learning for endoscopic instrument segmentation
Language: English
Date: 24 September 2024
Place of Publication: Darmstadt
Year of primary publication: 2021
Place of primary publication: Berlin ; Heidelberg
Publisher: Springer International Publishing
Journal or Publication Title: International Journal of Computer Assisted Radiology and Surgery
Volume of the journal: 16
Issue Number: 5
DOI: 10.26083/tuprints-00023530
Corresponding Links:
Origin: Secondary publication DeepGreen
Abstract:

Segmentation of surgical instruments in endoscopic video streams is essential for automated surgical scene understanding and process modeling. However, relying on fully supervised deep learning for this task is challenging because manual annotation occupies valuable time of the clinical experts.

We introduce a teacher–student learning approach that learns jointly from annotated simulation data and unlabeled real data to tackle the challenges in simulation-to-real unsupervised domain adaptation for endoscopic image segmentation.

Empirical results on three datasets highlight the effectiveness of the proposed framework over current approaches for the endoscopic instrument segmentation task. Additionally, we provide analysis of major factors affecting the performance on all datasets to highlight the strengths and failure modes of our approach.

We show that our proposed approach can successfully exploit the unlabeled real endoscopic video frames and improve generalization performance over pure simulation-based training and the previous state-of-the-art. This takes us one step closer to effective segmentation of surgical instrument in the annotation scarce setting.

Uncontrolled Keywords: Surgical instrument segmentation, Simulation-based learning, Self-supervision, Consistency learning, Self-ensembling, Unsupervised domain adaptation
Status: Publisher's Version
URN: urn:nbn:de:tuda-tuprints-235309
Classification DDC: 000 Generalities, computers, information > 004 Computer science
Divisions: 20 Department of Computer Science > Interactive Graphics Systems
Date Deposited: 24 Sep 2024 09:11
Last Modified: 24 Sep 2024 09:12
SWORD Depositor: Deep Green
URI: https://tuprints.ulb.tu-darmstadt.de/id/eprint/23530
PPN:
Export:
Actions (login required)
View Item View Item