Purpose: Segmentation of surgical instruments in endoscopic video streams is essential for automated surgical scene understanding and process modeling. However, relying on fully supervised deep learning for this task is challenging because manual annotation occupies valuable time of the clinical experts. Methods: We introduce a teacher–student learning approach that learns jointly from annotated simulation data and unlabeled real data to tackle the challenges in simulation-to-real unsupervised domain adaptation for endoscopic image segmentation. Results: Empirical results on three datasets highlight the effectiveness of the proposed framework over current approaches for the endoscopic instrument segmentation task. Additionally, we provide analysis of major factors affecting the performance on all datasets to highlight the strengths and failure modes of our approach. Conclusions: We show that our proposed approach can successfully exploit the unlabeled real endoscopic video frames and improve generalization performance over pure simulation-based training and the previous state-of-the-art. This takes us one step closer to effective segmentation of surgical instrument in the annotation scarce setting.
CITATION STYLE
Sahu, M., Mukhopadhyay, A., & Zachow, S. (2021). Simulation-to-real domain adaptation with teacher–student learning for endoscopic instrument segmentation. International Journal of Computer Assisted Radiology and Surgery, 16(5), 849–859. https://doi.org/10.1007/s11548-021-02383-4
Mendeley helps you to discover research relevant for your work.