A deep learning framework for real-time 3D model registration in robot-assisted laparoscopic surgery

30Citations
Citations of this article
48Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Introduction: The current study presents a deep learning framework to determine, in real-time, position and rotation of a target organ from an endoscopic video. These inferred data are used to overlay the 3D model of patient's organ over its real counterpart. The resulting augmented video flow is streamed back to the surgeon as a support during laparoscopic robot-assisted procedures. Methods: This framework exploits semantic segmentation and, thereafter, two techniques, based on Convolutional Neural Networks and motion analysis, were used to infer the rotation. Results: The segmentation shows optimal accuracies, with a mean IoU score greater than 80% in all tests. Different performance levels are obtained for rotation, depending on the surgical procedure. Discussion: Even if the presented methodology has various degrees of precision depending on the testing scenario, this work sets the first step for the adoption of deep learning and augmented reality to generalise the automatic registration process.

Author supplied keywords

Cite

CITATION STYLE

APA

Padovan, E., Marullo, G., Tanzi, L., Piazzolla, P., Moos, S., Porpiglia, F., & Vezzetti, E. (2022). A deep learning framework for real-time 3D model registration in robot-assisted laparoscopic surgery. International Journal of Medical Robotics and Computer Assisted Surgery, 18(3). https://doi.org/10.1002/rcs.2387

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free