Surgical phase recognition using movement data from video imagery and location sensor data

6Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Abstract The automatic recognition of surgical phases has strong potential to help medical staff understand individual and group patterns, optimize work flows, and identify potential work flow risks that lead to adverse medical events in an operating room. In this chapter, we investigate the performance of context recognition on the movement of operating room staff throughout their work environment, which was measured by imaging and tracking. We employed an optical flow algorithm and trajectory clustering techniques to extract movement characteristics of surgical staff from video imagery and time-stamped location data collected by an ultrasonic location aware system, respectively. Then we applied a Support Vector Machine to time-stamped location data, optical flow estimates, trajectory clusters, and combinations of these three data to examine the intraoperative context recognition rate. Our results show that the integration of both video imagery and location sensor data improves context awareness of neurosurgical operations.

Cite

CITATION STYLE

APA

Nara, A., Allen, C., & Izumi, K. (2017). Surgical phase recognition using movement data from video imagery and location sensor data. In Advances in Geographic Information Science (pp. 229–237). Springer Heidelberg. https://doi.org/10.1007/978-3-319-22786-3_21

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free