Graphics Processor Unit (GPU) accelerated shallow transparent layer detection in Optical Coherence Tomographic (OCT) images for real-time corneal surgical guidance

2Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.
Get full text

Abstract

An image analysis algorithm is described that utilizes a Graphics Processor Unit (GPU) to detect in real-time the most shallow subsurface tissue layer present in an OCT image obtained by a prototype SDOCT corneal imaging system. The system has a scanning depth range of 6mm and can acquire 15 volumes per second at the cost of lower resolution and signal-to-noise ratio (SNR) than diagnostic OCT scanners. To the best of our knowledge, we are the first to experiment with non-median percentile filtering for simultaneous noise reduction and feature enhancement in OCT images, and we believe we are the first to implement any form of non-median percentile filtering on a GPU. The algorithm was applied to five different test images. On an average, it took ~0.5 milliseconds to preprocess an image with a 20th-percentile filter, and ~1.7 milliseconds for our second-stage algorithm to detect the faintly imaged transparent surface.

Cite

CITATION STYLE

APA

Mathai, T. S., Galeotti, J., Horvath, S., & Stetten, G. (2014). Graphics Processor Unit (GPU) accelerated shallow transparent layer detection in Optical Coherence Tomographic (OCT) images for real-time corneal surgical guidance. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 8678. https://doi.org/10.1007/978-3-319-10437-9_1

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free