Deep learning algorithms as tools for automated image classification have recently experienced rapid growth in imaging-dependent medical specialties, including ophthalmology. However, only a few algorithms tailored to specific health conditions have been able to achieve regulatory approval for autonomous diagnosis. There is now an international effort to establish optimized thresholds for algorithm performance benchmarking in a rapidly evolving artificial intelligence field. This review examines the largest deep learning studies in glaucoma, with special focus on identifying recurrent challenges and limitations within these studies which preclude widespread clinical deployment. We focus on the 3 most common input modalities when diagnosing glaucoma, namely, fundus photographs, spectral domain optical coherence tomography scans, and standard automated perimetry data. We then analyze 3 major challenges present in all studies: defining the algorithm output of glaucoma, determining reliable ground truth datasets, and compiling representative training datasets.
CITATION STYLE
Lee, E. B., Wang, S. Y., & Chang, R. T. (2021, May 1). Interpreting Deep Learning Studies in Glaucoma: Unresolved Challenges. Asia-Pacific Journal of Ophthalmology. Lippincott Williams and Wilkins. https://doi.org/10.1097/APO.0000000000000395
Mendeley helps you to discover research relevant for your work.