Extracting decision-making features from the unstructured eye movements of clinicians on glaucoma OCT reports and developing AI models to classify expertise

This study aimed to investigate the eye movement patterns of ophthalmologists with varying expertise levels during the assessment of optical coherence tomography (OCT) reports for glaucoma detection. Objectives included evaluating eye gaze metrics and patterns as a function of ophthalmic education,...

Full description

Bibliographic Details
Published in:Frontiers in Medicine
Main Authors: Michelle Akerman, Sanmati Choudhary, Jeffrey M. Liebmann, George A. Cioffi, Royce W. S. Chen, Kaveri A. Thakoor
Format: Article
Language:English
Published: Frontiers Media S.A. 2023-09-01
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/fmed.2023.1251183/full
_version_ 1850365669726486528
author Michelle Akerman
Sanmati Choudhary
Jeffrey M. Liebmann
George A. Cioffi
Royce W. S. Chen
Kaveri A. Thakoor
Kaveri A. Thakoor
Kaveri A. Thakoor
author_facet Michelle Akerman
Sanmati Choudhary
Jeffrey M. Liebmann
George A. Cioffi
Royce W. S. Chen
Kaveri A. Thakoor
Kaveri A. Thakoor
Kaveri A. Thakoor
author_sort Michelle Akerman
collection DOAJ
container_title Frontiers in Medicine
description This study aimed to investigate the eye movement patterns of ophthalmologists with varying expertise levels during the assessment of optical coherence tomography (OCT) reports for glaucoma detection. Objectives included evaluating eye gaze metrics and patterns as a function of ophthalmic education, deriving novel features from eye-tracking, and developing binary classification models for disease detection and expertise differentiation. Thirteen ophthalmology residents, fellows, and clinicians specializing in glaucoma participated in the study. Junior residents had less than 1 year of experience, while senior residents had 2–3 years of experience. The expert group consisted of fellows and faculty with over 3 to 30+ years of experience. Each participant was presented with a set of 20 Topcon OCT reports (10 healthy and 10 glaucomatous) and was asked to determine the presence or absence of glaucoma and rate their confidence of diagnosis. The eye movements of each participant were recorded as they diagnosed the reports using a Pupil Labs Core eye tracker. Expert ophthalmologists exhibited more refined and focused eye fixations, particularly on specific regions of the OCT reports, such as the retinal nerve fiber layer (RNFL) probability map and circumpapillary RNFL b-scan. The binary classification models developed using the derived features demonstrated high accuracy up to 94.0% in differentiating between expert and novice clinicians. The derived features and trained binary classification models hold promise for improving the accuracy of glaucoma detection and distinguishing between expert and novice ophthalmologists. These findings have implications for enhancing ophthalmic education and for the development of effective diagnostic tools.
format Article
id doaj-art-eecfeb1f58994879b81df7bf0bf80d36
institution Directory of Open Access Journals
issn 2296-858X
language English
publishDate 2023-09-01
publisher Frontiers Media S.A.
record_format Article
spelling doaj-art-eecfeb1f58994879b81df7bf0bf80d362025-08-19T23:03:26ZengFrontiers Media S.A.Frontiers in Medicine2296-858X2023-09-011010.3389/fmed.2023.12511831251183Extracting decision-making features from the unstructured eye movements of clinicians on glaucoma OCT reports and developing AI models to classify expertiseMichelle Akerman0Sanmati Choudhary1Jeffrey M. Liebmann2George A. Cioffi3Royce W. S. Chen4Kaveri A. Thakoor5Kaveri A. Thakoor6Kaveri A. Thakoor7Department of Biomedical Engineering, Columbia University, New York, NY, United StatesDepartment of Computer Science, Columbia University, New York, NY, United StatesEdward S. Harkness Eye Institute, Department of Ophthalmology, Columbia University Irving Medical Center, New York, NY, United StatesEdward S. Harkness Eye Institute, Department of Ophthalmology, Columbia University Irving Medical Center, New York, NY, United StatesEdward S. Harkness Eye Institute, Department of Ophthalmology, Columbia University Irving Medical Center, New York, NY, United StatesDepartment of Biomedical Engineering, Columbia University, New York, NY, United StatesDepartment of Computer Science, Columbia University, New York, NY, United StatesEdward S. Harkness Eye Institute, Department of Ophthalmology, Columbia University Irving Medical Center, New York, NY, United StatesThis study aimed to investigate the eye movement patterns of ophthalmologists with varying expertise levels during the assessment of optical coherence tomography (OCT) reports for glaucoma detection. Objectives included evaluating eye gaze metrics and patterns as a function of ophthalmic education, deriving novel features from eye-tracking, and developing binary classification models for disease detection and expertise differentiation. Thirteen ophthalmology residents, fellows, and clinicians specializing in glaucoma participated in the study. Junior residents had less than 1 year of experience, while senior residents had 2–3 years of experience. The expert group consisted of fellows and faculty with over 3 to 30+ years of experience. Each participant was presented with a set of 20 Topcon OCT reports (10 healthy and 10 glaucomatous) and was asked to determine the presence or absence of glaucoma and rate their confidence of diagnosis. The eye movements of each participant were recorded as they diagnosed the reports using a Pupil Labs Core eye tracker. Expert ophthalmologists exhibited more refined and focused eye fixations, particularly on specific regions of the OCT reports, such as the retinal nerve fiber layer (RNFL) probability map and circumpapillary RNFL b-scan. The binary classification models developed using the derived features demonstrated high accuracy up to 94.0% in differentiating between expert and novice clinicians. The derived features and trained binary classification models hold promise for improving the accuracy of glaucoma detection and distinguishing between expert and novice ophthalmologists. These findings have implications for enhancing ophthalmic education and for the development of effective diagnostic tools.https://www.frontiersin.org/articles/10.3389/fmed.2023.1251183/fulleye-trackingfixationsoptical coherence tomographyglaucomaunsupervised clusteringneural networks
spellingShingle Michelle Akerman
Sanmati Choudhary
Jeffrey M. Liebmann
George A. Cioffi
Royce W. S. Chen
Kaveri A. Thakoor
Kaveri A. Thakoor
Kaveri A. Thakoor
Extracting decision-making features from the unstructured eye movements of clinicians on glaucoma OCT reports and developing AI models to classify expertise
eye-tracking
fixations
optical coherence tomography
glaucoma
unsupervised clustering
neural networks
title Extracting decision-making features from the unstructured eye movements of clinicians on glaucoma OCT reports and developing AI models to classify expertise
title_full Extracting decision-making features from the unstructured eye movements of clinicians on glaucoma OCT reports and developing AI models to classify expertise
title_fullStr Extracting decision-making features from the unstructured eye movements of clinicians on glaucoma OCT reports and developing AI models to classify expertise
title_full_unstemmed Extracting decision-making features from the unstructured eye movements of clinicians on glaucoma OCT reports and developing AI models to classify expertise
title_short Extracting decision-making features from the unstructured eye movements of clinicians on glaucoma OCT reports and developing AI models to classify expertise
title_sort extracting decision making features from the unstructured eye movements of clinicians on glaucoma oct reports and developing ai models to classify expertise
topic eye-tracking
fixations
optical coherence tomography
glaucoma
unsupervised clustering
neural networks
url https://www.frontiersin.org/articles/10.3389/fmed.2023.1251183/full
work_keys_str_mv AT michelleakerman extractingdecisionmakingfeaturesfromtheunstructuredeyemovementsofcliniciansonglaucomaoctreportsanddevelopingaimodelstoclassifyexpertise
AT sanmatichoudhary extractingdecisionmakingfeaturesfromtheunstructuredeyemovementsofcliniciansonglaucomaoctreportsanddevelopingaimodelstoclassifyexpertise
AT jeffreymliebmann extractingdecisionmakingfeaturesfromtheunstructuredeyemovementsofcliniciansonglaucomaoctreportsanddevelopingaimodelstoclassifyexpertise
AT georgeacioffi extractingdecisionmakingfeaturesfromtheunstructuredeyemovementsofcliniciansonglaucomaoctreportsanddevelopingaimodelstoclassifyexpertise
AT roycewschen extractingdecisionmakingfeaturesfromtheunstructuredeyemovementsofcliniciansonglaucomaoctreportsanddevelopingaimodelstoclassifyexpertise
AT kaveriathakoor extractingdecisionmakingfeaturesfromtheunstructuredeyemovementsofcliniciansonglaucomaoctreportsanddevelopingaimodelstoclassifyexpertise
AT kaveriathakoor extractingdecisionmakingfeaturesfromtheunstructuredeyemovementsofcliniciansonglaucomaoctreportsanddevelopingaimodelstoclassifyexpertise
AT kaveriathakoor extractingdecisionmakingfeaturesfromtheunstructuredeyemovementsofcliniciansonglaucomaoctreportsanddevelopingaimodelstoclassifyexpertise