Retinal toxicity among long-term users of Hydroxychloroquine manifests with loss in the Ellipsoid zone (EZ) detectable on SD-OCT imaging. This work reports an automatic deep-learning algorithm to detect and segment EZ loss in SD-OCT. The proposed model predicts EZ loss map, in a dual network architecture that operates in parallel combining scan-by-scan detections in horizontal and vertical directions. The combined model demonstrated the best overall performance with F1 score = 0.91 ± 0.07, improving the performance compared to individual models. Automatic methods for EZ loss detection could provide a useful tool to facilitate screening of patients for evidence of toxicity.
Spatial alignment of longitudinally acquired retinal images is necessary for the development of image-based metrics identifying structural features associated with disease progression in diseases such as age-related macular degeneration (AMD). This work develops and evaluates a feature-based registration framework for accurate and robust registration of retinal images. Methods: Two feature-based registration approaches were investigated for the alignment of fundus auto-fluorescence images. The first method used conventional SIFT local feature descriptors to solve for the geometric transformation between two corresponding point-sets. The second method used a deep-learning approach with a network architecture mirroring the feature localization and matching process of the conventional method. The methods were validated using clinical images acquired in an ongoing longitudinal study of AMD and consisted of 75 patients (145 eyes) with 4 year follow up imaging. In the deep-learning method, 113 image pairs were used during training (with the ground truth provided from manually verified SIFT feature registration) and 20 image pairs were used for testing (with the ground truth provided from manual landmark annotation). Results: Conventional method using SIFT features demonstrated target registration error (mean ± std) = 0.05 ± 0.04 mm, substantially improving the alignment from the initialization with error = 0.34 ± 0.22 mm. The deep-learning method, on the other hand, exhibited error = 0.10 ± 0.07 mm. While both methods improved upon the initial misalignment, SIFT method showed the best overall geometric accuracy. However, deep-learning method exhibited robust performance (error = 0.15 ± 0.09 mm) in the 7% of cases that SIFT method exhibited failures (error = 3.71 ± 6.36 mm). Conclusion: While both methods demonstrated successful performance, SIFT method exhibited the best overall geometric accuracy whereas deep-learning method was superior in terms of robustness. Achieving accurate and robust registration is essential in large-scale studies investigating factors underlying retinal disease progression such as in AMD.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.