Skip to main content
Erschienen in: Graefe's Archive for Clinical and Experimental Ophthalmology 11/2020

26.08.2020 | Glaucoma

Prediction of visual field from swept-source optical coherence tomography using deep learning algorithms

verfasst von: Keunheung Park, Jinmi Kim, Sangyoon Kim, Jonghoon Shin

Erschienen in: Graefe's Archive for Clinical and Experimental Ophthalmology | Ausgabe 11/2020

Einloggen, um Zugang zu erhalten

Abstract

Purpose

To develop a deep learning method to predict visual field (VF) from wide-angle swept-source optical coherence tomography (SS-OCT) and compare the performance of three Google Inception architectures.

Methods

Three deep learning models (with Inception-ResNet-v2, Inception-v3, and Inception-v4) were trained to predict 24-2 VF from the macular ganglion cell-inner plexiform layer and the peripapillary retinal nerve fibre layer map obtained by SS-OCT. The prediction performance of the three models was evaluated by using the root mean square error (RMSE) between the actual and predicted VF. The performance was also compared among different glaucoma severities and Garway-Heath sectorizations.

Results

The training dataset comprised images of 2220 eyes from 1120 subjects, and the test dataset was obtained from another 305 subjects (305 eyes). In all subjects, the global prediction errors (RMSEs) were 4.44 ± 2.09 dB, 4.78 ± 2.38 dB, and 4.85 ± 2.66 dB for the Inception-ResNet-v2, Inception-v3, and Inception-v4 architectures, respectively, and the prediction error of Inception-ResNet-v2 was significantly lower than the other two (P < 0.001). As glaucoma progressed, the prediction error of all three architectures significantly worsened to 6.59 dB, 7.33 dB, and 7.79 dB, respectively. In the analysis of sectors, the nasal sector had the lowest prediction error, followed by the superotemporal sector.

Conclusions

Inception-ResNet-v2 achieved the best performance, and the global prediction error (RMSE) was 4.44 dB. As glaucoma progressed, the prediction error became larger. This method may help clinicians determine VF, particularly for patients who are unable to undergo a physical VF test.
Anhänge
Nur mit Berechtigung zugänglich
Literatur
1.
Zurück zum Zitat Quigley HA, Katz J, Derick RJ et al (1992) An evaluation of optic disc and nerve fiber layer examinations in monitoring progression of early glaucoma damage. Ophthalmology 99:19–28CrossRef Quigley HA, Katz J, Derick RJ et al (1992) An evaluation of optic disc and nerve fiber layer examinations in monitoring progression of early glaucoma damage. Ophthalmology 99:19–28CrossRef
2.
Zurück zum Zitat Artes PH, Chauhan BC (2005) Longitudinal changes in the visual field and optic disc in glaucoma. Prog Retin Eye Res 24:333–354CrossRef Artes PH, Chauhan BC (2005) Longitudinal changes in the visual field and optic disc in glaucoma. Prog Retin Eye Res 24:333–354CrossRef
3.
Zurück zum Zitat Fogagnolo P, Sangermani C, Oddone F et al (2011) Long-term perimetric fluctuation in patients with different stages of glaucoma. Br J Ophthalmol 95:189–193CrossRef Fogagnolo P, Sangermani C, Oddone F et al (2011) Long-term perimetric fluctuation in patients with different stages of glaucoma. Br J Ophthalmol 95:189–193CrossRef
5.
Zurück zum Zitat Langerhorst CT, Van den Berg T, Van Spronsen R, Greve EL (1985) Results of a fluctuation analysis and defect volume program for automated static threshold perimetry with the scoperimeter. In: Sixth International Visual Field Symposium. Springer, pp 1–6 Langerhorst CT, Van den Berg T, Van Spronsen R, Greve EL (1985) Results of a fluctuation analysis and defect volume program for automated static threshold perimetry with the scoperimeter. In: Sixth International Visual Field Symposium. Springer, pp 1–6
6.
Zurück zum Zitat Brenton RS, Argus WA (1987) Fluctuations on the Humphrey and Octopus perimeters. Invest Ophthalmol Vis Sci 28:767–771PubMed Brenton RS, Argus WA (1987) Fluctuations on the Humphrey and Octopus perimeters. Invest Ophthalmol Vis Sci 28:767–771PubMed
7.
Zurück zum Zitat Gürses-Özden R, Teng C, Vessani R et al (2004) Macular and retinal nerve fiber layer thickness measurement reproducibility using optical coherence tomography (OCT-3). J Glaucoma 13:238CrossRef Gürses-Özden R, Teng C, Vessani R et al (2004) Macular and retinal nerve fiber layer thickness measurement reproducibility using optical coherence tomography (OCT-3). J Glaucoma 13:238CrossRef
8.
Zurück zum Zitat Blumenthal EZ, Williams JM, Weinreb RN et al (2000) Reproducibility of nerve fiber layer thickness measurements by use of optical coherence tomography. Ophthalmology 107:2278–2282CrossRef Blumenthal EZ, Williams JM, Weinreb RN et al (2000) Reproducibility of nerve fiber layer thickness measurements by use of optical coherence tomography. Ophthalmology 107:2278–2282CrossRef
14.
Zurück zum Zitat Garway-Heath DF, Poinoosawmy D, Fitzke FW, Hitchings RA (2000) Mapping the visual field to the optic disc in normal tension glaucoma eyes. Ophthalmology 107:1809–1815CrossRef Garway-Heath DF, Poinoosawmy D, Fitzke FW, Hitchings RA (2000) Mapping the visual field to the optic disc in normal tension glaucoma eyes. Ophthalmology 107:1809–1815CrossRef
15.
Zurück zum Zitat Szegedy C, Vanhoucke V, Ioffe S, et al (2016) Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 2818–2826 Szegedy C, Vanhoucke V, Ioffe S, et al (2016) Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 2818–2826
16.
Zurück zum Zitat He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 770–778 He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 770–778
17.
Zurück zum Zitat Bianco S, Cadene R, Celona L, Napoletano P (2018) Benchmark analysis of representative deep neural network architectures. IEEE Access 6:64270–64277CrossRef Bianco S, Cadene R, Celona L, Napoletano P (2018) Benchmark analysis of representative deep neural network architectures. IEEE Access 6:64270–64277CrossRef
18.
Zurück zum Zitat Foster PJ, Buhrmann R, Quigley HA, Johnson GJ (2002) The definition and classification of glaucoma in prevalence surveys. Br J Ophthalmol 86:238–242CrossRef Foster PJ, Buhrmann R, Quigley HA, Johnson GJ (2002) The definition and classification of glaucoma in prevalence surveys. Br J Ophthalmol 86:238–242CrossRef
21.
Zurück zum Zitat Dubitzky W, Granzow M, Berrar DP (2007) Fundamentals of data mining in genomics and proteomics. Springer Science & Business Media, pp 177–180 Dubitzky W, Granzow M, Berrar DP (2007) Fundamentals of data mining in genomics and proteomics. Springer Science & Business Media, pp 177–180
22.
Zurück zum Zitat Jonas JB, Budde WM, Lang P (1998) Neuroretinal rim width ratios in morphological glaucoma diagnosis. Br J Ophthalmol 82:1366–1371CrossRef Jonas JB, Budde WM, Lang P (1998) Neuroretinal rim width ratios in morphological glaucoma diagnosis. Br J Ophthalmol 82:1366–1371CrossRef
24.
Zurück zum Zitat Srivastava RK, Greff K, Schmidhuber J (2015) Highway networks. arXiv preprint arXiv:150500387 Srivastava RK, Greff K, Schmidhuber J (2015) Highway networks. arXiv preprint arXiv:150500387
25.
Zurück zum Zitat Li J, Fang F, Mei K, Zhang G (2018) Multi-scale residual network for image super-resolution. In: Proceedings of the European Conference on Computer Vision (ECCV). pp 517–532 Li J, Fang F, Mei K, Zhang G (2018) Multi-scale residual network for image super-resolution. In: Proceedings of the European Conference on Computer Vision (ECCV). pp 517–532
26.
Zurück zum Zitat Zhang Y, Tian Y, Kong Y, et al (2018) Residual dense network for image super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 2472–2481 Zhang Y, Tian Y, Kong Y, et al (2018) Residual dense network for image super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 2472–2481
27.
Zurück zum Zitat Li K, Bare B, Yan B, et al (2018) HNSR: highway networks based deep convolutional neural networks model for single image super-resolution. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). pp 1478–1482 Li K, Bare B, Yan B, et al (2018) HNSR: highway networks based deep convolutional neural networks model for single image super-resolution. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). pp 1478–1482
29.
Zurück zum Zitat Hood DC, De Cuir N, Blumberg DM et al (2016) A single wide-field OCT protocol can provide compelling information for the diagnosis of early glaucoma. Transl Vis Sci Technol 5:4–4CrossRef Hood DC, De Cuir N, Blumberg DM et al (2016) A single wide-field OCT protocol can provide compelling information for the diagnosis of early glaucoma. Transl Vis Sci Technol 5:4–4CrossRef
32.
Zurück zum Zitat Kawano J, Tomidokoro A, Mayama C et al (2006) Correlation between hemifield visual field damage and corresponding parapapillary atrophy in normal-tension glaucoma. Am J Ophthalmol 142:40–45.e1CrossRef Kawano J, Tomidokoro A, Mayama C et al (2006) Correlation between hemifield visual field damage and corresponding parapapillary atrophy in normal-tension glaucoma. Am J Ophthalmol 142:40–45.e1CrossRef
Metadaten
Titel
Prediction of visual field from swept-source optical coherence tomography using deep learning algorithms
verfasst von
Keunheung Park
Jinmi Kim
Sangyoon Kim
Jonghoon Shin
Publikationsdatum
26.08.2020
Verlag
Springer Berlin Heidelberg
Erschienen in
Graefe's Archive for Clinical and Experimental Ophthalmology / Ausgabe 11/2020
Print ISSN: 0721-832X
Elektronische ISSN: 1435-702X
DOI
https://doi.org/10.1007/s00417-020-04909-z

Weitere Artikel der Ausgabe 11/2020

Graefe's Archive for Clinical and Experimental Ophthalmology 11/2020 Zur Ausgabe

Neu im Fachgebiet Augenheilkunde

Update Augenheilkunde

Bestellen Sie unseren Fach-Newsletter und bleiben Sie gut informiert.