Automated Classification of Inherited Retinal Diseases in Optical Coherence Tomography Images Using Few-shot Learning

ZHAO Qi MAI Si Wei LI Qian HUANG Guan Chong GAO Ming Chen YANG Wen Li WANG Ge MA Ya LI Lei PENG Xiao Yan

ZHAO Qi, MAI Si Wei, LI Qian, HUANG Guan Chong, GAO Ming Chen, YANG Wen Li, WANG Ge, MA Ya, LI Lei, PENG Xiao Yan. Automated Classification of Inherited Retinal Diseases in Optical Coherence Tomography Images Using Few-shot Learning[J]. Biomedical and Environmental Sciences, 2023, 36(5): 431-440. doi: 10.3967/bes2023.052
Citation: ZHAO Qi, MAI Si Wei, LI Qian, HUANG Guan Chong, GAO Ming Chen, YANG Wen Li, WANG Ge, MA Ya, LI Lei, PENG Xiao Yan. Automated Classification of Inherited Retinal Diseases in Optical Coherence Tomography Images Using Few-shot Learning[J]. Biomedical and Environmental Sciences, 2023, 36(5): 431-440. doi: 10.3967/bes2023.052

doi: 10.3967/bes2023.052

Automated Classification of Inherited Retinal Diseases in Optical Coherence Tomography Images Using Few-shot Learning

Funds: This study was supported by National Natural Science Foundation of China [No.82171073]
More Information
    Author Bio:

    ZHAO Qi, male, born in 1978, MD, majoring in ocular fundus diseases and medical image analysis

    Corresponding author: PENG Xiao Yan, MD, PhD, Tel: 86-10-58265900, E-mail: 74000041@ccmu.edu.cn
    • 关键词:
    •  / 
    •  / 
    •  / 
    •  / 
    •  / 
    •  / 
    •  
  • Figure  1.  OCT images of three categories in IRDs. (A) Cone/cone-rod lesions: disruption of photoreceptor layers with thinned sensory retina at the fovea. (B) Rod-cone lesions: disruption of photoreceptor layers with thinned sensory retina at the areas outside the fovea with relatively preserved structure at the fovea. (C) Extensive lesions: extensive disruption of photoreceptor layers with thinned sensory retina. IRDs: inherited retinal disorders.

    Figure  2.  Overview of the proposed method.

    Figure  3.  Network architecture of the model. (A) Teacher model. (B) Student model.

    Figure  4.  ROC curves for FSL model and baseline model compared with retinal specialists. (A–C) Test1–3 of FSL model. (D–F) Test1–3 of baseline model. IRDs: inherited retinal disorders; FSL: few-shot learning; AUC: area under curve; CI: confidence interval.

    Figure  5.  Representative OCT images and corresponding heatmaps. (A) Example of cone/cone-rod lesions of IRDs and its corresponding superimposed heatmap. (B) Example of rod-cone lesions of IRDs and its corresponding superimposed heatmap. (C) Example of extensive lesions of IRDs and its corresponding superimposed heatmap. (D) Example of normal and its corresponding superimposed heatmap. (E) Example of control and its corresponding superimposed heatmap. IRDs: inherited retinal disorders.

    Table  1.   Composition of the target dataset

    VariablesCirrus systemSpectralis systemTotal
    SubjectsImagesSubjectsImagesSubjectsImages
    IRDs4173938387791,126
    Cone/cone-rod lesions17271513022401
    Rod-cone lesions12222914421366
    Extensive lesions122462411336359
    Normal252651826843533
    Control353113234767658
    Total1011,315881,0021892,317
      Note. IRDs: inherited retinal disorders.
    下载: 导出CSV

    Table  2.   Summary of model performance in the classification of OCT images

    VariablesFSL modelBaseline model
    Accuracy
    (95% CI)
    Sensitivity
    (95% CI)
    Specificity
    (95% CI)
    F1 Score
    (95% CI)
    Accuracy
    (95% CI)
    Sensitivity
    (95% CI)
    Specificity
    (95% CI)
    F1 Score
    (95% CI)
    Test1 (n = 599)Test1 (n = 599)
    IRDs
    Cone/cone-rod lesions0.975
    (0.958–0.985)
    0.988
    (0.927–0.999)
    0.973
    (0.954–0.984)
    0.919
    (0.844–0.966)
    0.939
    (0.916–0.957)
    0.860
    (0.794–0.907)
    0.970
    (0.947–0.983)
    0.887
    (0.825–0.930)
    Rod-cone lesions0.992
    (0.980–0.997
    0.984
    (0.939–0.997)
    0.994
    (0.980–0.998)
    0.981
    (0.939–0.997)
    0.934
    (0.910–0.952)
    0.586
    (0.476–0.689)
    0.994
    (0.981–0.998)
    0.723
    (0.597–0.816)
    Extensive lesions0.967
    (0.948–0.979)
    0.855
    (0.774–0.911)
    0.994
    (0.980–0.998)
    0.909
    (0.835–0.953)
    0.954
    (0.934–0.969)
    0.967
    (0.913–0.990)
    0.951
    (0.927–0.968)
    0.897
    (0.833–0.944)
    Normal0.995
    (0.984–0.999)
    0.990
    (0.935–0.999)
    0.996
    (0.984–0.999)
    0.984
    (0.935–0.999)
    0.966
    (0.948–0.979)
    0.933
    (0.854–0.972)
    0.972
    (0.953–0.984)
    0.892
    (0.807–0.944)
    Control0.987
    (0.973–0.994)
    0.971
    (0.930–0.989)
    0.993
    (0.978–0.998)
    0.977
    (0.937–0.993)
    0.944
    (0.922–0.961)
    0.947
    (0.889–0.976)
    0.943
    (0.918–0.962)
    0.883
    (0.818–0.931)
    Total0.983
    (0.978–0.987)
    0.957
    (0.936–0.971)
    0.990
    (0.984–0.993)
    0.957
    (0.938–0.972)
    0.948
    (0.939–0.955)
    0.872
    (0.842–0.897)
    0.967
    (0.958–0.973)
    0.870
    (0.840–0.896)
    Test2 (n = 610)Test2 (n = 610)
    IRDs
    Cone/cone-rod lesions0.967
    (0.949–0.979)
    0.954
    (0.890–0.983)
    0.970
    (0.950–0.983)
    0.912
    (0.839–0.954)
    0.943
    (0.921–0.960)
    0.804
    (0.713–0.872)
    0.973
    (0.954–0.985)
    0.835
    (0.746–0.894)
    Rod-cone lesions0.993
    (0.982–0.998)
    0.984
    (0.905–0.999)
    0.995
    (0.983–0.999)
    0.969
    (0.884–0.995)
    0.955
    (0.934–0.970)
    0.767
    (0.637–0.862)
    0.976
    (0.958–0.986)
    0.773
    (0.650–0.873)
    Extensive lesions0.967
    (0.949–0.979
    0.839
    (0.745–0.904)
    0.990
    (0.976–0.996)
    0.886
    (0.796–0.941)
    0.948
    (0.926–0.964)
    0.864
    (0.782–0.919)
    0.967
    (0.946–0.980)
    0.860
    (0.809–0.938)
    Normal0.980
    (0.965–0.989)
    0.899
    (0.823–0.946)
    0.998
    (0.987–0.999)
    0.942
    (0.874–0.976)
    0.941
    (0.919–0.958)
    0.849
    (0.763–0.909)
    0.961
    (0.939–0.975)
    0.837
    (0.755–0.902)
    Control0.962
    (0.943–0.975)
    0.966
    (0.932–0.984)
    0.960
    (0.933–0.977)
    0.952
    (0.917–0.976)
    0.928
    (0.904–0.947)
    0.935
    (0.890–0.962)
    0.924
    (0.892–0.948)
    0.903
    (0.857–0.939)
    Total0.974
    (0.968–0.979)
    0.934
    (0.911–0.952)
    0.984
    (0.978–0.988)
    0.935
    (0.913–0.954)
    0.943
    (0.934–0.951)
    0.866
    (0.835–0.892)
    0.962
    (0.954–0.969)
    0.859
    (0.828–0.885)
    Test3 (n = 594)Test3 (n = 594)
    IRDs
    Cone/cone-rod Lesions0.988
    (0.975–0.995)
    0.987
    (0.922–0.999)
    0.988
    (0.973–0.995)
    0.957
    (0.888–0.990)
    0.958
    (0.938–0.972)
    0.816
    (0.730–0.880)
    0.992
    (0.977–0.997)
    0.882
    (0.805–0.937)
    Rod-Cone Lesions0.992
    (0.979–0.997)
    0.979
    (0.918–0.996)
    0.994
    (0.981–0.998)
    0.974
    (0.918–0.996)
    0.985
    (0.970–0.993)
    0.933
    (0.830–0.978)
    0.991
    (0.977–0.997)
    0.926
    (0.830–0.978)
    Extensive Lesions0.983
    (0.968–0.991)
    0.906
    (0.786–0.965)
    0.991
    (0.977–0.997)
    0.906
    (0.786–0.965)
    0.983
    (0.968–0.991)
    0.929
    (0.855–0.969)
    0.994
    (0.981–0.998)
    0.948
    (0.878–0.981)
    Normal0.980
    (0.964–0.989)
    0.899
    (0.823–0.946)
    0.997
    (0.987–0.999)
    0.942
    (0.876–0.976)
    0.941
    (0.918–0.958)
    0.827
    (0.752–0.884)
    0.976
    (0.956–0.987)
    0.868
    (0.799–0.921)
    Control0.968
    (0.950–0.980)
    0.957
    (0.913–0.980)
    0.973
    (0.951–0.986)
    0.949
    (0.907–0.976)
    0.902
    (0.875–0.925)
    0.934
    (0.885–0.964)
    0.888
    (0.853–0.916)
    0.854
    (0.796–0.899)
    Total0.982
    (0.977–0.987)
    0.948
    (0.924–0.965)
    0.989
    (0.984–0.993)
    0.949
    (0.926–0.966)
    0.954
    (0.946–0.961)
    0.886
    (0.856–0.909)
    0.971
    (0.963–0.977)
    0.885
    (0.857–0.909)
      Note. IRDs: inherited retinal disorders; FSL: few-shot learning; CI: confidence interval.
    下载: 导出CSV
  • [1] Hanany M, Rivolta C, Sharon D. Worldwide carrier frequency and genetic prevalence of autosomal recessive inherited retinal diseases. Proc Natl Acad Sci USA, 2020; 117, 2710−6. doi:  10.1073/pnas.1913179117
    [2] Sundaramurthi H, Moran A, Perpetuini AC, et al. Emerging drug therapies for inherited retinal dystrophies. Adv Exp Med Biol, 2019; 1185, 263−7.
    [3] Botto C, Rucli M, Tekinsoy MD, et al. Early and late stage gene therapy interventions for inherited retinal degenerations. Prog Retin Eye Res, 2022; 86, 100975. doi:  10.1016/j.preteyeres.2021.100975
    [4] Zhang SJ, Wang LF, Xiao Z, et al. Analysis of radial peripapillary capillary density in patients with bietti crystalline dystrophy by optical coherence tomography angiography. Biomed Environ Sci, 2022; 35, 107−14.
    [5] Kim M, Yun J, Cho Y, et al. Deep learning in medical imaging. Neurospine, 2019; 16, 657−68. doi:  10.14245/ns.1938396.198
    [6] Abràmoff MD, Lou YY, Erginay A, et al. Improved automated detection of diabetic retinopathy on a publicly available dataset through integration of deep learning. Invest Ophthalmol Vis Sci, 2016; 57, 5200−6. doi:  10.1167/iovs.16-19964
    [7] Grassmann F, Mengelkamp J, Brandl C, et al. A deep learning algorithm for prediction of age-related eye disease study severity scale for age-related macular degeneration from color fundus photography. Ophthalmology, 2018; 125, 1410−20. doi:  10.1016/j.ophtha.2018.02.037
    [8] Liu HR, Li L, Wormstone IM, et al. Development and validation of a deep learning system to detect glaucomatous optic neuropathy using fundus photographs. JAMA Ophthalmol, 2019; 137, 1353−60. doi:  10.1001/jamaophthalmol.2019.3501
    [9] Kermany D, Zhang K, Goldbaum M. Labeled optical coherence tomography (OCT) and chest X-ray images for classification. Mendeley Data, 2018; 2.
    [10] Srinivasan PP, Kim LA, Mettu PS, et al. Fully automated detection of diabetic macular edema and dry age-related macular degeneration from optical coherence tomography images. Biomed Opt Express, 2014; 5, 3568−77. doi:  10.1364/BOE.5.003568
    [11] Hinton G, Vinyals O, Dean J. Distilling the knowledge in a neural network. Comput Sci, 2015; 14, 38−39.
    [12] Wang L, Yoon KJ. Knowledge distillation and student-teacher learning for visual intelligence: a review and new outlooks. IEEE Trans Pattern Anal Mach Intell, 2022; 44, 3048−68. doi:  10.1109/TPAMI.2021.3055564
    [13] Sankar S, Sidibé D, Cheung Y, et al. Classification of SD-OCT volumes for DME detection: an anomaly detection approach. Medical Imaging 2016: Computer-Aided Diagnosis, 2016; 9785, 688−93.
    [14] Anantrasirichai N, Achim A, Morgan JE, et al. SVM-based texture classification in optical coherence tomography. In: 2013 IEEE 10th International Symposium on Biomedical Imaging. IEEE. 2013, 1332−5.
    [15] Liu YY, Chen M, Ishikawa H, et al. Automated macular pathology diagnosis in retinal OCT images using multi-scale spatial pyramid and local binary patterns in texture and shape encoding. Med Image Anal, 2011; 15, 748−59. doi:  10.1016/j.media.2011.06.005
    [16] Hussain A, Bhuiyan A, Luu CD, et al. Classification of healthy and diseased retina using SD-OCT imaging and Random Forest algorithm. PLoS One, 2018; 13, e0198281. doi:  10.1371/journal.pone.0198281
    [17] Wang J, Deng GH, Li WY, et al. Deep learning for quality assessment of retinal OCT images. Biomed Opt Express, 2019; 10, 6057−72. doi:  10.1364/BOE.10.006057
    [18] Ji QG, He WJ, Huang J, et al. Efficient deep learning-based automated pathology identification in retinal optical coherence tomography images. Algorithms, 2018; 11, 88. doi:  10.3390/a11060088
    [19] Shih FY, Patel H. Deep learning classification on optical coherence tomography retina images. Intern J Pattern Recognit Artif Intell, 2020; 34, 2052002. doi:  10.1142/S0218001420520023
    [20] Fang LY, Wang C, Li ST, et al. Automatic classification of retinal three-dimensional optical coherence tomography images using principal component analysis network with composite kernels. J Biomed Opt, 2017; 22, 116011.
    [21] Ji QG, Huang J, He WJ, et al. Optimized deep convolutional neural networks for identification of macular diseases from optical coherence tomography images. Algorithms, 2019; 12, 51. doi:  10.3390/a12030051
    [22] Karri SPK, Chakraborty D, Chatterjee J. Transfer learning based classification of optical coherence tomography images with diabetic macular edema and dry age-related macular degeneration. Biomed Opt Express, 2017; 8, 579−92. doi:  10.1364/BOE.8.000579
    [23] Lu W, Tong Y, Yu Y, et al. Deep learning-based automated classification of multi-categorical abnormalities from optical coherence tomography images. Transl Vis Sci Technol, 2018; 7, 41. doi:  10.1167/tvst.7.6.41
    [24] Motozawa N, An GZ, Takagi S, et al. Optical coherence tomography-based deep-learning models for classifying normal and age-related macular degeneration and exudative and non-exudative age-related macular degeneration changes. Ophthalmol Ther, 2019; 8, 527−39. doi:  10.1007/s40123-019-00207-y
    [25] Fujinami-Yokokawa Y, Pontikos N, Yang LZ, et al. Prediction of causative genes in inherited retinal disorders from spectral-domain optical coherence tomography utilizing deep learning techniques. J Ophthalmol, 2019; 2019, 1691064.
    [26] Stevenson CH, Hong SC, Ogbuehi KC. Development of an artificial intelligence system to classify pathology and clinical features on retinal fundus images. Clin Exp Ophthalmol, 2019; 47, 484−9. doi:  10.1111/ceo.13433
    [27] Shah M, Roomans Ledo A, Rittscher J. Automated classification of normal and Stargardt disease optical coherence tomography images using deep learning. Acta Ophthalmol, 2020; 98, e715−21.
    [28] LeCun Y, Bottou L, Bengio Y, et al. Gradient-based learning applied to document recognition. Proc IEEE, 1998; 86, 2278−324. doi:  10.1109/5.726791
    [29] Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. In: Proceedings of the 3rd International Conference on Learning Representations. 2015.
    [30] Lake BM, Salakhutdinov R, Tenenbaum JB. One-shot learning by inverting a compositional causal process. In: Proceedings of the 26th International Conference on Neural Information Processing Systems. Curran Associates Inc. 2013, 2526−34.
    [31] Jankowski N, Duch W, Grąbczewski K. Meta-learning in computational intelligence. Springer. 2011.
    [32] Burlina P, Paul W, Mathew P, et al. Low-shot deep learning of diabetic retinopathy with potential applications to address artificial intelligence bias in retinal diagnostics and rare ophthalmic diseases. JAMA Ophthalmol, 2020; 138, 1070−7. doi:  10.1001/jamaophthalmol.2020.3269
    [33] Yoo TK, Choi JY, Kim HK. Feasibility study to improve deep learning in OCT diagnosis of rare retinal diseases with few-shot classification. Med Biol Eng Comput, 2021; 59, 401−15. doi:  10.1007/s11517-021-02321-1
    [34] Xu JG, Shen JX, Wan C, et al. A few-shot learning-based retinal vessel segmentation method for assisting in the central serous chorioretinopathy laser surgery. Front Med (Lausanne), 2022; 9, 821565.
    [35] Quellec G, Lamard M, Conze PH, et al. Automatic detection of rare pathologies in fundus photographs using few-shot learning. Med Image Anal, 2020; 61, 101660. doi:  10.1016/j.media.2020.101660
    [36] Kermany DS, Goldbaum M, Cai WJ, et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell, 2018; 172, 1122−31.e9. doi:  10.1016/j.cell.2018.02.010
    [37] Abbasi S, Hajabdollahi M, Karimi N, et al. Modeling teacher-student techniques in deep neural networks for knowledge distillation. In: 2020 International Conference on Machine Vision and Image Processing (MVIP). IEEE. 2020, 1−6.
    [38] Yoo TK, Choi JY, Seo JG, et al. The possibility of the combination of OCT and fundus images for improving the diagnostic accuracy of deep learning for age-related macular degeneration: a preliminary experiment. Med Biol Eng Comput, 2019; 57, 677−87. doi:  10.1007/s11517-018-1915-z
    [39] Miri MS, Abràmoff MD, Lee K, et al. Multimodal segmentation of optic disc and cup from SD-OCT and color fundus photographs using a machine-learning graph-based approach. IEEE Trans Med Imaging, 2015; 34, 1854−66. doi:  10.1109/TMI.2015.2412881
  • [1] CHEN Qian Wei, HUANG Xue Zan, DING Yu, ZHU Feng Ren, WANG Jia, ZOU Yuan Jie, DU Yuan Zhen, ZHANG Ya Jun, HUI Zi Wen, ZHU Feng Lin, MU Min.  Predicting the Risk of Arterial Stiffness in Coal Miners Based on Different Machine Learning Models . Biomedical and Environmental Sciences, 2024, 37(1): 108-111. doi: 10.3967/bes2024.009
    [2] WANG Dan Dan, WANG An Xin, ZHANG Xiao Li, WEI Wen Bin, WU Shou Ling, ZHAO Xing Quan.  Relationship of Retinal Nerve Fiber Layer Thickness and Retinal Vessel Calibers with Cognitive Impairment in the Asymptomatic Polyvascular Abnormalities Population . Biomedical and Environmental Sciences, 2024, 37(2): 196-203. doi: 10.3967/bes2024.020
    [3] WANG Xiao Ping, LI Ze Yan, ZHANG Meng, LIU Hong Yong.  Machine-learning-assisted Investigation into the Relationship between the Built Environment, Behavior, and Physical Health of the Elderly in China . Biomedical and Environmental Sciences, 2023, 36(10): 987-990. doi: 10.3967/bes2023.125
    [4] DING Zhong Ao, ZHANG Li Ying, LI Rui Ying, NIU Miao Miao, ZHAO Bo, DONG Xiao Kang, LIU Xiao Tian, HOU Jian, MAO Zhen Xing, WANG Chong Jian.  Contribution of Ambient Air Pollution on Risk Assessment of Type 2 Diabetes Mellitus via Explainable Machine Learning . Biomedical and Environmental Sciences, 2023, 36(6): 557-560. doi: 10.3967/bes2023.069
    [5] ZHANG Xiao Li, YU Sheng Nan, QU Ruo Di, ZHAO Qiu Yi, PAN Wei Zhe, CHEN Xu Shen, ZHANG Qian, LIU Yan, LI Jia, GAO Yi, LYU Yi, YAN Xiao Yan, LI Ben, REN Xue Feng, QIU Yu Lan.  Mechanism of Learning and Memory Impairment in Rats Exposed to Arsenic and/or Fluoride Based on Microbiome and Metabolome . Biomedical and Environmental Sciences, 2023, 36(3): 253-268. doi: 10.3967/bes2023.028
    [6] YANG Zhen Zhong, WU Da Wei, DING Bai, MA Hong Bo, FEI Jin Xue, CHENG Liang, ZHAO Ya Li.  Long-Term, Low-Level Microwave Radiation Impairs Learning and Memory via Synbindin: Molecular Basis and Underlying Mechanism . Biomedical and Environmental Sciences, 2022, 35(6): 552-557. doi: 10.3967/bes2022.074
    [7] WU Jie Wen, JIAO Xiao Kang, DU Xin Hui, JIAO Zeng Tao, LIANG Zuo Ru, PANG Ming Fan, JI Han Ran, CHENG Zhi Da, CAI Kang Ning, QI Xiao Peng.  Assessment of the Benefits of Targeted Interventions for Pandemic Control in China Based on Machine Learning Method and Web Service for COVID-19 Policy Simulation . Biomedical and Environmental Sciences, 2022, 35(5): 412-418. doi: 10.3967/bes2022.057
    [8] ZHANG Sheng Juan, WANG Li Fei, XIAO Zhe, LIU Zhi Qiang, XING Chen, LI Qian, SUN Hui Jing, YANG Zan Zhang, LYU Li Na, PENG Xiao Yan.  Analysis of Radial Peripapillary Capillary Density in Patients with Bietti Crystalline Dystrophy by Optical Coherence Tomography Angiography . Biomedical and Environmental Sciences, 2022, 35(2): 107-114. doi: 10.3967/bes2022.015
    [9] LI Xiao, LIU Chao Fei, GUAN Li, WEI Shu, YANG Xin, LI Shu Qiang.  Deep Learning in Chest Radiography: Detection of Pneumoconiosis . Biomedical and Environmental Sciences, 2021, 34(10): 842-845. doi: 10.3967/bes2021.116
    [10] LI Jing, LI Bao Yu, WEI Zi Jian, ZHAO Yu Zhu, LI Tan Shi.  Application Research on Gated Recurrent Unit Deep Learning Prediction and Graded Early Warning of Emergency Department Visits Based on Meteorological Environmental Data . Biomedical and Environmental Sciences, 2020, 33(10): 817-820. doi: 10.3967/bes2020.111
    [11] ZHAO Qi, YANG Wen Li, WANG Xiao Na, Wang Ruikang K, YOU Qi Sheng, CHU Zhong Di, XIN Chen, ZHANG Meng Yu, LI Dong Jun, WANG Zi Yang, CHEN Wei, LI Yi Feng, CUI Rui, SHEN Lin, WEI Wen Bin.  Repeatability and Reproducibility of Quantitative Assessment of the Retinal Microvasculature Using Optical Coherence Tomography Angiography Based on Optical Microangiography . Biomedical and Environmental Sciences, 2018, 31(6): 407-412. doi: 10.3967/bes2018.054
    [12] WANG Xiao Na, YOU Qi Sheng, ZHAO Hui Ying, PENG Xiao Yan.  Optical Coherence Tomography Features of Tuberculous Serpiginous-like Choroiditis and Serpiginous Choroiditis . Biomedical and Environmental Sciences, 2018, 31(5): 327-334. doi: 10.3967/bes2018.043
    [13] DENG Min, YUAN Jing, CHEN Wei Tian, Queenie CHAN, James F GRIFFITH, WANG Yi Xiang.  Evaluation of Glycosaminoglycan in the Lumbar Disc Using Chemical Exchange Saturation Transfer MR at 3.0 Tesla:Reproducibility and Correlation with Disc Degeneration . Biomedical and Environmental Sciences, 2016, 29(1): 47-55. doi: 10.3967/bes2016.005
    [14] CHEN Yao, LI Qian, ZOU Ye, ZHOU Zhao Xiang, FENG Wei Wei, BAO Yong Tuan, MA Rui Hong, JI Peng Cheng, WU Jiang, YANG Liu Qing, WU Xiang Yang.  Protective Effect of Mulberry Extract against Pb-induced Learning and Memory Deficits in Mice . Biomedical and Environmental Sciences, 2014, 27(1): 70-73. doi: 10.3967/bes2014.019
    [15] XIU-JING CAO, MING WANG, WEI-HENG CHEN, DA-MIAO ZHU, JIA-QI SHE, DI-YUN RUAN.  Effects of Chronic Administration of Melatonin on Spatial Learning Ability and Long-term Potentiation in Lead-exposed and Control Rats . Biomedical and Environmental Sciences, 2009, 22(1): 70-75.
    [16] JIE YANG, LI-NA JIANG, ZHEN-LI YUAN, YU-FEI ZHENG, LU WANG, MIN JI, ZHI-QIANG SHEN, XIN-WEI WANG, QIANG MA, ZHU-GE XI, JUN-WEN LI.  Impacts of Passive Smoking on Learning and Memory Ability of Mouse Offsprings and Intervention by Antioxidants . Biomedical and Environmental Sciences, 2008, 21(2): 144-149.
    [17] YING XIAO, LING WANG, RUO-JUN XU, ZHEN-YU CHEN.  DHA Depletion in Rat Brain Is Associated With Impairment on Spatial Learning and Memory . Biomedical and Environmental Sciences, 2006, 19(6): 474-480.
    [18] LI SUN, ZHENG-YAN ZHAO, JIAN HU, XIE-LAI ZHOU.  Potential Association of Lead Exposure During Early Development of Mice With Alteration of Hippocampus Nitric Oxide Levels and Learning Memory . Biomedical and Environmental Sciences, 2005, 18(6): 375-378.
    [19] JUAN HAN, HAN-RONG WU, YI-ZHEN YU, SEN-BEI YANG, YONG-MEI HUANG.  Study on Self-consciousness of Children With Learning Disabilities and Related Factors . Biomedical and Environmental Sciences, 2005, 18(3): 207-210.
    [20] Bin Yao, HAN-RONG WU.  Risk Factors of Learning Disabilities in Chinese Children in Wuhan . Biomedical and Environmental Sciences, 2003, 16(4): 392-397.
  • 加载中
图(5) / 表ll (2)
计量
  • 文章访问数:  456
  • HTML全文浏览量:  204
  • PDF下载量:  44
  • 被引次数: 0
出版历程
  • 收稿日期:  2022-11-23
  • 录用日期:  2023-02-23
  • 刊出日期:  2023-05-20

Automated Classification of Inherited Retinal Diseases in Optical Coherence Tomography Images Using Few-shot Learning

doi: 10.3967/bes2023.052
    基金项目:  This study was supported by National Natural Science Foundation of China [No.82171073]
    作者简介:

    ZHAO Qi, male, born in 1978, MD, majoring in ocular fundus diseases and medical image analysis

    通讯作者: PENG Xiao Yan, MD, PhD, Tel: 86-10-58265900, E-mail: 74000041@ccmu.edu.cn

English Abstract

ZHAO Qi, MAI Si Wei, LI Qian, HUANG Guan Chong, GAO Ming Chen, YANG Wen Li, WANG Ge, MA Ya, LI Lei, PENG Xiao Yan. Automated Classification of Inherited Retinal Diseases in Optical Coherence Tomography Images Using Few-shot Learning[J]. Biomedical and Environmental Sciences, 2023, 36(5): 431-440. doi: 10.3967/bes2023.052
Citation: ZHAO Qi, MAI Si Wei, LI Qian, HUANG Guan Chong, GAO Ming Chen, YANG Wen Li, WANG Ge, MA Ya, LI Lei, PENG Xiao Yan. Automated Classification of Inherited Retinal Diseases in Optical Coherence Tomography Images Using Few-shot Learning[J]. Biomedical and Environmental Sciences, 2023, 36(5): 431-440. doi: 10.3967/bes2023.052
    • The incidence of visual impairment and blindness caused by inherited retinal diseases (IRDs) is increasing because of prolonged life expectancy. A recent analysis shows approximately 5.5 million people worldwide have IRDs [1]. In recent years, several advancements have been made in the diagnosis and treatment of IRDs, including drug and gene therapies[2-4]. Spectral-domain optical coherence tomography (SD-OCT) has played a crucial role in the diagnosis, progression surveillance, strategy exploration, and response assessment of treatment in patients with IRDs. However, in some cases, the recognition, interpretation, and comparison of minor changes in IRDs, as shown by OCT, could be difficult and time-consuming for retinal specialists.

      Artificial intelligence technology is gaining popularity in retinal imaging owing to its enhanced processing power, massive data, and novel algorithms[5]. Recently, automated image analysis has been successfully applied to detect changes in fundus and OCT images of multiple retinal diseases, such as diabetic retinopathy[6], age-related macular degeneration (AMD)[7], and glaucoma[8]. These diseases are highly prevalent among populations and enable the acquisition of large volumes of training data for traditional machine learning (ML) approaches, including deep learning (DL).

      In contrast, for rare diseases such as IRDs, acquiring a large volume of high-quality representative data from the patient cohort is challenging. Furthermore, the datasets require accompanying annotations generated by specialists, which is time-consuming. The scarcity of samples and laborious work hinder the application of image classification.

      Therefore, researchers have imitated the fact that human beings can learn quickly and proposed few-shot learning (FSL). Unlike traditional networks, FSL aims to address the problem of data limitation when learning new tasks. This avoids the problems of parameter overfitting and a low model generalization performance. This shows greater adaptability to unknown conditions and provides a reliable solution for areas where samples are scarce.

      This study aimed to design a diagnosis FSL model for the classification of OCT images in patients with IRDs when only a very limited number of samples can be collected.

    • A publicly available dataset of OCT images from the Cell dataset[9] and BOE dataset[10] was selected as the auxiliary dataset for training. The Cell dataset contains four classifications, including choroidal neovascularization (CNV), diabetic macular edema (DME), drusen, and normal, which total 109,309 samples with two image resolutions, 1,536 × 496 and 1,024 × 496 pixels. The BOE dataset has 3,231 samples, including dry AMD, DME, and normal, with three types of image resolution: 1,024 × 496, 768 × 496, and 512 × 496 pixels.

      SD-OCT images of the target dataset were acquired using the Cirrus HD-OCT 5000 system (Carl Zeiss Meditec Inc., Dublin, USA) and the Heidelberg Spectralis system (Heidelberg Engineering, Heidelberg, Germany). An OCT scan containing the center of the fovea from each macular OCT was selected as the input data. Images from the Cirrus system were scanned at a length of 6 mm, and the image resolution was 1,180 × 786 pixels. The images from the Spectralis system were scanned at a length of 6 or 9 mm, and the image resolution was 1,024 × 496 and 1,536 × 496 pixels, respectively.

      Two certified retinal specialists graded all OCT images in the clinical datasets separately. The diagnosis of IRDs was based on both clinical and genetic detections. The OCT images of IRDs were classified into three types based on morphological features: cone/cone-rod lesions, that is, disruption of photoreceptor layers with thinned sensory retina at the fovea; rod-cone lesions, that is, disruption of photoreceptor layers with thinned sensory retina at the areas outside the fovea with relatively preserved structure at the fovea; and extensive lesions, that is, extensive disruption of photoreceptor layers with thinned sensory retina (Figure 1).

      Figure 1.  OCT images of three categories in IRDs. (A) Cone/cone-rod lesions: disruption of photoreceptor layers with thinned sensory retina at the fovea. (B) Rod-cone lesions: disruption of photoreceptor layers with thinned sensory retina at the areas outside the fovea with relatively preserved structure at the fovea. (C) Extensive lesions: extensive disruption of photoreceptor layers with thinned sensory retina. IRDs: inherited retinal disorders.

      This study was conducted following the principles of the Declaration of Helsinki. Approval was granted by the Ethics Committee of the Beijing Tongren Hospital, Capital Medical University.

    • The proposed pipeline consisted of three parts: data preprocessing, training the teacher model, and training the student model. Data preprocessing mainly includes image angle adjustments and vertical pixel column movements. The teacher model was first trained on the auxiliary OCT datasets with a four-class classification designed to learn the nuances of each disease and then transferred to the target OCT dataset to classify the five target classes. The student model was trained using the soft label provided by the teacher model based on knowledge distillation (KD) [11] and the hard label from annotation (Figure 2).

      Figure 2.  Overview of the proposed method.

    • The original OCT images show different angles, noise distribution, and size diversity because of the acquisition machine and the patient. This will distract the neural network from the focal area and increase the training time due to useless data input during training. We used a random inversion of the images in the process, which did not destroy the hierarchical information of the OCT images, and cropped the given images to a random size and aspect ratio to increase the generalization ability of the model. The experiments were all preprocessed and resized to 224 × 224 pixels.

    • The ResNet-50, a convolutional neural network (CNN) framework, was chosen as the teacher model, which is the backbone structure for absorbing and learning the information from the auxiliary dataset, specifically the textures, patterns, and pixel distributions in the end-level convolutional layers. After learning using the same type of dataset, we froze the parameters of the teacher model, except for the last fully connected layer, to transfer the task to the target domain.

    • To overcome the obstacles caused by the lack of training data, we used the combination of KD and student–teacher learning[12] for knowledge transfer. ResNet-18 was used as the student model to be trained from scratch to adapt to the target data with a smaller number of samples.

      $$ \begin{aligned} {\mathcal{L}} (x;W)=& \alpha \cdot H\left(y,\sigma \left({z}_{s};T=1\right)\right)+\beta \\ & \cdot H\left(\sigma \left({z}_{t};T=\tau \right),\sigma \left({z}_{s};T=\tau \right)\right) \end{aligned}$$ (1)

      where in Equation 1, α and $\beta $ control the balance of the information coming from the two sources, which generally add up to 1. H is the loss function, σ is the softmax function parameterized by the temperature T, zs is the logits from the student network, and zt is the logits from the teacher network. τ denotes the temperature of the adapted softmax function, and each probability pi of class i in the batch is calculated from logits zi as follows:

      $$ {p}_{i}=\frac{\mathrm{exp}\left(\frac{{z}_{i}}{T}\right)}{\sum _{j} \mathrm{e}\mathrm{x}\mathrm{p}\left(\frac{{z}_{j}}{T}\right)} $$ (2)

      where T in Equation 2 increases, the probability distribution of the output becomes “softer,” which means that the differences among the probabilities of each class decrease and more information will be provided.

      Figure 3 illustrates the network architecture of the model. The model was trained and tested using the Python (version 3.10.7) programming language with the PyTorch (version 1.8.1) library as the backend. The computer used in this study was equipped with an NVIDIA GeForce RTX 2070 8 GB graphics processing unit, 32 GB random access memory, and an Intel Core 11th Gen Intel(R) Core(TM) i9-11900 @ 2.50GHz central processing unit.

      Figure 3.  Network architecture of the model. (A) Teacher model. (B) Student model.

    • To fully evaluate the performance of the proposed method, we used various evaluation metrics, namely accuracy, sensitivity, specificity, and F1 score, which are defined as follows:

      $$ \text{Accuracy}=\frac{TP+TN}{TP+TN+FN+FP} $$ (3)
      $$ \text{Sensitivity}\text{}=\frac{TP}{TP+FN} $$ (4)
      $$ \text{Specificity}\text{}=\frac{TN}{TN+FP} $$ (5)
      $$ \mathrm{F}1=\frac{2TP}{2 {TP}+FN+FP} $$ (6)

      where TP, TN, FP, and FN denote true positives, true negatives, false positives, and false negatives, respectively, and are measured according to the confusion matrix.

      Receiver operating characteristic (ROC) curves were also used to display the performance of the FSL model for the classification of OCT images. Heatmaps were used to show the regions of interest (ROI) in the model.

      The process was repeated three times with random assignment of participants to the training and testing sets to control for selection bias, given the relatively small sample size. The metrics were calculated for each training/testing process for each category.

    • A total of 2,317 images from 189 participants were included in this study as the target dataset, of which 1,126 images of 79 participants were IRDs, 533 images of 43 participants were normal samples, and 658 images of 67 participants were controls, including CNV, DME, macular hole, epiretinal membrane, and retinal detachment (Table 1). The images were randomly split in a 3:1 ratio into the training and testing sets.

      Table 1.  Composition of the target dataset

      VariablesCirrus systemSpectralis systemTotal
      SubjectsImagesSubjectsImagesSubjectsImages
      IRDs4173938387791,126
      Cone/cone-rod lesions17271513022401
      Rod-cone lesions12222914421366
      Extensive lesions122462411336359
      Normal252651826843533
      Control353113234767658
      Total1011,315881,0021892,317
        Note. IRDs: inherited retinal disorders.
    • The FSL model achieved better performance than the baseline model. The baseline model was adapted from ResNet-18, which has the same structure as the student model. The baseline model was added to demonstrate that excluding the advantages of the model performance itself, our proposed method is the main contributor to improving the final results. In the three testing sets, the FSL model achieved a total accuracy ranging from 0.974 [95% confidence interval (CI) 0.968–0.979] to 0.983 (95% CI 0.978–0.987), total sensitivity from 0.934 (95% CI 0.911–0.952) to 0.957 (95% CI 0.936–0.971), total specificity from 0.984 (95% CI 0.978–0.988) to 0.990 (95% CI 0.984–0.993), and total F1 score from 0.935 (95% CI 0.913–0.954) to 0.957 (95% CI 0.938–0.972). The baseline model achieved total accuracy ranging from 0.943 (95% CI 0.934–0.951) to 0.954 (95% CI 0.946–0.961), total sensitivity from 0.866 (95% CI 0.835–0.892) to 0.886 (95% CI 0.856–0.909), total specificity from 0.962 (95% CI 0.954–0.969) to 0.971 (95% CI 0.963–0.977), and total F1 score from 0.859 (95% CI 0.828–0.885) to 0.885 (95% CI 0.857–0.909) (Table 2).

      Table 2.  Summary of model performance in the classification of OCT images

      VariablesFSL modelBaseline model
      Accuracy
      (95% CI)
      Sensitivity
      (95% CI)
      Specificity
      (95% CI)
      F1 Score
      (95% CI)
      Accuracy
      (95% CI)
      Sensitivity
      (95% CI)
      Specificity
      (95% CI)
      F1 Score
      (95% CI)
      Test1 (n = 599)Test1 (n = 599)
      IRDs
      Cone/cone-rod lesions0.975
      (0.958–0.985)
      0.988
      (0.927–0.999)
      0.973
      (0.954–0.984)
      0.919
      (0.844–0.966)
      0.939
      (0.916–0.957)
      0.860
      (0.794–0.907)
      0.970
      (0.947–0.983)
      0.887
      (0.825–0.930)
      Rod-cone lesions0.992
      (0.980–0.997
      0.984
      (0.939–0.997)
      0.994
      (0.980–0.998)
      0.981
      (0.939–0.997)
      0.934
      (0.910–0.952)
      0.586
      (0.476–0.689)
      0.994
      (0.981–0.998)
      0.723
      (0.597–0.816)
      Extensive lesions0.967
      (0.948–0.979)
      0.855
      (0.774–0.911)
      0.994
      (0.980–0.998)
      0.909
      (0.835–0.953)
      0.954
      (0.934–0.969)
      0.967
      (0.913–0.990)
      0.951
      (0.927–0.968)
      0.897
      (0.833–0.944)
      Normal0.995
      (0.984–0.999)
      0.990
      (0.935–0.999)
      0.996
      (0.984–0.999)
      0.984
      (0.935–0.999)
      0.966
      (0.948–0.979)
      0.933
      (0.854–0.972)
      0.972
      (0.953–0.984)
      0.892
      (0.807–0.944)
      Control0.987
      (0.973–0.994)
      0.971
      (0.930–0.989)
      0.993
      (0.978–0.998)
      0.977
      (0.937–0.993)
      0.944
      (0.922–0.961)
      0.947
      (0.889–0.976)
      0.943
      (0.918–0.962)
      0.883
      (0.818–0.931)
      Total0.983
      (0.978–0.987)
      0.957
      (0.936–0.971)
      0.990
      (0.984–0.993)
      0.957
      (0.938–0.972)
      0.948
      (0.939–0.955)
      0.872
      (0.842–0.897)
      0.967
      (0.958–0.973)
      0.870
      (0.840–0.896)
      Test2 (n = 610)Test2 (n = 610)
      IRDs
      Cone/cone-rod lesions0.967
      (0.949–0.979)
      0.954
      (0.890–0.983)
      0.970
      (0.950–0.983)
      0.912
      (0.839–0.954)
      0.943
      (0.921–0.960)
      0.804
      (0.713–0.872)
      0.973
      (0.954–0.985)
      0.835
      (0.746–0.894)
      Rod-cone lesions0.993
      (0.982–0.998)
      0.984
      (0.905–0.999)
      0.995
      (0.983–0.999)
      0.969
      (0.884–0.995)
      0.955
      (0.934–0.970)
      0.767
      (0.637–0.862)
      0.976
      (0.958–0.986)
      0.773
      (0.650–0.873)
      Extensive lesions0.967
      (0.949–0.979
      0.839
      (0.745–0.904)
      0.990
      (0.976–0.996)
      0.886
      (0.796–0.941)
      0.948
      (0.926–0.964)
      0.864
      (0.782–0.919)
      0.967
      (0.946–0.980)
      0.860
      (0.809–0.938)
      Normal0.980
      (0.965–0.989)
      0.899
      (0.823–0.946)
      0.998
      (0.987–0.999)
      0.942
      (0.874–0.976)
      0.941
      (0.919–0.958)
      0.849
      (0.763–0.909)
      0.961
      (0.939–0.975)
      0.837
      (0.755–0.902)
      Control0.962
      (0.943–0.975)
      0.966
      (0.932–0.984)
      0.960
      (0.933–0.977)
      0.952
      (0.917–0.976)
      0.928
      (0.904–0.947)
      0.935
      (0.890–0.962)
      0.924
      (0.892–0.948)
      0.903
      (0.857–0.939)
      Total0.974
      (0.968–0.979)
      0.934
      (0.911–0.952)
      0.984
      (0.978–0.988)
      0.935
      (0.913–0.954)
      0.943
      (0.934–0.951)
      0.866
      (0.835–0.892)
      0.962
      (0.954–0.969)
      0.859
      (0.828–0.885)
      Test3 (n = 594)Test3 (n = 594)
      IRDs
      Cone/cone-rod Lesions0.988
      (0.975–0.995)
      0.987
      (0.922–0.999)
      0.988
      (0.973–0.995)
      0.957
      (0.888–0.990)
      0.958
      (0.938–0.972)
      0.816
      (0.730–0.880)
      0.992
      (0.977–0.997)
      0.882
      (0.805–0.937)
      Rod-Cone Lesions0.992
      (0.979–0.997)
      0.979
      (0.918–0.996)
      0.994
      (0.981–0.998)
      0.974
      (0.918–0.996)
      0.985
      (0.970–0.993)
      0.933
      (0.830–0.978)
      0.991
      (0.977–0.997)
      0.926
      (0.830–0.978)
      Extensive Lesions0.983
      (0.968–0.991)
      0.906
      (0.786–0.965)
      0.991
      (0.977–0.997)
      0.906
      (0.786–0.965)
      0.983
      (0.968–0.991)
      0.929
      (0.855–0.969)
      0.994
      (0.981–0.998)
      0.948
      (0.878–0.981)
      Normal0.980
      (0.964–0.989)
      0.899
      (0.823–0.946)
      0.997
      (0.987–0.999)
      0.942
      (0.876–0.976)
      0.941
      (0.918–0.958)
      0.827
      (0.752–0.884)
      0.976
      (0.956–0.987)
      0.868
      (0.799–0.921)
      Control0.968
      (0.950–0.980)
      0.957
      (0.913–0.980)
      0.973
      (0.951–0.986)
      0.949
      (0.907–0.976)
      0.902
      (0.875–0.925)
      0.934
      (0.885–0.964)
      0.888
      (0.853–0.916)
      0.854
      (0.796–0.899)
      Total0.982
      (0.977–0.987)
      0.948
      (0.924–0.965)
      0.989
      (0.984–0.993)
      0.949
      (0.926–0.966)
      0.954
      (0.946–0.961)
      0.886
      (0.856–0.909)
      0.971
      (0.963–0.977)
      0.885
      (0.857–0.909)
        Note. IRDs: inherited retinal disorders; FSL: few-shot learning; CI: confidence interval.
    • The performance of the FSL model was compared with that of retinal specialists using ROC curve plots. The AUCs of the FSL model were higher for most sub-classifications (Figure 4).

      Figure 4.  ROC curves for FSL model and baseline model compared with retinal specialists. (A–C) Test1–3 of FSL model. (D–F) Test1–3 of baseline model. IRDs: inherited retinal disorders; FSL: few-shot learning; AUC: area under curve; CI: confidence interval.

    • The heatmaps demonstrated that the FSL model made the classification based on the correct lesion for diagnosis (Figure 5).

      Figure 5.  Representative OCT images and corresponding heatmaps. (A) Example of cone/cone-rod lesions of IRDs and its corresponding superimposed heatmap. (B) Example of rod-cone lesions of IRDs and its corresponding superimposed heatmap. (C) Example of extensive lesions of IRDs and its corresponding superimposed heatmap. (D) Example of normal and its corresponding superimposed heatmap. (E) Example of control and its corresponding superimposed heatmap. IRDs: inherited retinal disorders.

    • This preliminary study demonstrates the effective use of the FSL model for classifying OCT images from patients with IRDs and normal and control participants with a smaller volume of data.

      Previous methods for diagnosing ocular diseases such as AMD and DME through OCT images are based on traditional ML methods such as principal component analysis (PCA)[13, 14], support vector machine (SVM)[10, 15], and random forest[16]. Other studies have focused on DL methods, including improving existing mature and pretrained frameworks, such as Inception-v3[17, 18], VGG-16[17, 19], PCANet [20], GoogLeNet [21, 22], ResNet[21, 23], and DenseNet[21] to classify OCT images or unify multiple networks to make the classification more robust for diagnosing, for example, four-parallel-ResNet system[23] and multi-stage network[24].

      For the success of these algorithms, particularly for tasks such as image recognition, most models are overparameterized to extract the most salient features and ensure generalization. However, the performance of the models is heavily dependent on the very large and high-quality labels of the training datasets. Some studies have reported relevant results. Fujinami-Yokokawa et al.[25] used Inception-v3 CNN[26] to classify OCT images from patients with three different IRDs and obtained a relatively high accuracy. Shah et al.[27] used a new CNN with a simple architecture similar to LeNet[28], which was more effective than the VGG-19 CNN[29] in differentiating normal OCT images from images of patients with Stargardt disease. However, varying degrees of overfitting and reduced training accuracy were observed. These are the disadvantages of large DL algorithms for small sample data.

      Traditional network frames and large-volume data are no longer suitable for identifying rare diseases, such as IRDs, that belong to the small-sample classification problem. Therefore, the concept of FSL has been proposed to solve such problems[30, 31]. FSL combines limited supervision information (small samples) with prior knowledge (unlabeled or weakly labeled samples, other datasets and labels, other models, etc.) such that the model can effectively learn information from small samples. Recently, a few FSL methods, such as contrastive self-supervision learning[32], generative adversarial networks (GAN)[33], integrated frameworks [34], and unsupervised probabilistic models [35], have been used for the segmentation and classification of fundus images.

      This is the first time that FSL has been used to classify OCT images of IRDs. Our model is based on a student–teacher learning framework, an appealing paradigm for semi-supervised learning in medical image classification. The images were normalized using conventional preprocessing to reduce the effect of noise on the model during training. The teacher model is designed to adapt to the target dataset based on TL[36], an approach to increase the performance of DL classifiers, in which a large multiclass image classifier is pretrained on a natural image dataset and then retrained on the smaller medical imaging dataset of interest. The resulting model has fewer parameters to be trained than the original model and is “fine-tuned” for the medical imaging task. In the process from the teacher model to the student model, the KD method[11] that we chose was soft-target training in which the teacher generates soft labels[37] for the training data, and these soft targets are used in combination with one-hot encoded labels (hard targets) to train the student model. KD refers to a method that helps the training process of a smaller student network under the supervision of a larger teacher network. Unlike other compression methods, KD can downsize a network regardless of the structural difference between the teacher and student networks [12]. With this framework, the student model could increase the training speed, prevent overfitting, result in more generalized models, and maintain a relatively simple structure simultaneously.

      Our FSL model attained higher total accuracy, sensitivity, specificity, and F1 score than the baseline model and achieved better performance for most subclassifications. The differences in the sensitivity and F1 scores between the two models were more apparent. Higher sensitivity demonstrated that the model could achieve better classification results. Moreover, the F1 score is the harmonic mean of precision and recall, indicating the robust performance of the model. In addition, an interesting phenomenon exists, in which the specificity of both models was greater than the sensitivity, suggesting that the classification of models tended to be conservative. This may also be more secure for potential clinical applications. The ROC analysis also showed that the FSL model had higher AUCs for most subclassifications. The heatmaps demonstrated that the regions highlighted with warmer colors represented areas more critical for final class determination. The ROIs were precisely captured by the FSL model, and the results were compatible with the judgment of the retinal specialist. In this study, we trained and tested OCT images from different devices and sizes, and proved that the model had good generalizability. In summary, our FSL model could perform comparably to the advanced large teacher model and in a fraction of the time, making ideal candidates run in clinical settings where the runtime is important, and computing resources are often limited.

      Several recent studies have reported that the combination of multiple modalities and imaging techniques can improve diagnostic performance. Yoo et al.[38] designed an algorithm based on the combination of fundus photographs and OCT images, which increased the diagnostic accuracy of AMD compared with the data alone. Miri et al.[39] used complementary information from fundus photographs and OCT volumes to segment the optic disc and cup boundaries. Fundus and OCT images are recognized as the most influential biomarkers of retinal diseases. However, previous studies have paid little attention to the function of a diagnostic model that combines the fundus with OCT images. Images from different devices can provide unique and complementary information. Our study needs to adopt a multimodal process to combine various types of imaging for a more precise performance in the future.

      This study had some limitations. The clinical datasets included only patients with a sample size, which led to the existence of a potential similarity in the case of multiple images from the same patient. Future research, including genetic data with sufficient patient numbers to represent the significant phenotypic and genetic heterogeneity of IRDs, may help with phenotype–genotype correlation. Second, the cross-sectional study could not identify the sequence of intraretinal structural changes in the transitional zone over time, and misclassification due to disease progression may occur. Incorporating temporal data into the model may allow the leading disease front to be automatically identified and monitored.

      In this preliminary study, we demonstrated the potential of the FSL model to differentiate OCT images of IRDs from healthy individuals and patients with other common macular diseases. Furthermore, the model can also distinguish between different types of IRDs using a dataset smaller than that traditionally used. In real-life clinical settings, the general principle demonstrated in this study and similar network architectures can be applied to other retinal diseases with a low prevalence.

    • The authors have no conflict of interest to declare.

参考文献 (39)

目录

    /

    返回文章
    返回