Citation: HE X, LI XH, PENG J, et al. Multi-label fundus disease classification using dual-branch deep learning: an intelligent diagnosis framework inspired by traditional Chinese medicine Five Wheels theory. Digital Chinese Medicine, 2026, 9(1): 80-90. DOI: 10.1016/j.dcmed.2026.02.007
Citation: Citation: HE X, LI XH, PENG J, et al. Multi-label fundus disease classification using dual-branch deep learning: an intelligent diagnosis framework inspired by traditional Chinese medicine Five Wheels theory. Digital Chinese Medicine, 2026, 9(1): 80-90. DOI: 10.1016/j.dcmed.2026.02.007

Multi-label fundus disease classification using dual-branch deep learning: an intelligent diagnosis framework inspired by traditional Chinese medicine Five Wheels theory

  • Objective To develop a dual-branch deep learning framework for accurate multi-label classification of fundus diseases, addressing the key limitations of insufficient complementary feature extraction and inadequate cross-modal feature fusion in existing automated diagnostic methods.
    Methods The fundus multi-label classification dataset with 12 disease categories (FMLC-12) dataset was constructed by integrating complementary samples from Ocular Disease Intelligent Recognition (ODIR) and Retinal Fundus Multi-Disease Image Dataset (RFMiD), yielding 6 936 fundus images across 12 retinal pathology categories, and the framework was validated on both FMLC-12 and ODIR. Inspired by the holistic multi-regional assessment principle of the Five Wheels theory in traditional Chinese medicine (TCM) ophthalmology, the dual-branch multi-label network (DBMNet) was developed as a novel framework integrating complementary visual feature extraction with pathological correlation modeling. The architecture employed a TransNeXt backbone within a dual-branch design: one branch processed red-green-blue (RGB) images to capture color-dependent features, such as vascular patterns and lesion morphology, while the other processed grayscale-converted images to enhance subtle textural details and contrast variations. A feature interaction module (FIM) effectively integrated the multi-scale features from both branches. Comprehensive ablation studies were conducted to evaluate the contributions of the dual-branch architecture and the FIM. The performance of DBMNet was compared against four state-of-the-art methods, including EfficientNet Ensemble, transfer learning-based convolutional neural network (CNN), BFENet, and EyeDeep-Net, using mean average precision (mAP), F1-score, and Cohen's kappa coefficient.
    Results The dual-branch architecture improved mAP by 15.44 percentage points over the single-branch TransNeXt baseline, increasing from 34.41% to 44.24%, and the addition of FIM further boosted mAP to 49.85%. On FMLC-12, DBMNet achieved an mAP of 49.85%, a Cohen’s kappa coefficient of 62.14%, and an F1-score of 70.21%. Compared with BFENet (mAP: 45.42%, kappa: 46.64%, F1-score: 71.34%), DBMNet outperformed it by 4.43 percentage points in mAP and 15.50 percentage points in kappa, while BFENet achieved a marginally higher F1-score. On ODIR, DBMNet achieved an F1-score of 85.50%, comparable to state-of-the-art methods.
    Conclusion DBMNet effectively integrates RGB and grayscale visual modalities through a dual-branch architecture, significantly improving multi-label fundus disease classification. The framework not only addresses the issue of insufficient feature fusion in existing methods but also demonstrates outstanding performance in balancing detection across both common and rare diseases, providing a promising and clinically applicable pathway for standardized, intelligent fundus disease classification.
  • loading

Catalog

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return