LMeRAN: Label Masking-Enhanced Residual Attention Network for Multi-Label Chest X-Ray Disease Aided Diagnosis

Chest X-ray (CXR) imaging is essential for diagnosing thoracic diseases, and computer-aided diagnosis (CAD) systems have made substantial progress in automating the interpretation of CXR images. However, some existing methods often overemphasize local features while neglecting global context, limiti...

Full description

Saved in:
Bibliographic Details
Published inSensors (Basel, Switzerland) Vol. 25; no. 18; p. 5676
Main Authors Fu, Hongping, Song, Chao, Qu, Xiaolong, Li, Dongmei, Zhang, Lei
Format Journal Article
LanguageEnglish
Published Switzerland MDPI AG 11.09.2025
Subjects
Online AccessGet full text
ISSN1424-8220
1424-8220
DOI10.3390/s25185676

Cover

More Information
Summary:Chest X-ray (CXR) imaging is essential for diagnosing thoracic diseases, and computer-aided diagnosis (CAD) systems have made substantial progress in automating the interpretation of CXR images. However, some existing methods often overemphasize local features while neglecting global context, limiting their ability to capture the broader pathological landscape. Moreover, most methods fail to model label correlations, leading to insufficient utilization of prior knowledge. To address these limitations, we propose a novel multi-label CXR image classification framework, termed the Label Masking-enhanced Residual Attention Network (LMeRAN). Specifically, LMeRAN introduces an original label-specific residual attention to capture disease-relevant information effectively. By integrating multi-head self-attention with average pooling, the model dynamically assigns higher weights to critical lesion areas while retaining global contextual features. In addition, LMeRAN employs a label mask training strategy, enabling the model to learn complex label dependencies from partially available label information. Experiments conducted on the large-scale public dataset ChestX-ray14 demonstrate that LMeRAN achieves the highest mean AUC value of 0.825, resulting in an increase of 3.1% to 8.0% over several advanced baselines. To enhance interpretability, we also visualize the lesion regions relied upon by the model for classification, providing clearer insights into the model’s decision-making process.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
content type line 23
ISSN:1424-8220
1424-8220
DOI:10.3390/s25185676