Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
Addendum
Announcement
Announcements
Author’ response
Author’s reply
Authors' response
Authors#x2019; response
Book Received
Book Review
Book Reviews
Books Received
Centenary Review Article
Clinical Image
Clinical Images
Commentary
Communicable Diseases - Original Articles
Correspondence
Correspondence, Letter to Editor
Correspondences
Correspondences & Authors’ Responses
Corrigendum
Corrrespondence
Critique
Current Issue
Editorial
Editorial Podcast
Errata
Erratum
FORM IV
GUIDELINES
Health Technology Innovation
IAA CONSENSUS DOCUMENT
Innovations
Letter to Editor
Malnutrition & Other Health Issues - Original Articles
Media & News
Notice of Retraction
Obituary
Original Article
Original Articles
Panel of Reviewers (2006)
Panel of Reviewers (2007)
Panel of Reviewers (2009) Guidelines for Contributors
Perspective
Policy
Policy Document
Policy Guidelines
Policy, Review Article
Policy: Correspondence
Policy: Editorial
Policy: Mapping Review
Policy: Original Article
Policy: Perspective
Policy: Process Paper
Policy: Scoping Review
Policy: Special Report
Policy: Systematic Review
Policy: Viewpoint
Practice
Practice: Authors’ response
Practice: Book Review
Practice: Clinical Image
Practice: Commentary
Practice: Correspondence
Practice: Letter to Editor
Practice: Method
Practice: Obituary
Practice: Original Article
Practice: Pages From History of Medicine
Practice: Perspective
Practice: Review Article
Practice: Short Note
Practice: Short Paper
Practice: Special Report
Practice: Student IJMR
Practice: Systematic Review
Pratice, Original Article
Pratice, Review Article
Pratice, Short Paper
Programme
Programme, Correspondence, Letter to Editor
Programme: Authors’ response
Programme: Commentary
Programme: Correspondence
Programme: Editorial
Programme: Original Article
Programme: Originial Article
Programme: Perspective
Programme: Rapid Review
Programme: Review Article
Programme: Short Paper
Programme: Special Report
Programme: Status Paper
Programme: Systematic Review
Programme: Viewpoint
Protocol
Public Notice
Research Brief
Research Correspondence
Retraction
Review Article
Reviewers
Short Paper
Some Forthcoming Scientific Events
Special Article
Special Opinion Paper
Special Report
Special Section Nutrition & Food Security
Status Paper
Status Report
Strategy
Student IJMR
Systematic Article
Systematic Review
Systematic Review & Meta-Analysis
View Point
Viewpoint
White Paper
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
Addendum
Announcement
Announcements
Author’ response
Author’s reply
Authors' response
Authors#x2019; response
Book Received
Book Review
Book Reviews
Books Received
Centenary Review Article
Clinical Image
Clinical Images
Commentary
Communicable Diseases - Original Articles
Correspondence
Correspondence, Letter to Editor
Correspondences
Correspondences & Authors’ Responses
Corrigendum
Corrrespondence
Critique
Current Issue
Editorial
Editorial Podcast
Errata
Erratum
FORM IV
GUIDELINES
Health Technology Innovation
IAA CONSENSUS DOCUMENT
Innovations
Letter to Editor
Malnutrition & Other Health Issues - Original Articles
Media & News
Notice of Retraction
Obituary
Original Article
Original Articles
Panel of Reviewers (2006)
Panel of Reviewers (2007)
Panel of Reviewers (2009) Guidelines for Contributors
Perspective
Policy
Policy Document
Policy Guidelines
Policy, Review Article
Policy: Correspondence
Policy: Editorial
Policy: Mapping Review
Policy: Original Article
Policy: Perspective
Policy: Process Paper
Policy: Scoping Review
Policy: Special Report
Policy: Systematic Review
Policy: Viewpoint
Practice
Practice: Authors’ response
Practice: Book Review
Practice: Clinical Image
Practice: Commentary
Practice: Correspondence
Practice: Letter to Editor
Practice: Method
Practice: Obituary
Practice: Original Article
Practice: Pages From History of Medicine
Practice: Perspective
Practice: Review Article
Practice: Short Note
Practice: Short Paper
Practice: Special Report
Practice: Student IJMR
Practice: Systematic Review
Pratice, Original Article
Pratice, Review Article
Pratice, Short Paper
Programme
Programme, Correspondence, Letter to Editor
Programme: Authors’ response
Programme: Commentary
Programme: Correspondence
Programme: Editorial
Programme: Original Article
Programme: Originial Article
Programme: Perspective
Programme: Rapid Review
Programme: Review Article
Programme: Short Paper
Programme: Special Report
Programme: Status Paper
Programme: Systematic Review
Programme: Viewpoint
Protocol
Public Notice
Research Brief
Research Correspondence
Retraction
Review Article
Reviewers
Short Paper
Some Forthcoming Scientific Events
Special Article
Special Opinion Paper
Special Report
Special Section Nutrition & Food Security
Status Paper
Status Report
Strategy
Student IJMR
Systematic Article
Systematic Review
Systematic Review & Meta-Analysis
View Point
Viewpoint
White Paper
View/Download PDF

Translate this page into:

Original Article
163 (
2
); 174-181
doi:
10.25259/IJMR_1854_2025

Artificial intelligence (AI)-driven ensemble model for comprehensive chest X-ray abnormality detection and deployment

Department of AI and Digital Twin Division, Institute for Plasma Research, Ahmadabad, Gujarat, India
Divisions of Informatics and Data Centre, Indian Council of Medical Research, New Delhi, India
Divisions of Delivery Research, Indian Council of Medical Research, New Delhi, India
Department of Respiratory Medicine, Critical Care Medicine, University College of Medical Sciences, University of Delhi and Guru Teg Bahadur Hospital, New Delhi, India
Department of Radiology, University College of Medical Sciences, University of Delhi, New Delhi, India
Department of Paediatrics, University College of Medical Sciences, University of Delhi, New Delhi, India
Department of TB and Respiratory Diseases, National Institute of Tuberculosis and Respiratory Diseases, New Delhi, India
Department of Microbiology, Vallabhbhai Patel Chest Institute, All India Institute of Medical Sciences, New Delhi, India
Department of Pulmonary Medicine, Vallabhbhai Patel Chest Institute, All India Institute of Medical Sciences, New Delhi, India
Department of Pulmonary, Critical care and Sleep Medicine, All India Institute of Medical Sciences, New Delhi, India
Department of Paediatric Immunology, ICMR-National Institute of Immunohaematology, Mumbai, India
Department of Radiology, King Edward Memorial Hospital & Seth Govardhandas Sunderdas Medical College, Mumbai, India
Department of Pulmonary Medicine, King Edward Memorial Hospital & Seth Govardhandas Sunderdas Medical College, Mumbai, India
Department of Paediatrics, Grant Government Medical College & Sir J.J. Group of Hospitals, Mumbai, India
Department of Radiology, Bai Jerbai Wadia Hospital for Children, Mumbai, India
Department of Paediatric infectious diseases, Bai Jerbai Wadia Hospital for Children, Mumbai, India
Department of Radiodiagnosis, Mahatma Gandhi Institute of Medical Sciences, Wardha, India
Department of Microbiology, Mahatma Gandhi Institute of Medical Sciences, Wardha, India
Department of Respiratory Medicine, Indira Gandhi Govt. Medical College & Hospital, Nagpur, Maharashtra, India
Department of Paediatrics, All India Institute of Medical Sciences, Bhopal, India
Department of Community Medicine, Dr Laxminarayan Pandey Government Medical College, Ratlam, Madhya Pradesh, India
Department of Paediatrics, King George’s Medical University, Lucknow, Uttar Pradesh, India
Department of Respiratory Medicine, Agartala Government Medical College and G. B. Pant Hospital, Agartala, Tripura, Tamil Nadu, India
Department of Statistics and Probability, ICMR-National Institute for Research in Tuberculosis, Chennai, Tamil Nadu, India
Department of Public Health Research, #ICMR-Regional Medical Research Center, Bhubaneshwar, India
Department of Pulmonary Medicine, Srirama Chandra Bhanja Medical College and Hospital, Cuttack, Odisha, India
Department of Virology Laboratory, National Institute for Implementation Research on Non-Communicable Diseases, Jodhpur, Rajasthan, India
Department of Pulmonary Medicine, Dr. S. N. Medical College, Jodhpur, Rajasthan, India
Department of Radiology, St. John Medical College, Bengaluru, Karnataka, India
Department of Community Medicine, St. John Medical College, Bengaluru, Karnataka, India
Department of Pulmonary and Critical Care, Postgraduate Institute of Medical Sciences, Rohtak, Haryana, India

For correspondence: Dr Manjula Singh, Division of Delivery Research, Indian Council of Medical Research, Delhi 110 029, Professor, Academy of Scientific and Industrial Research, Ghaziabad, Uttar Pradesh, India e-mail: drmanjulasb@outlook.com; drmanjulasb@gmail.com

Licence
This is an open-access article distributed under the terms of the Creative Commons Attribution-Non Commercial-Share Alike 4.0 License, which allows others to remix, transform, and build upon the work non-commercially, as long as the author is credited and the new creations are licensed under the identical terms.

How to cite this article: Abhishek A, Chalga MS, Yadav RM, Agarwal K, Vohra V, Tayade A, et al. Artificial intelligence (AI)-driven ensemble model for comprehensive chest X-ray abnormality detection and deployment. Indian J Med Res. 2026;163:174-81. doi: 10.25259/IJMR_1854_2025.

Abstract

Background and objectives

Chest X-rays (CXR) are widely used for screening of thoracic abnormalities, particularly for tuberculosis (TB) in public health settings. However, the lack of trained radiologists in peripheral areas limits timely interpretation. This study presents the development and validation of DeepCXR v1.1, an artificial intelligence (AI)-powered tool designed to identify radiological chest abnormalities without relying on metadata or clinical inputs, making it ideal for large-scale screening programmes.

Methods

In present multicentric study, AI tool was trained on over 282,000 annotated data points from 54,000 CXR images (36,500 abnormal and 17,500 normal) collected from children and adults from 18 centres across 11 States in India. The tool employs a multi-model ensemble architecture including lung segmentation and lesion-specific models to classify images as normal or abnormal. The tool was validated on multiple datasets, and the final independent validation was done on 13927 CXR images collected prospectively from patients coming to the outpatient clinics of the departments of Medicine and Chest of participating centres.

Results

The tool demonstrated strong generalisability across training and validation datasets, achieving sensitivity of 92.2% [95% confidence interval (CI) 91.6, 92.7] and specificity of 77.4% (95% CI 76.1, 78.6) in a blind prospective validation. Its performance was independently validated by expert committees and health technology assessment panels. Advanced post-processing modules were integrated to enhance detection accuracy, particularly for complex anatomical regions such as the heart and diaphragm.

Interpretation and conclusions

DeepCXR v1.1, an indigenously developed AI tool for detecting abnormalities in chest X-rays offers a scalable, interpretable, and robust solution for augmenting radiological screening and improving early disease detection. The tool’s ability to function offline on basic hardware further supports its use in resource-limited settings.

Keywords

Abnormality detection
Artificial intelligence
Chest X-Ray
Lung segmentation
Tuberculosis

Tuberculosis (TB) remains one of the leading causes of death from a single infectious agent globally.1,2 Though microbiological confirmation remains the gold standard, in its absence diagnosis is often based on a combination of clinical evaluation and radiological findings on chest X-rays (CXR), which remain the most widely available diagnostic modality even in rural and hard-to-reach areas.3 However, while CXRs are available, their interpretation requires trained radiologists, whose availability is highly constrained in many parts of the world,4 leading to diagnostic delays.

Interpretation of CXR image is also time-intensive, susceptible to intra and inter-observer variability up-to 20-30%,5 and often inconsistent, especially under high volume public health programmes.6 Artificial intelligence (AI) tools offer a promising solution of rapid, automated chest abnormalities detection. Deep learning AI models especially those using convolutional neural networks (CNNs),7 when trained on sufficiently large and diverse datasets, annotated specifically for the purpose of training an AI algorithm can deliver high sensitivity and specificity in detecting some conditions such as tuberculosis, pneumonia, lung cancer, and interstitial lung diseases.810

Despite these advancements, challenges exist to detect early-stage subtle lesions and distinguishing between morphologically similar disease entities.11 Available models have limited scope restricted to detection of limited TB lesions or few conditions like pneumonia, often excluding paediatric cases, thus limiting real-world applicability.12 Our study aimed to develop a clinically deployable AI model as screening tool for differentiating abnormal and normal X-rays, as per requirements of screening programmes, specifically, India’s National TB Elimination Program (NTEP)13 by enabling scalable, reliable and cost-effective screening by categorising CXRs as normal or abnormal in high TB burden settings.

Methods

This study was initiated under India TB Research Consortium (ITRC), Indian Council of Medical Research (ICMR) in collaboration with Institute of Plasma Research (IPR) and participating sites for data collection.

Study setting and sample size

This multicentric study was conducted between May 2022 and December 2024. The training and validation datasets were collected from 18 sites across 11 States in India after obtaining ethical clearance from participating sites (details of training strategies is provided as Supplementary Material). The sample size was independently determined for tuberculosis and non-tuberculous lung diseases in both paediatric and adult populations, considering the prevalence of tuberculosis and other lung diseases. Sample size for radiological evidence of one particular presentation of tuberculosis was calculated as 1900 X-rays, considering the prevalence of tuberculosis as 5% as per National TB prevalence survey 2019-2021.14 Considering 20 different presentations of tuberculosis, the total sample size was 38,000 X-rays, which was further subdivided into adult (90%) and paediatric (10%) groups as per the prevalence in these groups15 rounded to the nearest thousand. For non-TB diseases, sample size equivalent to that considered for the different TB presentations were considered for the training dataset.

Supplementary Material

Normal chest X-rays (reported normal by the radiologist) were collected from individuals without any chest complaint and coming for medical examination or pre-anaesthetic check-up. Abnormal chest X-rays were obtained from patients having confirmed disease (TB including various presentation as confirmed by gold standard tests and other lung diseases-based on respective gold/reference standard) having defined lesions as per the disease variant or the disease.

Some of the X-rays were available as archived data with site investigators or were collected from medical record department (MRD) of the respective Institute. These X-rays were obtained when patients visited the outpatients or when hospitalised for diagnosis. Some of the X-rays were collected prospectively from patients visiting the OPD for follow up or with fresh chest complaints with some carrying X-ray images done outside the participating Institutes. All CXRs were included only after the quality of the X-rays was satisfactory. One of the participating sites uploaded few CXRs from field study. These X-rays were taken from handheld X-rays machines.

Most of the X-rays were taken in the hospital from X-ray machines located in the Radiology department of the participating hospitals (Supplementary Table I). All X-ray machines used by sites had AERB (atomic energy regulatory board) approval and used the similar defined parameters for taking chest X-rays.

Supplementary Table I

Training dataset composition

The dataset used to train the AI model comprised of total 54,000 chest X-ray (CXR) images, (36,500 abnormal and 17,500 normal), sourced from multiple Indian states (Supplementary Fig. 1).

Supplementary Fig. 1

The dataset included 9,930 images from children (including neonates) <18 yr and 44,070 from adults. From abnormal images, a total of 282,000 bounding box (BB) annotations were generated, representing the entire spectrum of pulmonary and extrapulmonary disease presentations - both tuberculous and non-tuberculous. With an average width of 131.96 and height of 171.96 pixels, BB size distribution, (Supplementary Fig. 1B), indicated a right-skewed distribution, suggesting a larger number of relatively small BBs, with fewer instances of extremely large BBs. These large BBs indicate, late-stage pleural effusion and structurally distorted cases. Normal images were used to balance the incorrect prediction.

Validation dataset

Comprised of 20972 CXR images (13583 normal and 7389 abnormal) (Supplementary Fig. 1E).

Data pre-processing and standardization

A tool Conv.4.1 was developed for pre-processing naïve CXR images to resize them to standard 850x700 pixels using contrast limited adaptive histogram equalisation (CLAHE) and enhance subtle features like early infiltrates and small nodules, reducing computational load while preserving spatial detail for accurate lesion annotation.

Data annotation

Pre-processed images were imported into label.me annotation tool.16 Text was entered for demarcation of regions of interest and suspected lesions by radiologists/pulmonary physicians as per reference taxonomy using polygon feature. The site radiologists and pulmonary physicians were trained on identification of lesions and annotations. All identified abnormal lesions were categorised under a single ‘abnormal’ category.

Data augmentation

To increase model robustness and reduce overfitting, augmentation techniques were applied ±2 degrees per image (resulting in 2 additional samples) inclusive of horizontal flipping; hue, contrast, and saturation adjustments and random square cropping (scale 0.2 to 1.5).

Model architecture

The tool comprised an ensemble of six deep learning models tailored to specific tasks within CXR interpretation. Table I shows parameters for segmentation and detection model.

Table I. Comparative tabulation used for training the ensemble of AI models
Feature Model 0 Model 1
Feature extractor Inception-ResNet-v2 ResNet-101
Image resizer Fixed 600x600 Fixed 600x600
Training steps 500,000 1,800,000
Learning rate Cosine decay (0.005base) Cosine decay (0.005 base, lower warmup)
Batch size 4 8
Data augmentation Flip only Flip, hue, contrast, saturation, crop
Instance segmentation Yes (MaskR-CNN) No (Faster R-CNN only)
Dropout No Yes

Training strategy

Region proposal and classification

This was executed in two stages. In the first stage, the region proposal network (RPN) generated up to 300 region proposals per image. These proposals were filtered using an intersection over union (IoU) threshold of 0.7 to eliminate overlapping regions. The loss functions for this stage included a localisation loss weighted at 2.0 and an objectless loss weighted at 1.0. In second stage, a Mask R-CNN-based predictor performed BB classification and localisation. This stage used fully connected (FC) layers with L2 regularisation and applied a Softmax function for classification. To handle duplicate detections, batch non-maximum suppression (NMS) was applied with an IoU threshold of 0.6. The system allows a maximum of 100 detections per class and up to 300 detections per image.

Development flowchart

The training of AI model ( Fig. 1) was done on CXR with confirmed diagnosis. After scanning all images and annotation files for errors, a total of 36,500 abnormal and 17,500 normal images were identified, ensuring sufficient dataset for model training using preprocessing and augmentation techniques. The dataset was then split into training and testing sets, maintaining an 80-20% ratio.

Flowchart depicting detailed steps of AI model training process.
Fig. 1.
Flowchart depicting detailed steps of AI model training process.

The AI model was trained for minimum 500,000 epochs, with the test set used to identify the best-fit model and prevent overfitting ( Table I). Supplementary Figures 2, 3 and 4 depict the custom approach to identify lesions using AI prediction.

Supplementary Fig. 2

Supplementary Fig. 3

Supplementary Fig. 4

Finally, the model was validated using a separate set of true positive and true negative images. The training process was repeated with different hyperparameter configurations until the model achieved an acceptable accuracy.

Training and evaluation losses

Supplementary Figure 5 represents the training and evaluation loss trends over multiple epochs during model training indicating how well the model was learning.

Supplementary Fig. 5

The trend observed suggested that the model generalised well, as both losses decrease overtime. Additionally, applying regularisation techniques like dropout or increasing the dataset size helped enhance generalisation and prevent over-fitting, where the model memorises training data instead of learning general patterns or underfitting i.e model is not learning enough from the data.

Heat-map visualization

The generated heat-map overlay (Supplementary Fig. 6) on CXR highlights regions of interest where the current AI model has detected potential abnormalities. The colour intensity varies, with red and yellow areas indicating highly probable regions with potential pathological findings, while blue and green regions represent lower activity areas. The visualisation effectively demonstrates how the model localises critical areas in the radiograph, aiding in interpretability and clinical decision-making.

Supplementary Fig. 6

Operational consideration

This AI system is operational both online and offline, making it deployable on high-end GPUs as well as basic laptops. Its efficiency, rapid processing, and minimal infrastructure requirements make it particularly suitable for integration into public health screening programs in low-resource settings.

Analysis

The receiver operating characteristic (ROC) curve evaluated the model’s performance by plotting the true positive rate against the false positive rate across thresholds, with the area under the curve (AUC) quantifying overall class discrimination ability ( Fig. 2). The F1 score is the mean of precision and recall, providing a balanced measure of accuracy that emphasises correct identification of positive cases while accounting for both false positives and false negatives.

Receiver operating characteristic (ROC) curve for determining optimal threshold for evaluating the model’s performance with the area under the curve (AUC) quantifying overall class discrimination ability.
Fig. 2.
Receiver operating characteristic (ROC) curve for determining optimal threshold for evaluating the model’s performance with the area under the curve (AUC) quantifying overall class discrimination ability.

Statistical analysis

The tool’s performance in terms of sensitivity, specificity, along with 95% confidence intervals was calculated by comparison with radiologist’s observation as normal or abnormal using SPSS software (IBM India Ltd., Bengaluru).

Results

The AI model’s performance was evaluated using three distinct test datasets summarised in Table II.

Table II. Performance metrics of AI tool evaluated on different datasets for identification of abnormality in chest X-rays
Model DeepCXR Version 1.1 Number of images Sensitivity % (95% CI) Specificity % (95% CI) Accuracy (%) F1 score Precision (%)
Dataset Case 1:(Controlled internal Dataset) 3048 96.07 95.80 95.94 95.81 95.55
Dataset Case 2:(Validation dataset 1) 20971 94.2 (93.7, 94.8) 88.3 (87.8, 88.9) 90.4 89.80 91
Dataset Case 3:(Validation dataset 2) 13927 92.2 (91.6, 92.7) 77.4 (76.1, 78.6) 87.4 88.46 85

Case1: Controlled internal data set and curated test set for threshold selection

The first dataset was used to determine the ROC curve ( Fig. 2) and establish an optimal threshold for the base model. The CXR images used included true normal from patients screened for physical fitness (pre-employment fitness, fitness for surgery) and the abnormal from patients with confirmed diseases which were verified by a panel of three experts. A cut-off value of 0.3 was found to be most suitable for distinguishing abnormal images. Additionally, for grouped classification, any value above 0.75 was considered significant. The confusion matrix for this dataset is detailed in Supplementary Table II A.

Supplementary Table IIA

Case 2: Large-scale validation dataset

The validation dataset consisted of 20,971 images, including 13,583 normal images and 7388 abnormal images verified by a panel of three experts. The model demonstrated strong performance with a sensitivity of 94.2 % and specificity of 88.3% (Supplementary Table II B).

Supplementary Table IIB

Case 3: Independent validation dataset on prospectively collected data

The AI tool was validated on a larger dataset collected prospectively after obtaining consent from patients coming to the OPD of Medicine Department/Chest clinic of participating centres. It included 13927 CXR images and the analysis of AI results compared to radiologists’ reading was done by an independent statistician which yielded a sensitivity and specificity of 92.2% (95% CI 91.6, 92.7) and 77.4% (95% CI 76.1, 78.6), respectively (Supplementary Table II C). The developing and validation team were blinded to this dataset.

Supplementary Table IIC

These results highlight the robustness of the AI model in detecting abnormalities across different datasets.

Discussion

Despite growing number of AI tools developed for healthcare, their performance in terms of specificity and diagnostic accuracy has restricted their uptake for public health use including the AI tools developed for screening of chest X-rays for differentiating normal /abnormal and identification of TB. Our AI tool (DeepCXR v1.1) consistently demonstrated sensitivity and specificity of 92.2% and 77.4%; and 94.2% and 88.3% on two independent larger validation datasets of 13,927 and 20,971 CXRs, respectively, as per the WHO Target Product Profile17 for screening better, than any of the currently available AI tools globally.

Our tool’s better performance is due to its training on an extensive diverse dataset of digital/scanned radiographs, in terms of disease spectrum and geographical variations, age related changes, anatomical variants18 and across all age groups ensuring its generalisability for use in real world settings which added to the strength of our tool in comparison to other commercially available tools.8,19 Limitations of other AI tools reported in various studies are comparatively lower specificity, ability to detect fewer common lesions only20,21 possibly due to training on smaller number20 and limited spectrum of presentations, reliance on publicly available datasets,22 or lack of diversity owing to data from single centre.23

In direct comparison with other AI systems as reported in studies, DeepCXR v1.1 not only outperformed them but also addressed limitations of other AI tools as reported by Mosquera et al19 (86.0% sensitivity; 88.0% specificity on small dataset of 1,064 images), and Jin et al8 (88.5% sensitivity; 72.3% specificity on 6,006 images; AUC 0.87) A.A.M. Jasmi et al24 in their work evaluated qXRver 2.1 with 1.3M CXRs, the tool achieved high sensitivity (96.2%) and NPV (99.9%) but low PPV (5.1%) due to low prevalence. In another smaller study by Blake et al25 (993 CXRs), qXRver 4.0 showed near-perfect sensitivity (99.7%) and NPV (99.3%) but modest specificity (67.4%). AI model by Anderson et al,22 reported comparable sensitivity of 90.8% and specificity of 88.7% on a publicly available dataset of 20,000 images only from adults thereby limiting its use only to adult population. Also, it does not classify lesions across tuberculous, non-tuberculous and healed lesions, especially post-TB sequelae which have been included in our study. In addition, our tool also generates a fully automated report like other AI tools26 and can be integrated into clinical workflows to assist radiologists, thus reducing reporting delays.

Our AI model offers several notable strengths. Its high-resolution segmentation of lung zones enhances precision in localising abnormalities, allowing for more accurate detection. By employing model specialisation, the system confidently identifies a diverse range of pathologies, including subtle lesions, frequently overlooked by radiologists. Robust post-processing layers further refine accuracy by filtering-out false positives using anatomical knowledge, such as recognising costochondral junctions, rotated films, and diaphragm variations. Additionally, an image quality assessment module filters out poor-quality scans, ensuring reliable performance even in peripheral settings with varied imaging equipment. Finally, the system incorporates explainable AI through heatmap overlays, providing clinicians with transparent and interpretable evidence of model predictions, which facilitates trust and supports clinical adoption. A key strength of the system lies in its ensemble-based architecture, which integrates six specialised AI models, each optimised to detect specific abnormalities such as pleural effusion, infiltrates, small nodules, cardiomegaly, and TB-related changes. The ensemble design allows the system to overcome inherent limitations of single-model approaches by leveraging complementary strengths of different networks. This enhanced sensitivity to detect abnormalities, reduced false negatives increasing specificity. Additionally, a dedicated segmentation model isolates the lung region, enabling more focused analysis reducing interference from anatomical structures.

The AI tool is designed for flexible deployment across varied settings. In online mode, using an NVIDIA RTX 4090 GPU, the model processes images in ⁓30 sec each. In offline mode, on a standard Intel i3 laptop, inference takes about 50 sec per image. This dual-mode capability ensures that the tool can be effectively used in both urban hospitals and rural, resource-constrained environments without relying on cloud connectivity.

Despite being unique in view of its training on a vast diverse dataset, the tool has some limitations, as performance of tool for detecting abnormality in X-rays with antero-posterior view may be limited. Although trained, AI may not be able to differentiate between active and healed lesions; hence, needs clinical correlation. Sensitivity may be low for very early parenchymal lesions, retrocardiac and supraclavicular lesions. The tool has low confidence in detecting early-stage disease. All these has been added as disclaimer in the report by AI tool.

In addition, we would like to highlight here that the performance of the tool was validated in comparison to Radiologists reporting which itself varied across 18 sites due to known inter-radiologists variations and might have resulted in lower specificity during independent validation as compared to a dataset (case 1 and 2) wherein the radiologists annotations were checked and verified by a group of experts. Therefore, validations against reporting by one radiologist would have been better for comparison.

To conclude, this indigenous AI-based chest abnormality detection tool (DeepCXR v1.1) has demonstrated high accuracy for detecting abnormalities in Chest X-rays and can be used as screening tool for large-scale screening specially in resource constrained settings. Its use for mass screening specially for TB will help in faster identification of abnormal CXRs which can be then subjected to confirmatory tests leading to early diagnosis thus reducing need for confirmatory tests in entire population in high TB endemic areas and reducing transmission by early initiation of treatment, helping National programme to achieve Goal of TB Elimination.

Acknowledgment

Authors acknowledge the following: ICMR-Expert Committee on TB Diagnostics and Monitoring Committee: Dr. V. M. Katoch (Former Secy, DHR & DG, ICMR; Chairman); Dr. D. Behera (Former Dean, PGIMER, Chandigarh; Co-Chair); Dr Chakravarthy Bhagvati (Professor, School of Computer and Information Sciences, University of Hyderabad), Dr. Rohit Sarin (Former Director, NITRD and Technical Advisor NTEP, Govt. of India); Dr. A.P. Dubey (Former Director Prof. and HOD MAMC); Dr. Sarman Singh (Former Director AIIMS Bhopal); Dr. R.M Pandey (Ex-HOD, Biostatistics, AIIMS and Dr. A.S. Paintal, Chair of ICMR); Dr. Camilla Rodriques (Sr. Consultant, Hinduja Hospital, Mumbai); Dr. R.P. Joshi (former DDG-TB, NTEP, Delhi); Dr. Urvashi B. Singh (DDG-TB, CTD, MoHFW); Dr. Ravi. P. Singh (Sr. Scientist E1, CSIR, Pune); Dr. Prabha Desikan (Former Director BMHRC, Bhopal); Dr. Ahilandeshwari Prasad (Prof. Radiology, Dr RML Hospital); Dr. Munish Guleria (Prof. Dept of Radiology, RMLH) and Dr.Tavpritesh Sethi (IIT, Delhi) for their constant support throughout the study. Dr R.M. Pandey (A.S. Paintal Chair, ICMR Hqrs.) for independent evaluation of the performance of the AI tool. Authors wish to thank Dr Rajiv Bahl (Secretary DHR and DG, ICMR) for his technical inputs and support and also wish to acknowledge Director, IPR and Dean R&D and administration for their inputs. Ms Meenakshi Bakshi for guidance on quality management.

Author contributions

AA, AS, MS: Development of the AI tool; MSC: Development of the portal for data collection, study design, data cleaning, writing of the report; RMY, AK, SG: Study design, data collection, data cleaning, study coordination; KA, VV, AT, MVB, PM, HM, PVB, SB, AD, DP, CP, SP, JT, PN, AA, IS, SSM, AG, RR, SSM, AM, CRC, MP, RM, BD, AS, RKM, KC, DC, PKS, JKP: Enrolment of participants in the study, data collection, data cleaning at their respective sites; AT, VV, KA, MP: Annotation of the study as central annotation team; MS: Conceptualised the study, study design, wrote the study protocol, contributed overall coordination, data curation and report writing, had full access to all the data in the study, manuscript writing, edited the article throughout all stages and had final responsibility for the decision to submit for publication. All authors edited the article throughout all stages and approved the final printed version of manuscript.

Financial support and sponsorship

This study was funded by India TB Research Consortium, Indian Council of Medical Research, New Delhi (Grant number 5/8/5/37/ITRC/AI tool-Chest X-rays/2021/ECD-I).

Conflicts of Interest

Indian Council of Medical Research is the sponsor of the study and was involved in conceptualization, study design, coordination, monitoring and convening meetings to ensure compliance to protocol and timely completion, however, had no role in participant enrolment and data acquisition. The authors declare that they have no competing interests; no financial relationships with any organizations that might have an interest in the submitted work; no other relationships or activities that could appear to have influenced the submitted work.

Use of Artificial Intelligence (AI)-Assisted Technology for manuscript preparation

The authors confirm that there was no use of AI-assisted technology for assisting in the writing of the manuscript and no images were manipulated using AI.

References

  1. World Health Organization. Global tuberculosis report. 2023. Available from: https://www.who.int/publications/i/item/9789240089695, accessed on September 25, 2025.
  2. , , , , , , et al. Tuberculosis. Nat Rev Dis Primers.. 2016;2:16076.
    [CrossRef] [PubMed] [Google Scholar]
  3. , , , . Assessing and comparing chest radiograph interpretation in the Department of Internal Medicine at the University of the Witwatersrand Medical School, according to seniority. Afr J Thorac Crit Care Med.. 2023;29 10.7196/AJTCCM.2023.v29i1.265
    [CrossRef] [Google Scholar]
  4. . Training for rural radiology and imaging in Sub-Saharan Africa: Addressing the mismatch between services and population. J Clin Imaging Sci.. 2012;2:37.
    [CrossRef] [PubMed] [PubMed Central] [Google Scholar]
  5. , ed. Toman’s tuberculosis : case detection, treatment, and monitoring : Questions and answers (2nd ed). Geneva: World Health Organization; .
  6. , . Deep learning at chest radiography: Automated classification of pulmonary tuberculosis by using convolutional neural networks. Radiology.. 2017;284:574-82.
    [CrossRef] [PubMed] [Google Scholar]
  7. , , , , . Explainable artificial intelligence in deep learning–based detection of aortic elongation on chest X-ray images. Eur Heart J - Digital Health.. 2024;5:524-3.
    [Google Scholar]
  8. , , , , , , et al. Diagnostic effect of artificial intelligence solution for referable thoracic abnormalities on chest radiography: A multicenter respiratory outpatient diagnostic cohort study. Eur Radiol.. 2022;32:3469-7.
    [CrossRef] [PubMed] [PubMed Central] [Google Scholar]
  9. , , , , , , et al. Performance of a deep-learning algorithm for referable thoracic abnormalities on chest radiographs: A multicenter study of a health screening cohort. PLoS One.. 2021;16:e0246472.
    [CrossRef] [PubMed] [PubMed Central] [Google Scholar]
  10. , . Application of artificial intelligence in chest radiograph interpretation: Is the time ripe yet? J Clin Sci. 2024;13:157-8.
    [CrossRef] [Google Scholar]
  11. , , , , . CheXclusion: Fairness gaps in deep chest X-ray classifiers. Pac Symp Biocomput.. 2021;26:232-43.
    [CrossRef] [PubMed] [Google Scholar]
  12. , , , , , , et al. The application of artificial intelligence and radiomics in lung cancer. Precis Clin Med.. 2020;3:214-27.
    [CrossRef] [PubMed] [PubMed Central] [Google Scholar]
  13. Central Tuberculosis Division. Ministry of Health and Family Welfare, Government of India. National Strategic Plan for Tuberculosis Elimination 2017–2025. Available from: https://tbcindia.mohfw.gov.in/wp-content/uploads/2023/05/National-Strategic-Plan-2017-25.pdf, accessed on September 3, 2025.
  14. Central Tuberculosis Division. Ministry of Health and Family Welfare, Government of India. Indian Council of Medical Research, ICMR-National National Institute for Research in Tuberculosis. National TB Prevalence Survey in India (2019 – 2021): Summary report. Available from: https://tbcindia.mohfw.gov.in/wp-content/uploads/2023/05/25032022161020NATBPSReport.pdf, accessed on September 3, 2025.
  15. , , , , . Profile of respiratory problems in patients presenting to a referral pulmonary clinic. Lung India.. 2008;25:4-7.
    [CrossRef] [PubMed] [Google Scholar]
  16. Wada, K. Image polygonal annotation with python (Polygon, Rectangle, Circle, Line, Point and Image-Level Flag Annotation). 2018. Available from: https://github.com/wkentaro/labelme, accessed on September 25, 2025.
  17. , . Complete spectrum of X-ray images from all chest conditions crucial for training to develop robust artificial intelligence model for TB triage. Indian J Tuberc 2026 doi: 10.1016/j.ijtb.2026.01.002
    [CrossRef] [Google Scholar]
  18. , , , , , , et al. Chest x-ray automated triage: A semiologic approach designed for clinical implementation, exploiting different types of labels through a combination of four deep learning architectures. Comput Methods Programs Biomed.. 2021;206:106130.
    [CrossRef] [PubMed] [Google Scholar]
  19. , , , , , , et al. Development and validation of a deep learning-based automatic detection algorithm for active pulmonary tuberculosis on chest radiographs. Clin Infect Dis.. 2019;69:739-47.
    [CrossRef] [PubMed] [PubMed Central] [Google Scholar]
  20. , , , , , , et al. Development and validation of deep learning-based automatic detection algorithm for malignant pulmonary nodules on chest radiographs. Radiology.. 2019;290:218-2.
    [CrossRef] [PubMed] [Google Scholar]
  21. , , , , , , et al. Deep learning improves physician accuracy in the comprehensive detection of abnormalities on chest X-rays. Sci Rep.. 2024;14:25151.
    [CrossRef] [PubMed] [PubMed Central] [Google Scholar]
  22. , , , , , . ChestX-ray8: Hospital-scale chest X-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. Proc IEEE Conf Comput Vis Pattern Recognit. 2017:3462-71.
    [CrossRef] [Google Scholar]
  23. , , , , , , et al. Post-deployment performance of a deep learning algorithm for normal and abnormal chest X-ray classification: A study at visa screening centers in the United Arab Emirates. Eur J Radiol Open.. 2024;13:100606.
    [CrossRef] [PubMed] [PubMed Central] [Google Scholar]
  24. , , , , , , et al. Using artificial intelligence to stratify normal versus abnormal chest X-rays: External validation of a deep learning algorithm at East Kent Hospitals university NHS foundation trust. Diagnostics (Basel).. 2023;13:3408.
    [CrossRef] [PubMed] [PubMed Central] [Google Scholar]
  25. , , , , . Automatic report generation for chest X-ray images via adversarial reinforcement learning. IEEE Access.. 2021;9:21236-50.
    [CrossRef] [Google Scholar]
Show Sections
Scroll to Top