Skip to main content

Deep Learning at Chest Radiography: Automated Classification of Pulmonary Tuberculosis by Using Convolutional Neural Networks

In this paper, the authors explore the use of Deep Convolutional Neural Networls (DCNN) in classifying Tuberculosis (TB) in chest radiographs. One of the advantages of deep learning is its ability to excel with high dimensional datasets, such as images, which can be represented at multiple levels. 

Dataset

Four deidentified HIPAA-compliant datasets were used in this study that were exempted from review by the institutional review board, which consisted of 1007 posteroanterior chest radiographs.

DCNN Models and Training

AlexNet and GoogLeNet models, including pre-trained (on ImageNet from Caffe Model Zoo) and untrained models were used in the study. It was found that the AUCs of the pretrained networks were greater. The following solver parameters were used for training: 120 epochs; base learning rate for untrained models and for pretrained models, 0.01 and 0.001, respectively with stochastic gradient descent.  Both of the DCNNs in this studied used dropout or model regularization strategies to help overcome overfitting.

Data Augmentation

The following data augmentation techniques further increased the performance

  1. Random cropping of 227x227
  2. Mean subtraction and mirror images
  3. Rotation of 90, 180 and 270.
  4. Contrast Limited Adaptive Histogram Equalization processing

Ensembling


The ensembling technique was used to increase the AUC even further. Ensembles were performed by taking different weighted averages of the probability scores generated by the classifiers 
The best performing ensemble model had an AUC of 0.99. Refer below the table borrowed from the paper for complete results

The sensitivity of pre-trained AlexNet was 92.0% and the specificity was 94.7%. The sensitivity of pre-trained GoogLeNet was 92.0% and the specificity was 98.7%. The sensitivity of the ensemble was 97.3% and the specificity was 94.7%. 

Radiologist-augmented approach

This is were the paper takes turn to beyond the realms of deep learning, were they use a certain human to classify the images were the models fail.

For cases where the AlexNet and GoogLeNet classifiers had disagreement, an independent board-certified cardiothoracic radiologist (B.S., with 18 years of experience) blindly interpreted the images as either having manifestations of TB or as normal. This resulted in a sensitivity of 97.3% and a specificity of 100%.

Comments

Popular Posts

BLIP: Bootstrapping Language-Image Pretraining for Unified Vision-Language Understanding

BLIP is a new vision-language model proposed by Microsoft Research Asia in 2022. It introduces a bootstrapping method to learn from noisy image-text pairs scraped from the web. The BLIP Framework BLIP consists of three key components: MED  - A multimodal encoder-decoder model that can encode images, text, and generate image-grounded text. Captioner  - Fine-tuned on COCO to generate captions for web images. Filter  - Fine-tuned on COCO to filter noisy image-text pairs. The pretraining process follows these steps: Collect noisy image-text pairs from the web. Pretrain MED on this data. Finetune captioner and filter on the COCO dataset. Use captioner to generate new captions for web images. Filter noisy pairs using the filter model. Repeat the process by pretraining on a cleaned dataset. This bootstrapping allows BLIP to learn from web-scale noisy data in a self-supervised manner. Innovations in BLIP Some interesting aspects of BLIP: Combines encoder-decoder capability in one...

Learning to Read Chest X-Rays: Recurrent Neural Feedback Model for Automated Image Annotation

In this paper , the authors present a deep learning model to detect disease from chest x-ray images. A convolutional neural network (CNN) is trained to detect the disease names. Recurrent neural networks (RNNs) are then trained to describe the contexts of a detected disease, based on the deep CNN features. CNN Models used and Dataset CNNs encode input images effectively. In this paper, the authors experiment with a Network in Network (NIN) model and GoogLeNet model. The dataset contains 3,955 radiology reports and 7,470 associated chest x-rays. 71% of the dataset accounts for normal cases (no disease). The data set was balanced by augmenting training images by randomly cropping 224x224 images from the original 256x256 size image. Adaptability of Transfer learning Since this boils down to a classification problem on a small dataset, transfer learning is a technique that comes to our mind. The authors experimented this with ImageNet trained models. ImageNet trained CN...

Chest X-Ray Analysis of Tuberculosis by Deep Learning with Segmentation and Augmentation

In this paper , the authors explore the efficiency of lung segmentation, lossless and lossy data augmentation in  computer-aided diagnosis (CADx) of tuberculosis using deep convolutional neural networks applied to a small and not well-balanced Chest X-ray (CXR) dataset. Dataset Shenzhen Hospital (SH) dataset of CXR images was acquired from Shenzhen No. 3 People's Hospital in Shenzhen, China. It contains normal and abnormal CXR images with marks of tuberculosis. Methodology Based on previous literature, attempts to perform training for such small CXR datasets without any pre-processing failed to see good results. So the authors attempted segmenting the lung images before being inputted to the model. This gave demonstrated a more successful training and an increase in prediction accuracy. To perform lung segmentation, i.e. to cut the left and right lung fields from the lung parts in standard CXRs, manually prepared masks were used. The dataset was split into 8:1:1...