← Back to Portfolio

🏆 Tuberculosis Cough Classification

2025 | Ongoing — AI-powered TB Detection from Cough Sounds
Paul G. Allen School of Computer Science & Engineering • Harborview Medical Center
TB Cough Classification
Machine Learning Healthcare AI Signal Processing

Project Overview

Developing an AI system to detect tuberculosis from cough sounds, making TB screening more accessible in resource-limited settings and helping healthcare workers identify cases earlier to save lives.

Detailed Introduction

Motivation. Tuberculosis (TB) remains one of the world's deadliest infectious diseases, with over 10 million new cases and 1.5 million deaths annually. Early detection is crucial for effective treatment and preventing transmission. However, traditional diagnostic methods like sputum microscopy and chest X-rays require specialized equipment and trained personnel, making them inaccessible in many resource-limited settings.

Key project objectives

Background: Global TB Challenge

Tuberculosis disproportionately affects low- and middle-income countries, where access to diagnostic tools is limited. The World Health Organization estimates that 3 million TB cases go undiagnosed each year, contributing to continued transmission and poor outcomes. Our project aims to address this gap by leveraging the ubiquity of mobile devices and the distinct acoustic patterns of TB-related coughs.

Technical approach

Our system uses deep learning techniques to analyze cough sound characteristics, including:

Data collection & privacy

  1. Ethical approval: IRB approval for audio data collection from TB patients and healthy controls.
  2. Data acquisition: High-quality audio recordings in controlled clinical environments.
  3. Privacy protection: De-identification of audio samples and secure data storage protocols.
  4. Quality control: Expert validation of TB diagnosis and audio quality assessment.

Model development strategy

We follow a systematic approach to model development:

  1. Baseline models: Start with traditional machine learning approaches (SVM, Random Forest) for comparison.
  2. Deep learning exploration: Implement CNN and RNN architectures optimized for audio classification.
  3. Ensemble methods: Combine multiple models to improve overall performance and robustness.
  4. Cross-validation: Rigorous evaluation using stratified k-fold cross-validation and holdout test sets.

Evaluation & validation

Our evaluation framework includes:

Team & roles

Core investigators

  • MS June Lee — Project Lead, Machine Learning & Signal Processing
  • PhD Alex Ching — Audio Processing & Model Architecture
  • PhD Cynthia Dong — Clinical Validation & Data Analysis
  • Prof. Shwetak N. Patel — Faculty Advisor, Ubiquitous Computing
  • MD David Horne — Clinical Partner, Harborview Medical Center

Collaborators

  • Clinical informatics & data governance
  • Audio engineering & signal processing
  • Mobile app development
  • Global health & implementation research

Implementation & deployment

Timeline & milestones

Months 0–3: Data collection setup, IRB approval, initial audio preprocessing pipeline.
Months 4–9: Model development, baseline performance evaluation, feature engineering optimization.
Months 10–15: Clinical validation, mobile app development, pilot testing in clinical settings.
Months 16+: Field deployment, performance monitoring, iterative improvements based on user feedback.

Expected outcomes

Challenges & mitigation

Global impact potential

This project has the potential to revolutionize TB screening in resource-limited settings by providing an accessible, cost-effective, and non-invasive screening tool. By leveraging the ubiquity of mobile devices, we can bring advanced diagnostic capabilities to underserved populations and contribute to global TB elimination efforts.

Contact & collaboration

If you're interested in collaborating, contributing data, or learning more about this project, please contact: