Learning from Imperfect Annotations: Robust and Efficient Deep Learning for Medical Imaging Analysis
Supervisory Team
Introduction
This project addresses the critical challenge of imperfect annotations in medical imaging datasets, including scarce labels, weak supervision, and noisy or incorrect annotations. These imperfections significantly limit the development and deployment of robust and generalisable AI models in clinical practice. This PhD project aims to explore novel deep learning techniques that can learn effectively under these imperfect annotation conditions, reducing the dependency on exhaustive expert labelling and improving model trustworthiness. The project will also investigate how these techniques can generalise across different medical imaging modalities and clinical tasks, ensuring broader applicability and real-world impact.
Aim
- To develop robust and annotation-efficient deep learning methods that can handle scarce, weak, and noisy labels in medical imaging datasets and generalise across diverse imaging domains.
Objectives
-
Investigate and categorise types of annotation imperfections (e.g., label noise, sparsity, uncertainty) common in medical imaging.
-
Develop deep learning models that are resilient to label noise and capable of learning from unreliable or conflicting annotations.
-
Explore annotation-efficient learning strategies such as self-supervised learning, one/few-shot learning, and active learning.
-
Ensure generalizability of the proposed methods across different medical imaging modalities (e.g., radiology, pathology, dermatology) and tasks (e.g., classification, segmentation, detection).
-
Collaborate with clinical experts to validate the practicality and reliability of the developed solutions.
Methodology
-
Noise-Robust Learning: noise modelling, loss correction, and confident learning.
-
Self-Supervised & Semi-Supervised Learning: contrastive learning, pretext tasks.
-
Few-Shot & One-Shot Learning: metric learning, prototype-based networks.
-
Active Learning: selecting informative samples for annotation.
-
Cross-Domain Validation: evaluating across modalities and tasks.
-
Evaluation: benchmarking on public and clinical datasets with expert validation.
Timeline
-
Year 1: Literature review, data acquisition and cleaning
-
Year 2: Model development and internal validation
-
Year 3: Clinical testing, dissemination, thesis completion
Expected Outcomes
-
Reduce annotation burden in medical imaging, lowering costs and time for dataset creation.
-
Enable scalable and accessible AI tools that can be deployed in real-world clinical settings.
-
Develop robust deep learning models resistant to noisy and limited data.
-
Accelerate safe AI deployment in healthcare.
-
Create generalisable models across multiple medical domains.
Contact
Dr. Eman Alajrami