thesis defense - computer science: mohammed maaz inayatullah sibhai
please join the computer science department for the upcoming thesis defense:
presenter: mohammed maaz inayatullah sibhai
thesis title: uncertainty-guided transformer learning for trustworthy medical image classification
abstract: reliable medical image classification is fundamental for the safe use of deep learning in clinical decision support. the state-of-the-art deep learning models, such as medical vision transformers perform well in medical image segmentation. these models often present unreliable probability estimates and do not have built-in ways to explicitly handle uncertainty or interpretability. these issues become especially problematic when inputs are ambiguous or datasets are not uniformly distributed, which are common in real-world clinical settings.
this study contributes in extending the architecture of medical transformer (medformer), a hierarchical medical vision transformer guided by uncertainty and prototypes, to improve trustworthiness without reducing feature representation. the model uses per-token evidential uncertainty estimation via a dirichlet approach, enabling explicit measurement of uncertainty and spatial localization. instead of just acting as a post-hoc diagnostic tool, uncertainty actively guides feature routing and refinement during training, decreasing unreliable updates in uncertain regions. additionally, prototype-based learning is incorporated to maintain a structured, class-specific geometry in the embedding space and support similarity-based, interpretable decisions grounded in visual patterns.
the proposed model has been tested on various medical imaging types, including mammography, breast ultrasound, brain tumor mri, and breast histopathology, providing a thorough testing across different dataset contexts. experiments show that, while classification accuracy improvements vary across datasets, the method reliably improves calibration, reduces overconfidence, and enhances selective prediction compared to the baseline medformer. these results indicate that integrating uncertainty estimation and prototype-based regularisation into transformer-based representation learning can greatly boost the reliability and explainability of medical image classifiers, supporting the development of trustworthy ai systems for clinical use.
experimental results show that the proposed model improves calibration, reduces overconfidence, and enhances selective prediction across all evaluated datasets compared to the baseline medformer. the accuracies are reported in the selected benchmark datasets, with larger improvements in modalities with clearer visual cues and more modest changes in mammography due to inherent ambiguity. overall, uncertainty-guided routing and prototype-based learning improve trustworthiness without sacrificing discriminative performance.
committee members:
dr. saad b. ahmed (supervisor, committee chair), dr. abedalrhman alkhateeb (co-supervisor), dr. garima bajwa, dr. farhan ghaffar (electrical & computer engineering)
please contact grad.compsci@lakeheadu.ca for the zoom link. everyone is welcome.
