dc.description.abstract |
Domain shift is a significant challenge for deep learning-based medical image analysis, particularly in the case of thoracic disease classification from chest X-rays (CXRs). Over the last decade, many domain adaptation (DA) and domain-generalization (DG) approaches for thoracic disease classification have been proposed. However, these methods do not explicitly regularize the content and style characteristics of the extracted domain-invariant features. Recent findings have shown that deep-learning models display a strong bias toward styles (i.e., uninformative textures) rather than content, in stark contrast to the human vision system. Therefore, domain-agnostic models for pathology diagnosis from CXR images should extract domain-invariant features that are style-invariant and content-biased. In this thesis, we propose a style randomization module (SRM) at the image level that randomly samples the style statistics parameters from a set constructed by the possible value range of a CXR image to create a more diversified augmented dataset compared to previous methods where available training data is utilized for sampling style parameters. We also employ an SRM at the feature level that uses learnable parameters as style embeddings to manipulate style while keeping the content intact. The two SRMs work together hierarchically to create rich, diversified, style-perturbed features on the fly during training. In addition, we utilize the Frobenius Norm loss between the global semantic features and Kullback-Leibler divergence loss between the predictive distributions, respectively, of two versions of the same CXR image (original CXR image and style-perturbed CXR image) to tweak the framework’s sensitivity toward content markers for accurate predictions. Extensive experiments with large-scale thoracic disease datasets demonstrate that our proposed pipeline is more robust in the presence of domain shift and achieves state-of-the-art performance. Our proposed method, trained on CheXpert and MIMIC-CXR datasets, achieves a mean percentage AUC score of 77.21±0.30, 87.74±0.48, 82.40±0.13 on the unseen domain test datasets, named BRAX, VinDr-CXR, and NIH chest X-ray14 respectively, compared to 75.78±0.45, 87.24±0.57, 82.16±0.21 from the former best models on five-fold cross-validation with statistically significant results for the thoracic disease classification task. |
en_US |