Sammendrag
Deep Learning in modern Artificial Intelligence (AI) has witnessed unprecedented success on a variety of domains over the past decade, ranging from computer vision to natural language reasoning tasks. This success is owed primarily to the availability of large, annotated datasets, the existence of powerful mathematical models, and the mechanism to train large models on such data with advanced resources of compute.
However, this success has led to increased scrutiny on the failure points of models trained on suspect data. Issues such as model and data bias, reliance on spurious correlations, and poor generalization capability on challenging test data, to name a few, have surfaced in the research community.
As a result, it seems imperative to diagnose such systems for generalization performance on challenging test data, and uncovering potential biases hidden in datasets.
In this thesis, we address these key challenges through the following directions:
first, in the generalization capabilities with limited labeled data - few-shot learning, semi-supervised learning, and unsupervised learning.
Second, towards bias discovery in existing models and datasets, particularly in unsupervised group robust learning, and debiased synthetic data generation.
Our two broad directions are encapsulated by a common challenge: the paucity of labeled data, since manually annotating large datasets is a time consuming and expensive process for humans. This motivation is relevant today due to the exponential growth in the sizes of models and datasets in use. It is becoming more and more intractable for humans to
annotate billions of data points, leading to large benchmark datasets that are not well calibrated with human expectations on fairness. These issues, if left unchecked, are inevitably exacerbated when models train on such datasets.
We consider these two directions, i.e. model generalization with limited labels, and the existence of biased data, to be two sides of the same coin, and thus coin the framework encapsulating such research
as Model and Data Diagnosis. This work proposes novel contributions in few-shot learning, semi-supervised learning, unsupervised learning, and in data diagnosis and debiasing techniques. Further, we show that
model and data diagnosis do not necessarily exist as disparate entities, and can be viewed in a co-dependent context. Finally, this thesis hopes to amplify the scrutiny surrounding
model capabilities, however impressive, trained on datasets, however vast.
Har del(er)
Paper I: Chakraborty, R., Sletten, A. & Kampffmeyer, M. (2024). ExMap: Leveraging Explainability Heatmaps for Unsupervised Group Robustness to Spurious Correlations. 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 12017-12026. Also available at http://dx.doi.org/10.1109/CVPR52733.2024.01142.
Paper II: Trosten, D.J., Chakraborty, R., Løkse, S., Wickstrøm, K., Jenssen, R. & Kampffmeyer, M.C. (2023). Hubs and Hyperspheres: Reducing Hubness and Improving Transductive Few-shot Learning with Hyperspherical Embeddings. 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 7527-7536. Also available at https://doi.org/10.1109/CVPR52729.2023.00727. Accepted manuscript version available in Munin at https://hdl.handle.net/10037/32935.
Paper III: Chakraborty, R., Wang, O., Gao, J., Zhang, C., Zheng, R. & de la Torre, F. Visual Data Diagnosis and Debiasing with Concept Graphs. (Manuscript under review). Also available on arXiv at https://doi.org/10.48550/arXiv.2409.18055.
Paper IV: Chakraborty, R., Ricaud,B., Jenssen, R. & Kampffmeyer, M. On Disentangled Representations and the Oversmoothing Problem in Graph Convolutional Networks. (Manuscript).