dc.contributor.advisor | Michael, Kampffmeyer | |
dc.contributor.author | Chakraborty, Rwiddhi | |
dc.date.accessioned | 2024-12-02T09:09:58Z | |
dc.date.available | 2024-12-02T09:09:58Z | |
dc.date.issued | 2024-12-13 | |
dc.description.abstract | Deep Learning in modern Artificial Intelligence (AI) has witnessed unprecedented success on a variety of domains over the past decade, ranging from computer vision to natural language reasoning tasks. This success is owed primarily to the availability of large, annotated datasets, the existence of powerful mathematical models, and the mechanism to train large models on such data with advanced resources of compute.
However, this success has led to increased scrutiny on the failure points of models trained on suspect data. Issues such as model and data bias, reliance on spurious correlations, and poor generalization capability on challenging test data, to name a few, have surfaced in the research community.
As a result, it seems imperative to diagnose such systems for generalization performance on challenging test data, and uncovering potential biases hidden in datasets.
In this thesis, we address these key challenges through the following directions:
first, in the generalization capabilities with limited labeled data - few-shot learning, semi-supervised learning, and unsupervised learning.
Second, towards bias discovery in existing models and datasets, particularly in unsupervised group robust learning, and debiased synthetic data generation.
Our two broad directions are encapsulated by a common challenge: the paucity of labeled data, since manually annotating large datasets is a time consuming and expensive process for humans. This motivation is relevant today due to the exponential growth in the sizes of models and datasets in use. It is becoming more and more intractable for humans to
annotate billions of data points, leading to large benchmark datasets that are not well calibrated with human expectations on fairness. These issues, if left unchecked, are inevitably exacerbated when models train on such datasets.
We consider these two directions, i.e. model generalization with limited labels, and the existence of biased data, to be two sides of the same coin, and thus coin the framework encapsulating such research
as Model and Data Diagnosis. This work proposes novel contributions in few-shot learning, semi-supervised learning, unsupervised learning, and in data diagnosis and debiasing techniques. Further, we show that
model and data diagnosis do not necessarily exist as disparate entities, and can be viewed in a co-dependent context. Finally, this thesis hopes to amplify the scrutiny surrounding
model capabilities, however impressive, trained on datasets, however vast. | en_US |
dc.description.doctoraltype | ph.d. | en_US |
dc.description.popularabstract | Artificial Intelligence (AI) has significantly impacted the way humans live and interact with each other in the modern world. How reliable are modern AI models and datasets? What biases do they contain? How well do they perform on challenging scenarios? How can we describe their behaviour better with respect to human intuition? These are the concepts of model and data diagnosis that we investigate in this thesis. | en_US |
dc.identifier.isbn | 978-82-8236-602-1 (trykk) | |
dc.identifier.isbn | 978-82-8236-603-8 (PdF) | |
dc.identifier.uri | https://hdl.handle.net/10037/35867 | |
dc.language.iso | eng | en_US |
dc.publisher | UiT Norges arktiske universitet | en_US |
dc.publisher | UiT The Arctic University of Norway | en_US |
dc.relation.haspart | <p>Paper I: Chakraborty, R., Sletten, A. & Kampffmeyer, M. (2024). ExMap: Leveraging Explainability Heatmaps for Unsupervised Group Robustness to Spurious Correlations. <i>2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>, Seattle, WA, USA, 12017-12026. Also available at <a href=http://dx.doi.org/10.1109/CVPR52733.2024.01142>http://dx.doi.org/10.1109/CVPR52733.2024.01142</a>.
<p>Paper II: Trosten, D.J., Chakraborty, R., Løkse, S., Wickstrøm, K., Jenssen, R. & Kampffmeyer, M.C. (2023). Hubs and Hyperspheres: Reducing Hubness and Improving Transductive Few-shot Learning with Hyperspherical Embeddings. <i>2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>, Vancouver, BC, Canada, 7527-7536. Also available at <a href=https://doi.org/10.1109/CVPR52729.2023.00727>https://doi.org/10.1109/CVPR52729.2023.00727</a>. Accepted manuscript version available in Munin at <a href=https://hdl.handle.net/10037/32935>https://hdl.handle.net/10037/32935</a>.
<p>Paper III: Chakraborty, R., Wang, O., Gao, J., Zhang, C., Zheng, R. & de la Torre, F. Visual Data Diagnosis and Debiasing with Concept Graphs. (Manuscript under review). Also available on arXiv at <a href=https://doi.org/10.48550/arXiv.2409.18055>https://doi.org/10.48550/arXiv.2409.18055</a>.
<p>Paper IV: Chakraborty, R., Ricaud,B., Jenssen, R. & Kampffmeyer, M. On Disentangled Representations and the Oversmoothing Problem in Graph Convolutional Networks. (Manuscript). | en_US |
dc.rights.accessRights | openAccess | en_US |
dc.rights.holder | Copyright 2024 The Author(s) | |
dc.rights.uri | https://creativecommons.org/licenses/by-nc-sa/4.0 | en_US |
dc.rights | Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0) | en_US |
dc.title | Model and Data Diagnosis under Limited Supervision in Modern AI | en_US |
dc.type | Doctoral thesis | en_US |
dc.type | Doktorgradsavhandling | en_US |