Show simple item record

dc.contributor.advisorMichael, Kampffmeyer
dc.contributor.authorChakraborty, Rwiddhi
dc.date.accessioned2024-12-02T09:09:58Z
dc.date.available2024-12-02T09:09:58Z
dc.date.issued2024-12-13
dc.description.abstractDeep Learning in modern Artificial Intelligence (AI) has witnessed unprecedented success on a variety of domains over the past decade, ranging from computer vision to natural language reasoning tasks. This success is owed primarily to the availability of large, annotated datasets, the existence of powerful mathematical models, and the mechanism to train large models on such data with advanced resources of compute. However, this success has led to increased scrutiny on the failure points of models trained on suspect data. Issues such as model and data bias, reliance on spurious correlations, and poor generalization capability on challenging test data, to name a few, have surfaced in the research community. As a result, it seems imperative to diagnose such systems for generalization performance on challenging test data, and uncovering potential biases hidden in datasets. In this thesis, we address these key challenges through the following directions: first, in the generalization capabilities with limited labeled data - few-shot learning, semi-supervised learning, and unsupervised learning. Second, towards bias discovery in existing models and datasets, particularly in unsupervised group robust learning, and debiased synthetic data generation. Our two broad directions are encapsulated by a common challenge: the paucity of labeled data, since manually annotating large datasets is a time consuming and expensive process for humans. This motivation is relevant today due to the exponential growth in the sizes of models and datasets in use. It is becoming more and more intractable for humans to annotate billions of data points, leading to large benchmark datasets that are not well calibrated with human expectations on fairness. These issues, if left unchecked, are inevitably exacerbated when models train on such datasets. We consider these two directions, i.e. model generalization with limited labels, and the existence of biased data, to be two sides of the same coin, and thus coin the framework encapsulating such research as Model and Data Diagnosis. This work proposes novel contributions in few-shot learning, semi-supervised learning, unsupervised learning, and in data diagnosis and debiasing techniques. Further, we show that model and data diagnosis do not necessarily exist as disparate entities, and can be viewed in a co-dependent context. Finally, this thesis hopes to amplify the scrutiny surrounding model capabilities, however impressive, trained on datasets, however vast.en_US
dc.description.doctoraltypeph.d.en_US
dc.description.popularabstractArtificial Intelligence (AI) has significantly impacted the way humans live and interact with each other in the modern world. How reliable are modern AI models and datasets? What biases do they contain? How well do they perform on challenging scenarios? How can we describe their behaviour better with respect to human intuition? These are the concepts of model and data diagnosis that we investigate in this thesis.en_US
dc.identifier.isbn978-82-8236-602-1 (trykk)
dc.identifier.isbn978-82-8236-603-8 (PdF)
dc.identifier.urihttps://hdl.handle.net/10037/35867
dc.language.isoengen_US
dc.publisherUiT Norges arktiske universiteten_US
dc.publisherUiT The Arctic University of Norwayen_US
dc.relation.haspart<p>Paper I: Chakraborty, R., Sletten, A. & Kampffmeyer, M. (2024). ExMap: Leveraging Explainability Heatmaps for Unsupervised Group Robustness to Spurious Correlations. <i>2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>, Seattle, WA, USA, 12017-12026. Also available at <a href=http://dx.doi.org/10.1109/CVPR52733.2024.01142>http://dx.doi.org/10.1109/CVPR52733.2024.01142</a>. <p>Paper II: Trosten, D.J., Chakraborty, R., Løkse, S., Wickstrøm, K., Jenssen, R. & Kampffmeyer, M.C. (2023). Hubs and Hyperspheres: Reducing Hubness and Improving Transductive Few-shot Learning with Hyperspherical Embeddings. <i>2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>, Vancouver, BC, Canada, 7527-7536. Also available at <a href=https://doi.org/10.1109/CVPR52729.2023.00727>https://doi.org/10.1109/CVPR52729.2023.00727</a>. Accepted manuscript version available in Munin at <a href=https://hdl.handle.net/10037/32935>https://hdl.handle.net/10037/32935</a>. <p>Paper III: Chakraborty, R., Wang, O., Gao, J., Zhang, C., Zheng, R. & de la Torre, F. Visual Data Diagnosis and Debiasing with Concept Graphs. (Manuscript under review). Also available on arXiv at <a href=https://doi.org/10.48550/arXiv.2409.18055>https://doi.org/10.48550/arXiv.2409.18055</a>. <p>Paper IV: Chakraborty, R., Ricaud,B., Jenssen, R. & Kampffmeyer, M. On Disentangled Representations and the Oversmoothing Problem in Graph Convolutional Networks. (Manuscript).en_US
dc.rights.accessRightsopenAccessen_US
dc.rights.holderCopyright 2024 The Author(s)
dc.rights.urihttps://creativecommons.org/licenses/by-nc-sa/4.0en_US
dc.rightsAttribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0)en_US
dc.titleModel and Data Diagnosis under Limited Supervision in Modern AIen_US
dc.typeDoctoral thesisen_US
dc.typeDoktorgradsavhandlingen_US


File(s) in this item

Thumbnail
Thumbnail

This item appears in the following collection(s)

Show simple item record

Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0)
Except where otherwise noted, this item's license is described as Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0)