A learnable distillation approach for model-agnostic explainability with multimodal applications (Record no. 428394)
[ view plain ]
000 -LEADER | |
---|---|
fixed length control field | 04211nam a22003017a 4500 |
008 - FIXED-LENGTH DATA ELEMENTS--GENERAL INFORMATION | |
fixed length control field | 230526b |||||||| |||| 00| 0 eng d |
041 ## - LANGUAGE CODE | |
Language code of text/sound track or separate title | en |
082 ## - DEWEY DECIMAL CLASSIFICATION NUMBER | |
Classification number | 621 |
Item number | DEB |
100 ## - MAIN ENTRY--PERSONAL NAME | |
Personal name | Bhattacharya, Debarpan |
245 ## - TITLE STATEMENT | |
Title | A learnable distillation approach for model-agnostic explainability with multimodal applications |
260 ## - PUBLICATION, DISTRIBUTION, ETC. (IMPRINT) | |
Place of publication, distribution, etc | Bengaluru : |
Name of publisher, distributor, etc | IISc , |
Date of publication, distribution, etc | 2023 . |
300 ## - PHYSICAL DESCRIPTION | |
Extent | xv, 66p. |
Other physical details | col. ill. |
Dimensions | 29.1cm *20.5 cm |
Accompanying material | e-Thesis |
Size of unit | 35.4Mb |
500 ## - GENERAL NOTE | |
General note | Include bibliographical references and index. |
502 ## - DISSERTATION NOTE | |
Dissertation note | MTech (Res); 2023; Electrical engineering |
520 ## - SUMMARY, ETC. | |
Summary, etc | Deep neural networks are the most widely used examples of sophisticated mapping functions from feature space to class labels. In the recent years, several high impact decisions in domains such as finance, healthcare, law and autonomous driving, are made with deep models. In these tasks, the model decisions lack interpretability, and pose difficulties in making the models accountable. Hence, there is a strong demand for developing explainable approaches which can elicit how the deep neural architecture, despite the astounding performance improvements observed in all fields, including computer vision, natural language processing, generates the output decisions. The current frameworks for explainability of deep models are based on gradients (eg. GradCAM, guided-gradCAM, Integrated gradients etc) or based on locally linear assumptions (eg. LIME). Some of these approaches require the knowledge of the deep model architecture, which may be restrictive in many applications. Further, most of the prior works in the literature highlight the results on a set of small number of examples to illustrate the performance of these XAI methods, often lacking statistical evaluation. This thesis proposes a new approach for explainability based on mask estimation approaches, called the Distillation Approach for Model-agnostic Explainability (DAME). The DAME is a saliency-based explainability model that is post-hoc, model-agnostic (applicable to any black box architecture), and requires only query access to black box. The DAME is a student-teacher modeling approach, where the teacher model is the original model for which the explainability is sought, while the student model is the mask estimation model. The input sample is augmented with various data augmentation techniques to produce numerous samples in the immediate vicinity of the input. Using these samples, the mask estimation model is learnt to generate the saliency map of the input sample for predicting the labels. A distillation loss is used to train the DAME model, and the student model tries to locally approximate the original model. Once the DAME model is trained, the DAME generates a region of the input (either in space or in time domain for images and audio samples, respectively) that best explains the model predictions. We also propose an evaluation framework, for both image and audio tasks, where the XAI models are evaluated in a statistical framework on a set of held-out of examples with the Intersection-over-Union (IoU) metric. We have validated the DAME model for vision, audio and biomedical tasks. Firstly, we deploy the DAME for explaining a ResNet-50 classifier pre-trained on ImageNet dataset for the object recognition task. Secondly, we explain the predictions made by ResNet-50 classifier fine-tuned on Environmental Sound Classification (ESC-10) dataset for the audio event classification task. Finally, we validate the DAME model on the COVID-19 classification task using cough audio recordings. In these tasks, the DAME model is shown to outperform existing benchmarks for explainable modeling. The thesis concludes with a discussion on the limitations of the DAME approach along with the potential future directions. |
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM | |
Topical term or geographic name as entry element | Post-hoc explainability |
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM | |
Topical term or geographic name as entry element | Model-agnostic explainability |
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM | |
Topical term or geographic name as entry element | Interpretability |
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM | |
Topical term or geographic name as entry element | Saliency |
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM | |
Topical term or geographic name as entry element | Medical XAI |
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM | |
Topical term or geographic name as entry element | Artificial intelligence |
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM | |
Topical term or geographic name as entry element | Distillation approach |
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM | |
Topical term or geographic name as entry element | Model-agnostic explainability |
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM | |
Topical term or geographic name as entry element | COVID-19 |
700 ## - ADDED ENTRY--PERSONAL NAME | |
Personal name | Ganapathy, Sriram advised |
856 ## - ELECTRONIC LOCATION AND ACCESS | |
Uniform Resource Identifier | https://etd.iisc.ac.in/handle/2005/6108 |
942 ## - ADDED ENTRY ELEMENTS (KOHA) | |
Koha item type | Thesis |
No items available.