
Select an Action

A Study of Interpretability Mechanisms for Deep Networks
Title:
A Study of Interpretability Mechanisms for Deep Networks
Author:
Kokate, Apurva, author. (orcid)0000-0003-2353-4171
ISBN:
9780438074392
Personal Author:
Physical Description:
1 electronic resource (104 pages)
General Note:
Source: Masters Abstracts International, Volume: 57-06M(E).
Includes supplementary digital materials.
Advisors: Soumik Sarkar; Jin Tian Committee members: Forrest S. Bao; Chinmay Hegde.
Abstract:
Deep neural networks are traditionally considered to be "black-box" models where it is generally difficult to interpret a certain decision made by such models given a test instance. However, as deep learning is increasingly becoming the tool of choice in making many safety-critical and time-critical decisions such as perception for self-driving cars, the machine learning community has been extremely interested recently to build interpretation mechanisms for these so called black box deep learning models primarily to build users' trust with the models. Many such mechanisms have been developed to explain behavior of deep models such as convolutional neural networks (CNNs) and provide visual interpretations of their classification decisions. However, there is still no consensus in the community about the specific goals and performance metrics for the interpretability mechanisms. In this thesis, we review the recent literature to arrive at a formal definition for the "Interpretability-problem" for CNNs with the help of different axioms. We observe that many recently proposed mechanisms do not adhere to the axioms of interpretability and hence not quite robust in performance. In this context, we propose a framework to test the interpretation algorithms under model perturbation and data perturbation. This framework tests the "sensitivity" of the algorithms and helps in evaluating "implementation invariance", which are desired characteristics for any interpretability mechanism. We demonstrate our framework using two well-known algorithms namely "Saliency Maps" and "Grad-CAM" and introduce a new interpretability technique called "Forward-Backward Interpretability algorithm" that provides a systematic framework for visualizing information flow in deep networks. Finally, we also present visualization and interpretability results for an impactful scientific application involving microstructure-property mapping in material science.
Local Note:
School code: 0097
Added Corporate Author:
Available:*
Shelf Number | Item Barcode | Shelf Location | Status |
|---|---|---|---|
| XX(690998.1) | 690998-1001 | Proquest E-Thesis Collection | Searching... |
On Order
Select a list
Make this your default list.
The following items were successfully added.
There was an error while adding the following items. Please try again.
:
Select An Item
Data usage warning: You will receive one text message for each title you selected.
Standard text messaging rates apply.


