Eylem Seç
A Study of Interpretability Mechanisms for Deep Networks
Başlık:
A Study of Interpretability Mechanisms for Deep Networks
Yazar:
Kokate, Apurva, author. (orcid)0000-0003-2353-4171
ISBN:
9780438074392
Yazar Ek Girişi:
Fiziksel Tanımlama:
1 electronic resource (104 pages)
Genel Not:
Source: Masters Abstracts International, Volume: 57-06M(E).
Includes supplementary digital materials.
Advisors: Soumik Sarkar; Jin Tian Committee members: Forrest S. Bao; Chinmay Hegde.
Özet:
Deep neural networks are traditionally considered to be "black-box" models where it is generally difficult to interpret a certain decision made by such models given a test instance. However, as deep learning is increasingly becoming the tool of choice in making many safety-critical and time-critical decisions such as perception for self-driving cars, the machine learning community has been extremely interested recently to build interpretation mechanisms for these so called black box deep learning models primarily to build users' trust with the models. Many such mechanisms have been developed to explain behavior of deep models such as convolutional neural networks (CNNs) and provide visual interpretations of their classification decisions. However, there is still no consensus in the community about the specific goals and performance metrics for the interpretability mechanisms. In this thesis, we review the recent literature to arrive at a formal definition for the "Interpretability-problem" for CNNs with the help of different axioms. We observe that many recently proposed mechanisms do not adhere to the axioms of interpretability and hence not quite robust in performance. In this context, we propose a framework to test the interpretation algorithms under model perturbation and data perturbation. This framework tests the "sensitivity" of the algorithms and helps in evaluating "implementation invariance", which are desired characteristics for any interpretability mechanism. We demonstrate our framework using two well-known algorithms namely "Saliency Maps" and "Grad-CAM" and introduce a new interpretability technique called "Forward-Backward Interpretability algorithm" that provides a systematic framework for visualizing information flow in deep networks. Finally, we also present visualization and interpretability results for an impactful scientific application involving microstructure-property mapping in material science.
Notlar:
School code: 0097
Tüzel Kişi Ek Girişi:
Mevcut:*
Yer Numarası | Demirbaş Numarası | Shelf Location | Lokasyon / Statüsü / İade Tarihi |
---|---|---|---|
XX(690998.1) | 690998-1001 | Proquest E-Tez Koleksiyonu | Arıyor... |
On Order
Liste seç
Bunu varsayılan liste yap.
Öğeler başarıyla eklendi
Öğeler eklenirken hata oldu. Lütfen tekrar deneyiniz.
:
Select An Item
Data usage warning: You will receive one text message for each title you selected.
Standard text messaging rates apply.