Shao, Xiaoting (2022)
Explaining and Interactively Debugging Deep Models.
Technische Universität Darmstadt
doi: 10.26083/tuprints-00021868
Ph.D. Thesis, Primary publication, Publisher's Version
Text
Dissertation_signed.pdf Copyright Information: CC BY-SA 4.0 International - Creative Commons, Attribution ShareAlike. Download (30MB) |
Item Type: | Ph.D. Thesis | ||||
---|---|---|---|---|---|
Type of entry: | Primary publication | ||||
Title: | Explaining and Interactively Debugging Deep Models | ||||
Language: | English | ||||
Referees: | Kersting, Prof. Dr. Kristian ; Teso, Prof. Dr. Stefano | ||||
Date: | 2022 | ||||
Place of Publication: | Darmstadt | ||||
Collation: | xix, 174 Seiten | ||||
Date of oral examination: | 20 July 2022 | ||||
DOI: | 10.26083/tuprints-00021868 | ||||
Abstract: | Artificial Intelligence (AI) has made a huge impact on our everyday lives. As a dominant branch of AI since the 1990s, Machine Learning (ML) has been applied to a wide range of scenarios, including image recognition, speech recognition, fraud detection, recommendation systems, time series prediction and self-driving cars. Deep learning, backed up by Deep Neural Networks (DNNs), is a major subfield of machine learning. DNNs are good at approximating smooth functions, i.e., learning a mapping from inputs to outputs, which is also known as the predictive or supervised learning approach. Sometimes, one is not interested in a specific predictive task, but rather in finding interesting patterns in the data. In this case, a descriptive or unsupervised learning approach is needed, and the task can be formalized as density estimation. Deep probabilistic models have gained popularity for density estimation because they maintain a good balance between expressivity and tractability, whereas classical probabilistic models face an inherent trade-off. Deep neural networks and deep probabilistic models are both deep models in the sense that they are composed of multiple layers of computation units. They are essentially computation graphs and consequently, it is hard for humans to understand the underlying decision logic behind their behavior. Despite the representational and predictive power deep models have demonstrated in many complex problems, their opaqueness is a common reason for concern. In this thesis, we provide insights into deep models using high-level interpretations and explanations of why particular decisions are made. Explanations that contradict our intuitions or prior knowledge on the underlying domain can expose a potential concern, which may imply some desiderata of ML systems are not met. For example, a deep model may obtain high predictive accuracy by exploiting a spurious correlation in the dataset, which can lead to a lack of robustness, or unfairness if the spurious correlation is linked to a protected attribute. Built on the framework of Explanatory Interactive Machine Learning (XIL), we propose to interactively improve deep models based on the explanations we get. This way, we put users in the training loop and take user feedback on explanations as additional training signals. As an effect, the model can learn the rules that align with our intuitions or prior knowledge. |
||||
Alternative Abstract: |
|
||||
Status: | Publisher's Version | ||||
URN: | urn:nbn:de:tuda-tuprints-218689 | ||||
Classification DDC: | 000 Generalities, computers, information > 004 Computer science | ||||
Divisions: | 20 Department of Computer Science > Artificial Intelligence and Machine Learning | ||||
Date Deposited: | 19 Aug 2022 09:44 | ||||
Last Modified: | 22 Sep 2022 06:09 | ||||
URI: | https://tuprints.ulb.tu-darmstadt.de/id/eprint/21868 | ||||
PPN: | 499076680 | ||||
Export: |
View Item |