Model-Agnostic Explanations and Evaluation of Machine Learning

dc.contributor.advisorGuestrin, Carlos
dc.contributor.authorCorreia Ribeiro, Marco Tulio
dc.date.accessioned2019-02-22T17:04:28Z
dc.date.available2019-02-22T17:04:28Z
dc.date.issued2019-02-22
dc.date.submitted2018
dc.descriptionThesis (Ph.D.)--University of Washington, 2018
dc.description.abstractDespite many successes, complex machine learning systems are limited in their impact due to several issues regarding communication with humans: they are functionally black boxes, hard to debug and hard to evaluate properly. This communication is crucial though: humans are the ones who train, deploy and use machine learning models, and thus have to make trust and evaluation decisions. Furthermore, it is humans who try to improve these models, and having an understanding of their behavior is very valuable for this purpose. This dissertation addresses this communication problem by presenting model-agnostic explanations and evaluation, which improve the interaction between humans and any machine learning model. Specifically, we present: (1) Local Interpretable Model-Agnostic Explanations (LIME), an explanation technique that can explain any black box model by approximating it locally with a linear model, (2) Anchors, model-agnostic explanations that represent sufficient conditions for predictions, (3) Semantically Equivalent Adversaries and Adversarial Rules (SEAs and SEARs), semantic-preserving perturbations and rules that unearth brittleness bugs in text models, and (4) Implication Consistency, a new kind of evaluation metric that considers the relationship between model outputs in order to measure higher level thinking. We demonstrate that these contributions enable efficient communication between machine learning models and humans, empowering humans to better evaluate, improve, and assess trust in models.
dc.embargo.termsOpen Access
dc.format.mimetypeapplication/pdf
dc.identifier.otherCorreiaRibeiro_washington_0250E_19379.pdf
dc.identifier.urihttp://hdl.handle.net/1773/43353
dc.language.isoen_US
dc.rightsnone
dc.subjecthuman-computer interaction
dc.subjectinterpretability
dc.subjectmachine learning
dc.subjectComputer science
dc.subject.otherComputer science and engineering
dc.titleModel-Agnostic Explanations and Evaluation of Machine Learning
dc.typeThesis

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
CorreiaRibeiro_washington_0250E_19379.pdf
Size:
9.87 MB
Format:
Adobe Portable Document Format