IIIT-Delhi Institutional Repository

A bayesian perspective to explainable AI

Show simple item record

dc.contributor.author Saini, Aditya
dc.contributor.author Prasad, Ranjitha (Advisor)
dc.date.accessioned 2023-04-06T11:58:42Z
dc.date.available 2023-04-06T11:58:42Z
dc.date.issued 2021-07
dc.identifier.uri http://repository.iiitd.edu.in/xmlui/handle/123456789/1101
dc.description.abstract The LIME(Local Interpretable Model-Agnostic Explanations) framework is a popular technique for providing simple to understand post-hoc model explanations. However, it su ers from the perplexing issue of inconsistent explanations. Attributed primarily to the sampling strategy used in the framework, this issue renders LIME completely useless for safety-critical domains like healthcare and robotics, where the notion of trustworthiness and consistency are of the utmost importance. In this work, a novel modi cation of LIME based on Gaussian process priors and Bayesian optimization based sampling(BO-LIME) is proposed. We illustrate the performance of the proposed technique on two real-world datasets, and demonstrate the superior stability of BOLIME using mean standard deviation of the explanations as the evaluation metric. Furthermore, we demonstrate that the proposed technique is able to generate faithful explanations using much fewer surrogate samples as compared to LIME. en_US
dc.language.iso en_US en_US
dc.publisher IIIT-Delhi en_US
dc.subject Explainable AI en_US
dc.subject LIME en_US
dc.subject Inconsistent Explanations en_US
dc.subject Gaussian Processes en_US
dc.subject Bayesian optimization en_US
dc.subject Acquisition functions en_US
dc.title A bayesian perspective to explainable AI en_US
dc.type Other en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search Repository


Advanced Search

Browse

My Account