Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1910.02065
Cited By
Can I Trust the Explainer? Verifying Post-hoc Explanatory Methods
4 October 2019
Oana-Maria Camburu
Eleonora Giunchiglia
Jakob N. Foerster
Thomas Lukasiewicz
Phil Blunsom
FAtt
AAML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Can I Trust the Explainer? Verifying Post-hoc Explanatory Methods"
18 / 18 papers shown
Title
Feature Importance Depends on Properties of the Data: Towards Choosing the Correct Explanations for Your Data and Decision Trees based Models
Célia Wafa Ayad
Thomas Bonnier
Benjamin Bosch
Sonali Parbhoo
Jesse Read
FAtt
XAI
103
0
0
11 Feb 2025
Clairvoyance: A Pipeline Toolkit for Medical Time Series
Daniel Jarrett
Jinsung Yoon
Ioana Bica
Zhaozhi Qian
A. Ercole
M. Schaar
AI4TS
26
35
0
28 Oct 2023
Logic-Based Explainability in Machine Learning
Sasha Rubin
LRM
XAI
47
39
0
24 Oct 2022
On Computing Relevant Features for Explaining NBCs
Yacine Izza
Sasha Rubin
36
5
0
11 Jul 2022
Eliminating The Impossible, Whatever Remains Must Be True
Jinqiang Yu
Alexey Ignatiev
Peter J. Stuckey
Nina Narodytska
Sasha Rubin
19
23
0
20 Jun 2022
Towards a consistent interpretation of AIOps models
Yingzhe Lyu
Gopi Krishnan Rajbahadur
Dayi Lin
Boyuan Chen
Zhen Ming
Z. Jiang
AI4CE
22
19
0
04 Feb 2022
Post-Hoc Explanations Fail to Achieve their Purpose in Adversarial Contexts
Sebastian Bordt
Michèle Finck
Eric Raidl
U. V. Luxburg
AILaw
39
77
0
25 Jan 2022
Knowledge-Grounded Self-Rationalization via Extractive and Natural Language Explanations
Bodhisattwa Prasad Majumder
Oana-Maria Camburu
Thomas Lukasiewicz
Julian McAuley
25
35
0
25 Jun 2021
A Framework for Evaluating Post Hoc Feature-Additive Explainers
Zachariah Carmichael
Walter J. Scheirer
FAtt
46
4
0
15 Jun 2021
Prompting Contrastive Explanations for Commonsense Reasoning Tasks
Bhargavi Paranjape
Julian Michael
Marjan Ghazvininejad
Luke Zettlemoyer
Hannaneh Hajishirzi
ReLM
LRM
22
66
0
12 Jun 2021
On Efficiently Explaining Graph-Based Classifiers
Xuanxiang Huang
Yacine Izza
Alexey Ignatiev
Sasha Rubin
FAtt
34
37
0
02 Jun 2021
SAT-Based Rigorous Explanations for Decision Lists
Alexey Ignatiev
Sasha Rubin
XAI
25
44
0
14 May 2021
Why do you think that? Exploring Faithful Sentence-Level Rationales Without Supervision
Max Glockner
Ivan Habernal
Iryna Gurevych
LRM
27
25
0
07 Oct 2020
Explaining Deep Neural Networks
Oana-Maria Camburu
XAI
FAtt
28
26
0
04 Oct 2020
NILE : Natural Language Inference with Faithful Natural Language Explanations
Sawan Kumar
Partha P. Talukdar
XAI
LRM
11
159
0
25 May 2020
Evaluating and Aggregating Feature-based Model Explanations
Umang Bhatt
Adrian Weller
J. M. F. Moura
XAI
33
218
0
01 May 2020
e-SNLI: Natural Language Inference with Natural Language Explanations
Oana-Maria Camburu
Tim Rocktaschel
Thomas Lukasiewicz
Phil Blunsom
LRM
260
620
0
04 Dec 2018
Learning Attitudes and Attributes from Multi-Aspect Reviews
Julian McAuley
J. Leskovec
Dan Jurafsky
200
296
0
15 Oct 2012
1