Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2202.01602
Cited By
The Disagreement Problem in Explainable Machine Learning: A Practitioner's Perspective
3 February 2022
Satyapriya Krishna
Tessa Han
Alex Gu
Steven Wu
S. Jabbari
Himabindu Lakkaraju
Re-assign community
ArXiv
PDF
HTML
Papers citing
"The Disagreement Problem in Explainable Machine Learning: A Practitioner's Perspective"
50 / 105 papers shown
Title
Integrating Explainable AI in Medical Devices: Technical, Clinical and Regulatory Insights and Recommendations
Dima Alattal
Asal Khoshravan Azar
P. Myles
Richard Branson
Hatim Abdulhussein
Allan Tucker
29
0
0
10 May 2025
Why and How LLMs Hallucinate: Connecting the Dots with Subsequence Associations
Yiyou Sun
Y. Gai
Lijie Chen
Abhilasha Ravichander
Yejin Choi
D. Song
HILM
57
0
0
17 Apr 2025
A constraints-based approach to fully interpretable neural networks for detecting learner behaviors
Juan D. Pinto
Luc Paquette
43
0
0
10 Apr 2025
Fourier Feature Attribution: A New Efficiency Attribution Method
Zechen Liu
Feiyang Zhang
Wei Song
X. Li
Wei Wei
FAtt
57
0
0
02 Apr 2025
CFIRE: A General Method for Combining Local Explanations
Sebastian Müller
Vanessa Toborek
Tamás Horváth
Christian Bauckhage
FAtt
48
0
0
01 Apr 2025
Detecting Systematic Weaknesses in Vision Models along Predefined Human-Understandable Dimensions
Sujan Sai Gannamaneni
Rohil Prakash Rao
Michael Mock
Maram Akila
Stefan Wrobel
AAML
136
0
0
17 Feb 2025
Building Bridges, Not Walls -- Advancing Interpretability by Unifying Feature, Data, and Model Component Attribution
Shichang Zhang
Tessa Han
Usha Bhalla
Hima Lakkaraju
FAtt
147
0
0
17 Feb 2025
Feature Importance Depends on Properties of the Data: Towards Choosing the Correct Explanations for Your Data and Decision Trees based Models
Célia Wafa Ayad
Thomas Bonnier
Benjamin Bosch
Sonali Parbhoo
Jesse Read
FAtt
XAI
98
0
0
11 Feb 2025
A Tale of Two Imperatives: Privacy and Explainability
Supriya Manna
Niladri Sett
91
0
0
30 Dec 2024
From Flexibility to Manipulation: The Slippery Slope of XAI Evaluation
Kristoffer Wickstrøm
Marina M.-C. Höhne
Anna Hedström
AAML
79
2
0
07 Dec 2024
EXAGREE: Towards Explanation Agreement in Explainable Machine Learning
Sichao Li
Quanling Deng
Amanda S. Barnard
37
0
0
04 Nov 2024
Feature Responsiveness Scores: Model-Agnostic Explanations for Recourse
Seung Hyun Cheon
Anneke Wernerfelt
Sorelle A. Friedler
Berk Ustun
FaML
FAtt
45
0
0
29 Oct 2024
Explainability in AI Based Applications: A Framework for Comparing Different Techniques
Arne Grobrugge
Nidhi Mishra
Johannes Jakubik
G. Satzger
99
1
0
28 Oct 2024
Explainable News Summarization -- Analysis and mitigation of Disagreement Problem
Seema Aswani
Sujala D. Shetty
31
0
0
24 Oct 2024
An Ontology-Enabled Approach For User-Centered and Knowledge-Enabled Explanations of AI Systems
Shruthi Chari
29
0
0
23 Oct 2024
Backward Compatibility in Attributive Explanation and Enhanced Model Training Method
Ryuta Matsuno
36
0
0
05 Aug 2024
On the Evaluation Consistency of Attribution-based Explanations
Jiarui Duan
Haoling Li
Haofei Zhang
Hao Jiang
Mengqi Xue
Li Sun
Mingli Song
Jie Song
XAI
46
0
0
28 Jul 2024
Exploring the Plausibility of Hate and Counter Speech Detectors with Explainable AI
Adrian Jaques Böck
D. Slijepcevic
Matthias Zeppelzauer
42
0
0
25 Jul 2024
MOUNTAINEER: Topology-Driven Visual Analytics for Comparing Local Explanations
Parikshit Solunke
Vitória Guardieiro
Joao Rulff
Peter Xenopoulos
G. Chan
Brian Barr
L. G. Nonato
Claudio Silva
29
1
0
21 Jun 2024
ChaosMining: A Benchmark to Evaluate Post-Hoc Local Attribution Methods in Low SNR Environments
Ge Shi
Ziwen Kan
J. Smucny
Ian Davidson
54
0
0
17 Jun 2024
Are Objective Explanatory Evaluation metrics Trustworthy? An Adversarial Analysis
Prithwijit Chowdhury
M. Prabhushankar
Ghassan AlRegib
Mohamed Deriche
28
0
0
12 Jun 2024
Exploring the Relationship Between Feature Attribution Methods and Model Performance
Priscylla Silva
Claudio T. Silva
L. G. Nonato
FAtt
22
1
0
22 May 2024
SIDEs: Separating Idealization from Deceptive Explanations in xAI
Emily Sullivan
49
2
0
25 Apr 2024
T-Explainer: A Model-Agnostic Explainability Framework Based on Gradients
Evandro S. Ortigossa
Fábio F. Dias
Brian Barr
Claudio T. Silva
L. G. Nonato
FAtt
54
2
0
25 Apr 2024
A Guide to Feature Importance Methods for Scientific Inference
F. K. Ewald
Ludwig Bothmann
Marvin N. Wright
B. Bischl
Giuseppe Casalicchio
Gunnar Konig
30
10
0
19 Apr 2024
Toward Understanding the Disagreement Problem in Neural Network Feature Attribution
Niklas Koenen
Marvin N. Wright
FAtt
34
5
0
17 Apr 2024
CNN-based explanation ensembling for dataset, representation and explanations evaluation
Weronika Hryniewska-Guzik
Luca Longo
P. Biecek
FAtt
43
0
0
16 Apr 2024
Visual Concept Connectome (VCC): Open World Concept Discovery and their Interlayer Connections in Deep Models
M. Kowal
Richard P. Wildes
Konstantinos G. Derpanis
GNN
30
8
0
02 Apr 2024
The Role of Syntactic Span Preferences in Post-Hoc Explanation Disagreement
Jonathan Kamp
Lisa Beinborn
Antske Fokkens
30
1
0
28 Mar 2024
Connecting Algorithmic Fairness to Quality Dimensions in Machine Learning in Official Statistics and Survey Production
Patrick Oliver Schenk
Christoph Kern
FaML
24
0
0
14 Feb 2024
InterpretCC: Intrinsic User-Centric Interpretability through Global Mixture of Experts
Vinitra Swamy
Syrielle Montariol
Julian Blackwell
Jibril Frej
Martin Jaggi
Tanja Kaser
31
3
0
05 Feb 2024
A comprehensive study on fidelity metrics for XAI
Miquel Miró-Nicolau
Antoni Jaume-i-Capó
Gabriel Moyà Alcover
30
11
0
19 Jan 2024
An Axiomatic Approach to Model-Agnostic Concept Explanations
Zhili Feng
Michal Moshkovitz
Dotan Di Castro
J. Zico Kolter
LRM
23
0
0
12 Jan 2024
Decoupling Pixel Flipping and Occlusion Strategy for Consistent XAI Benchmarks
Stefan Blücher
Johanna Vielhaben
Nils Strodthoff
AAML
61
20
0
12 Jan 2024
Can Physician Judgment Enhance Model Trustworthiness? A Case Study on Predicting Pathological Lymph Nodes in Rectal Cancer
Kazuma Kobayashi
Yasuyuki Takamizawa
M. Miyake
Sono Ito
Lin Gu
Tatsuya Nakatsuka
Yu Akagi
Tatsuya Harada
Y. Kanemitsu
Ryuji Hamamoto
30
2
0
15 Dec 2023
A novel post-hoc explanation comparison metric and applications
Shreyan Mitra
Leilani H. Gilpin
FAtt
31
0
0
17 Nov 2023
The Disagreement Problem in Faithfulness Metrics
Brian Barr
Noah Fatsi
Leif Hancox-Li
Peter Richter
Daniel Proano
Caleb Mok
36
4
0
13 Nov 2023
Sample based Explanations via Generalized Representers
Che-Ping Tsai
Chih-Kuan Yeh
Pradeep Ravikumar
FAtt
39
8
0
27 Oct 2023
How Well Do Feature-Additive Explainers Explain Feature-Additive Predictors?
Zachariah Carmichael
Walter J. Scheirer
FAtt
30
4
0
27 Oct 2023
Can Large Language Models Explain Themselves? A Study of LLM-Generated Self-Explanations
Shiyuan Huang
Siddarth Mamidanna
Shreedhar Jangam
Yilun Zhou
Leilani H. Gilpin
LRM
MILM
ELM
35
66
0
17 Oct 2023
The Thousand Faces of Explainable AI Along the Machine Learning Life Cycle: Industrial Reality and Current State of Research
Thomas Decker
Ralf Gross
Alexander Koebler
Michael Lebacher
Ronald Schnitzer
Stefan H. Weber
31
2
0
11 Oct 2023
AttributionLab: Faithfulness of Feature Attribution Under Controllable Environments
Yang Zhang
Yawei Li
Hannah Brown
Mina Rezaei
Bernd Bischl
Philip H. S. Torr
Ashkan Khakzar
Kenji Kawaguchi
OOD
50
1
0
10 Oct 2023
Dynamic Top-k Estimation Consolidates Disagreement between Feature Attribution Methods
Jonathan Kamp
Lisa Beinborn
Antske Fokkens
FAtt
33
1
0
09 Oct 2023
IDTraffickers: An Authorship Attribution Dataset to link and connect Potential Human-Trafficking Operations on Text Escort Advertisements
V. Saxena
Benjamin Bashpole
Gijs Van Dijck
Gerasimos Spanakis
40
2
0
09 Oct 2023
The Blame Problem in Evaluating Local Explanations, and How to Tackle it
Amir Hossein Akhavan Rahnama
ELM
FAtt
28
4
0
05 Oct 2023
Pixel-Grounded Prototypical Part Networks
Zachariah Carmichael
Suhas Lohit
A. Cherian
Michael J. Jones
Walter J. Scheirer
33
11
0
25 Sep 2023
Goodhart's Law Applies to NLP's Explanation Benchmarks
Jennifer Hsia
Danish Pruthi
Aarti Singh
Zachary Chase Lipton
26
6
0
28 Aug 2023
A Dual-Perspective Approach to Evaluating Feature Attribution Methods
Yawei Li
Yanglin Zhang
Kenji Kawaguchi
Ashkan Khakzar
Bernd Bischl
Mina Rezaei
FAtt
XAI
39
0
0
17 Aug 2023
Analyzing Chain-of-Thought Prompting in Large Language Models via Gradient-based Feature Attributions
Skyler Wu
Eric Meng Shen
Charumathi Badrinath
Jiaqi Ma
Himabindu Lakkaraju
LRM
30
26
0
25 Jul 2023
Dissenting Explanations: Leveraging Disagreement to Reduce Model Overreliance
Omer Reingold
J. Shen
Aditi Talati
FAtt
LRM
22
3
0
14 Jul 2023
1
2
3
Next