ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1911.02508
  4. Cited By
Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation
  Methods

Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation Methods

6 November 2019
Dylan Slack
Sophie Hilgard
Emily Jia
Sameer Singh
Himabindu Lakkaraju
    FAtt
    AAML
    MLAU
ArXivPDFHTML

Papers citing "Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation Methods"

50 / 133 papers shown
Title
A Fast Kernel-based Conditional Independence test with Application to Causal Discovery
A Fast Kernel-based Conditional Independence test with Application to Causal Discovery
Oliver Schacht
Biwei Huang
12
0
0
16 May 2025
Enhanced Photonic Chip Design via Interpretable Machine Learning Techniques
Enhanced Photonic Chip Design via Interpretable Machine Learning Techniques
Lirandë Pira
Airin Antony
Nayanthara Prathap
Daniel Peace
Jacquiline Romero
9
0
0
14 May 2025
SHAP-based Explanations are Sensitive to Feature Representation
Hyunseung Hwang
Andrew Bell
Joao Fonseca
Venetia Pliatsika
Julia Stoyanovich
Steven Euijong Whang
TDI
FAtt
32
0
0
13 May 2025
Explanations as Bias Detectors: A Critical Study of Local Post-hoc XAI Methods for Fairness Exploration
Explanations as Bias Detectors: A Critical Study of Local Post-hoc XAI Methods for Fairness Exploration
Vasiliki Papanikou
Danae Pla Karidi
E. Pitoura
Emmanouil Panagiotou
Eirini Ntoutsi
33
0
0
01 May 2025
Probabilistic Stability Guarantees for Feature Attributions
Probabilistic Stability Guarantees for Feature Attributions
Helen Jin
Anton Xue
Weiqiu You
Surbhi Goel
Eric Wong
27
0
0
18 Apr 2025
Towards Responsible and Trustworthy Educational Data Mining: Comparing Symbolic, Sub-Symbolic, and Neural-Symbolic AI Methods
Towards Responsible and Trustworthy Educational Data Mining: Comparing Symbolic, Sub-Symbolic, and Neural-Symbolic AI Methods
Danial Hooshyar
Eve Kikas
Yeongwook Yang
Gustav Šír
Raija Hamalainen
T. Karkkainen
Roger Azevedo
59
1
0
01 Apr 2025
ExpProof : Operationalizing Explanations for Confidential Models with ZKPs
ExpProof : Operationalizing Explanations for Confidential Models with ZKPs
Chhavi Yadav
Evan Monroe Laufer
Dan Boneh
Kamalika Chaudhuri
91
0
0
06 Feb 2025
The Effect of Similarity Measures on Accurate Stability Estimates for Local Surrogate Models in Text-based Explainable AI
The Effect of Similarity Measures on Accurate Stability Estimates for Local Surrogate Models in Text-based Explainable AI
Christopher Burger
Charles Walter
Thai Le
AAML
148
1
0
20 Jan 2025
Feature Responsiveness Scores: Model-Agnostic Explanations for Recourse
Feature Responsiveness Scores: Model-Agnostic Explanations for Recourse
Seung Hyun Cheon
Anneke Wernerfelt
Sorelle A. Friedler
Berk Ustun
FaML
FAtt
45
0
0
29 Oct 2024
A prototype-based model for set classification
A prototype-based model for set classification
Mohammad Mohammadi
Sreejita Ghosh
VLM
112
1
0
25 Aug 2024
Need of AI in Modern Education: in the Eyes of Explainable AI (xAI)
Need of AI in Modern Education: in the Eyes of Explainable AI (xAI)
Supriya Manna
Dionis Barcari
45
3
0
31 Jul 2024
The US Algorithmic Accountability Act of 2022 vs. The EU Artificial
  Intelligence Act: What can they learn from each other?
The US Algorithmic Accountability Act of 2022 vs. The EU Artificial Intelligence Act: What can they learn from each other?
Jakob Mokander
Prathm Juneja
David S. Watson
Luciano Floridi
34
54
0
07 Jul 2024
Efficient and Accurate Explanation Estimation with Distribution Compression
Efficient and Accurate Explanation Estimation with Distribution Compression
Hubert Baniecki
Giuseppe Casalicchio
Bernd Bischl
Przemyslaw Biecek
FAtt
46
3
0
26 Jun 2024
Explaining Predictions by Characteristic Rules
Explaining Predictions by Characteristic Rules
Amr Alkhatib
Henrik Bostrom
Michalis Vazirgiannis
29
5
0
31 May 2024
Manifold Integrated Gradients: Riemannian Geometry for Feature
  Attribution
Manifold Integrated Gradients: Riemannian Geometry for Feature Attribution
Eslam Zaher
Maciej Trzaskowski
Quan Nguyen
Fred Roosta
AAML
29
4
0
16 May 2024
Stability of Explainable Recommendation
Stability of Explainable Recommendation
Sairamvinay Vijayaraghavan
Prasant Mohapatra
AAML
38
1
0
03 May 2024
Robust Explainable Recommendation
Robust Explainable Recommendation
Sairamvinay Vijayaraghavan
Prasant Mohapatra
AAML
23
0
0
03 May 2024
Why You Should Not Trust Interpretations in Machine Learning:
  Adversarial Attacks on Partial Dependence Plots
Why You Should Not Trust Interpretations in Machine Learning: Adversarial Attacks on Partial Dependence Plots
Xi Xin
Giles Hooker
Fei Huang
AAML
46
6
0
29 Apr 2024
On the Value of Labeled Data and Symbolic Methods for Hidden Neuron
  Activation Analysis
On the Value of Labeled Data and Symbolic Methods for Hidden Neuron Activation Analysis
Abhilekha Dalal
R. Rayan
Adrita Barua
Eugene Y. Vasserman
Md Kamruzzaman Sarker
Pascal Hitzler
30
4
0
21 Apr 2024
RankingSHAP -- Listwise Feature Attribution Explanations for Ranking Models
RankingSHAP -- Listwise Feature Attribution Explanations for Ranking Models
Maria Heuss
Maarten de Rijke
Avishek Anand
124
1
0
24 Mar 2024
Improving deep learning with prior knowledge and cognitive models: A
  survey on enhancing explainability, adversarial robustness and zero-shot
  learning
Improving deep learning with prior knowledge and cognitive models: A survey on enhancing explainability, adversarial robustness and zero-shot learning
F. Mumuni
A. Mumuni
AAML
37
5
0
11 Mar 2024
Are Classification Robustness and Explanation Robustness Really Strongly
  Correlated? An Analysis Through Input Loss Landscape
Are Classification Robustness and Explanation Robustness Really Strongly Correlated? An Analysis Through Input Loss Landscape
Tiejin Chen
Wenwang Huang
Linsey Pang
Dongsheng Luo
Hua Wei
OOD
49
0
0
09 Mar 2024
Attention Meets Post-hoc Interpretability: A Mathematical Perspective
Attention Meets Post-hoc Interpretability: A Mathematical Perspective
Gianluigi Lopardo
F. Precioso
Damien Garreau
16
4
0
05 Feb 2024
Black-Box Access is Insufficient for Rigorous AI Audits
Black-Box Access is Insufficient for Rigorous AI Audits
Stephen Casper
Carson Ezell
Charlotte Siegmann
Noam Kolt
Taylor Lynn Curtis
...
Michael Gerovitch
David Bau
Max Tegmark
David M. Krueger
Dylan Hadfield-Menell
AAML
34
78
0
25 Jan 2024
3VL: Using Trees to Improve Vision-Language Models' Interpretability
3VL: Using Trees to Improve Vision-Language Models' Interpretability
Nir Yellinek
Leonid Karlinsky
Raja Giryes
CoGe
VLM
49
4
0
28 Dec 2023
On the Relationship Between Interpretability and Explainability in
  Machine Learning
On the Relationship Between Interpretability and Explainability in Machine Learning
Benjamin Leblanc
Pascal Germain
FaML
29
0
0
20 Nov 2023
Explaining black boxes with a SMILE: Statistical Model-agnostic
  Interpretability with Local Explanations
Explaining black boxes with a SMILE: Statistical Model-agnostic Interpretability with Local Explanations
Koorosh Aslansefat
Mojgan Hashemian
M. Walker
Mohammed Naveed Akram
Ioannis Sorokos
Y. Papadopoulos
FAtt
AAML
19
2
0
13 Nov 2023
Beyond XAI:Obstacles Towards Responsible AI
Beyond XAI:Obstacles Towards Responsible AI
Yulu Pi
37
2
0
07 Sep 2023
Explaining Black-Box Models through Counterfactuals
Explaining Black-Box Models through Counterfactuals
Patrick Altmeyer
A. V. Deursen
Cynthia C. S. Liem
CML
LRM
37
2
0
14 Aug 2023
LUCID-GAN: Conditional Generative Models to Locate Unfairness
LUCID-GAN: Conditional Generative Models to Locate Unfairness
Andres Algaba
Carmen Mazijn
Carina E. A. Prunkl
J. Danckaert
Vincent Ginis
SyDa
39
1
0
28 Jul 2023
Probabilistic Constrained Reinforcement Learning with Formal
  Interpretability
Probabilistic Constrained Reinforcement Learning with Formal Interpretability
Yanran Wang
Qiuchen Qian
David E. Boyle
16
4
0
13 Jul 2023
The future of human-centric eXplainable Artificial Intelligence (XAI) is
  not post-hoc explanations
The future of human-centric eXplainable Artificial Intelligence (XAI) is not post-hoc explanations
Vinitra Swamy
Jibril Frej
Tanja Kaser
34
14
0
01 Jul 2023
BELLA: Black box model Explanations by Local Linear Approximations
BELLA: Black box model Explanations by Local Linear Approximations
N. Radulovic
Albert Bifet
Fabian M. Suchanek
FAtt
37
1
0
18 May 2023
Explainability in AI Policies: A Critical Review of Communications,
  Reports, Regulations, and Standards in the EU, US, and UK
Explainability in AI Policies: A Critical Review of Communications, Reports, Regulations, and Standards in the EU, US, and UK
L. Nannini
Agathe Balayn
A. Smith
21
37
0
20 Apr 2023
Interpretable (not just posthoc-explainable) heterogeneous survivor
  bias-corrected treatment effects for assignment of postdischarge
  interventions to prevent readmissions
Interpretable (not just posthoc-explainable) heterogeneous survivor bias-corrected treatment effects for assignment of postdischarge interventions to prevent readmissions
Hongjing Xia
Joshua C. Chang
S. Nowak
Sonya Mahajan
R. Mahajan
Ted L. Chang
Carson C. Chow
38
1
0
19 Apr 2023
An Interpretable Loan Credit Evaluation Method Based on Rule
  Representation Learner
An Interpretable Loan Credit Evaluation Method Based on Rule Representation Learner
Zi-yu Chen
Xiaomeng Wang
Yuanjiang Huang
Tao Jia
35
1
0
03 Apr 2023
Don't be fooled: label leakage in explanation methods and the importance
  of their quantitative evaluation
Don't be fooled: label leakage in explanation methods and the importance of their quantitative evaluation
N. Jethani
A. Saporta
Rajesh Ranganath
FAtt
29
10
0
24 Feb 2023
Function Composition in Trustworthy Machine Learning: Implementation
  Choices, Insights, and Questions
Function Composition in Trustworthy Machine Learning: Implementation Choices, Insights, and Questions
Manish Nagireddy
Moninder Singh
Samuel C. Hoffman
Evaline Ju
K. Ramamurthy
Kush R. Varshney
30
1
0
17 Feb 2023
A novel approach to generate datasets with XAI ground truth to evaluate
  image models
A novel approach to generate datasets with XAI ground truth to evaluate image models
Miquel Miró-Nicolau
Antoni Jaume-i-Capó
Gabriel Moyà Alcover
22
4
0
11 Feb 2023
Variational Information Pursuit for Interpretable Predictions
Variational Information Pursuit for Interpretable Predictions
Aditya Chattopadhyay
Kwan Ho Ryan Chan
B. Haeffele
D. Geman
René Vidal
DRL
24
10
0
06 Feb 2023
Faithful Chain-of-Thought Reasoning
Faithful Chain-of-Thought Reasoning
Qing Lyu
Shreya Havaldar
Adam Stein
Li Zhang
D. Rao
Eric Wong
Marianna Apidianaki
Chris Callison-Burch
ReLM
LRM
41
207
0
31 Jan 2023
Don't Lie to Me: Avoiding Malicious Explanations with STEALTH
Don't Lie to Me: Avoiding Malicious Explanations with STEALTH
Lauren Alvarez
Tim Menzies
34
2
0
25 Jan 2023
Rank-LIME: Local Model-Agnostic Feature Attribution for Learning to Rank
Rank-LIME: Local Model-Agnostic Feature Attribution for Learning to Rank
Tanya Chowdhury
Razieh Rahimi
James Allan
FAtt
34
18
0
24 Dec 2022
Robust Explanation Constraints for Neural Networks
Robust Explanation Constraints for Neural Networks
Matthew Wicker
Juyeon Heo
Luca Costabello
Adrian Weller
FAtt
29
18
0
16 Dec 2022
On the Limit of Explaining Black-box Temporal Graph Neural Networks
On the Limit of Explaining Black-box Temporal Graph Neural Networks
Minh Nhat Vu
My T. Thai
16
4
0
02 Dec 2022
Understanding and Enhancing Robustness of Concept-based Models
Understanding and Enhancing Robustness of Concept-based Models
Sanchit Sinha
Mengdi Huai
Jianhui Sun
Aidong Zhang
AAML
34
18
0
29 Nov 2022
Towards More Robust Interpretation via Local Gradient Alignment
Towards More Robust Interpretation via Local Gradient Alignment
Sunghwan Joo
Seokhyeon Jeong
Juyeon Heo
Adrian Weller
Taesup Moon
FAtt
30
5
0
29 Nov 2022
Testing the effectiveness of saliency-based explainability in NLP using
  randomized survey-based experiments
Testing the effectiveness of saliency-based explainability in NLP using randomized survey-based experiments
Adel Rahimi
Shaurya Jain
FAtt
13
0
0
25 Nov 2022
Explaining Image Classifiers with Multiscale Directional Image
  Representation
Explaining Image Classifiers with Multiscale Directional Image Representation
Stefan Kolek
Robert Windesheim
Héctor Andrade-Loarca
Gitta Kutyniok
Ron Levie
29
4
0
22 Nov 2022
CRAFT: Concept Recursive Activation FacTorization for Explainability
CRAFT: Concept Recursive Activation FacTorization for Explainability
Thomas Fel
Agustin Picard
Louis Bethune
Thibaut Boissin
David Vigouroux
Julien Colin
Rémi Cadène
Thomas Serre
19
102
0
17 Nov 2022
123
Next