ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2004.03685
  4. Cited By
Towards Faithfully Interpretable NLP Systems: How should we define and
  evaluate faithfulness?

Towards Faithfully Interpretable NLP Systems: How should we define and evaluate faithfulness?

7 April 2020
Alon Jacovi
Yoav Goldberg
    XAI
ArXivPDFHTML

Papers citing "Towards Faithfully Interpretable NLP Systems: How should we define and evaluate faithfulness?"

50 / 381 papers shown
Title
FLARE: Faithful Logic-Aided Reasoning and Exploration
FLARE: Faithful Logic-Aided Reasoning and Exploration
Erik Arakelyan
Pasquale Minervini
Pat Verga
Patrick Lewis
Isabelle Augenstein
ReLM
LRM
69
2
0
14 Oct 2024
Faithful Interpretation for Graph Neural Networks
Faithful Interpretation for Graph Neural Networks
Lijie Hu
Tianhao Huang
Lu Yu
Wanyu Lin
Tianhang Zheng
Di Wang
31
1
0
09 Oct 2024
Is the MMI Criterion Necessary for Interpretability? Degenerating
  Non-causal Features to Plain Noise for Self-Rationalization
Is the MMI Criterion Necessary for Interpretability? Degenerating Non-causal Features to Plain Noise for Self-Rationalization
Wei Liu
Zhiying Deng
Zhongyu Niu
Jun Wang
Haozhao Wang
YuanKai Zhang
Ruixuan Li
42
4
0
08 Oct 2024
Mechanistic?
Mechanistic?
Naomi Saphra
Sarah Wiegreffe
AI4CE
29
10
0
07 Oct 2024
Explanation sensitivity to the randomness of large language models: the
  case of journalistic text classification
Explanation sensitivity to the randomness of large language models: the case of journalistic text classification
Jérémie Bogaert
Marie-Catherine de Marneffe
Antonin Descampe
Louis Escouflaire
Cedrick Fairon
François-Xavier Standaert
24
1
0
07 Oct 2024
Zero-Shot Fact Verification via Natural Logic and Large Language Models
Zero-Shot Fact Verification via Natural Logic and Large Language Models
Marek Strong
Rami Aly
Andreas Vlachos
LRM
37
0
0
04 Oct 2024
F-Fidelity: A Robust Framework for Faithfulness Evaluation of Explainable AI
F-Fidelity: A Robust Framework for Faithfulness Evaluation of Explainable AI
Xu Zheng
Farhad Shirani
Zhuomin Chen
Chaohao Lin
Wei Cheng
Wenbo Guo
Dongsheng Luo
AAML
38
0
0
03 Oct 2024
COOL: Efficient and Reliable Chain-Oriented Objective Logic with Neural Networks Feedback Control for Program Synthesis
COOL: Efficient and Reliable Chain-Oriented Objective Logic with Neural Networks Feedback Control for Program Synthesis
Jipeng Han
39
0
0
02 Oct 2024
Faithfulness and the Notion of Adversarial Sensitivity in NLP
  Explanations
Faithfulness and the Notion of Adversarial Sensitivity in NLP Explanations
Supriya Manna
Niladri Sett
AAML
29
2
0
26 Sep 2024
Explainable AI needs formal notions of explanation correctness
Explainable AI needs formal notions of explanation correctness
Stefan Haufe
Rick Wilming
Benedict Clark
Rustam Zhumagambetov
Danny Panknin
Ahcène Boubekki
XAI
33
1
0
22 Sep 2024
Prompts Are Programs Too! Understanding How Developers Build Software Containing Prompts
Prompts Are Programs Too! Understanding How Developers Build Software Containing Prompts
Jenny T Liang
Melissa Lin
Nikitha Rao
Brad A. Myers
83
6
0
19 Sep 2024
Watch Your Steps: Observable and Modular Chains of Thought
Watch Your Steps: Observable and Modular Chains of Thought
Cassandra A. Cohen
William W. Cohen
LRM
36
1
0
17 Sep 2024
"The Data Says Otherwise"-Towards Automated Fact-checking and
  Communication of Data Claims
"The Data Says Otherwise"-Towards Automated Fact-checking and Communication of Data Claims
Yu Fu
Shunan Guo
Jane Hoffswell
Victor S. Bursztyn
Ryan A. Rossi
J. Stasko
31
2
0
16 Sep 2024
DRIVE: Dependable Robust Interpretable Visionary Ensemble Framework in
  Autonomous Driving
DRIVE: Dependable Robust Interpretable Visionary Ensemble Framework in Autonomous Driving
Songning Lai
Tianlang Xue
Hongru Xiao
Lijie Hu
Jiemin Wu
Ninghui Feng
Runwei Guan
Haicheng Liao
Zhenning Li
Yutao Yue
36
4
0
16 Sep 2024
Cross-Refine: Improving Natural Language Explanation Generation by
  Learning in Tandem
Cross-Refine: Improving Natural Language Explanation Generation by Learning in Tandem
Qianli Wang
Tatiana Anikina
Nils Feldhus
Simon Ostermann
Sebastian Möller
Vera Schmitt
LRM
38
1
0
11 Sep 2024
An Evaluation of Explanation Methods for Black-Box Detectors of
  Machine-Generated Text
An Evaluation of Explanation Methods for Black-Box Detectors of Machine-Generated Text
Loris Schoenegger
Yuxi Xia
Benjamin Roth
FAtt
46
0
0
26 Aug 2024
Counterfactuals As a Means for Evaluating Faithfulness of Attribution Methods in Autoregressive Language Models
Counterfactuals As a Means for Evaluating Faithfulness of Attribution Methods in Autoregressive Language Models
Sepehr Kamahi
Yadollah Yaghoobzadeh
53
0
0
21 Aug 2024
Visual Agents as Fast and Slow Thinkers
Visual Agents as Fast and Slow Thinkers
Guangyan Sun
Mingyu Jin
Zhenting Wang
Cheng-Long Wang
Siqi Ma
Qifan Wang
Ying Nian Wu
Ying Nian Wu
Dongfang Liu
Dongfang Liu
LLMAG
LRM
79
13
0
16 Aug 2024
Normalized AOPC: Fixing Misleading Faithfulness Metrics for Feature
  Attribution Explainability
Normalized AOPC: Fixing Misleading Faithfulness Metrics for Feature Attribution Explainability
Joakim Edin
Andreas Geert Motzfeldt
Casper L. Christensen
Tuukka Ruotsalo
Lars Maaløe
Maria Maistro
40
3
0
15 Aug 2024
In2Core: Leveraging Influence Functions for Coreset Selection in
  Instruction Finetuning of Large Language Models
In2Core: Leveraging Influence Functions for Coreset Selection in Instruction Finetuning of Large Language Models
Ayrton San Joaquin
Bin Wang
Zhengyuan Liu
Nicholas Asher
Brian Lim
Philippe Muller
Nancy Chen
42
0
0
07 Aug 2024
Data Debugging is NP-hard for Classifiers Trained with SGD
Data Debugging is NP-hard for Classifiers Trained with SGD
Zizheng Guo
Pengyu Chen
Yanzhang Fu
Xuelong Li
28
0
0
02 Aug 2024
Faithful and Plausible Natural Language Explanations for Image Classification: A Pipeline Approach
Faithful and Plausible Natural Language Explanations for Image Classification: A Pipeline Approach
Adam Wojciechowski
Mateusz Lango
Ondrej Dusek
FAtt
51
0
0
30 Jul 2024
BEExAI: Benchmark to Evaluate Explainable AI
BEExAI: Benchmark to Evaluate Explainable AI
Samuel Sithakoul
Sara Meftah
Clément Feutry
49
8
0
29 Jul 2024
On Behalf of the Stakeholders: Trends in NLP Model Interpretability in the Era of LLMs
On Behalf of the Stakeholders: Trends in NLP Model Interpretability in the Era of LLMs
Nitay Calderon
Roi Reichart
42
13
0
27 Jul 2024
Exploring the Plausibility of Hate and Counter Speech Detectors with
  Explainable AI
Exploring the Plausibility of Hate and Counter Speech Detectors with Explainable AI
Adrian Jaques Böck
D. Slijepcevic
Matthias Zeppelzauer
44
0
0
25 Jul 2024
Transformer Circuit Faithfulness Metrics are not Robust
Transformer Circuit Faithfulness Metrics are not Robust
Joseph Miller
Bilal Chughtai
William Saunders
53
7
0
11 Jul 2024
A Practical Review of Mechanistic Interpretability for Transformer-Based Language Models
A Practical Review of Mechanistic Interpretability for Transformer-Based Language Models
Daking Rai
Yilun Zhou
Shi Feng
Abulhair Saparov
Ziyu Yao
85
22
0
02 Jul 2024
Towards Understanding Sensitive and Decisive Patterns in Explainable AI:
  A Case Study of Model Interpretation in Geometric Deep Learning
Towards Understanding Sensitive and Decisive Patterns in Explainable AI: A Case Study of Model Interpretation in Geometric Deep Learning
Jiajun Zhu
Siqi Miao
Rex Ying
Pan Li
45
1
0
30 Jun 2024
Evaluating Human Alignment and Model Faithfulness of LLM Rationale
Evaluating Human Alignment and Model Faithfulness of LLM Rationale
Mohsen Fayyaz
Fan Yin
Jiao Sun
Nanyun Peng
65
3
0
28 Jun 2024
Rethinking harmless refusals when fine-tuning foundation models
Rethinking harmless refusals when fine-tuning foundation models
Florin Pop
Judd Rosenblatt
Diogo Schwerz de Lucena
Michael Vaiana
18
0
0
27 Jun 2024
The Illusion of Competence: Evaluating the Effect of Explanations on
  Users' Mental Models of Visual Question Answering Systems
The Illusion of Competence: Evaluating the Effect of Explanations on Users' Mental Models of Visual Question Answering Systems
Judith Sieker
Simeon Junker
R. Utescher
Nazia Attari
H. Wersing
Hendrik Buschmeier
Sina Zarrieß
30
1
0
27 Jun 2024
A look under the hood of the Interactive Deep Learning Enterprise
  (No-IDLE)
A look under the hood of the Interactive Deep Learning Enterprise (No-IDLE)
Daniel Sonntag
Michael Barz
Thiago S. Gouvêa
VLM
52
4
0
27 Jun 2024
On the Role of Visual Grounding in VQA
On the Role of Visual Grounding in VQA
Daniel Reich
Tanja Schultz
21
1
0
26 Jun 2024
This actually looks like that: Proto-BagNets for local and global
  interpretability-by-design
This actually looks like that: Proto-BagNets for local and global interpretability-by-design
K. Djoumessi
B. Bah
Laura Kühlewein
Philipp Berens
Lisa M. Koch
FAtt
21
0
0
21 Jun 2024
Model Internals-based Answer Attribution for Trustworthy
  Retrieval-Augmented Generation
Model Internals-based Answer Attribution for Trustworthy Retrieval-Augmented Generation
Jirui Qi
Gabriele Sarti
Raquel Fernández
Arianna Bisazza
RALM
45
6
0
19 Jun 2024
An Unsupervised Approach to Achieve Supervised-Level Explainability in
  Healthcare Records
An Unsupervised Approach to Achieve Supervised-Level Explainability in Healthcare Records
Joakim Edin
Maria Maistro
Lars Maaløe
Lasse Borgholt
Jakob Drachmann Havtorn
Tuukka Ruotsalo
FAtt
42
2
0
13 Jun 2024
Why Would You Suggest That? Human Trust in Language Model Responses
Why Would You Suggest That? Human Trust in Language Model Responses
Manasi Sharma
H. Siu
Rohan R. Paleja
Jaime D. Peña
LRM
42
6
0
04 Jun 2024
XPrompt:Explaining Large Language Model's Generation via Joint Prompt
  Attribution
XPrompt:Explaining Large Language Model's Generation via Joint Prompt Attribution
Yurui Chang
Bochuan Cao
Yujia Wang
Jinghui Chen
Lu Lin
LRM
32
0
0
30 May 2024
Explainable Molecular Property Prediction: Aligning Chemical Concepts
  with Predictions via Language Models
Explainable Molecular Property Prediction: Aligning Chemical Concepts with Predictions via Language Models
Zhenzhong Wang
Zehui Lin
Wanyu Lin
Ming Yang
Minggang Zeng
Kay Chen Tan
28
3
0
25 May 2024
Dissociation of Faithful and Unfaithful Reasoning in LLMs
Dissociation of Faithful and Unfaithful Reasoning in LLMs
Evelyn Yee
Alice Li
Chenyu Tang
Yeon Ho Jung
R. Paturi
Leon Bergen
LRM
32
4
0
23 May 2024
Towards a Unified Framework for Evaluating Explanations
Towards a Unified Framework for Evaluating Explanations
Juan D. Pinto
Luc Paquette
33
1
0
22 May 2024
Why do explanations fail? A typology and discussion on failures in XAI
Why do explanations fail? A typology and discussion on failures in XAI
Clara Bove
Thibault Laugel
Marie-Jeanne Lesot
C. Tijus
Marcin Detyniecki
33
2
0
22 May 2024
TimeX++: Learning Time-Series Explanations with Information Bottleneck
TimeX++: Learning Time-Series Explanations with Information Bottleneck
Zichuan Liu
Tianchun Wang
Jimeng Shi
Xu Zheng
Zhuomin Chen
Lei Song
Wenqian Dong
J. Obeysekera
Farhad Shirani
Dongsheng Luo
AI4TS
37
8
0
15 May 2024
Challenges and Opportunities in Text Generation Explainability
Challenges and Opportunities in Text Generation Explainability
Kenza Amara
Rita Sevastjanova
Mennatallah El-Assady
SILM
48
2
0
14 May 2024
Interpretability Needs a New Paradigm
Interpretability Needs a New Paradigm
Andreas Madsen
Himabindu Lakkaraju
Siva Reddy
Sarath Chandar
39
4
0
08 May 2024
ACORN: Aspect-wise Commonsense Reasoning Explanation Evaluation
ACORN: Aspect-wise Commonsense Reasoning Explanation Evaluation
Ana Brassard
Benjamin Heinzerling
Keito Kudo
Keisuke Sakaguchi
Kentaro Inui
LRM
39
0
0
08 May 2024
Do Vision & Language Decoders use Images and Text equally? How Self-consistent are their Explanations?
Do Vision & Language Decoders use Images and Text equally? How Self-consistent are their Explanations?
Letitia Parcalabescu
Anette Frank
MLLM
CoGe
VLM
84
3
0
29 Apr 2024
Mechanistic Interpretability for AI Safety -- A Review
Mechanistic Interpretability for AI Safety -- A Review
Leonard Bereska
E. Gavves
AI4CE
45
118
0
22 Apr 2024
The Probabilities Also Matter: A More Faithful Metric for Faithfulness
  of Free-Text Explanations in Large Language Models
The Probabilities Also Matter: A More Faithful Metric for Faithfulness of Free-Text Explanations in Large Language Models
Noah Y. Siegel
Oana-Maria Camburu
N. Heess
Maria Perez-Ortiz
26
8
0
04 Apr 2024
On the Faithfulness of Vision Transformer Explanations
On the Faithfulness of Vision Transformer Explanations
Junyi Wu
Weitai Kang
Hao Tang
Yuan Hong
Yan Yan
32
6
0
01 Apr 2024
Previous
12345678
Next