Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2210.07455
Cited By
Controlling Bias Exposure for Fair Interpretable Predictions
14 October 2022
Zexue He
Yu-Xiang Wang
Julian McAuley
Bodhisattwa Prasad Majumder
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Controlling Bias Exposure for Fair Interpretable Predictions"
20 / 20 papers shown
Title
UNO Arena for Evaluating Sequential Decision-Making Capability of Large Language Models
Zhanyue Qin
Haochuan Wang
Deyuan Liu
Ziyang Song
Cunhang Fan
...
Zhen Lei
Zhiying Tu
Dianhui Chu
Xiaoyan Yu
Dianbo Sui
ELM
LRM
56
1
0
24 Jun 2024
Deconstructing The Ethics of Large Language Models from Long-standing Issues to New-emerging Dilemmas
Chengyuan Deng
Yiqun Duan
Xin Jin
Heng Chang
Yijun Tian
...
Kuofeng Gao
Sihong He
Jun Zhuang
Lu Cheng
Haohan Wang
AILaw
43
16
0
08 Jun 2024
The Life Cycle of Large Language Models: A Review of Biases in Education
Jinsook Lee
Yann Hicke
Renzhe Yu
Christopher A. Brooks
René F. Kizilcec
AI4Ed
34
1
0
03 Jun 2024
A Note on Bias to Complete
Jia Xu
Mona Diab
47
2
0
18 Feb 2024
Self-Debiasing Large Language Models: Zero-Shot Recognition and Reduction of Stereotypes
Isabel O. Gallegos
Ryan A. Rossi
Joe Barrow
Md Mehrab Tanjim
Tong Yu
Hanieh Deilamsalehy
Ruiyi Zhang
Sungchul Kim
Franck Dernoncourt
24
19
0
03 Feb 2024
Deciphering Compatibility Relationships with Textual Descriptions via Extraction and Explanation
Yu-Xiang Wang
Zexue He
Zhankui He
Hao Xu
Julian McAuley
11
3
0
17 Dec 2023
Tackling Bias in Pre-trained Language Models: Current Trends and Under-represented Societies
Vithya Yogarajan
Gillian Dobbie
Te Taka Keegan
R. Neuwirth
ALM
43
11
0
03 Dec 2023
Bias A-head? Analyzing Bias in Transformer-Based Language Model Attention Heads
Yi Yang
Hanyu Duan
Ahmed Abbasi
John P. Lalor
Kar Yan Tam
13
5
0
17 Nov 2023
Evaluating Bias and Fairness in Gender-Neutral Pretrained Vision-and-Language Models
Laura Cabello
Emanuele Bugliarello
Stephanie Brandl
Desmond Elliott
23
7
0
26 Oct 2023
On the Interplay between Fairness and Explainability
Stephanie Brandl
Emanuele Bugliarello
Ilias Chalkidis
FaML
27
4
0
25 Oct 2023
MedEval: A Multi-Level, Multi-Task, and Multi-Domain Medical Benchmark for Language Model Evaluation
Zexue He
Yu-Xiang Wang
An Yan
Yao Liu
Eric Y. Chang
Amilcare Gentili
Julian McAuley
Chun-Nan Hsu
ELM
83
14
0
21 Oct 2023
Bias and Fairness in Large Language Models: A Survey
Isabel O. Gallegos
Ryan A. Rossi
Joe Barrow
Md Mehrab Tanjim
Sungchul Kim
Franck Dernoncourt
Tong Yu
Ruiyi Zhang
Nesreen Ahmed
AILaw
21
490
0
02 Sep 2023
Label Denoising through Cross-Model Agreement
Yu-Xiang Wang
Xin Xin
Zaiqiao Meng
J. Jose
Fuli Feng
NoLa
34
1
0
27 Aug 2023
Sociodemographic Bias in Language Models: A Survey and Forward Path
Vipul Gupta
Pranav Narayanan Venkit
Shomir Wilson
R. Passonneau
42
20
0
13 Jun 2023
Controlling Learned Effects to Reduce Spurious Correlations in Text Classifiers
Parikshit Bansal
Amit Sharma
CML
24
5
0
26 May 2023
Interpretable Outlier Summarization
Yu Wang
Lei Cao
Yizhou Yan
Samuel Madden
16
0
0
11 Mar 2023
Synthetic Pre-Training Tasks for Neural Machine Translation
Zexue He
Graeme W. Blackwood
Rameswar Panda
Julian McAuley
Rogerio Feris
21
3
0
19 Dec 2022
InterFair: Debiasing with Natural Language Feedback for Fair Interpretable Predictions
Bodhisattwa Prasad Majumder
Zexue He
Julian McAuley
8
5
0
14 Oct 2022
Evaluating Debiasing Techniques for Intersectional Biases
Shivashankar Subramanian
Xudong Han
Timothy Baldwin
Trevor Cohn
Lea Frermann
95
49
0
21 Sep 2021
The Woman Worked as a Babysitter: On Biases in Language Generation
Emily Sheng
Kai-Wei Chang
Premkumar Natarajan
Nanyun Peng
214
616
0
03 Sep 2019
1