Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2210.07440
Cited By
InterFair: Debiasing with Natural Language Feedback for Fair Interpretable Predictions
14 October 2022
Bodhisattwa Prasad Majumder
Zexue He
Julian McAuley
Re-assign community
ArXiv
PDF
HTML
Papers citing
"InterFair: Debiasing with Natural Language Feedback for Fair Interpretable Predictions"
6 / 6 papers shown
Title
Mitigating Gender Bias in Code Large Language Models via Model Editing
Zhanyue Qin
Haochuan Wang
Zecheng Wang
Deyuan Liu
Cunhang Fan
Zhao Lv
Zhiying Tu
Dianhui Chu
Dianbo Sui
KELM
26
1
0
10 Oct 2024
Potential and Challenges of Model Editing for Social Debiasing
Jianhao Yan
Futing Wang
Yafu Li
Yue Zhang
KELM
68
9
0
21 Feb 2024
Self-Debiasing Large Language Models: Zero-Shot Recognition and Reduction of Stereotypes
Isabel O. Gallegos
Ryan A. Rossi
Joe Barrow
Md Mehrab Tanjim
Tong Yu
Hanieh Deilamsalehy
Ruiyi Zhang
Sungchul Kim
Franck Dernoncourt
24
19
0
03 Feb 2024
Bias and Fairness in Large Language Models: A Survey
Isabel O. Gallegos
Ryan A. Rossi
Joe Barrow
Md Mehrab Tanjim
Sungchul Kim
Franck Dernoncourt
Tong Yu
Ruiyi Zhang
Nesreen Ahmed
AILaw
26
490
0
02 Sep 2023
Sociodemographic Bias in Language Models: A Survey and Forward Path
Vipul Gupta
Pranav Narayanan Venkit
Shomir Wilson
R. Passonneau
42
20
0
13 Jun 2023
Fast Model Editing at Scale
E. Mitchell
Charles Lin
Antoine Bosselut
Chelsea Finn
Christopher D. Manning
KELM
230
343
0
21 Oct 2021
1