Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2410.17519
Cited By
v1
v2 (latest)
Large Language Models Still Exhibit Bias in Long Text
23 October 2024
Wonje Jeung
Dongjae Jeon
Ashkan Yousefpour
Jonghyun Choi
ALM
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Large Language Models Still Exhibit Bias in Long Text"
16 / 16 papers shown
Title
SAFEPATH: Preventing Harmful Reasoning in Chain-of-Thought via Early Alignment
Wonje Jeung
Sangyeon Yoon
Minsuk Kahng
Albert No
LRM
LLMSV
165
1
0
20 May 2025
SEPS: A Separability Measure for Robust Unlearning in LLMs
Wonje Jeung
Sangyeon Yoon
Albert No
MU
VLM
220
1
0
20 May 2025
GenderBench: Evaluation Suite for Gender Biases in LLMs
Matúš Pikuliak
79
0
0
17 May 2025
Improving Conversational Recommendation Systems via Bias Analysis and Language-Model-Enhanced Data Augmentation
Xi Wang
Hossein A. Rahmani
Jiqun Liu
Emine Yilmaz
73
12
0
25 Oct 2023
GPT-4 Technical Report
OpenAI OpenAI
OpenAI Josh Achiam
Steven Adler
Sandhini Agarwal
Lama Ahmad
...
Shengjia Zhao
Tianhao Zheng
Juntang Zhuang
William Zhuk
Barret Zoph
LLMAG
MLLM
1.5K
14,761
0
15 Mar 2023
Mitigating Language-Dependent Ethnic Bias in BERT
Jaimeen Ahn
Alice Oh
218
101
0
13 Sep 2021
Collecting a Large-Scale Gender Bias Dataset for Coreference Resolution and Machine Translation
Shahar Levy
Koren Lazar
Gabriel Stanovsky
67
70
0
08 Sep 2021
Finetuned Language Models Are Zero-Shot Learners
Jason W. Wei
Maarten Bosma
Vincent Zhao
Kelvin Guu
Adams Wei Yu
Brian Lester
Nan Du
Andrew M. Dai
Quoc V. Le
ALM
UQCV
251
3,789
0
03 Sep 2021
Process for Adapting Language Models to Society (PALMS) with Values-Targeted Datasets
Irene Solaiman
Christy Dennison
110
226
0
18 Jun 2021
RedditBias: A Real-World Resource for Bias Evaluation and Debiasing of Conversational Language Models
Soumya Barikeri
Anne Lauscher
Ivan Vulić
Goran Glavaš
96
184
0
07 Jun 2021
Measuring and Reducing Gendered Correlations in Pre-trained Models
Kellie Webster
Xuezhi Wang
Ian Tenney
Alex Beutel
Emily Pitler
Ellie Pavlick
Jilin Chen
Ed Chi
Slav Petrov
FaML
84
260
0
12 Oct 2020
UnQovering Stereotyping Biases via Underspecified Questions
Tao Li
Tushar Khot
Daniel Khashabi
Ashish Sabharwal
Vivek Srikumar
71
138
0
06 Oct 2020
Detecting Emergent Intersectional Biases: Contextualized Word Embeddings Contain a Distribution of Human-like Biases
W. Guo
Aylin Caliskan
39
245
0
06 Jun 2020
On Measuring and Mitigating Biased Inferences of Word Embeddings
Sunipa Dev
Tao Li
J. M. Phillips
Vivek Srikumar
81
174
0
25 Aug 2019
Measuring Bias in Contextualized Word Representations
Keita Kurita
Nidhi Vyas
Ayush Pareek
A. Black
Yulia Tsvetkov
116
453
0
18 Jun 2019
RACE: Large-scale ReAding Comprehension Dataset From Examinations
Guokun Lai
Qizhe Xie
Hanxiao Liu
Yiming Yang
Eduard H. Hovy
ELM
203
1,359
0
15 Apr 2017
1