Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2410.02657
Cited By
Hate Personified: Investigating the role of LLMs in content moderation
3 October 2024
Sarah Masud
Sahajpreet Singh
Viktor Hangya
Alexander Fraser
Tanmoy Chakraborty
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Hate Personified: Investigating the role of LLMs in content moderation"
6 / 6 papers shown
Title
Can Prompting LLMs Unlock Hate Speech Detection across Languages? A Zero-shot and Few-shot Study
Faeze Ghorbanpour
Daryna Dementieva
Alexander Fraser
51
0
0
09 May 2025
Out of Sight Out of Mind, Out of Sight Out of Mind: Measuring Bias in Language Models Against Overlooked Marginalized Groups in Regional Contexts
Fatma Elsafoury
David Hartmann
34
0
0
17 Apr 2025
Selective Demonstration Retrieval for Improved Implicit Hate Speech Detection
Yumin Kim
Hwanhee Lee
38
0
0
16 Apr 2025
LLM-C3MOD: A Human-LLM Collaborative System for Cross-Cultural Hate Speech Moderation
Junyeong Park
Seogyeong Jeong
Shri Kiran Srinivasan
Yohan Lee
Alice Oh
57
1
0
10 Mar 2025
Revealing Hidden Mechanisms of Cross-Country Content Moderation with Natural Language Processing
Neemesh Yadav
Jiarui Liu
Francesco Ortu
Roya Ensafi
Zhijing Jin
Rada Mihalcea
41
0
0
07 Mar 2025
Extreme Speech Classification in the Era of LLMs: Exploring Open-Source and Proprietary Models
Sarthak Mahajan
Nimmi Rangaswamy
57
0
0
24 Feb 2025
1