Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2405.17713
Cited By
AI Alignment with Changing and Influenceable Reward Functions
28 May 2024
Micah Carroll
Davis Foote
Anand Siththaranjan
Stuart J. Russell
Anca Dragan
Re-assign community
ArXiv
PDF
HTML
Papers citing
"AI Alignment with Changing and Influenceable Reward Functions"
4 / 4 papers shown
Title
A Survey on Personalized Alignment -- The Missing Piece for Large Language Models in Real-World Applications
Jian Guan
Jian Wu
Jia-Nan Li
Chuanqi Cheng
Wei Wu
LM&MA
108
1
0
21 Mar 2025
Aligning Generalisation Between Humans and Machines
Filip Ilievski
Barbara Hammer
F. V. Harmelen
Benjamin Paassen
S. Saralajew
...
Vered Shwartz
Gabriella Skitalinskaya
Clemens Stachl
Gido M. van de Ven
T. Villmann
226
1
0
23 Nov 2024
Problem Solving Through Human-AI Preference-Based Cooperation
Subhabrata Dutta
Timo Kaufmann
Goran Glavaš
Ivan Habernal
Kristian Kersting
Frauke Kreuter
Mira Mezini
Iryna Gurevych
Eyke Hüllermeier
Hinrich Schuetze
135
1
0
14 Aug 2024
Right Now, Wrong Then: Non-Stationary Direct Preference Optimization under Preference Drift
Seongho Son
William Bankes
Sayak Ray Chowdhury
Brooks Paige
Ilija Bogunovic
71
4
0
26 Jul 2024
1