ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2410.07826
  4. Cited By
Fine-Tuning Language Models for Ethical Ambiguity: A Comparative Study
  of Alignment with Human Responses

Fine-Tuning Language Models for Ethical Ambiguity: A Comparative Study of Alignment with Human Responses

10 October 2024
Pranav Senthilkumar
Visshwa Balasubramanian
Prisha Jain
Aneesa Maity
Jonathan Lu
Kevin Zhu
ArXiv (abs)PDFHTML

Papers citing "Fine-Tuning Language Models for Ethical Ambiguity: A Comparative Study of Alignment with Human Responses"

2 / 2 papers shown
Title
Direct Preference Optimization: Your Language Model is Secretly a Reward
  Model
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Rafael Rafailov
Archit Sharma
E. Mitchell
Stefano Ermon
Christopher D. Manning
Chelsea Finn
ALM
387
4,139
0
29 May 2023
Aligning AI With Shared Human Values
Aligning AI With Shared Human Values
Dan Hendrycks
Collin Burns
Steven Basart
Andrew Critch
Jingkai Li
Basel Alomair
Jacob Steinhardt
145
569
0
05 Aug 2020
1