ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2204.02515
  4. Cited By
Inferring Rewards from Language in Context

Inferring Rewards from Language in Context

5 April 2022
Jessy Lin
Daniel Fried
Dan Klein
Anca Dragan
    LM&Ro
ArXivPDFHTML

Papers citing "Inferring Rewards from Language in Context"

41 / 41 papers shown
Title
NatSGLD: A Dataset with Speech, Gesture, Logic, and Demonstration for Robot Learning in Natural Human-Robot Interaction
NatSGLD: A Dataset with Speech, Gesture, Logic, and Demonstration for Robot Learning in Natural Human-Robot Interaction
Snehesh Shrestha
Yantian Zha
Saketh Banagiri
Ge Gao
Yiannis Aloimonos
Cornelia Fermüller
41
0
0
23 Feb 2025
Generalist Virtual Agents: A Survey on Autonomous Agents Across Digital Platforms
Minghe Gao
Wendong Bu
Bingchen Miao
Yang Wu
Yunfei Li
Juncheng Billy Li
Siliang Tang
Qi Wu
Yueting Zhuang
Meng Wang
LM&Ro
39
3
0
17 Nov 2024
Need Help? Designing Proactive AI Assistants for Programming
Need Help? Designing Proactive AI Assistants for Programming
Valerie Chen
Alan Zhu
Sebastian Zhao
Hussein Mozannar
David Sontag
Ameet Talwalkar
37
5
0
06 Oct 2024
Controllable Traffic Simulation through LLM-Guided Hierarchical
  Chain-of-Thought Reasoning
Controllable Traffic Simulation through LLM-Guided Hierarchical Chain-of-Thought Reasoning
Zhiyuan Liu
Leheng Li
Yuning Wang
Haotian Lin
Zhizhe Liu
Lei He
Jianqiang Wang
LRM
31
2
0
23 Sep 2024
Beyond Preferences in AI Alignment
Beyond Preferences in AI Alignment
Tan Zhi-Xuan
Micah Carroll
Matija Franklin
Hal Ashton
35
16
0
30 Aug 2024
Problem Solving Through Human-AI Preference-Based Cooperation
Problem Solving Through Human-AI Preference-Based Cooperation
Subhabrata Dutta
Timo Kaufmann
Goran Glavas
Ivan Habernal
Kristian Kersting
Frauke Kreuter
Mira Mezini
Iryna Gurevych
Eyke Hüllermeier
Hinrich Schuetze
95
1
0
14 Aug 2024
Building Machines that Learn and Think with People
Building Machines that Learn and Think with People
Katherine M. Collins
Ilia Sucholutsky
Umang Bhatt
Kartik Chandra
Lionel Wong
...
Mark K. Ho
Vikash K. Mansinghka
Adrian Weller
Joshua B. Tenenbaum
Thomas L. Griffiths
47
30
0
22 Jul 2024
BigCodeBench: Benchmarking Code Generation with Diverse Function Calls and Complex Instructions
BigCodeBench: Benchmarking Code Generation with Diverse Function Calls and Complex Instructions
Terry Yue Zhuo
Minh Chien Vu
Jenny Chim
Han Hu
Wenhao Yu
...
David Lo
Daniel Fried
Xiaoning Du
H. D. Vries
Leandro von Werra
71
129
0
22 Jun 2024
LGR2: Language Guided Reward Relabeling for Accelerating Hierarchical Reinforcement Learning
LGR2: Language Guided Reward Relabeling for Accelerating Hierarchical Reinforcement Learning
Utsav Singh
Pramit Bhattacharyya
Vinay P. Namboodiri
LM&Ro
42
1
0
09 Jun 2024
Representational Alignment Supports Effective Machine Teaching
Representational Alignment Supports Effective Machine Teaching
Ilia Sucholutsky
Katherine M. Collins
Maya Malaviya
Nori Jacoby
Weiyang Liu
...
J. Tenenbaum
Brad Love
Z. Pardos
Adrian Weller
Thomas L. Griffiths
54
3
0
06 Jun 2024
Pragmatic Feature Preferences: Learning Reward-Relevant Preferences from
  Human Input
Pragmatic Feature Preferences: Learning Reward-Relevant Preferences from Human Input
Andi Peng
Yuying Sun
Tianmin Shu
David Abel
38
3
0
23 May 2024
Bayesian Preference Elicitation with Language Models
Bayesian Preference Elicitation with Language Models
Kunal Handa
Yarin Gal
Ellie Pavlick
Noah D. Goodman
Jacob Andreas
Alex Tamkin
Belinda Z. Li
34
12
0
08 Mar 2024
Is this the real life? Is this just fantasy? The Misleading Success of
  Simulating Social Interactions With LLMs
Is this the real life? Is this just fantasy? The Misleading Success of Simulating Social Interactions With LLMs
Xuhui Zhou
Zhe Su
Tiwalayo Eisape
Hyunwoo J. Kim
Maarten Sap
34
37
0
08 Mar 2024
Pragmatic Instruction Following and Goal Assistance via Cooperative
  Language-Guided Inverse Planning
Pragmatic Instruction Following and Goal Assistance via Cooperative Language-Guided Inverse Planning
Zhi-Xuan Tan
Lance Ying
Vikash K. Mansinghka
Joshua B. Tenenbaum
38
22
0
27 Feb 2024
Professional Agents -- Evolving Large Language Models into Autonomous
  Experts with Human-Level Competencies
Professional Agents -- Evolving Large Language Models into Autonomous Experts with Human-Level Competencies
Zhixuan Chu
Yan Wang
Feng Zhu
Lu Yu
Longfei Li
Jinjie Gu
LLMAG
20
8
0
06 Feb 2024
Preference-Conditioned Language-Guided Abstraction
Preference-Conditioned Language-Guided Abstraction
Andi Peng
Andreea Bobu
Belinda Z. Li
T. Sumers
Ilia Sucholutsky
Nishanth Kumar
Thomas L. Griffiths
Julie A. Shah
24
12
0
05 Feb 2024
Computational Experiments Meet Large Language Model Based Agents: A
  Survey and Perspective
Computational Experiments Meet Large Language Model Based Agents: A Survey and Perspective
Qun Ma
Xiao Xue
Deyu Zhou
Xiangning Yu
Donghua Liu
...
Yifan Shen
Peilin Ji
Juanjuan Li
Gang Wang
Wanpeng Ma
AI4CE
LM&Ro
LLMAG
13
7
0
01 Feb 2024
RePLan: Robotic Replanning with Perception and Language Models
RePLan: Robotic Replanning with Perception and Language Models
Marta Skreta
Zihan Zhou
Jia Lin Yuan
Kourosh Darvish
Alán Aspuru-Guzik
Animesh Garg
LM&Ro
LRM
37
26
0
08 Jan 2024
Auto MC-Reward: Automated Dense Reward Design with Large Language Models
  for Minecraft
Auto MC-Reward: Automated Dense Reward Design with Large Language Models for Minecraft
Hao Li
Xue Yang
Zhaokai Wang
Xizhou Zhu
Jie Zhou
Yu Qiao
Xiaogang Wang
Hongsheng Li
Lewei Lu
Jifeng Dai
35
32
0
14 Dec 2023
FoMo Rewards: Can we cast foundation models as reward functions?
FoMo Rewards: Can we cast foundation models as reward functions?
Ekdeep Singh Lubana
Johann Brehmer
P. D. Haan
Taco S. Cohen
OffRL
LRM
48
2
0
06 Dec 2023
Make a Donut: Hierarchical EMD-Space Planning for Zero-Shot Deformable Manipulation with Tools
Make a Donut: Hierarchical EMD-Space Planning for Zero-Shot Deformable Manipulation with Tools
Yang You
Bokui Shen
Congyue Deng
Haoran Geng
Songlin Wei
He-Nan Wang
Leonidas J. Guibas
26
1
0
05 Nov 2023
LanguageMPC: Large Language Models as Decision Makers for Autonomous Driving
LanguageMPC: Large Language Models as Decision Makers for Autonomous Driving
Hao Sha
Yao Mu
Yuxuan Jiang
Li Chen
Chenfeng Xu
Ping Luo
Shengbo Eben Li
Masayoshi Tomizuka
Wei Zhan
Mingyu Ding
114
159
0
04 Oct 2023
Improve the efficiency of deep reinforcement learning through semantic
  exploration guided by natural language
Improve the efficiency of deep reinforcement learning through semantic exploration guided by natural language
Zhourui Guo
Meng Yao
Yang Yu
Qiyue Yin
OnRL
28
1
0
21 Sep 2023
Amortizing Pragmatic Program Synthesis with Rankings
Yewen Pu
Saujas Vaduguru
Priyan Vaithilingam
Elena L. Glassman
Daniel Fried
20
3
0
01 Sep 2023
Open Problems and Fundamental Limitations of Reinforcement Learning from
  Human Feedback
Open Problems and Fundamental Limitations of Reinforcement Learning from Human Feedback
Stephen Casper
Xander Davies
Claudia Shi
T. Gilbert
Jérémy Scheurer
...
Erdem Biyik
Anca Dragan
David M. Krueger
Dorsa Sadigh
Dylan Hadfield-Menell
ALM
OffRL
47
470
0
27 Jul 2023
Language to Rewards for Robotic Skill Synthesis
Language to Rewards for Robotic Skill Synthesis
Wenhao Yu
Nimrod Gileadi
Chuyuan Fu
Sean Kirmani
Kuang-Huei Lee
...
N. Heess
Dorsa Sadigh
Jie Tan
Yuval Tassa
F. Xia
LM&Ro
35
269
0
14 Jun 2023
Decision-Oriented Dialogue for Human-AI Collaboration
Decision-Oriented Dialogue for Human-AI Collaboration
Jessy Lin
Nicholas Tomlin
Jacob Andreas
J. Eisner
LLMAG
22
26
0
31 May 2023
A Reminder of its Brittleness: Language Reward Shaping May Hinder
  Learning for Instruction Following Agents
A Reminder of its Brittleness: Language Reward Shaping May Hinder Learning for Instruction Following Agents
Sukai Huang
N. Lipovetzky
Trevor Cohn
30
2
0
26 May 2023
Inverse Preference Learning: Preference-based RL without a Reward
  Function
Inverse Preference Learning: Preference-based RL without a Reward Function
Joey Hejna
Dorsa Sadigh
OffRL
24
48
0
24 May 2023
Training Language Models with Language Feedback at Scale
Training Language Models with Language Feedback at Scale
Jérémy Scheurer
Jon Ander Campos
Tomasz Korbak
Jun Shern Chan
Angelica Chen
Kyunghyun Cho
Ethan Perez
ALM
39
101
0
28 Mar 2023
Reward Design with Language Models
Reward Design with Language Models
Minae Kwon
Sang Michael Xie
Kalesha Bullard
Dorsa Sadigh
LM&Ro
25
200
0
27 Feb 2023
Grounding Language Models to Images for Multimodal Inputs and Outputs
Grounding Language Models to Images for Multimodal Inputs and Outputs
Jing Yu Koh
Ruslan Salakhutdinov
Daniel Fried
MLLM
28
117
0
31 Jan 2023
Distilling Internet-Scale Vision-Language Models into Embodied Agents
Distilling Internet-Scale Vision-Language Models into Embodied Agents
T. Sumers
Kenneth Marino
Arun Ahuja
Rob Fergus
Ishita Dasgupta
LM&Ro
26
24
0
29 Jan 2023
Large Language Models as Fiduciaries: A Case Study Toward Robustly
  Communicating With Artificial Intelligence Through Legal Standards
Large Language Models as Fiduciaries: A Case Study Toward Robustly Communicating With Artificial Intelligence Through Legal Standards
John J. Nay
ELM
AILaw
27
15
0
24 Jan 2023
Define, Evaluate, and Improve Task-Oriented Cognitive Capabilities for
  Instruction Generation Models
Define, Evaluate, and Improve Task-Oriented Cognitive Capabilities for Instruction Generation Models
Lingjun Zhao
Khanh Nguyen
Hal Daumé
ELM
30
6
0
21 Dec 2022
Pragmatics in Language Grounding: Phenomena, Tasks, and Modeling
  Approaches
Pragmatics in Language Grounding: Phenomena, Tasks, and Modeling Approaches
Daniel Fried
Nicholas Tomlin
Jennifer Hu
Roma Patel
Aida Nematzadeh
19
6
0
15 Nov 2022
Overcoming Referential Ambiguity in Language-Guided Goal-Conditioned
  Reinforcement Learning
Overcoming Referential Ambiguity in Language-Guided Goal-Conditioned Reinforcement Learning
Hugo Caselles-Dupré
Olivier Sigaud
Mohamed Chetouani
17
2
0
26 Sep 2022
Law Informs Code: A Legal Informatics Approach to Aligning Artificial
  Intelligence with Humans
Law Informs Code: A Legal Informatics Approach to Aligning Artificial Intelligence with Humans
John J. Nay
ELM
AILaw
88
27
0
14 Sep 2022
How to talk so AI will learn: Instructions, descriptions, and autonomy
How to talk so AI will learn: Instructions, descriptions, and autonomy
T. Sumers
Robert D. Hawkins
Mark K. Ho
Thomas L. Griffiths
Dylan Hadfield-Menell
LM&Ro
28
20
0
16 Jun 2022
Training Language Models with Language Feedback
Training Language Models with Language Feedback
Jérémy Scheurer
Jon Ander Campos
Jun Shern Chan
Angelica Chen
Kyunghyun Cho
Ethan Perez
ALM
36
47
0
29 Apr 2022
Speaker-Follower Models for Vision-and-Language Navigation
Speaker-Follower Models for Vision-and-Language Navigation
Daniel Fried
Ronghang Hu
Volkan Cirik
Anna Rohrbach
Jacob Andreas
Louis-Philippe Morency
Taylor Berg-Kirkpatrick
Kate Saenko
Dan Klein
Trevor Darrell
LM&Ro
LRM
257
496
0
07 Jun 2018
1