ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2011.06709
  4. Cited By
Active Reinforcement Learning: Observing Rewards at a Cost
v1v2 (latest)

Active Reinforcement Learning: Observing Rewards at a Cost

13 November 2020
David M. Krueger
Jan Leike
Owain Evans
J. Salvatier
ArXiv (abs)PDFHTML

Papers citing "Active Reinforcement Learning: Observing Rewards at a Cost"

10 / 10 papers shown
Title
Batch Active Learning of Reward Functions from Human Preferences
Batch Active Learning of Reward Functions from Human Preferences
Erdem Biyik
Nima Anari
Dorsa Sadigh
85
9
0
24 Feb 2024
Reinforcement Learning from Human Feedback with Active Queries
Reinforcement Learning from Human Feedback with Active Queries
Kaixuan Ji
Jiafan He
Quanquan Gu
100
19
0
14 Feb 2024
Scientific Discovery and the Cost of Measurement -- Balancing
  Information and Cost in Reinforcement Learning
Scientific Discovery and the Cost of Measurement -- Balancing Information and Cost in Reinforcement Learning
C. Bellinger
Andriy Drozdyuk
Mark Crowley
Isaac Tamblyn
OffRL
33
7
0
14 Dec 2021
Reinforcement Learning for Selective Key Applications in Power Systems:
  Recent Advances and Future Challenges
Reinforcement Learning for Selective Key Applications in Power Systems: Recent Advances and Future Challenges
Xin Chen
Guannan Qu
Yujie Tang
S. Low
Na Li
77
237
0
27 Jan 2021
AI Research Considerations for Human Existential Safety (ARCHES)
AI Research Considerations for Human Existential Safety (ARCHES)
Andrew Critch
David M. Krueger
109
53
0
30 May 2020
Active Measure Reinforcement Learning for Observation Cost Minimization
Active Measure Reinforcement Learning for Observation Cost Minimization
C. Bellinger
Rory Coles
Mark Crowley
Isaac Tamblyn
OffRL
35
24
0
26 May 2020
Scalable agent alignment via reward modeling: a research direction
Scalable agent alignment via reward modeling: a research direction
Jan Leike
David M. Krueger
Tom Everitt
Miljan Martic
Vishal Maini
Shane Legg
124
420
0
19 Nov 2018
Batch Active Preference-Based Learning of Reward Functions
Batch Active Preference-Based Learning of Reward Functions
Erdem Biyik
Dorsa Sadigh
115
113
0
10 Oct 2018
Active Reinforcement Learning with Monte-Carlo Tree Search
Active Reinforcement Learning with Monte-Carlo Tree Search
Sebastian Schulze
Owain Evans
57
14
0
13 Mar 2018
Trial without Error: Towards Safe Reinforcement Learning via Human
  Intervention
Trial without Error: Towards Safe Reinforcement Learning via Human Intervention
William Saunders
Girish Sastry
Andreas Stuhlmuller
Owain Evans
OffRL
74
231
0
17 Jul 2017
1