ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2205.06811
  4. Cited By
Nearly Optimal Algorithms for Linear Contextual Bandits with Adversarial
  Corruptions

Nearly Optimal Algorithms for Linear Contextual Bandits with Adversarial Corruptions

13 May 2022
Jiafan He
Dongruo Zhou
Tong Zhang
Quanquan Gu
ArXivPDFHTML

Papers citing "Nearly Optimal Algorithms for Linear Contextual Bandits with Adversarial Corruptions"

42 / 42 papers shown
Title
Provably Efficient Multi-Objective Bandit Algorithms under Preference-Centric Customization
Provably Efficient Multi-Objective Bandit Algorithms under Preference-Centric Customization
Linfeng Cao
Ming Shi
Ness B. Shroff
44
0
0
20 Feb 2025
Linear Bandits with Partially Observable Features
Wonyoung Hedge Kim
Sungwoo Park
G. Iyengar
A. Zeevi
Min Hwan Oh
53
0
0
10 Feb 2025
Catoni Contextual Bandits are Robust to Heavy-tailed Rewards
Catoni Contextual Bandits are Robust to Heavy-tailed Rewards
Chenlu Ye
Yujia Jin
Alekh Agarwal
Tong Zhang
98
0
0
04 Feb 2025
A Model Selection Approach for Corruption Robust Reinforcement Learning
A Model Selection Approach for Corruption Robust Reinforcement Learning
Chen-Yu Wei
Christoph Dann
Julian Zimmert
80
44
0
31 Dec 2024
Robust Thompson Sampling Algorithms Against Reward Poisoning Attacks
Robust Thompson Sampling Algorithms Against Reward Poisoning Attacks
Yinglun Xu
Zhiwei Wang
Gagandeep Singh
AAML
23
0
0
25 Oct 2024
How Does Variance Shape the Regret in Contextual Bandits?
How Does Variance Shape the Regret in Contextual Bandits?
Zeyu Jia
Jian Qian
Alexander Rakhlin
Chen-Yu Wei
35
4
0
16 Oct 2024
Corruption-Robust Linear Bandits: Minimax Optimality and Gap-Dependent
  Misspecification
Corruption-Robust Linear Bandits: Minimax Optimality and Gap-Dependent Misspecification
Haolin Liu
Artin Tajdini
Andrew Wagenmaker
Chen-Yu Wei
29
0
0
10 Oct 2024
Strategic Linear Contextual Bandits
Strategic Linear Contextual Bandits
Thomas Kleine Buening
Aadirupa Saha
Christos Dimitrakakis
Haifeng Xu
38
0
0
01 Jun 2024
Non-stochastic Bandits With Evolving Observations
Non-stochastic Bandits With Evolving Observations
Yogev Bar-On
Yishay Mansour
27
1
0
27 May 2024
Learning from Imperfect Human Feedback: a Tale from Corruption-Robust
  Dueling
Learning from Imperfect Human Feedback: a Tale from Corruption-Robust Dueling
Yuwei Cheng
Fan Yao
Xuefeng Liu
Haifeng Xu
38
1
0
18 May 2024
Improved Bound for Robust Causal Bandits with Linear Models
Improved Bound for Robust Causal Bandits with Linear Models
Zirui Yan
Arpan Mukherjee
Burak Varici
A. Tajer
25
1
0
13 May 2024
Nearly Optimal Algorithms for Contextual Dueling Bandits from Adversarial Feedback
Nearly Optimal Algorithms for Contextual Dueling Bandits from Adversarial Feedback
Qiwei Di
Jiafan He
Quanquan Gu
29
1
0
16 Apr 2024
Incentivized Learning in Principal-Agent Bandit Games
Incentivized Learning in Principal-Agent Bandit Games
Antoine Scheid
D. Tiapkin
Etienne Boursier
Aymeric Capitaine
El-Mahdi El-Mhamdi
Eric Moulines
Michael I. Jordan
Alain Durmus
35
6
0
06 Mar 2024
LC-Tsallis-INF: Generalized Best-of-Both-Worlds Linear Contextual
  Bandits
LC-Tsallis-INF: Generalized Best-of-Both-Worlds Linear Contextual Bandits
Masahiro Kato
Shinji Ito
34
0
0
05 Mar 2024
Corruption-Robust Offline Two-Player Zero-Sum Markov Games
Corruption-Robust Offline Two-Player Zero-Sum Markov Games
Andi Nika
Debmalya Mandal
Adish Singla
Goran Radanović
OffRL
32
1
0
04 Mar 2024
Towards Robust Model-Based Reinforcement Learning Against Adversarial
  Corruption
Towards Robust Model-Based Reinforcement Learning Against Adversarial Corruption
Chen Ye
Jiafan He
Quanquan Gu
Tong Zhang
43
5
0
14 Feb 2024
Reinforcement Learning from Human Feedback with Active Queries
Reinforcement Learning from Human Feedback with Active Queries
Kaixuan Ji
Jiafan He
Quanquan Gu
20
17
0
14 Feb 2024
Best-of-Both-Worlds Linear Contextual Bandits
Best-of-Both-Worlds Linear Contextual Bandits
Masahiro Kato
Shinji Ito
43
0
0
27 Dec 2023
Best-of-Both-Worlds Algorithms for Linear Contextual Bandits
Best-of-Both-Worlds Algorithms for Linear Contextual Bandits
Yuko Kuroki
Alberto Rumi
Taira Tsuchiya
Fabio Vitale
Nicolò Cesa-Bianchi
31
5
0
24 Dec 2023
Robust Best-arm Identification in Linear Bandits
Robust Best-arm Identification in Linear Bandits
Wei Wang
Sattar Vakili
Ilija Bogunovic
15
0
0
08 Nov 2023
Federated Linear Bandits with Finite Adversarial Actions
Federated Linear Bandits with Finite Adversarial Actions
Li Fan
Ruida Zhou
Chao Tian
Cong Shen
FedML
94
2
0
02 Nov 2023
Robust Causal Bandits for Linear Models
Robust Causal Bandits for Linear Models
Zirui Yan
Arpan Mukherjee
Burak Varici
A. Tajer
CML
22
4
0
30 Oct 2023
Efficient Algorithms for Generalized Linear Bandits with Heavy-tailed
  Rewards
Efficient Algorithms for Generalized Linear Bandits with Heavy-tailed Rewards
Bo Xue
Yimu Wang
Yuanyu Wan
Jinfeng Yi
Lijun Zhang
33
3
0
28 Oct 2023
Corruption-Robust Offline Reinforcement Learning with General Function
  Approximation
Corruption-Robust Offline Reinforcement Learning with General Function Approximation
Chen Ye
Rui Yang
Quanquan Gu
Tong Zhang
OffRL
33
17
0
23 Oct 2023
Online Corrupted User Detection and Regret Minimization
Online Corrupted User Detection and Regret Minimization
Zhiyong Wang
Jize Xie
Tong Yu
Shuai Li
J. C. Lui
24
6
0
07 Oct 2023
Optimal Sample Selection Through Uncertainty Estimation and Its
  Application in Deep Learning
Optimal Sample Selection Through Uncertainty Estimation and Its Application in Deep Learning
Yong Lin
Chen Liu
Chen Ye
Qing Lian
Yuan Yao
Tong Zhang
27
4
0
05 Sep 2023
Corruption-Robust Lipschitz Contextual Search
Corruption-Robust Lipschitz Contextual Search
Shiliang Zuo
19
1
0
26 Jul 2023
Robust Lipschitz Bandits to Adversarial Corruptions
Robust Lipschitz Bandits to Adversarial Corruptions
Yue Kang
Cho-Jui Hsieh
T. C. Lee
AAML
24
8
0
29 May 2023
Learning to Seek: Multi-Agent Online Source Seeking Against
  Non-Stochastic Disturbances
Learning to Seek: Multi-Agent Online Source Seeking Against Non-Stochastic Disturbances
Bin Du
Kun Qian
Christian G. Claudel
Dengfeng Sun
16
0
0
29 Apr 2023
On the Interplay Between Misspecification and Sub-optimality Gap in
  Linear Contextual Bandits
On the Interplay Between Misspecification and Sub-optimality Gap in Linear Contextual Bandits
Weitong Zhang
Jiafan He
Zhiyuan Fan
Q. Gu
92
5
0
16 Mar 2023
Best-of-three-worlds Analysis for Linear Bandits with
  Follow-the-regularized-leader Algorithm
Best-of-three-worlds Analysis for Linear Bandits with Follow-the-regularized-leader Algorithm
Fang-yuan Kong
Canzhe Zhao
Shuai Li
35
11
0
13 Mar 2023
Variance-Dependent Regret Bounds for Linear Bandits and Reinforcement
  Learning: Adaptivity and Computational Efficiency
Variance-Dependent Regret Bounds for Linear Bandits and Reinforcement Learning: Adaptivity and Computational Efficiency
Heyang Zhao
Jiafan He
Dongruo Zhou
Tong Zhang
Quanquan Gu
24
27
0
21 Feb 2023
Near-Optimal Adversarial Reinforcement Learning with Switching Costs
Near-Optimal Adversarial Reinforcement Learning with Switching Costs
Ming Shi
Yitao Liang
Ness B. Shroff
28
2
0
08 Feb 2023
Nearly Minimax Optimal Reinforcement Learning for Linear Markov Decision
  Processes
Nearly Minimax Optimal Reinforcement Learning for Linear Markov Decision Processes
Jiafan He
Heyang Zhao
Dongruo Zhou
Quanquan Gu
OffRL
51
53
0
12 Dec 2022
Corruption-Robust Algorithms with Uncertainty Weighting for Nonlinear
  Contextual Bandits and Markov Decision Processes
Corruption-Robust Algorithms with Uncertainty Weighting for Nonlinear Contextual Bandits and Markov Decision Processes
Chen Ye
Wei Xiong
Quanquan Gu
Tong Zhang
23
29
0
12 Dec 2022
Contexts can be Cheap: Solving Stochastic Contextual Bandits with Linear
  Bandit Algorithms
Contexts can be Cheap: Solving Stochastic Contextual Bandits with Linear Bandit Algorithms
Osama A. Hanna
Lin F. Yang
Christina Fragouli
27
11
0
08 Nov 2022
Collaborative Linear Bandits with Adversarial Agents: Near-Optimal
  Regret Bounds
Collaborative Linear Bandits with Adversarial Agents: Near-Optimal Regret Bounds
A. Mitra
Arman Adibi
George J. Pappas
Hamed Hassani
44
6
0
06 Jun 2022
Computationally Efficient Horizon-Free Reinforcement Learning for Linear
  Mixture MDPs
Computationally Efficient Horizon-Free Reinforcement Learning for Linear Mixture MDPs
Dongruo Zhou
Quanquan Gu
73
43
0
23 May 2022
A Robust Phased Elimination Algorithm for Corruption-Tolerant Gaussian
  Process Bandits
A Robust Phased Elimination Algorithm for Corruption-Tolerant Gaussian Process Bandits
Ilija Bogunovic
Zihan Li
Andreas Krause
Jonathan Scarlett
17
7
0
03 Feb 2022
Learning Stochastic Shortest Path with Linear Function Approximation
Learning Stochastic Shortest Path with Linear Function Approximation
Steffen Czolbe
Jiafan He
Adrian V. Dalca
Quanquan Gu
36
30
0
25 Oct 2021
Near Optimal Adversarial Attacks on Stochastic Bandits and Defenses with
  Smoothed Responses
Near Optimal Adversarial Attacks on Stochastic Bandits and Defenses with Smoothed Responses
Shiliang Zuo
AAML
11
0
0
21 Aug 2020
Weighted Linear Bandits for Non-Stationary Environments
Weighted Linear Bandits for Non-Stationary Environments
Yoan Russac
Claire Vernade
Olivier Cappé
82
101
0
19 Sep 2019
1