ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2310.18955
17
13

Optimal Algorithms for Online Convex Optimization with Adversarial Constraints

29 October 2023
Abhishek Sinha
Rahul Vaze
ArXivPDFHTML
Abstract

A well-studied generalization of the standard online convex optimization (OCO) is constrained online convex optimization (COCO). In COCO, on every round, a convex cost function and a convex constraint function are revealed to the learner after the action for that round is chosen. The objective is to design an online policy that simultaneously achieves a small regret while ensuring a small cumulative constraint violation (CCV) against an adaptive adversary interacting over a horizon of length TTT. A long-standing open question in COCO is whether an online policy can simultaneously achieve O(T)O(\sqrt{T})O(T​) regret and O(T)O(\sqrt{T})O(T​) CCV without any restrictive assumptions. For the first time, we answer this in the affirmative and show that an online policy can simultaneously achieve O(T)O(\sqrt{T})O(T​) regret and O~(T)\tilde{O}(\sqrt{T})O~(T​) CCV. Furthermore, in the case of strongly convex cost and convex constraint functions, the regret guarantee can be improved to O(log⁡T)O(\log T)O(logT) while keeping the CCV bound the same as above. We establish these results by effectively combining the adaptive regret bound of the AdaGrad algorithm with Lyapunov optimization - a classic tool from control theory. Surprisingly, the analysis is short and elegant.

View on arXiv
Comments on this paper