ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.06709
16
0

Beyond O~(T)\tilde{O}(\sqrt{T})O~(T​) Constraint Violation for Online Convex Optimization with Adversarial Constraints

10 May 2025
Abhishek Sinha
Rahul Vaze
ArXivPDFHTML
Abstract

We revisit the Online Convex Optimization problem with adversarial constraints (COCO) where, in each round, a learner is presented with a convex cost function and a convex constraint function, both of which may be chosen adversarially. The learner selects actions from a convex decision set in an online fashion, with the goal of minimizing both regret and the cumulative constraint violation (CCV) over a horizon of TTT rounds. The best-known policy for this problem achieves O(T)O(\sqrt{T})O(T​) regret and O~(T)\tilde{O}(\sqrt{T})O~(T​) CCV. In this paper, we present a surprising improvement that achieves a significantly smaller CCV by trading it off with regret. Specifically, for any bounded convex cost and constraint functions, we propose an online policy that achieves O~(dT+Tβ)\tilde{O}(\sqrt{dT}+ T^\beta)O~(dT​+Tβ) regret and O~(dT1−β)\tilde{O}(dT^{1-\beta})O~(dT1−β) CCV, where ddd is the dimension of the decision set and β∈[0,1]\beta \in [0,1]β∈[0,1] is a tunable parameter. We achieve this result by first considering the special case of Constrained Expert\textsf{Constrained Expert}Constrained Expert problem where the decision set is a probability simplex and the cost and constraint functions are linear. Leveraging a new adaptive small-loss regret bound, we propose an efficient policy for the Constrained Expert\textsf{Constrained Expert}Constrained Expert problem, that attains O(Tln⁡N+Tβ)O(\sqrt{T\ln N}+T^{\beta})O(TlnN​+Tβ) regret and O~(T1−βln⁡N)\tilde{O}(T^{1-\beta} \ln N)O~(T1−βlnN) CCV, where NNN is the number of experts. The original problem is then reduced to the Constrained Expert\textsf{Constrained Expert}Constrained Expert problem via a covering argument. Finally, with an additional smoothness assumption, we propose an efficient gradient-based policy attaining O(Tmax⁡(12,β))O(T^{\max(\frac{1}{2},\beta)})O(Tmax(21​,β)) regret and O~(T1−β)\tilde{O}(T^{1-\beta})O~(T1−β) CCV.

View on arXiv
@article{sinha2025_2505.06709,
  title={ Beyond $\tilde{O}(\sqrt{T})$ Constraint Violation for Online Convex Optimization with Adversarial Constraints },
  author={ Abhishek Sinha and Rahul Vaze },
  journal={arXiv preprint arXiv:2505.06709},
  year={ 2025 }
}
Comments on this paper