ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2405.11454
29
1

Comparisons Are All You Need for Optimizing Smooth Functions

19 May 2024
Chenyi Zhang
Tongyang Li
    AAML
ArXivPDFHTML
Abstract

When optimizing machine learning models, there are various scenarios where gradient computations are challenging or even infeasible. Furthermore, in reinforcement learning (RL), preference-based RL that only compares between options has wide applications, including reinforcement learning with human feedback in large language models. In this paper, we systematically study optimization of a smooth function f ⁣:Rn→Rf\colon\mathbb{R}^n\to\mathbb{R}f:Rn→R only assuming an oracle that compares function values at two points and tells which is larger. When fff is convex, we give two algorithms using O~(n/ϵ)\tilde{O}(n/\epsilon)O~(n/ϵ) and O~(n2)\tilde{O}(n^{2})O~(n2) comparison queries to find an ϵ\epsilonϵ-optimal solution, respectively. When fff is nonconvex, our algorithm uses O~(n/ϵ2)\tilde{O}(n/\epsilon^2)O~(n/ϵ2) comparison queries to find an ϵ\epsilonϵ-approximate stationary point. All these results match the best-known zeroth-order algorithms with function evaluation queries in nnn dependence, thus suggest that \emph{comparisons are all you need for optimizing smooth functions using derivative-free methods}. In addition, we also give an algorithm for escaping saddle points and reaching an ϵ\epsilonϵ-second order stationary point of a nonconvex fff, using O~(n1.5/ϵ2.5)\tilde{O}(n^{1.5}/\epsilon^{2.5})O~(n1.5/ϵ2.5) comparison queries.

View on arXiv
Comments on this paper