ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2404.06720
  4. Cited By
Gradient Descent is Pareto-Optimal in the Oracle Complexity and Memory
  Tradeoff for Feasibility Problems

Gradient Descent is Pareto-Optimal in the Oracle Complexity and Memory Tradeoff for Feasibility Problems

10 April 2024
Moise Blanchard
ArXivPDFHTML

Papers citing "Gradient Descent is Pareto-Optimal in the Oracle Complexity and Memory Tradeoff for Feasibility Problems"

1 / 1 papers shown
Title
Efficient Convex Optimization Requires Superlinear Memory
Efficient Convex Optimization Requires Superlinear Memory
A. Marsden
Vatsal Sharan
Aaron Sidford
Gregory Valiant
29
14
0
29 Mar 2022
1