ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2010.00892
  4. Cited By
Variance-Reduced Methods for Machine Learning

Variance-Reduced Methods for Machine Learning

2 October 2020
Robert Mansel Gower
Mark Schmidt
Francis R. Bach
Peter Richtárik
ArXiv (abs)PDFHTML

Papers citing "Variance-Reduced Methods for Machine Learning"

50 / 59 papers shown
Title
AutoSGD: Automatic Learning Rate Selection for Stochastic Gradient Descent
AutoSGD: Automatic Learning Rate Selection for Stochastic Gradient Descent
Nikola Surjanovic
Alexandre Bouchard-Côté
Trevor Campbell
20
0
0
27 May 2025
LLaDA 1.5: Variance-Reduced Preference Optimization for Large Language Diffusion Models
LLaDA 1.5: Variance-Reduced Preference Optimization for Large Language Diffusion Models
Fengqi Zhu
Rongzhen Wang
Shen Nie
Xiaolu Zhang
Chunwei Wu
...
Jun Zhou
Jianfei Chen
Yankai Lin
Ji-Rong Wen
Chongxuan Li
190
2
0
25 May 2025
Variance Reduction Methods Do Not Need to Compute Full Gradients: Improved Efficiency through Shuffling
Variance Reduction Methods Do Not Need to Compute Full Gradients: Improved Efficiency through Shuffling
Daniil Medyakov
Gleb Molodtsov
S. Chezhegov
Alexey Rebrikov
Aleksandr Beznosikov
147
0
0
21 Feb 2025
Analysis of regularized federated learning
Analysis of regularized federated learning
Langming Liu
Dingxuan Zhou
FedML
31
0
0
03 Nov 2024
Efficient Optimization Algorithms for Linear Adversarial Training
Efficient Optimization Algorithms for Linear Adversarial Training
Antônio H. Ribeiro
Thomas B. Schon
Dave Zahariah
Francis Bach
AAML
107
2
0
16 Oct 2024
Gap-Dependent Bounds for Q-Learning using Reference-Advantage Decomposition
Gap-Dependent Bounds for Q-Learning using Reference-Advantage Decomposition
Zhong Zheng
Haochen Zhang
Lingzhou Xue
OffRL
136
4
0
10 Oct 2024
Dataset Distillation via Knowledge Distillation: Towards Efficient Self-Supervised Pre-Training of Deep Networks
Dataset Distillation via Knowledge Distillation: Towards Efficient Self-Supervised Pre-Training of Deep Networks
S. Joshi
Jiayi Ni
Baharan Mirzasoleiman
DD
180
2
0
03 Oct 2024
On the SAGA algorithm with decreasing step
On the SAGA algorithm with decreasing step
Luis Fredes
Bernard Bercu
Eméric Gbaguidi
51
1
0
02 Oct 2024
On the Convergence of FedProx with Extrapolation and Inexact Prox
On the Convergence of FedProx with Extrapolation and Inexact Prox
Hanmin Li
Peter Richtárik
FedML
88
1
0
02 Oct 2024
An Effective Dynamic Gradient Calibration Method for Continual Learning
An Effective Dynamic Gradient Calibration Method for Continual Learning
Weichen Lin
Jiaxiang Chen
Ru Huang
Huihua Ding
CLL
94
0
0
30 Jul 2024
SOREL: A Stochastic Algorithm for Spectral Risks Minimization
SOREL: A Stochastic Algorithm for Spectral Risks Minimization
Yuze Ge
Rujun Jiang
65
0
0
19 Jul 2024
Unbiased least squares regression via averaged stochastic gradient
  descent
Unbiased least squares regression via averaged stochastic gradient descent
Nabil Kahalé
78
0
0
26 Jun 2024
Communication-Efficient Adaptive Batch Size Strategies for Distributed
  Local Gradient Methods
Communication-Efficient Adaptive Batch Size Strategies for Distributed Local Gradient Methods
Tim Tsz-Kit Lau
Weijian Li
Chenwei Xu
Han Liu
Mladen Kolar
84
1
0
20 Jun 2024
Demystifying SGD with Doubly Stochastic Gradients
Demystifying SGD with Doubly Stochastic Gradients
Kyurae Kim
Joohwan Ko
Yian Ma
Jacob R. Gardner
140
2
0
03 Jun 2024
Double Variance Reduction: A Smoothing Trick for Composite Optimization
  Problems without First-Order Gradient
Double Variance Reduction: A Smoothing Trick for Composite Optimization Problems without First-Order Gradient
Hao Di
Haishan Ye
Yueling Zhang
Xiangyu Chang
Guang Dai
Ivor W. Tsang
123
1
0
28 May 2024
A Unified Theory of Stochastic Proximal Point Methods without Smoothness
A Unified Theory of Stochastic Proximal Point Methods without Smoothness
Peter Richtárik
Abdurakhmon Sadiev
Yury Demidovich
80
4
0
24 May 2024
Streamlining in the Riemannian Realm: Efficient Riemannian Optimization
  with Loopless Variance Reduction
Streamlining in the Riemannian Realm: Efficient Riemannian Optimization with Loopless Variance Reduction
Yury Demidovich
Grigory Malinovsky
Peter Richtárik
83
3
0
11 Mar 2024
LoCoDL: Communication-Efficient Distributed Learning with Local Training and Compression
LoCoDL: Communication-Efficient Distributed Learning with Local Training and Compression
Laurent Condat
Artavazd Maranjyan
Peter Richtárik
123
5
0
07 Mar 2024
AdAdaGrad: Adaptive Batch Size Schemes for Adaptive Gradient Methods
AdAdaGrad: Adaptive Batch Size Schemes for Adaptive Gradient Methods
Tim Tsz-Kit Lau
Han Liu
Mladen Kolar
ODL
70
6
0
17 Feb 2024
Critical Influence of Overparameterization on Sharpness-aware Minimization
Critical Influence of Overparameterization on Sharpness-aware Minimization
Sungbin Shin
Dongyeop Lee
Maksym Andriushchenko
Namhoon Lee
AAML
158
2
0
29 Nov 2023
A Large Deviations Perspective on Policy Gradient Algorithms
A Large Deviations Perspective on Policy Gradient Algorithms
Wouter Jongeneel
Daniel Kuhn
Mengmeng Li
55
1
0
13 Nov 2023
A Coefficient Makes SVRG Effective
A Coefficient Makes SVRG Effective
Yida Yin
Zhiqiu Xu
Zhiyuan Li
Trevor Darrell
Zhuang Liu
87
1
0
09 Nov 2023
Balancing Act: Constraining Disparate Impact in Sparse Models
Balancing Act: Constraining Disparate Impact in Sparse Models
Meraj Hashemizadeh
Juan Ramirez
Rohan Sukumaran
G. Farnadi
Simon Lacoste-Julien
Jose Gallego-Posada
60
6
0
31 Oct 2023
Coreset Markov Chain Monte Carlo
Coreset Markov Chain Monte Carlo
Naitong Chen
Trevor Campbell
47
4
0
25 Oct 2023
Variance Reduced Halpern Iteration for Finite-Sum Monotone Inclusions
Variance Reduced Halpern Iteration for Finite-Sum Monotone Inclusions
Xu Cai
Ahmet Alacaoglu
Jelena Diakonikolas
99
8
0
04 Oct 2023
Joint Sampling and Optimisation for Inverse Rendering
Joint Sampling and Optimisation for Inverse Rendering
Martin Balint
K. Myszkowski
Hans-Peter Seidel
Gurprit Singh
19
1
0
27 Sep 2023
Linear Convergence of Black-Box Variational Inference: Should We Stick
  the Landing?
Linear Convergence of Black-Box Variational Inference: Should We Stick the Landing?
Kyurae Kim
Yian Ma
Jacob R. Gardner
88
7
0
27 Jul 2023
Towards a Better Theoretical Understanding of Independent Subnetwork
  Training
Towards a Better Theoretical Understanding of Independent Subnetwork Training
Egor Shulgin
Peter Richtárik
AI4CE
106
6
0
28 Jun 2023
Adaptive Federated Learning with Auto-Tuned Clients
Adaptive Federated Learning with Auto-Tuned Clients
Junhyung Lyle Kim
Taha Toghani
César A. Uribe
Anastasios Kyrillidis
FedML
124
6
0
19 Jun 2023
$\textbf{A}^2\textbf{CiD}^2$: Accelerating Asynchronous Communication in
  Decentralized Deep Learning
A2CiD2\textbf{A}^2\textbf{CiD}^2A2CiD2: Accelerating Asynchronous Communication in Decentralized Deep Learning
Adel Nabli
Eugene Belilovsky
Edouard Oyallon
74
7
0
14 Jun 2023
Sketch-and-Project Meets Newton Method: Global $\mathcal O(k^{-2})$
  Convergence with Low-Rank Updates
Sketch-and-Project Meets Newton Method: Global O(k−2)\mathcal O(k^{-2})O(k−2) Convergence with Low-Rank Updates
Slavomír Hanzely
51
7
0
22 May 2023
Unified analysis of SGD-type methods
Unified analysis of SGD-type methods
Eduard A. Gorbunov
69
2
0
29 Mar 2023
Byzantine-Robust Loopless Stochastic Variance-Reduced Gradient
Byzantine-Robust Loopless Stochastic Variance-Reduced Gradient
Nikita Fedin
Eduard A. Gorbunov
44
2
0
08 Mar 2023
Stochastic Approximation Beyond Gradient for Signal Processing and
  Machine Learning
Stochastic Approximation Beyond Gradient for Signal Processing and Machine Learning
Aymeric Dieuleveut
G. Fort
Eric Moulines
Hoi-To Wai
74
12
0
22 Feb 2023
TAMUNA: Doubly Accelerated Distributed Optimization with Local Training,
  Compression, and Partial Participation
TAMUNA: Doubly Accelerated Distributed Optimization with Local Training, Compression, and Partial Participation
Laurent Condat
Ivan Agarský
Grigory Malinovsky
Peter Richtárik
FedML
108
4
0
20 Feb 2023
A One-Sample Decentralized Proximal Algorithm for Non-Convex Stochastic
  Composite Optimization
A One-Sample Decentralized Proximal Algorithm for Non-Convex Stochastic Composite Optimization
Tesi Xiao
Xuxing Chen
Krishnakumar Balasubramanian
Saeed Ghadimi
94
10
0
20 Feb 2023
Balance is Essence: Accelerating Sparse Training via Adaptive Gradient
  Correction
Balance is Essence: Accelerating Sparse Training via Adaptive Gradient Correction
Bowen Lei
Dongkuan Xu
Ruqi Zhang
Shuren He
Bani Mallick
117
6
0
09 Jan 2023
Xtreme Margin: A Tunable Loss Function for Binary Classification
  Problems
Xtreme Margin: A Tunable Loss Function for Binary Classification Problems
Rayan Wali
MQ
99
4
0
31 Oct 2022
Provably Doubly Accelerated Federated Learning: The First Theoretically
  Successful Combination of Local Training and Communication Compression
Provably Doubly Accelerated Federated Learning: The First Theoretically Successful Combination of Local Training and Communication Compression
Laurent Condat
Ivan Agarský
Peter Richtárik
FedML
113
17
0
24 Oct 2022
Joint control variate for faster black-box variational inference
Joint control variate for faster black-box variational inference
Xi Wang
Tomas Geffner
Justin Domke
BDLDRL
39
0
0
13 Oct 2022
RECAPP: Crafting a More Efficient Catalyst for Convex Optimization
RECAPP: Crafting a More Efficient Catalyst for Convex Optimization
Y. Carmon
A. Jambulapati
Yujia Jin
Aaron Sidford
125
12
0
17 Jun 2022
Variance Reduction is an Antidote to Byzantines: Better Rates, Weaker
  Assumptions and Communication Compression as a Cherry on the Top
Variance Reduction is an Antidote to Byzantines: Better Rates, Weaker Assumptions and Communication Compression as a Cherry on the Top
Eduard A. Gorbunov
Samuel Horváth
Peter Richtárik
Gauthier Gidel
AAML
38
0
0
01 Jun 2022
EF-BV: A Unified Theory of Error Feedback and Variance Reduction
  Mechanisms for Biased and Unbiased Compression in Distributed Optimization
EF-BV: A Unified Theory of Error Feedback and Variance Reduction Mechanisms for Biased and Unbiased Compression in Distributed Optimization
Laurent Condat
Kai Yi
Peter Richtárik
107
21
0
09 May 2022
A Statistical Analysis of Polyak-Ruppert Averaged Q-learning
A Statistical Analysis of Polyak-Ruppert Averaged Q-learning
Xiang Li
Wenhao Yang
Jiadong Liang
Zhihua Zhang
Michael I. Jordan
122
17
0
29 Dec 2021
LoSAC: An Efficient Local Stochastic Average Control Method for
  Federated Optimization
LoSAC: An Efficient Local Stochastic Average Control Method for Federated Optimization
Huiming Chen
Huandong Wang
Quanming Yao
Yong Li
Depeng Jin
Qiang Yang
FedML
54
5
0
15 Dec 2021
Nearly Optimal Linear Convergence of Stochastic Primal-Dual Methods for
  Linear Programming
Nearly Optimal Linear Convergence of Stochastic Primal-Dual Methods for Linear Programming
Haihao Lu
Jinwen Yang
52
6
0
10 Nov 2021
Gradient Descent on Infinitely Wide Neural Networks: Global Convergence
  and Generalization
Gradient Descent on Infinitely Wide Neural Networks: Global Convergence and Generalization
Francis R. Bach
Lénaïc Chizat
MLT
67
24
0
15 Oct 2021
Breaking the Sample Complexity Barrier to Regret-Optimal Model-Free
  Reinforcement Learning
Breaking the Sample Complexity Barrier to Regret-Optimal Model-Free Reinforcement Learning
Gen Li
Laixi Shi
Yuxin Chen
Yuejie Chi
OffRL
94
54
0
09 Oct 2021
MURANA: A Generic Framework for Stochastic Variance-Reduced Optimization
MURANA: A Generic Framework for Stochastic Variance-Reduced Optimization
Laurent Condat
Peter Richtárik
73
19
0
06 Jun 2021
Random Reshuffling with Variance Reduction: New Analysis and Better
  Rates
Random Reshuffling with Variance Reduction: New Analysis and Better Rates
Grigory Malinovsky
Alibek Sailanbayev
Peter Richtárik
49
21
0
19 Apr 2021
12
Next