ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2006.07897
  4. Cited By
Entropic gradient descent algorithms and wide flat minima

Entropic gradient descent algorithms and wide flat minima

14 June 2020
Fabrizio Pittorino
C. Lucibello
Christoph Feinauer
Gabriele Perugini
Carlo Baldassi
Elizaveta Demyanenko
R. Zecchina
    ODL
    MLT
ArXivPDFHTML

Papers citing "Entropic gradient descent algorithms and wide flat minima"

10 / 10 papers shown
Title
SEAL: Searching Expandable Architectures for Incremental Learning
SEAL: Searching Expandable Architectures for Incremental Learning
Matteo Gambella
Vicente Javier Castro Solar
Manuel Roveri
CLL
36
0
0
15 May 2025
Revisiting Confidence Estimation: Towards Reliable Failure Prediction
Revisiting Confidence Estimation: Towards Reliable Failure Prediction
Fei Zhu
Xu-Yao Zhang
Zhen Cheng
Cheng-Lin Liu
UQCV
57
10
0
05 Mar 2024
FlatNAS: optimizing Flatness in Neural Architecture Search for
  Out-of-Distribution Robustness
FlatNAS: optimizing Flatness in Neural Architecture Search for Out-of-Distribution Robustness
Matteo Gambella
Fabrizio Pittorino
Manuel Roveri
OOD
44
3
0
29 Feb 2024
Phase transitions in the mini-batch size for sparse and dense two-layer
  neural networks
Phase transitions in the mini-batch size for sparse and dense two-layer neural networks
Raffaele Marino
F. Ricci-Tersenghi
32
14
0
10 May 2023
Rethinking Confidence Calibration for Failure Prediction
Rethinking Confidence Calibration for Failure Prediction
Fei Zhu
Zhen Cheng
Xu-Yao Zhang
Cheng-Lin Liu
UQCV
22
39
0
06 Mar 2023
REPAIR: REnormalizing Permuted Activations for Interpolation Repair
REPAIR: REnormalizing Permuted Activations for Interpolation Repair
Keller Jordan
Hanie Sedghi
O. Saukh
R. Entezari
Behnam Neyshabur
MoMe
46
94
0
15 Nov 2022
Understanding the effect of sparsity on neural networks robustness
Understanding the effect of sparsity on neural networks robustness
Lukas Timpl
R. Entezari
Hanie Sedghi
Behnam Neyshabur
O. Saukh
51
12
0
22 Jun 2022
Learning through atypical "phase transitions" in overparameterized
  neural networks
Learning through atypical "phase transitions" in overparameterized neural networks
Carlo Baldassi
Clarissa Lauditi
Enrico M. Malatesta
R. Pacelli
Gabriele Perugini
R. Zecchina
36
26
0
01 Oct 2021
Memorizing without overfitting: Bias, variance, and interpolation in
  over-parameterized models
Memorizing without overfitting: Bias, variance, and interpolation in over-parameterized models
J. Rocks
Pankaj Mehta
23
41
0
26 Oct 2020
On Large-Batch Training for Deep Learning: Generalization Gap and Sharp
  Minima
On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima
N. Keskar
Dheevatsa Mudigere
J. Nocedal
M. Smelyanskiy
P. T. P. Tang
ODL
310
2,896
0
15 Sep 2016
1