ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2010.01412
  4. Cited By
Sharpness-Aware Minimization for Efficiently Improving Generalization

Sharpness-Aware Minimization for Efficiently Improving Generalization

3 October 2020
Pierre Foret
Ariel Kleiner
H. Mobahi
Behnam Neyshabur
    AAML
ArXivPDFHTML

Papers citing "Sharpness-Aware Minimization for Efficiently Improving Generalization"

50 / 867 papers shown
Title
Evaluating Bayes Error Estimators on Real-World Datasets with FeeBee
Evaluating Bayes Error Estimators on Real-World Datasets with FeeBee
Cédric Renggli
Luka Rimanic
Nora Hollenstein
Ce Zhang
17
10
0
30 Aug 2021
Partial Domain Adaptation without Domain Alignment
Partial Domain Adaptation without Domain Alignment
Weikai Li
Songcan Chen
26
13
0
29 Aug 2021
Bridged Adversarial Training
Bridged Adversarial Training
Hoki Kim
Woojin Lee
Sungyoon Lee
Jaewook Lee
AAML
GAN
16
9
0
25 Aug 2021
Towards Efficient and Data Agnostic Image Classification Training
  Pipeline for Embedded Systems
Towards Efficient and Data Agnostic Image Classification Training Pipeline for Embedded Systems
K. Prokofiev
V. Sovrasov
3DH
19
2
0
16 Aug 2021
Where do Models go Wrong? Parameter-Space Saliency Maps for
  Explainability
Where do Models go Wrong? Parameter-Space Saliency Maps for Explainability
Roman Levin
Manli Shu
Eitan Borgnia
Furong Huang
Micah Goldblum
Tom Goldstein
FAtt
AAML
20
10
0
03 Aug 2021
Taxonomizing local versus global structure in neural network loss
  landscapes
Taxonomizing local versus global structure in neural network loss landscapes
Yaoqing Yang
Liam Hodgkinson
Ryan Theisen
Joe Zou
Joseph E. Gonzalez
Kannan Ramchandran
Michael W. Mahoney
32
36
0
23 Jul 2021
Mediated Uncoupled Learning: Learning Functions without Direct
  Input-output Correspondences
Mediated Uncoupled Learning: Learning Functions without Direct Input-output Correspondences
Ikko Yamane
Junya Honda
Florian Yger
Masashi Sugiyama
SSL
FedML
OOD
16
1
0
16 Jul 2021
COVID Detection in Chest CTs: Improving the Baseline on COV19-CT-DB
COVID Detection in Chest CTs: Improving the Baseline on COV19-CT-DB
R. Miron
Cosmin Moisii
Sergiu-Andrei Dinu
Mihaela Breaban
16
6
0
10 Jul 2021
On Margins and Derandomisation in PAC-Bayes
On Margins and Derandomisation in PAC-Bayes
Felix Biggs
Benjamin Guedj
28
20
0
08 Jul 2021
Combining EfficientNet and Vision Transformers for Video Deepfake
  Detection
Combining EfficientNet and Vision Transformers for Video Deepfake Detection
D. Coccomini
Nicola Messina
Claudio Gennaro
Fabrizio Falchi
ViT
32
169
0
06 Jul 2021
Backward-Compatible Prediction Updates: A Probabilistic Approach
Backward-Compatible Prediction Updates: A Probabilistic Approach
Frederik Trauble
Julius von Kügelgen
Matthäus Kleindessner
Francesco Locatello
Bernhard Schölkopf
Peter V. Gehler
25
16
0
02 Jul 2021
ResViT: Residual vision transformers for multi-modal medical image
  synthesis
ResViT: Residual vision transformers for multi-modal medical image synthesis
Onat Dalmaz
Mahmut Yurt
Tolga Çukur
ViT
MedIm
32
338
0
30 Jun 2021
VOLO: Vision Outlooker for Visual Recognition
VOLO: Vision Outlooker for Visual Recognition
Li-xin Yuan
Qibin Hou
Zihang Jiang
Jiashi Feng
Shuicheng Yan
ViT
52
313
0
24 Jun 2021
Minimum sharpness: Scale-invariant parameter-robustness of neural
  networks
Minimum sharpness: Scale-invariant parameter-robustness of neural networks
Hikaru Ibayashi
Takuo Hamaguchi
Masaaki Imaizumi
25
5
0
23 Jun 2021
Graceful Degradation and Related Fields
Graceful Degradation and Related Fields
J. Dymond
31
4
0
21 Jun 2021
A Winning Hand: Compressing Deep Networks Can Improve
  Out-Of-Distribution Robustness
A Winning Hand: Compressing Deep Networks Can Improve Out-Of-Distribution Robustness
James Diffenderfer
Brian Bartoldson
Shreya Chaganti
Jize Zhang
B. Kailkhura
OOD
31
69
0
16 Jun 2021
Label Noise SGD Provably Prefers Flat Global Minimizers
Label Noise SGD Provably Prefers Flat Global Minimizers
Alexandru Damian
Tengyu Ma
Jason D. Lee
NoLa
29
113
0
11 Jun 2021
Learning distinct features helps, provably
Learning distinct features helps, provably
Firas Laakom
Jenni Raitoharju
Alexandros Iosifidis
Moncef Gabbouj
MLT
36
6
0
10 Jun 2021
Thompson Sampling with a Mixture Prior
Thompson Sampling with a Mixture Prior
Joey Hong
B. Kveton
Manzil Zaheer
Mohammad Ghavamzadeh
Craig Boutilier
13
12
0
10 Jun 2021
Digital Taxonomist: Identifying Plant Species in Community Scientists'
  Photographs
Digital Taxonomist: Identifying Plant Species in Community Scientists' Photographs
Riccardo de Lutio
Yihang She
Stefano Dáronco
Stefani A. Russo
P. Brun
Jan Dirk Wegner
Konrad Schindler
22
21
0
07 Jun 2021
Evaluating State-of-the-Art Classification Models Against Bayes
  Optimality
Evaluating State-of-the-Art Classification Models Against Bayes Optimality
Ryan Theisen
Huan Wang
L. Varshney
Caiming Xiong
R. Socher
13
9
0
07 Jun 2021
When Vision Transformers Outperform ResNets without Pre-training or
  Strong Data Augmentations
When Vision Transformers Outperform ResNets without Pre-training or Strong Data Augmentations
Xiangning Chen
Cho-Jui Hsieh
Boqing Gong
ViT
29
320
0
03 Jun 2021
Personalizing Pre-trained Models
Personalizing Pre-trained Models
Mina Khan
P. Srivatsa
Advait Rane
Shriram Chenniappa
A. Hazariwala
Pattie Maes
VLM
47
5
0
02 Jun 2021
Drawing Multiple Augmentation Samples Per Image During Training
  Efficiently Decreases Test Error
Drawing Multiple Augmentation Samples Per Image During Training Efficiently Decreases Test Error
Stanislav Fort
Andrew Brock
Razvan Pascanu
Soham De
Samuel L. Smith
21
31
0
27 May 2021
Using Early-Learning Regularization to Classify Real-World Noisy Data
Using Early-Learning Regularization to Classify Real-World Noisy Data
Alessio Galatolo
Alfred Nilsson
Roderick Karlemstrand
Yineng Wang
NoLa
11
1
0
27 May 2021
Blurs Behave Like Ensembles: Spatial Smoothings to Improve Accuracy,
  Uncertainty, and Robustness
Blurs Behave Like Ensembles: Spatial Smoothings to Improve Accuracy, Uncertainty, and Robustness
Namuk Park
S. Kim
UQCV
AAML
20
19
0
26 May 2021
Compressing Heavy-Tailed Weight Matrices for Non-Vacuous Generalization
  Bounds
Compressing Heavy-Tailed Weight Matrices for Non-Vacuous Generalization Bounds
John Y. Shin
16
5
0
23 May 2021
Understanding and Improvement of Adversarial Training for Network
  Embedding from an Optimization Perspective
Understanding and Improvement of Adversarial Training for Network Embedding from an Optimization Perspective
Lun Du
Xu Chen
Fei Gao
Qiang Fu
Kunqing Xie
Shi Han
Dongmei Zhang
18
12
0
17 May 2021
Vision Transformers are Robust Learners
Vision Transformers are Robust Learners
Sayak Paul
Pin-Yu Chen
ViT
28
304
0
17 May 2021
Boosting Light-Weight Depth Estimation Via Knowledge Distillation
Boosting Light-Weight Depth Estimation Via Knowledge Distillation
Junjie Hu
Chenyou Fan
Hualie Jiang
Xiyue Guo
Yuan Gao
Xiangyong Lu
Tin Lun Lam
15
26
0
13 May 2021
AdaBoost and robust one-bit compressed sensing
AdaBoost and robust one-bit compressed sensing
Geoffrey Chinot
Felix Kuchelmeister
Matthias Löffler
Sara van de Geer
32
5
0
05 May 2021
Exact Stochastic Second Order Deep Learning
Exact Stochastic Second Order Deep Learning
F. Mehouachi
C. Kasmi
ODL
14
0
0
08 Apr 2021
Shapley Explanation Networks
Shapley Explanation Networks
Rui Wang
Xiaoqian Wang
David I. Inouye
TDI
FAtt
19
44
0
06 Apr 2021
Defending Against Image Corruptions Through Adversarial Augmentations
Defending Against Image Corruptions Through Adversarial Augmentations
D. A. Calian
Florian Stimberg
Olivia Wiles
Sylvestre-Alvise Rebuffi
András Gyorgy
Timothy A. Mann
Sven Gowal
AAML
17
41
0
02 Apr 2021
Estimating the Generalization in Deep Neural Networks via Sparsity
Estimating the Generalization in Deep Neural Networks via Sparsity
Yang Zhao
Hao Zhang
40
2
0
02 Apr 2021
Facial expression and attributes recognition based on multi-task
  learning of lightweight neural networks
Facial expression and attributes recognition based on multi-task learning of lightweight neural networks
Andrey V. Savchenko
CVBM
3DH
24
128
0
31 Mar 2021
On the Adversarial Robustness of Vision Transformers
On the Adversarial Robustness of Vision Transformers
Rulin Shao
Zhouxing Shi
Jinfeng Yi
Pin-Yu Chen
Cho-Jui Hsieh
ViT
30
137
0
29 Mar 2021
Can Vision Transformers Learn without Natural Images?
Can Vision Transformers Learn without Natural Images?
Kodai Nakashima
Hirokatsu Kataoka
Asato Matsumoto
K. Iwata
Nakamasa Inoue
ViT
22
34
0
24 Mar 2021
How to decay your learning rate
How to decay your learning rate
Aitor Lewkowycz
36
24
0
23 Mar 2021
A Pilot Study For Fragment Identification Using 2D NMR and Deep Learning
A Pilot Study For Fragment Identification Using 2D NMR and Deep Learning
Stefan Kuhn
Eda Tumer
Simon Colreavy-Donnelly
R. Borges
24
13
0
18 Mar 2021
UPANets: Learning from the Universal Pixel Attention Networks
UPANets: Learning from the Universal Pixel Attention Networks
Ching-Hsun Tseng
Shin-Jye Lee
Jianxing Feng
Shengzhong Mao
Yuping Wu
Jia-Yu Shang
Mou-Chung Tseng
Xiao-Jun Zeng
14
15
0
15 Mar 2021
Cycle Self-Training for Domain Adaptation
Cycle Self-Training for Domain Adaptation
Hong Liu
Jianmin Wang
Mingsheng Long
36
174
0
05 Mar 2021
Formalizing Generalization and Robustness of Neural Networks to Weight
  Perturbations
Formalizing Generalization and Robustness of Neural Networks to Weight Perturbations
Yu-Lin Tsai
Chia-Yi Hsu
Chia-Mu Yu
Pin-Yu Chen
AAML
OOD
25
26
0
03 Mar 2021
Smoothness Analysis of Adversarial Training
Smoothness Analysis of Adversarial Training
Sekitoshi Kanai
Masanori Yamada
Hiroshi Takahashi
Yuki Yamanaka
Yasutoshi Ida
AAML
40
6
0
02 Mar 2021
Siamese Labels Auxiliary Learning
Siamese Labels Auxiliary Learning
Wenrui Gan
Zhulin Liu
Cheng Chen
Tong Zhang
17
1
0
27 Feb 2021
Multiplicative Reweighting for Robust Neural Network Optimization
Multiplicative Reweighting for Robust Neural Network Optimization
Noga Bar
Tomer Koren
Raja Giryes
OOD
NoLa
13
9
0
24 Feb 2021
ASAM: Adaptive Sharpness-Aware Minimization for Scale-Invariant Learning
  of Deep Neural Networks
ASAM: Adaptive Sharpness-Aware Minimization for Scale-Invariant Learning of Deep Neural Networks
Jungmin Kwon
Jeongseop Kim
Hyunseong Park
I. Choi
31
281
0
23 Feb 2021
The Uncanny Similarity of Recurrence and Depth
The Uncanny Similarity of Recurrence and Depth
Avi Schwarzschild
Arjun Gupta
Amin Ghiasi
Micah Goldblum
Tom Goldstein
25
10
0
22 Feb 2021
Learning Neural Network Subspaces
Learning Neural Network Subspaces
Mitchell Wortsman
Maxwell Horton
Carlos Guestrin
Ali Farhadi
Mohammad Rastegari
UQCV
27
85
0
20 Feb 2021
SWAD: Domain Generalization by Seeking Flat Minima
SWAD: Domain Generalization by Seeking Flat Minima
Junbum Cha
Sanghyuk Chun
Kyungjae Lee
Han-Cheol Cho
Seunghyun Park
Yunsung Lee
Sungrae Park
MoMe
216
423
0
17 Feb 2021
Previous
123...161718
Next