ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2207.12547
  4. Cited By
The BUTTER Zone: An Empirical Study of Training Dynamics in Fully
  Connected Neural Networks

The BUTTER Zone: An Empirical Study of Training Dynamics in Fully Connected Neural Networks

25 July 2022
Charles Edison Tripp
J. Perr-Sauer
L. Hayne
M. Lunacek
Jamil Gafur
    AI4CE
ArXivPDFHTML

Papers citing "The BUTTER Zone: An Empirical Study of Training Dynamics in Fully Connected Neural Networks"

14 / 14 papers shown
Title
Scaling Laws vs Model Architectures: How does Inductive Bias Influence
  Scaling?
Scaling Laws vs Model Architectures: How does Inductive Bias Influence Scaling?
Yi Tay
Mostafa Dehghani
Samira Abnar
Hyung Won Chung
W. Fedus
J. Rao
Sharan Narang
Vinh Q. Tran
Dani Yogatama
Donald Metzler
AI4CE
77
101
0
21 Jul 2022
TransNAS-Bench-101: Improving Transferability and Generalizability of
  Cross-Task Neural Architecture Search
TransNAS-Bench-101: Improving Transferability and Generalizability of Cross-Task Neural Architecture Search
Yawen Duan
Xin Chen
Hang Xu
Zewei Chen
Li Xiaodan
Tong Zhang
Zhenguo Li
ViT
41
70
0
25 May 2021
Prioritized Architecture Sampling with Monto-Carlo Tree Search
Prioritized Architecture Sampling with Monto-Carlo Tree Search
Xiu Su
Tao Huang
Yanxi Li
Shan You
Fei Wang
Chao Qian
Changshui Zhang
Chang Xu
24
49
0
22 Mar 2021
PMLB v1.0: An open source dataset collection for benchmarking machine
  learning methods
PMLB v1.0: An open source dataset collection for benchmarking machine learning methods
Joseph D. Romano
Trang T. Le
William La Cava
John T. Gregg
Daniel J. Goldberg
Natasha L. Ray
Praneel Chakraborty
Daniel Himmelstein
Weixuan Fu
J. Moore
GP
12
74
0
30 Nov 2020
Multiple Descent: Design Your Own Generalization Curve
Multiple Descent: Design Your Own Generalization Curve
Lin Chen
Yifei Min
M. Belkin
Amin Karbasi
DRL
49
61
0
03 Aug 2020
Scaling Laws for Neural Language Models
Scaling Laws for Neural Language Models
Jared Kaplan
Sam McCandlish
T. Henighan
Tom B. Brown
B. Chess
R. Child
Scott Gray
Alec Radford
Jeff Wu
Dario Amodei
435
4,662
0
23 Jan 2020
Fantastic Generalization Measures and Where to Find Them
Fantastic Generalization Measures and Where to Find Them
Yiding Jiang
Behnam Neyshabur
H. Mobahi
Dilip Krishnan
Samy Bengio
AI4CE
56
599
0
04 Dec 2019
A Constructive Prediction of the Generalization Error Across Scales
A Constructive Prediction of the Generalization Error Across Scales
Jonathan S. Rosenfeld
Amir Rosenfeld
Yonatan Belinkov
Nir Shavit
61
210
0
27 Sep 2019
Tabular Benchmarks for Joint Architecture and Hyperparameter
  Optimization
Tabular Benchmarks for Joint Architecture and Hyperparameter Optimization
Aaron Klein
Frank Hutter
24
92
0
13 May 2019
Reconciling modern machine learning practice and the bias-variance
  trade-off
Reconciling modern machine learning practice and the bias-variance trade-off
M. Belkin
Daniel J. Hsu
Siyuan Ma
Soumik Mandal
168
1,628
0
28 Dec 2018
Datasheets for Datasets
Datasheets for Datasets
Timnit Gebru
Jamie Morgenstern
Briana Vecchione
Jennifer Wortman Vaughan
Hanna M. Wallach
Hal Daumé
Kate Crawford
209
2,158
0
23 Mar 2018
Optimal approximation of continuous functions by very deep ReLU networks
Optimal approximation of continuous functions by very deep ReLU networks
Dmitry Yarotsky
111
293
0
10 Feb 2018
Implicit Regularization in Deep Learning
Implicit Regularization in Deep Learning
Behnam Neyshabur
32
145
0
06 Sep 2017
The Power of Depth for Feedforward Neural Networks
The Power of Depth for Feedforward Neural Networks
Ronen Eldan
Ohad Shamir
129
731
0
12 Dec 2015
1