ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1912.05671
  4. Cited By
Linear Mode Connectivity and the Lottery Ticket Hypothesis

Linear Mode Connectivity and the Lottery Ticket Hypothesis

11 December 2019
Jonathan Frankle
Gintare Karolina Dziugaite
Daniel M. Roy
Michael Carbin
    MoMe
ArXivPDFHTML

Papers citing "Linear Mode Connectivity and the Lottery Ticket Hypothesis"

50 / 172 papers shown
Title
Train ñ Trade: Foundations of Parameter Markets
Train ñ Trade: Foundations of Parameter Markets
Tzu-Heng Huang
Harit Vishwakarma
Frederic Sala
AIFin
37
2
0
07 Dec 2023
Efficient Rehearsal Free Zero Forgetting Continual Learning using
  Adaptive Weight Modulation
Efficient Rehearsal Free Zero Forgetting Continual Learning using Adaptive Weight Modulation
Yonatan Sverdlov
Shimon Ullman
34
0
0
26 Nov 2023
Language and Task Arithmetic with Parameter-Efficient Layers for
  Zero-Shot Summarization
Language and Task Arithmetic with Parameter-Efficient Layers for Zero-Shot Summarization
Alexandra Chronopoulou
Jonas Pfeiffer
Joshua Maynez
Xinyi Wang
Sebastian Ruder
Priyanka Agrawal
MoMe
26
16
0
15 Nov 2023
One is More: Diverse Perspectives within a Single Network for Efficient
  DRL
One is More: Diverse Perspectives within a Single Network for Efficient DRL
Yiqin Tan
Ling Pan
Longbo Huang
OffRL
43
0
0
21 Oct 2023
Model Merging by Uncertainty-Based Gradient Matching
Model Merging by Uncertainty-Based Gradient Matching
Nico Daheim
Thomas Möllenhoff
Edoardo Ponti
Iryna Gurevych
Mohammad Emtiyaz Khan
MoMe
FedML
37
45
0
19 Oct 2023
The Snowflake Hypothesis: Training Deep GNN with One Node One Receptive
  field
The Snowflake Hypothesis: Training Deep GNN with One Node One Receptive field
Kun Wang
Guohao Li
Shilong Wang
Guibin Zhang
Kaidi Wang
Yang You
Xiaojiang Peng
Keli Zhang
Yang Wang
42
8
0
19 Aug 2023
Layer-wise Linear Mode Connectivity
Layer-wise Linear Mode Connectivity
Linara Adilova
Maksym Andriushchenko
Michael Kamp
Asja Fischer
Martin Jaggi
FedML
FAtt
MoMe
43
15
0
13 Jul 2023
Distilled Pruning: Using Synthetic Data to Win the Lottery
Distilled Pruning: Using Synthetic Data to Win the Lottery
Luke McDermott
Daniel Cummings
SyDa
DD
34
1
0
07 Jul 2023
Improving Online Continual Learning Performance and Stability with
  Temporal Ensembles
Improving Online Continual Learning Performance and Stability with Temporal Ensembles
Albin Soutif--Cormerais
Antonio Carta
Joost van de Weijer
CLL
20
11
0
29 Jun 2023
Quantifying lottery tickets under label noise: accuracy, calibration,
  and complexity
Quantifying lottery tickets under label noise: accuracy, calibration, and complexity
V. Arora
Daniele Irto
Sebastian Goldt
G. Sanguinetti
52
2
0
21 Jun 2023
Make Pre-trained Model Reversible: From Parameter to Memory Efficient
  Fine-Tuning
Make Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuning
Baohao Liao
Shaomu Tan
Christof Monz
KELM
23
29
0
01 Jun 2023
Investigating how ReLU-networks encode symmetries
Investigating how ReLU-networks encode symmetries
Georg Bökman
Fredrik Kahl
31
6
0
26 May 2023
Sparse Weight Averaging with Multiple Particles for Iterative Magnitude
  Pruning
Sparse Weight Averaging with Multiple Particles for Iterative Magnitude Pruning
Moonseok Choi
Hyungi Lee
G. Nam
Juho Lee
40
2
0
24 May 2023
Subspace-Configurable Networks
Subspace-Configurable Networks
Dong Wang
O. Saukh
Xiaoxi He
Lothar Thiele
OOD
38
0
0
22 May 2023
Task Arithmetic in the Tangent Space: Improved Editing of Pre-Trained
  Models
Task Arithmetic in the Tangent Space: Improved Editing of Pre-Trained Models
Guillermo Ortiz-Jiménez
Alessandro Favero
P. Frossard
MoMe
51
113
0
22 May 2023
NTK-SAP: Improving neural network pruning by aligning training dynamics
NTK-SAP: Improving neural network pruning by aligning training dynamics
Yite Wang
Dawei Li
Ruoyu Sun
44
19
0
06 Apr 2023
On the Variance of Neural Network Training with respect to Test Sets and
  Distributions
On the Variance of Neural Network Training with respect to Test Sets and Distributions
Keller Jordan
OOD
24
11
0
04 Apr 2023
Exploring the Performance of Pruning Methods in Neural Networks: An
  Empirical Study of the Lottery Ticket Hypothesis
Exploring the Performance of Pruning Methods in Neural Networks: An Empirical Study of the Lottery Ticket Hypothesis
Eirik Fladmark
Muhammad Hamza Sajjad
Laura Brinkholm Justesen
28
2
0
26 Mar 2023
Generalization Matters: Loss Minima Flattening via Parameter
  Hybridization for Efficient Online Knowledge Distillation
Generalization Matters: Loss Minima Flattening via Parameter Hybridization for Efficient Online Knowledge Distillation
Tianli Zhang
Mengqi Xue
Jiangtao Zhang
Haofei Zhang
Yu Wang
Lechao Cheng
Mingli Song
Mingli Song
33
5
0
26 Mar 2023
Sparse-IFT: Sparse Iso-FLOP Transformations for Maximizing Training
  Efficiency
Sparse-IFT: Sparse Iso-FLOP Transformations for Maximizing Training Efficiency
Vithursan Thangarasa
Shreyas Saxena
Abhay Gupta
Sean Lie
41
3
0
21 Mar 2023
Achieving a Better Stability-Plasticity Trade-off via Auxiliary Networks
  in Continual Learning
Achieving a Better Stability-Plasticity Trade-off via Auxiliary Networks in Continual Learning
Sang-Ho Kim
Lorenzo Noci
Antonio Orvieto
Thomas Hofmann
CLL
37
35
0
16 Mar 2023
Understanding plasticity in neural networks
Understanding plasticity in neural networks
Clare Lyle
Zeyu Zheng
Evgenii Nikishin
Bernardo Avila-Pires
Razvan Pascanu
Will Dabney
AI4CE
45
98
0
02 Mar 2023
Average of Pruning: Improving Performance and Stability of
  Out-of-Distribution Detection
Average of Pruning: Improving Performance and Stability of Out-of-Distribution Detection
Zhen Cheng
Fei Zhu
Xu-Yao Zhang
Cheng-Lin Liu
MoMe
OODD
45
11
0
02 Mar 2023
DART: Diversify-Aggregate-Repeat Training Improves Generalization of
  Neural Networks
DART: Diversify-Aggregate-Repeat Training Improves Generalization of Neural Networks
Samyak Jain
Sravanti Addepalli
P. Sahu
Priyam Dey
R. Venkatesh Babu
MoMe
OOD
48
20
0
28 Feb 2023
Modular Deep Learning
Modular Deep Learning
Jonas Pfeiffer
Sebastian Ruder
Ivan Vulić
Edoardo Ponti
MoMe
OOD
34
73
0
22 Feb 2023
Considering Layerwise Importance in the Lottery Ticket Hypothesis
Considering Layerwise Importance in the Lottery Ticket Hypothesis
Benjamin Vandersmissen
José Oramas
37
1
0
22 Feb 2023
Simple Hardware-Efficient Long Convolutions for Sequence Modeling
Simple Hardware-Efficient Long Convolutions for Sequence Modeling
Daniel Y. Fu
Elliot L. Epstein
Eric N. D. Nguyen
A. Thomas
Michael Zhang
Tri Dao
Atri Rudra
Christopher Ré
25
52
0
13 Feb 2023
Quantum Neuron Selection: Finding High Performing Subnetworks With
  Quantum Algorithms
Quantum Neuron Selection: Finding High Performing Subnetworks With Quantum Algorithms
Tim Whitaker
33
1
0
12 Feb 2023
Knowledge is a Region in Weight Space for Fine-tuned Language Models
Knowledge is a Region in Weight Space for Fine-tuned Language Models
Almog Gueta
Elad Venezian
Colin Raffel
Noam Slonim
Yoav Katz
Leshem Choshen
36
50
0
09 Feb 2023
Why is the State of Neural Network Pruning so Confusing? On the
  Fairness, Comparison Setup, and Trainability in Network Pruning
Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Pruning
Huan Wang
Can Qin
Yue Bai
Yun Fu
37
20
0
12 Jan 2023
Training trajectories, mini-batch losses and the curious role of the
  learning rate
Training trajectories, mini-batch losses and the curious role of the learning rate
Mark Sandler
A. Zhmoginov
Max Vladymyrov
Nolan Miller
ODL
28
10
0
05 Jan 2023
Dataless Knowledge Fusion by Merging Weights of Language Models
Dataless Knowledge Fusion by Merging Weights of Language Models
Xisen Jin
Xiang Ren
Daniel Preoţiuc-Pietro
Pengxiang Cheng
FedML
MoMe
29
215
0
19 Dec 2022
Can We Find Strong Lottery Tickets in Generative Models?
Can We Find Strong Lottery Tickets in Generative Models?
Sangyeop Yeo
Yoojin Jang
Jy-yong Sohn
Dongyoon Han
Jaejun Yoo
20
6
0
16 Dec 2022
AP: Selective Activation for De-sparsifying Pruned Neural Networks
AP: Selective Activation for De-sparsifying Pruned Neural Networks
Shiyu Liu
Rohan Ghosh
Dylan Tan
Mehul Motani
AAML
26
0
0
09 Dec 2022
Editing Models with Task Arithmetic
Editing Models with Task Arithmetic
Gabriel Ilharco
Marco Tulio Ribeiro
Mitchell Wortsman
Suchin Gururangan
Ludwig Schmidt
Hannaneh Hajishirzi
Ali Farhadi
KELM
MoMe
MU
77
443
0
08 Dec 2022
ColD Fusion: Collaborative Descent for Distributed Multitask Finetuning
ColD Fusion: Collaborative Descent for Distributed Multitask Finetuning
Shachar Don-Yehiya
Elad Venezian
Colin Raffel
Noam Slonim
Yoav Katz
Leshem Choshen
MoMe
30
52
0
02 Dec 2022
The Effect of Data Dimensionality on Neural Network Prunability
The Effect of Data Dimensionality on Neural Network Prunability
Zachary Ankner
Alex Renda
Gintare Karolina Dziugaite
Jonathan Frankle
Tian Jin
36
5
0
01 Dec 2022
LU decomposition and Toeplitz decomposition of a neural network
LU decomposition and Toeplitz decomposition of a neural network
Yucong Liu
Simiao Jiao
Lek-Heng Lim
30
7
0
25 Nov 2022
Linear Interpolation In Parameter Space is Good Enough for Fine-Tuned
  Language Models
Linear Interpolation In Parameter Space is Good Enough for Fine-Tuned Language Models
Mark Rofin
Nikita Balagansky
Daniil Gavrilov
MoMe
KELM
38
5
0
22 Nov 2022
Mechanistic Mode Connectivity
Mechanistic Mode Connectivity
Ekdeep Singh Lubana
Eric J. Bigelow
Robert P. Dick
David M. Krueger
Hidenori Tanaka
34
45
0
15 Nov 2022
REPAIR: REnormalizing Permuted Activations for Interpolation Repair
REPAIR: REnormalizing Permuted Activations for Interpolation Repair
Keller Jordan
Hanie Sedghi
O. Saukh
R. Entezari
Behnam Neyshabur
MoMe
46
94
0
15 Nov 2022
AdaMix: Mixture-of-Adaptations for Parameter-efficient Model Tuning
Yaqing Wang
Sahaj Agarwal
Subhabrata Mukherjee
Xiaodong Liu
Jing Gao
Ahmed Hassan Awadallah
Jianfeng Gao
MoE
27
118
0
31 Oct 2022
Symmetries, flat minima, and the conserved quantities of gradient flow
Symmetries, flat minima, and the conserved quantities of gradient flow
Bo Zhao
I. Ganev
Robin Walters
Rose Yu
Nima Dehmamy
57
16
0
31 Oct 2022
Exploring Mode Connectivity for Pre-trained Language Models
Exploring Mode Connectivity for Pre-trained Language Models
Yujia Qin
Cheng Qian
Jing Yi
Weize Chen
Yankai Lin
Xu Han
Zhiyuan Liu
Maosong Sun
Jie Zhou
31
20
0
25 Oct 2022
lo-fi: distributed fine-tuning without communication
lo-fi: distributed fine-tuning without communication
Mitchell Wortsman
Suchin Gururangan
Shen Li
Ali Farhadi
Ludwig Schmidt
Michael G. Rabbat
Ari S. Morcos
34
24
0
19 Oct 2022
Pareto Manifold Learning: Tackling multiple tasks via ensembles of
  single-task models
Pareto Manifold Learning: Tackling multiple tasks via ensembles of single-task models
Nikolaos Dimitriadis
P. Frossard
Franccois Fleuret
36
25
0
18 Oct 2022
Advancing Model Pruning via Bi-level Optimization
Advancing Model Pruning via Bi-level Optimization
Yihua Zhang
Yuguang Yao
Parikshit Ram
Pu Zhao
Tianlong Chen
Min-Fong Hong
Yanzhi Wang
Sijia Liu
56
68
0
08 Oct 2022
Unmasking the Lottery Ticket Hypothesis: What's Encoded in a Winning
  Ticket's Mask?
Unmasking the Lottery Ticket Hypothesis: What's Encoded in a Winning Ticket's Mask?
Mansheej Paul
F. Chen
Brett W. Larsen
Jonathan Frankle
Surya Ganguli
Gintare Karolina Dziugaite
UQCV
42
38
0
06 Oct 2022
Stochastic optimization on matrices and a graphon McKean-Vlasov limit
Stochastic optimization on matrices and a graphon McKean-Vlasov limit
Zaïd Harchaoui
Sewoong Oh
Soumik Pal
Raghav Somani
Raghavendra Tripathi
36
2
0
02 Oct 2022
On the Factory Floor: ML Engineering for Industrial-Scale Ads
  Recommendation Models
On the Factory Floor: ML Engineering for Industrial-Scale Ads Recommendation Models
Rohan Anil
S. Gadanho
Danya Huang
Nijith Jacob
Zhuoshu Li
...
Cristina Pop
Kevin Regan
G. Shamir
Rakesh Shivanna
Qiqi Yan
3DV
29
41
0
12 Sep 2022
Previous
1234
Next