ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1803.00885
  4. Cited By
Essentially No Barriers in Neural Network Energy Landscape
v1v2v3v4v5 (latest)

Essentially No Barriers in Neural Network Energy Landscape

2 March 2018
Felix Dräxler
K. Veschgini
M. Salmhofer
Fred Hamprecht
    MoMe
ArXiv (abs)PDFHTML

Papers citing "Essentially No Barriers in Neural Network Energy Landscape"

50 / 295 papers shown
Title
Hidden Breakthroughs in Language Model Training
Hidden Breakthroughs in Language Model Training
Sara Kangaslahti
Elan Rosenfeld
Naomi Saphra
28
0
0
18 Jun 2025
Symmetry in Neural Network Parameter Spaces
Symmetry in Neural Network Parameter Spaces
Bo Zhao
Robin Walters
Rose Yu
27
0
0
16 Jun 2025
Circumventing Backdoor Space via Weight Symmetry
Circumventing Backdoor Space via Weight Symmetry
Jie Peng
Hongwei Yang
Jing Zhao
Hengji Dong
Hui He
Weizhe Zhang
Haoyu He
AAML
17
0
0
09 Jun 2025
Walking the Weight Manifold: a Topological Approach to Conditioning Inspired by Neuromodulation
Walking the Weight Manifold: a Topological Approach to Conditioning Inspired by Neuromodulation
Ari S. Benjamin
Kyle Daruwalla
Christian Pehle
Anthony M. Zador
55
0
0
29 May 2025
Understanding Mode Connectivity via Parameter Space Symmetry
Understanding Mode Connectivity via Parameter Space Symmetry
B. Zhao
Nima Dehmamy
Robin Walters
Rose Yu
236
8
0
29 May 2025
Decom-Renorm-Merge: Model Merging on the Right Space Improves Multitasking
Decom-Renorm-Merge: Model Merging on the Right Space Improves Multitasking
Yuatyong Chaichana
Thanapat Trachu
Peerat Limkonchotiwat
Konpat Preechakul
Tirasan Khandhawit
Ekapol Chuangsuwanich
MoMe
82
0
0
29 May 2025
SGD as Free Energy Minimization: A Thermodynamic View on Neural Network Training
SGD as Free Energy Minimization: A Thermodynamic View on Neural Network Training
Ildus Sadrtdinov
Ivan Klimov
E. Lobacheva
Dmitry Vetrov
35
0
0
29 May 2025
Update Your Transformer to the Latest Release: Re-Basin of Task Vectors
Update Your Transformer to the Latest Release: Re-Basin of Task Vectors
Filippo Rinaldi
Giacomo Capitani
Lorenzo Bonicelli
Donato Crisostomi
Federico Bolelli
E. Ficarra
Emanuele Rodolà
Simone Calderara
Angelo Porrello
26
0
0
28 May 2025
Benignity of loss landscape with weight decay requires both large overparametrization and initialization
Benignity of loss landscape with weight decay requires both large overparametrization and initialization
Etienne Boursier
Matthew Bowditch
Matthias Englert
R. Lazic
56
0
0
28 May 2025
Semantic Aware Linear Transfer by Recycling Pre-trained Language Models for Cross-lingual Transfer
Semantic Aware Linear Transfer by Recycling Pre-trained Language Models for Cross-lingual Transfer
Seungyoon Lee
Seongtae Hong
Hyeonseok Moon
Heuiseok Lim
KELM
90
0
0
16 May 2025
Low-Loss Space in Neural Networks is Continuous and Fully Connected
Low-Loss Space in Neural Networks is Continuous and Fully Connected
Yongding Tian
Zaid Al-Ars
Maksim Kitsak
P. Hofstee
3DPC
116
1
0
05 May 2025
Towards Quantifying the Hessian Structure of Neural Networks
Towards Quantifying the Hessian Structure of Neural Networks
Zhaorui Dong
Yushun Zhang
Zhi-Quan Luo
Jianfeng Yao
Ruoyu Sun
77
1
0
05 May 2025
Dynamic Fisher-weighted Model Merging via Bayesian Optimization
Dynamic Fisher-weighted Model Merging via Bayesian Optimization
Sanwoo Lee
Jiahao Liu
Qifan Wang
Jiadong Wang
Xunliang Cai
Yunfang Wu
MoMe
464
1
0
26 Apr 2025
A Combinatorial Theory of Dropout: Subnetworks, Graph Geometry, and Generalization
A Combinatorial Theory of Dropout: Subnetworks, Graph Geometry, and Generalization
Sahil Rajesh Dhayalkar
130
1
0
20 Apr 2025
Understanding Machine Unlearning Through the Lens of Mode Connectivity
Understanding Machine Unlearning Through the Lens of Mode Connectivity
Jiali Cheng
Hadi Amiri
MU
458
1
0
08 Apr 2025
Finding Stable Subnetworks at Initialization with Dataset Distillation
Finding Stable Subnetworks at Initialization with Dataset Distillation
Luke McDermott
Rahul Parhi
DD
87
0
0
23 Mar 2025
Analyzing the Role of Permutation Invariance in Linear Mode Connectivity
Keyao Zhan
Puheng Li
Lei Wu
MoMe
113
0
0
13 Mar 2025
SplatPose: Geometry-Aware 6-DoF Pose Estimation from Single RGB Image via 3D Gaussian Splatting
Linqi Yang
Xiongwei Zhao
Qihao Sun
Ke Wang
Ao Chen
Peng Kang
3DGS
138
0
0
07 Mar 2025
Paths and Ambient Spaces in Neural Loss Landscapes
Daniel Dold
Julius Kobialka
Nicolai Palm
Emanuel Sommer
David Rügamer
Oliver Durr
AI4CE
126
0
0
05 Mar 2025
Deep Learning is Not So Mysterious or Different
Deep Learning is Not So Mysterious or Different
Andrew Gordon Wilson
101
6
0
03 Mar 2025
Extrapolating and Decoupling Image-to-Video Generation Models: Motion Modeling is Easier Than You Think
Jie Tian
Xiaoye Qu
Zhenyi Lu
Xiaoye Qu
Sichen Liu
Yu Cheng
DiffMVGen
81
4
0
02 Mar 2025
Low-Rank and Sparse Model Merging for Multi-Lingual Speech Recognition and Translation
Low-Rank and Sparse Model Merging for Multi-Lingual Speech Recognition and Translation
Qiuming Zhao
Guangzhi Sun
Chao Zhang
MoMeVLM
457
1
0
24 Feb 2025
Rethinking Spiking Neural Networks from an Ensemble Learning Perspective
Rethinking Spiking Neural Networks from an Ensemble Learning Perspective
Lin Zuo
Yongqi Ding
Mengmeng Jing
Pei He
Hanpu Deng
100
3
0
20 Feb 2025
High-dimensional manifold of solutions in neural networks: insights from statistical physics
High-dimensional manifold of solutions in neural networks: insights from statistical physics
Enrico M. Malatesta
114
4
0
20 Feb 2025
Unveiling Mode Connectivity in Graph Neural Networks
Unveiling Mode Connectivity in Graph Neural Networks
Bingheng Li
Z. Chen
Haoyu Han
Shenglai Zeng
J. Liu
Jiliang Tang
88
1
0
18 Feb 2025
In-Context Parametric Inference: Point or Distribution Estimators?
In-Context Parametric Inference: Point or Distribution Estimators?
Sarthak Mittal
Yoshua Bengio
Nikolay Malkin
Guillaume Lajoie
130
0
0
17 Feb 2025
Dissecting a Small Artificial Neural Network
Dissecting a Small Artificial Neural Network
Xiguang Yang
Krish Arora
Michael Bachmann
89
0
0
03 Jan 2025
Parameter-Efficient Interventions for Enhanced Model Merging
Parameter-Efficient Interventions for Enhanced Model Merging
Marcin Osial
Daniel Marczak
Bartosz Zieliñski
MoMe
150
1
0
22 Dec 2024
Non-Uniform Parameter-Wise Model Merging
Non-Uniform Parameter-Wise Model Merging
Albert Manuel Orozco Camacho
Stefan Horoi
Guy Wolf
Eugene Belilovsky
MoMeFedML
142
0
0
20 Dec 2024
Implicit Neural Compression of Point Clouds
Implicit Neural Compression of Point Clouds
Hongning Ruan
Yulin Shao
Qianqian Yang
Liang Zhao
Zhaoyang Zhang
Dusit Niyato
3DPC
116
1
0
11 Dec 2024
Task Arithmetic Through The Lens Of One-Shot Federated Learning
Task Arithmetic Through The Lens Of One-Shot Federated Learning
Zhixu Tao
I. Mason
Sanjeev R. Kulkarni
Xavier Boix
MoMeFedML
131
7
0
27 Nov 2024
CopRA: A Progressive LoRA Training Strategy
CopRA: A Progressive LoRA Training Strategy
Zhan Zhuang
Xiequn Wang
Yulong Zhang
Wei Li
Yu Zhang
Ying Wei
116
1
0
30 Oct 2024
Efficient and Effective Weight-Ensembling Mixture of Experts for
  Multi-Task Model Merging
Efficient and Effective Weight-Ensembling Mixture of Experts for Multi-Task Model Merging
Li Shen
Anke Tang
Enneng Yang
G. Guo
Yong Luo
Lefei Zhang
Xiaochun Cao
Di Lin
Dacheng Tao
MoMe
83
9
0
29 Oct 2024
Subgraph Aggregation for Out-of-Distribution Generalization on Graphs
Subgraph Aggregation for Out-of-Distribution Generalization on Graphs
Bowen Liu
Haoyang Li
Shuning Wang
Shuo Nie
Shanghang Zhang
OODDCML
172
0
0
29 Oct 2024
Model merging with SVD to tie the Knots
Model merging with SVD to tie the Knots
George Stoica
Pratik Ramesh
B. Ecsedi
Leshem Choshen
Judy Hoffman
MoMe
99
19
0
25 Oct 2024
In Search of the Successful Interpolation: On the Role of Sharpness in
  CLIP Generalization
In Search of the Successful Interpolation: On the Role of Sharpness in CLIP Generalization
Alireza Abdollahpoorrostam
57
0
0
21 Oct 2024
Uncovering, Explaining, and Mitigating the Superficial Safety of
  Backdoor Defense
Uncovering, Explaining, and Mitigating the Superficial Safety of Backdoor Defense
Rui Min
Zeyu Qin
Nevin L. Zhang
Li Shen
Minhao Cheng
AAML
91
4
0
13 Oct 2024
What Matters for Model Merging at Scale?
What Matters for Model Merging at Scale?
Prateek Yadav
Tu Vu
Jonathan Lai
Alexandra Chronopoulou
Manaal Faruqui
Joey Tianyi Zhou
Tsendsuren Munkhdalai
MoMe
107
22
0
04 Oct 2024
Foldable SuperNets: Scalable Merging of Transformers with Different
  Initializations and Tasks
Foldable SuperNets: Scalable Merging of Transformers with Different Initializations and Tasks
Edan Kinderman
Itay Hubara
Haggai Maron
Daniel Soudry
MoMe
104
2
0
02 Oct 2024
Input Space Mode Connectivity in Deep Neural Networks
Input Space Mode Connectivity in Deep Neural Networks
Jakub Vrabel
Ori Shem-Ur
Yaron Oz
David Krueger
112
1
0
09 Sep 2024
Weight Scope Alignment: A Frustratingly Easy Method for Model Merging
Weight Scope Alignment: A Frustratingly Easy Method for Model Merging
Yichu Xu
Xin-Chun Li
Le Gan
De-Chuan Zhan
MoMe
91
0
0
22 Aug 2024
SMILE: Zero-Shot Sparse Mixture of Low-Rank Experts Construction From
  Pre-Trained Foundation Models
SMILE: Zero-Shot Sparse Mixture of Low-Rank Experts Construction From Pre-Trained Foundation Models
Anke Tang
Li Shen
Yong Luo
Shuai Xie
Han Hu
Lefei Zhang
Di Lin
Dacheng Tao
MoMe
104
4
0
19 Aug 2024
Enhancing material property prediction with ensemble deep graph
  convolutional networks
Enhancing material property prediction with ensemble deep graph convolutional networks
Chowdhury Mohammad Abid Rahman
Ghadendra B. Bhandari
Nasser M. Nasrabadi
Aldo H. Romero
P. Gyawali
AI4CE
84
4
0
26 Jul 2024
Flatness-aware Sequential Learning Generates Resilient Backdoors
Flatness-aware Sequential Learning Generates Resilient Backdoors
Hoang Pham
The-Anh Ta
Anh Tran
Khoa D. Doan
FedMLAAML
76
1
0
20 Jul 2024
Exploring End-to-end Differentiable Neural Charged Particle Tracking --
  A Loss Landscape Perspective
Exploring End-to-end Differentiable Neural Charged Particle Tracking -- A Loss Landscape Perspective
T. Kortus
Ralf Keidel
N.R. Gauger
133
0
0
18 Jul 2024
Harmony in Diversity: Merging Neural Networks with Canonical Correlation
  Analysis
Harmony in Diversity: Merging Neural Networks with Canonical Correlation Analysis
Stefan Horoi
Albert Manuel Orozco Camacho
Eugene Belilovsky
Guy Wolf
FedMLMoMe
58
10
0
07 Jul 2024
Adaptive Stochastic Weight Averaging
Adaptive Stochastic Weight Averaging
Caglar Demir
Arnab Sharma
Axel-Cyrille Ngonga Ngomo
MoMe
69
1
0
27 Jun 2024
MD tree: a model-diagnostic tree grown on loss landscape
MD tree: a model-diagnostic tree grown on loss landscape
Yefan Zhou
Jianlong Chen
Qinxue Cao
Konstantin Schürholt
Yaoqing Yang
102
2
0
24 Jun 2024
Landscaping Linear Mode Connectivity
Landscaping Linear Mode Connectivity
Sidak Pal Singh
Linara Adilova
Michael Kamp
Asja Fischer
Bernhard Scholkopf
Thomas Hofmann
120
6
0
24 Jun 2024
Twin-Merging: Dynamic Integration of Modular Expertise in Model Merging
Twin-Merging: Dynamic Integration of Modular Expertise in Model Merging
Zhenyi Lu
Chenghao Fan
Wei Wei
Xiaoye Qu
Dangyang Chen
Yu Cheng
MoMe
126
63
0
17 Jun 2024
123456
Next