ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.19827
12
0

Revisiting Glorot Initialization for Long-Range Linear Recurrences

26 May 2025
Noga Bar
Mariia Seleznova
Yotam Alexander
Gitta Kutyniok
Raja Giryes
ArXiv (abs)PDFHTML
Main:9 Pages
5 Figures
Bibliography:3 Pages
3 Tables
Appendix:5 Pages
Abstract

Proper initialization is critical for Recurrent Neural Networks (RNNs), particularly in long-range reasoning tasks, where repeated application of the same weight matrix can cause vanishing or exploding signals. A common baseline for linear recurrences is Glorot initialization, designed to ensure stable signal propagation--but derived under the infinite-width, fixed-length regime--an unrealistic setting for RNNs processing long sequences. In this work, we show that Glorot initialization is in fact unstable: small positive deviations in the spectral radius are amplified through time and cause the hidden state to explode. Our theoretical analysis demonstrates that sequences of length t=O(n)t = O(\sqrt{n})t=O(n​), where nnn is the hidden width, are sufficient to induce instability. To address this, we propose a simple, dimension-aware rescaling of Glorot that shifts the spectral radius slightly below one, preventing rapid signal explosion or decay. These results suggest that standard initialization schemes may break down in the long-sequence regime, motivating a separate line of theory for stable recurrent initialization.

View on arXiv
@article{bar2025_2505.19827,
  title={ Revisiting Glorot Initialization for Long-Range Linear Recurrences },
  author={ Noga Bar and Mariia Seleznova and Yotam Alexander and Gitta Kutyniok and Raja Giryes },
  journal={arXiv preprint arXiv:2505.19827},
  year={ 2025 }
}
Comments on this paper