ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.11836
  4. Cited By
SplInterp: Improving our Understanding and Training of Sparse Autoencoders

SplInterp: Improving our Understanding and Training of Sparse Autoencoders

17 May 2025
Jeremy Budd
Javier Ideami
Benjamin Macdowall Rynne
Keith Duggar
Randall Balestriero
ArXivPDFHTML

Papers citing "SplInterp: Improving our Understanding and Training of Sparse Autoencoders"

5 / 5 papers shown
Title
Projecting Assumptions: The Duality Between Sparse Autoencoders and Concept Geometry
Sai Sumedh R. Hindupur
Ekdeep Singh Lubana
Thomas Fel
Demba Ba
89
9
0
03 Mar 2025
Sparse Autoencoders Do Not Find Canonical Units of Analysis
Sparse Autoencoders Do Not Find Canonical Units of Analysis
Patrick Leask
Bart Bussmann
Michael T. Pearce
Joseph Isaac Bloom
Curt Tigges
Noura Al Moubayed
Lee D. Sharkey
Neel Nanda
88
13
0
07 Feb 2025
Sparse Autoencoders Can Interpret Randomly Initialized Transformers
Sparse Autoencoders Can Interpret Randomly Initialized Transformers
Thomas Heap
Tim Lawson
Lucy Farnik
Laurence Aitchison
57
16
0
29 Jan 2025
Toy Models of Superposition
Toy Models of Superposition
Nelson Elhage
Tristan Hume
Catherine Olsson
Nicholas Schiefer
T. Henighan
...
Sam McCandlish
Jared Kaplan
Dario Amodei
Martin Wattenberg
C. Olah
AAML
MILM
172
365
0
21 Sep 2022
Calculus of the exponent of Kurdyka-Łojasiewicz inequality and its
  applications to linear convergence of first-order methods
Calculus of the exponent of Kurdyka-Łojasiewicz inequality and its applications to linear convergence of first-order methods
Guoyin Li
Ting Kei Pong
138
295
0
09 Feb 2016
1