Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2404.02040
Cited By
Transformers as Transducers
2 April 2024
Lena Strobl
Dana Angluin
David Chiang
Jonathan Rawski
Ashish Sabharwal
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Transformers as Transducers"
7 / 7 papers shown
Title
A Little Depth Goes a Long Way: The Expressive Power of Log-Depth Transformers
William Merrill
Ashish Sabharwal
76
8
0
05 Mar 2025
What Algorithms can Transformers Learn? A Study in Length Generalization
Hattie Zhou
Arwen Bradley
Etai Littwin
Noam Razin
Omid Saremi
Josh Susskind
Samy Bengio
Preetum Nakkiran
68
122
0
24 Oct 2023
Towards Revealing the Mystery behind Chain of Thought: A Theoretical Perspective
Guhao Feng
Bohang Zhang
Yuntian Gu
Haotian Ye
Di He
Liwei Wang
LRM
82
246
0
24 May 2023
Formal Language Recognition by Hard Attention Transformers: Perspectives from Circuit Complexity
Sophie Hao
Dana Angluin
Robert Frank
43
78
0
13 Apr 2022
Overcoming a Theoretical Limitation of Self-Attention
David Chiang
Peter A. Cholak
62
83
0
24 Feb 2022
Saturated Transformers are Constant-Depth Threshold Circuits
William Merrill
Ashish Sabharwal
Noah A. Smith
59
105
0
30 Jun 2021
Theoretical Limitations of Self-Attention in Neural Sequence Models
Michael Hahn
49
271
0
16 Jun 2019
1