Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1612.04426
Cited By
Improving Neural Language Models with a Continuous Cache
13 December 2016
Edouard Grave
Armand Joulin
Nicolas Usunier
KELM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Improving Neural Language Models with a Continuous Cache"
22 / 72 papers shown
Title
Trellis Networks for Sequence Modeling
Shaojie Bai
J. Zico Kolter
V. Koltun
25
145
0
15 Oct 2018
Persistence pays off: Paying Attention to What the LSTM Gating Mechanism Persists
Giancarlo D. Salton
John D. Kelleher
KELM
RALM
26
6
0
10 Oct 2018
Adaptive Input Representations for Neural Language Modeling
Alexei Baevski
Michael Auli
32
388
0
28 Sep 2018
Information-Weighted Neural Cache Language Models for ASR
Lyan Verwimp
J. Pelemans
Hugo Van hamme
P. Wambacq
KELM
RALM
16
2
0
24 Sep 2018
Direct Output Connection for a High-Rank Language Model
Sho Takase
Jun Suzuki
Masaaki Nagata
18
36
0
30 Aug 2018
A Neural Model of Adaptation in Reading
Marten van Schijndel
Tal Linzen
27
62
0
29 Aug 2018
Universal Transformers
Mostafa Dehghani
Stephan Gouws
Oriol Vinyals
Jakob Uszkoreit
Lukasz Kaiser
42
745
0
10 Jul 2018
Relational recurrent neural networks
Adam Santoro
Ryan Faulkner
David Raposo
Jack W. Rae
Mike Chrzanowski
T. Weber
Daan Wierstra
Oriol Vinyals
Razvan Pascanu
Timothy Lillicrap
GNN
30
209
0
05 Jun 2018
A Simple Cache Model for Image Recognition
Emin Orhan
VLM
25
30
0
21 May 2018
Sharp Nearby, Fuzzy Far Away: How Neural Language Models Use Context
Urvashi Khandelwal
He He
Peng Qi
Dan Jurafsky
RALM
18
294
0
12 May 2018
Noisin: Unbiased Regularization for Recurrent Neural Networks
Adji Bousso Dieng
Rajesh Ranganath
Jaan Altosaar
David M. Blei
29
22
0
03 May 2018
Assessing Language Models with Scaling Properties
Shuntaro Takahashi
Kumiko Tanaka-Ishii
ELM
LRM
19
2
0
24 Apr 2018
Fast Parametric Learning with Activation Memorization
Jack W. Rae
Chris Dyer
Peter Dayan
Timothy Lillicrap
KELM
41
46
0
27 Mar 2018
An Analysis of Neural Language Modeling at Multiple Scales
Stephen Merity
N. Keskar
R. Socher
24
170
0
22 Mar 2018
An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling
Shaojie Bai
J. Zico Kolter
V. Koltun
DRL
42
4,731
0
04 Mar 2018
Memory-based Parameter Adaptation
Pablo Sprechmann
Siddhant M. Jayakumar
Jack W. Rae
Alexander Pritzel
Adria Puigdomenech Badia
Benigno Uria
Oriol Vinyals
Demis Hassabis
Razvan Pascanu
Charles Blundell
ODL
OOD
VLM
16
101
0
28 Feb 2018
Learning to Remember Translation History with a Continuous Cache
Zhaopeng Tu
Yang Liu
Shuming Shi
Tong Zhang
CLL
38
179
0
26 Nov 2017
Neural Language Modeling by Jointly Learning Syntax and Lexicon
Songlin Yang
Zhouhan Lin
Chin-Wei Huang
Aaron Courville
46
178
0
02 Nov 2017
Learning Differentially Private Recurrent Language Models
H. B. McMahan
Daniel Ramage
Kunal Talwar
Li Zhang
FedML
30
125
0
18 Oct 2017
Regularizing and Optimizing LSTM Language Models
Stephen Merity
N. Keskar
R. Socher
95
1,091
0
07 Aug 2017
Language Modeling with Gated Convolutional Networks
Yann N. Dauphin
Angela Fan
Michael Auli
David Grangier
107
2,365
0
23 Dec 2016
Multiplicative LSTM for sequence modelling
Ben Krause
Liang Lu
Iain Murray
Steve Renals
35
208
0
26 Sep 2016
Previous
1
2