ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1707.05589
  4. Cited By
On the State of the Art of Evaluation in Neural Language Models
v1v2 (latest)

On the State of the Art of Evaluation in Neural Language Models

18 July 2017
Gábor Melis
Chris Dyer
Phil Blunsom
ArXiv (abs)PDFHTML

Papers citing "On the State of the Art of Evaluation in Neural Language Models"

50 / 190 papers shown
Title
Pitfalls in the Evaluation of Sentence Embeddings
Pitfalls in the Evaluation of Sentence Embeddings
Steffen Eger
Andreas Rucklé
Iryna Gurevych
ELM
68
17
0
04 Jun 2019
On Network Design Spaces for Visual Recognition
On Network Design Spaces for Visual Recognition
Ilija Radosavovic
Justin Johnson
Saining Xie
Wan-Yen Lo
Piotr Dollár
104
136
0
30 May 2019
Efficient Neural Architecture Search via Proximal Iterations
Efficient Neural Architecture Search via Proximal Iterations
Quanming Yao
Ju Xu
Wei-Wei Tu
Zhanxing Zhu
114
104
0
30 May 2019
Meta-Surrogate Benchmarking for Hyperparameter Optimization
Meta-Surrogate Benchmarking for Hyperparameter Optimization
Aaron Klein
Zhenwen Dai
Frank Hutter
Neil D. Lawrence
Javier I. González
OffRL
84
36
0
30 May 2019
Gmail Smart Compose: Real-Time Assisted Writing
Gmail Smart Compose: Real-Time Assisted Writing
Mengzhao Chen
Benjamin Lee
G. Bansal
Yuan Cao
Shuyuan Zhang
...
Yinan Wang
Andrew M. Dai
Zhiwen Chen
Timothy Sohn
Yonghui Wu
81
208
0
17 May 2019
Deep Residual Output Layers for Neural Language Generation
Deep Residual Output Layers for Neural Language Generation
Nikolaos Pappas
James Henderson
89
7
0
14 May 2019
Mutual Information Scaling and Expressive Power of Sequence Models
Mutual Information Scaling and Expressive Power of Sequence Models
Huitao Shen
85
18
0
10 May 2019
Survey of Dropout Methods for Deep Neural Networks
Survey of Dropout Methods for Deep Neural Networks
Alex Labach
Hojjat Salehinejad
S. Valaee
80
151
0
25 Apr 2019
The Scientific Method in the Science of Machine Learning
The Scientific Method in the Science of Machine Learning
Jessica Zosa Forde
Michela Paganini
73
37
0
24 Apr 2019
Better Automatic Evaluation of Open-Domain Dialogue Systems with
  Contextualized Embeddings
Better Automatic Evaluation of Open-Domain Dialogue Systems with Contextualized Embeddings
Sarik Ghazarian
Johnny Tian-Zheng Wei
Aram Galstyan
Nanyun Peng
58
90
0
24 Apr 2019
Adversarial Dropout for Recurrent Neural Networks
Adversarial Dropout for Recurrent Neural Networks
Sungrae Park
Kyungwoo Song
Mingi Ji
Wonsung Lee
Il-Chul Moon
32
6
0
22 Apr 2019
Good-Enough Compositional Data Augmentation
Good-Enough Compositional Data Augmentation
Jacob Andreas
115
234
0
21 Apr 2019
Language Models with Transformers
Language Models with Transformers
Chenguang Wang
Mu Li
Alex Smola
102
122
0
20 Apr 2019
Looking Beyond Label Noise: Shifted Label Distribution Matters in
  Distantly Supervised Relation Extraction
Looking Beyond Label Noise: Shifted Label Distribution Matters in Distantly Supervised Relation Extraction
Qinyuan Ye
Liyuan Liu
Maosen Zhang
Xiang Ren
86
18
0
19 Apr 2019
Effective Estimation of Deep Generative Language Models
Effective Estimation of Deep Generative Language Models
Tom Pelsmaeker
Wilker Aziz
BDL
92
28
0
17 Apr 2019
Elucidating image-to-set prediction: An analysis of models, losses and
  datasets
Elucidating image-to-set prediction: An analysis of models, losses and datasets
Luis Villaseñor-Pineda
Amaia Salvador
M. Drozdzal
Adriana Romero
108
12
0
11 Apr 2019
Knowledge Distillation For Recurrent Neural Network Language Modeling
  With Trust Regularization
Knowledge Distillation For Recurrent Neural Network Language Modeling With Trust Regularization
Yangyang Shi
M. Hwang
X. Lei
Haoyu Sheng
152
25
0
08 Apr 2019
Unsupervised Recurrent Neural Network Grammars
Unsupervised Recurrent Neural Network Grammars
Yoon Kim
Alexander M. Rush
Lei Yu
A. Kuncoro
Chris Dyer
Gábor Melis
LRMRALMSSL
136
115
0
07 Apr 2019
Maybe Deep Neural Networks are the Best Choice for Modeling Source Code
Maybe Deep Neural Networks are the Best Choice for Modeling Source Code
Rafael-Michael Karampatsis
Charles Sutton
119
54
0
13 Mar 2019
Deep learning for molecular design - a review of the state of the art
Deep learning for molecular design - a review of the state of the art
Daniel C. Elton
Zois Boukouvalas
M. Fuge
Peter W. Chung
AI4CE3DV
103
328
0
11 Mar 2019
Breaking the Softmax Bottleneck via Learnable Monotonic Pointwise
  Non-linearities
Breaking the Softmax Bottleneck via Learnable Monotonic Pointwise Non-linearities
O. Ganea
Sylvain Gelly
Gary Bécigneul
Aliaksei Severyn
79
18
0
21 Feb 2019
Optimal Kronecker-Sum Approximation of Real Time Recurrent Learning
Optimal Kronecker-Sum Approximation of Real Time Recurrent Learning
Frederik Benzing
M. Gauy
Asier Mujika
A. Martinsson
Angelika Steger
126
23
0
11 Feb 2019
Comprehensive Evaluation of Deep Learning Architectures for Prediction
  of DNA/RNA Sequence Binding Specificities
Comprehensive Evaluation of Deep Learning Architectures for Prediction of DNA/RNA Sequence Binding Specificities
A. Trabelsi
Mohamed Chaabane
Asa Ben-Hur
ELM
45
118
0
29 Jan 2019
FastGRNN: A Fast, Accurate, Stable and Tiny Kilobyte Sized Gated
  Recurrent Neural Network
FastGRNN: A Fast, Accurate, Stable and Tiny Kilobyte Sized Gated Recurrent Neural Network
Aditya Kusupati
Manish Singh
Kush S. Bhatia
A. Kumar
Prateek Jain
Manik Varma
80
190
0
08 Jan 2019
A Tutorial on Deep Latent Variable Models of Natural Language
A Tutorial on Deep Latent Variable Models of Natural Language
Yoon Kim
Sam Wiseman
Alexander M. Rush
BDLVLM
121
42
0
17 Dec 2018
Distilling Information from a Flood: A Possibility for the Use of
  Meta-Analysis and Systematic Review in Machine Learning Research
Distilling Information from a Flood: A Possibility for the Use of Meta-Analysis and Systematic Review in Machine Learning Research
Peter Henderson
Emma Brunskill
AI4CE
60
3
0
03 Dec 2018
Learning to Reason with Third-Order Tensor Products
Learning to Reason with Third-Order Tensor Products
Imanol Schlag
Jürgen Schmidhuber
NAI
76
64
0
29 Nov 2018
ESPNetv2: A Light-weight, Power Efficient, and General Purpose
  Convolutional Neural Network
ESPNetv2: A Light-weight, Power Efficient, and General Purpose Convolutional Neural Network
Sachin Mehta
Mohammad Rastegari
Linda G. Shapiro
Hannaneh Hajishirzi
VLM
89
400
0
28 Nov 2018
Learning to Discover, Ground and Use Words with Segmental Neural
  Language Models
Learning to Discover, Ground and Use Words with Segmental Neural Language Models
Kazuya Kawakami
Chris Dyer
Phil Blunsom
VLM
57
11
0
23 Nov 2018
Jointly Learning to Label Sentences and Tokens
Jointly Learning to Label Sentences and Tokens
Marek Rei
Anders Søgaard
AI4TS
98
40
0
14 Nov 2018
Pitfalls of Graph Neural Network Evaluation
Pitfalls of Graph Neural Network Evaluation
Oleksandr Shchur
Maximilian Mumme
Aleksandar Bojchevski
Stephan Günnemann
GNN
178
1,375
0
14 Nov 2018
Language Modeling for Code-Switching: Evaluation, Integration of
  Monolingual Data, and Discriminative Training
Language Modeling for Code-Switching: Evaluation, Integration of Monolingual Data, and Discriminative Training
Hila Gonen
Yoav Goldberg
75
32
0
28 Oct 2018
Ordered Neurons: Integrating Tree Structures into Recurrent Neural
  Networks
Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks
Songlin Yang
Shawn Tan
Alessandro Sordoni
Aaron Courville
145
325
0
22 Oct 2018
Trellis Networks for Sequence Modeling
Trellis Networks for Sequence Modeling
Shaojie Bai
J. Zico Kolter
V. Koltun
93
146
0
15 Oct 2018
Deep Reinforcement Learning
Deep Reinforcement Learning
Yuxi Li
VLMOffRL
194
144
0
15 Oct 2018
A Survey and Critique of Multiagent Deep Reinforcement Learning
A Survey and Critique of Multiagent Deep Reinforcement Learning
Pablo Hernandez-Leal
Bilal Kartal
Matthew E. Taylor
OffRL
121
570
0
12 Oct 2018
Persistence pays off: Paying Attention to What the LSTM Gating Mechanism
  Persists
Persistence pays off: Paying Attention to What the LSTM Gating Mechanism Persists
Giancarlo D. Salton
John D. Kelleher
KELMRALM
140
6
0
10 Oct 2018
Information-Weighted Neural Cache Language Models for ASR
Information-Weighted Neural Cache Language Models for ASR
Lyan Verwimp
J. Pelemans
Hugo Van hamme
P. Wambacq
KELMRALM
20
2
0
24 Sep 2018
Can LSTM Learn to Capture Agreement? The Case of Basque
Can LSTM Learn to Capture Agreement? The Case of Basque
Shauli Ravfogel
Francis M. Tyers
Yoav Goldberg
76
43
0
11 Sep 2018
How clever is the FiLM model, and how clever can it be?
How clever is the FiLM model, and how clever can it be?
A. Kuhnle
Huiyuan Xie
Ann A. Copestake
68
6
0
09 Sep 2018
Direct Output Connection for a High-Rank Language Model
Direct Output Connection for a High-Rank Language Model
Sho Takase
Jun Suzuki
Masaaki Nagata
99
36
0
30 Aug 2018
Targeted Syntactic Evaluation of Language Models
Targeted Syntactic Evaluation of Language Models
Rebecca Marvin
Tal Linzen
99
417
0
27 Aug 2018
Dissecting Contextual Word Embeddings: Architecture and Representation
Dissecting Contextual Word Embeddings: Architecture and Representation
Matthew E. Peters
Mark Neumann
Luke Zettlemoyer
Wen-tau Yih
113
434
0
27 Aug 2018
Neural Architecture Optimization
Neural Architecture Optimization
Renqian Luo
Fei Tian
Tao Qin
Enhong Chen
Tie-Yan Liu
3DV
119
660
0
22 Aug 2018
State-of-the-art Chinese Word Segmentation with Bi-LSTMs
State-of-the-art Chinese Word Segmentation with Bi-LSTMs
Ji Ma
Kuzman Ganchev
David J. Weiss
71
102
0
20 Aug 2018
Improved Language Modeling by Decoding the Past
Improved Language Modeling by Decoding the Past
Siddhartha Brahma
BDLAI4TS
53
6
0
14 Aug 2018
REGMAPR - Text Matching Made Easy
REGMAPR - Text Matching Made Easy
Siddhartha Brahma
VLM
44
1
0
13 Aug 2018
TensorFuzz: Debugging Neural Networks with Coverage-Guided Fuzzing
TensorFuzz: Debugging Neural Networks with Coverage-Guided Fuzzing
Augustus Odena
Ian Goodfellow
AAML
76
323
0
28 Jul 2018
Towards Automated Deep Learning: Efficient Joint Neural Architecture and
  Hyperparameter Search
Towards Automated Deep Learning: Efficient Joint Neural Architecture and Hyperparameter Search
Arber Zela
Aaron Klein
Stefan Falkner
Frank Hutter
82
161
0
18 Jul 2018
Convergence guarantees for RMSProp and ADAM in non-convex optimization
  and an empirical comparison to Nesterov acceleration
Convergence guarantees for RMSProp and ADAM in non-convex optimization and an empirical comparison to Nesterov acceleration
Soham De
Anirbit Mukherjee
Enayat Ullah
79
102
0
18 Jul 2018
Previous
1234
Next