Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1905.05950
Cited By
v1
v2 (latest)
BERT Rediscovers the Classical NLP Pipeline
15 May 2019
Ian Tenney
Dipanjan Das
Ellie Pavlick
MILM
SSeg
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"BERT Rediscovers the Classical NLP Pipeline"
21 / 821 papers shown
Title
Low-Resource Parsing with Crosslingual Contextualized Representations
Phoebe Mulcaire
Jungo Kasai
Noah A. Smith
76
19
0
19 Sep 2019
Effective Use of Transformer Networks for Entity Tracking
Aditya Gupta
Greg Durrett
59
20
0
05 Sep 2019
The Bottom-up Evolution of Representations in the Transformer: A Study with Machine Translation and Language Modeling Objectives
Elena Voita
Rico Sennrich
Ivan Titov
302
187
0
03 Sep 2019
QuASE: Question-Answer Driven Sentence Encoding
Hangfeng He
Qiang Ning
Dan Roth
40
1
0
01 Sep 2019
Higher-order Comparisons of Sentence Encoder Representations
Mostafa Abdou
Artur Kulmizev
Felix Hill
D. Low
Anders Søgaard
75
17
0
01 Sep 2019
Adaptively Sparse Transformers
Gonçalo M. Correia
Vlad Niculae
André F. T. Martins
136
257
0
30 Aug 2019
Learning Latent Parameters without Human Response Patterns: Item Response Theory with Artificial Crowds
John P. Lalor
Hao Wu
Hong-ye Yu
94
45
0
29 Aug 2019
Does BERT agree? Evaluating knowledge of structure dependence through agreement relations
Geoff Bacon
T. Regier
55
21
0
26 Aug 2019
Visualizing and Understanding the Effectiveness of BERT
Y. Hao
Li Dong
Furu Wei
Ke Xu
150
186
0
15 Aug 2019
On Identifiability in Transformers
Gino Brunner
Yang Liu
Damian Pascual
Oliver Richter
Massimiliano Ciaramita
Roger Wattenhofer
ViT
77
189
0
12 Aug 2019
What BERT is not: Lessons from a new suite of psycholinguistic diagnostics for language models
Allyson Ettinger
120
610
0
31 Jul 2019
Leveraging Pre-trained Checkpoints for Sequence Generation Tasks
S. Rothe
Shashi Narayan
Aliaksei Severyn
SILM
146
438
0
29 Jul 2019
Theoretical Limitations of Self-Attention in Neural Sequence Models
Michael Hahn
94
276
0
16 Jun 2019
What Does BERT Look At? An Analysis of BERT's Attention
Kevin Clark
Urvashi Khandelwal
Omer Levy
Christopher D. Manning
MILM
289
1,609
0
11 Jun 2019
Analyzing the Structure of Attention in a Transformer Language Model
Jesse Vig
Yonatan Belinkov
85
371
0
07 Jun 2019
Visualizing and Measuring the Geometry of BERT
Andy Coenen
Emily Reif
Ann Yuan
Been Kim
Adam Pearce
F. Viégas
Martin Wattenberg
MILM
99
419
0
06 Jun 2019
How multilingual is Multilingual BERT?
Telmo Pires
Eva Schlinger
Dan Garrette
LRM
VLM
248
1,416
0
04 Jun 2019
Taming Pretrained Transformers for Extreme Multi-label Text Classification
Wei-Cheng Chang
Hsiang-Fu Yu
Kai Zhong
Yiming Yang
Inderjit Dhillon
75
20
0
07 May 2019
75 Languages, 1 Model: Parsing Universal Dependencies Universally
Dan Kondratyuk
Milan Straka
117
264
0
03 Apr 2019
Toward Fast and Accurate Neural Chinese Word Segmentation with Multi-Criteria Learning
Weipéng Huáng
Xingyi Cheng
Kunlong Chen
Taifeng Wang
Wei Chu
50
62
0
11 Mar 2019
Learning to Perform Role-Filler Binding with Schematic Knowledge
Catherine Chen
Qihong Lu
A. Beukers
Christopher A. Baldassano
K. A. Norman
57
10
0
24 Feb 2019
Previous
1
2
3
...
15
16
17