Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1905.05950
Cited By
v1
v2 (latest)
BERT Rediscovers the Classical NLP Pipeline
15 May 2019
Ian Tenney
Dipanjan Das
Ellie Pavlick
MILM
SSeg
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"BERT Rediscovers the Classical NLP Pipeline"
50 / 821 papers shown
Title
SimAlign: High Quality Word Alignments without Parallel Training Data using Static and Contextualized Embeddings
Masoud Jalili Sabet
Philipp Dufter
François Yvon
Hinrich Schütze
115
238
0
18 Apr 2020
Transform and Tell: Entity-Aware News Image Captioning
Alasdair Tran
A. Mathews
Lexing Xie
VLM
60
97
0
17 Apr 2020
Null It Out: Guarding Protected Attributes by Iterative Nullspace Projection
Shauli Ravfogel
Yanai Elazar
Hila Gonen
Michael Twiton
Yoav Goldberg
156
388
0
16 Apr 2020
What's so special about BERT's layers? A closer look at the NLP pipeline in monolingual and multilingual models
Wietse de Vries
Andreas van Cranenburgh
Malvina Nissim
MILM
SSeg
MoE
142
66
0
14 Apr 2020
Overestimation of Syntactic Representationin Neural Language Models
Jordan Kodner
Nitish Gupta
95
12
0
10 Apr 2020
On the Effect of Dropping Layers of Pre-trained Transformer Models
Hassan Sajjad
Fahim Dalvi
Nadir Durrani
Preslav Nakov
83
143
0
08 Apr 2020
Towards Evaluating the Robustness of Chinese BERT Classifiers
Wei Ping
Boyuan Pan
Xin Li
Yue Liu
AAML
77
8
0
07 Apr 2020
A Systematic Analysis of Morphological Content in BERT Models for Multiple Languages
Daniel Edmiston
55
32
0
06 Apr 2020
Unsupervised Domain Clusters in Pretrained Language Models
Roee Aharoni
Yoav Goldberg
103
252
0
05 Apr 2020
Deep Entity Matching with Pre-Trained Language Models
Yuliang Li
Jinfeng Li
Yoshihiko Suhara
A. Doan
W. Tan
VLM
108
391
0
01 Apr 2020
Understanding Cross-Lingual Syntactic Transfer in Multilingual Recurrent Neural Networks
Prajit Dhar
Arianna Bisazza
44
10
0
31 Mar 2020
Abstractive Text Summarization based on Language Model Conditioning and Locality Modeling
Dmitrii Aksenov
J. Moreno-Schneider
Peter Bourgonje
Robert Schwarzenberg
Leonhard Hennig
Georg Rehm
115
26
0
29 Mar 2020
Meta Fine-Tuning Neural Language Models for Multi-Domain Text Mining
Chengyu Wang
Minghui Qiu
Jun Huang
Xiaofeng He
AI4CE
101
24
0
29 Mar 2020
Probing Word Translations in the Transformer and Trading Decoder for Encoder Layers
Hongfei Xu
Josef van Genabith
Qiuhui Liu
Deyi Xiong
30
3
0
21 Mar 2020
Pre-trained Models for Natural Language Processing: A Survey
Xipeng Qiu
Tianxiang Sun
Yige Xu
Yunfan Shao
Ning Dai
Xuanjing Huang
LM&MA
VLM
395
1,500
0
18 Mar 2020
A Survey on Contextual Embeddings
Qi Liu
Matt J. Kusner
Phil Blunsom
276
151
0
16 Mar 2020
Investigating Entity Knowledge in BERT with Simple Neural End-To-End Entity Linking
Samuel Broscheit
OCL
105
123
0
11 Mar 2020
A Framework for Evaluation of Machine Reading Comprehension Gold Standards
Viktor Schlegel
Marco Valentino
André Freitas
Goran Nenadic
Riza Batista-Navarro
58
30
0
10 Mar 2020
jiant: A Software Toolkit for Research on General-Purpose Text Understanding Models
Yada Pruksachatkun
Philip Yeres
Haokun Liu
Jason Phang
Phu Mon Htut
Alex Jinpeng Wang
Ian Tenney
Samuel R. Bowman
SSeg
48
94
0
04 Mar 2020
DC-BERT: Decoupling Question and Document for Efficient Contextual Encoding
Yuyu Zhang
Ping Nie
Xiubo Geng
Arun Ramamurthy
Le Song
Daxin Jiang
81
61
0
28 Feb 2020
A Primer in BERTology: What we know about how BERT works
Anna Rogers
Olga Kovaleva
Anna Rumshisky
OffRL
146
1,511
0
27 Feb 2020
What BERT Sees: Cross-Modal Transfer for Visual Question Generation
Thomas Scialom
Patrick Bordes
Paul-Alexis Dray
Jacopo Staiano
Patrick Gallinari
59
6
0
25 Feb 2020
Fixed Encoder Self-Attention Patterns in Transformer-Based Machine Translation
Alessandro Raganato
Yves Scherrer
Jörg Tiedemann
100
92
0
24 Feb 2020
BERT's output layer recognizes all hidden layers? Some Intriguing Phenomena and a simple way to boost BERT
Wei-Tsung Kao
Tsung-Han Wu
Po-Han Chi
Chun-Cheng Hsieh
Hung-yi Lee
SSL
44
5
0
25 Jan 2020
AdaBERT: Task-Adaptive BERT Compression with Differentiable Neural Architecture Search
Daoyuan Chen
Yaliang Li
Minghui Qiu
Zhen Wang
Bofang Li
Bolin Ding
Hongbo Deng
Jun Huang
Wei Lin
Jingren Zhou
MQ
97
104
0
13 Jan 2020
oLMpics -- On what Language Model Pre-training Captures
Alon Talmor
Yanai Elazar
Yoav Goldberg
Jonathan Berant
LRM
107
304
0
31 Dec 2019
Siamese Networks for Large-Scale Author Identification
Chakaveh Saedi
Mark Dras
85
38
0
23 Dec 2019
BERTje: A Dutch BERT Model
Wietse de Vries
Andreas van Cranenburgh
Arianna Bisazza
Tommaso Caselli
Gertjan van Noord
Malvina Nissim
VLM
SSeg
98
295
0
19 Dec 2019
WaLDORf: Wasteless Language-model Distillation On Reading-comprehension
J. Tian
A. Kreuzer
Pai-Hung Chen
Hans-Martin Will
VLM
60
3
0
13 Dec 2019
Unsupervised Transfer Learning via BERT Neuron Selection
M. Valipour
E. Lee
Jaime R. Jamacaro
C. Bessega
58
5
0
10 Dec 2019
Modelling Semantic Categories using Conceptual Neighborhood
Zied Bouraoui
Jose Camacho-Collados
Luis Espinosa-Anke
Steven Schockaert
47
10
0
03 Dec 2019
Inducing Relational Knowledge from BERT
Zied Bouraoui
Jose Camacho-Collados
Steven Schockaert
100
167
0
28 Nov 2019
How Can We Know What Language Models Know?
Zhengbao Jiang
Frank F. Xu
Jun Araki
Graham Neubig
KELM
199
1,415
0
28 Nov 2019
Do Attention Heads in BERT Track Syntactic Dependencies?
Phu Mon Htut
Jason Phang
Shikha Bordia
Samuel R. Bowman
85
137
0
27 Nov 2019
Unsupervised Domain Adaptation of Language Models for Reading Comprehension
Kosuke Nishida
Kyosuke Nishida
Itsumi Saito
Hisako Asano
J. Tomita
102
26
0
25 Nov 2019
CamemBERT: a Tasty French Language Model
Louis Martin
Benjamin Muller
Pedro Ortiz Suarez
Yoann Dupont
Laurent Romary
Eric Villemonte de la Clergerie
Djamé Seddah
Benoît Sagot
145
981
0
10 Nov 2019
Learning to Few-Shot Learn Across Diverse Natural Language Classification Tasks
Trapit Bansal
Rishikesh Jha
Andrew McCallum
SSL
106
121
0
10 Nov 2019
Generalizing Natural Language Analysis through Span-relation Representations
Zhengbao Jiang
Wenyuan Xu
Jun Araki
Graham Neubig
81
60
0
10 Nov 2019
Multi-Sentence Argument Linking
Seth Ebner
Patrick Xia
Ryan Culkin
Kyle Rawlins
Benjamin Van Durme
HAI
113
163
0
09 Nov 2019
Reducing Sentiment Bias in Language Models via Counterfactual Evaluation
Po-Sen Huang
Huan Zhang
Ray Jiang
Robert Stanforth
Johannes Welbl
Jack W. Rae
Vishal Maini
Dani Yogatama
Pushmeet Kohli
109
217
0
08 Nov 2019
BERTs of a feather do not generalize together: Large variability in generalization across models with similar test set performance
R. Thomas McCoy
Junghyun Min
Tal Linzen
144
151
0
07 Nov 2019
What does BERT Learn from Multiple-Choice Reading Comprehension Datasets?
Chenglei Si
Shuohang Wang
Min-Yen Kan
Jing Jiang
88
53
0
28 Oct 2019
HUBERT Untangles BERT to Improve Transfer across NLP Tasks
M. Moradshahi
Hamid Palangi
M. Lam
P. Smolensky
Jianfeng Gao
143
16
0
25 Oct 2019
Enhancing the Transformer with Explicit Relational Encoding for Math Problem Solving
Imanol Schlag
P. Smolensky
Roland Fernandez
Nebojsa Jojic
Jürgen Schmidhuber
Jianfeng Gao
92
52
0
15 Oct 2019
Whatcha lookin' at? DeepLIFTing BERT's Attention in Question Answering
Ekaterina Arkhangelskaia
Sourav Dutta
AIMat
35
10
0
14 Oct 2019
exBERT: A Visual Analysis Tool to Explore Learned Representations in Transformers Models
Benjamin Hoover
Hendrik Strobelt
Sebastian Gehrmann
49
87
0
11 Oct 2019
Is Multilingual BERT Fluent in Language Generation?
Samuel Rönnqvist
Jenna Kanerva
T. Salakoski
Filip Ginter
79
72
0
09 Oct 2019
HuggingFace's Transformers: State-of-the-art Natural Language Processing
Thomas Wolf
Lysandre Debut
Victor Sanh
Julien Chaumond
Clement Delangue
...
Teven Le Scao
Sylvain Gugger
Mariama Drame
Quentin Lhoest
Alexander M. Rush
AI4CE
165
1,962
0
09 Oct 2019
On the use of BERT for Neural Machine Translation
Stéphane Clinchant
K. Jung
Vassilina Nikoulina
86
90
0
27 Sep 2019
Attention Interpretability Across NLP Tasks
Shikhar Vashishth
Shyam Upadhyay
Gaurav Singh Tomar
Manaal Faruqui
XAI
MILM
95
176
0
24 Sep 2019
Previous
1
2
3
...
15
16
17
Next