ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.05950
  4. Cited By
BERT Rediscovers the Classical NLP Pipeline
v1v2 (latest)

BERT Rediscovers the Classical NLP Pipeline

15 May 2019
Ian Tenney
Dipanjan Das
Ellie Pavlick
    MILMSSeg
ArXiv (abs)PDFHTML

Papers citing "BERT Rediscovers the Classical NLP Pipeline"

50 / 821 papers shown
Title
Better Neural Machine Translation by Extracting Linguistic Information
  from BERT
Better Neural Machine Translation by Extracting Linguistic Information from BERT
Hassan S. Shavarani
Anoop Sarkar
97
15
0
07 Apr 2021
What's the best place for an AI conference, Vancouver or ______: Why
  completing comparative questions is difficult
What's the best place for an AI conference, Vancouver or ______: Why completing comparative questions is difficult
Avishai Zagoury
Einat Minkov
Idan Szpektor
William W. Cohen
ELM
54
6
0
05 Apr 2021
Annotating Columns with Pre-trained Language Models
Annotating Columns with Pre-trained Language Models
Yoshihiko Suhara
Jinfeng Li
Yuliang Li
Dan Zhang
cCaugatay Demiralp
Chen Chen
W. Tan
LMTD
58
90
0
05 Apr 2021
Exploring the Role of BERT Token Representations to Explain Sentence
  Probing Results
Exploring the Role of BERT Token Representations to Explain Sentence Probing Results
Hosein Mohebbi
Ali Modarressi
Mohammad Taher Pilehvar
MILM
67
26
0
03 Apr 2021
Attention, please! A survey of Neural Attention Models in Deep Learning
Attention, please! A survey of Neural Attention Models in Deep Learning
Alana de Santana Correia
Esther Luna Colombini
HAI
128
198
0
31 Mar 2021
HiT: Hierarchical Transformer with Momentum Contrast for Video-Text
  Retrieval
HiT: Hierarchical Transformer with Momentum Contrast for Video-Text Retrieval
Song Liu
Haoqi Fan
Shengsheng Qian
Yiru Chen
Wenkui Ding
Zhongyuan Wang
113
147
0
28 Mar 2021
Leveraging pre-trained representations to improve access to
  untranscribed speech from endangered languages
Leveraging pre-trained representations to improve access to untranscribed speech from endangered languages
Nay San
Martijn Bartelds
Mitchell Browne
Lily Clifford
Fiona Gibson
...
Jane Simpson
Myfany Turpin
Maria Vollmer
Sasha Wilmoth
Dan Jurafsky
68
16
0
26 Mar 2021
Coordinate Constructions in English Enhanced Universal Dependencies:
  Analysis and Computational Modeling
Coordinate Constructions in English Enhanced Universal Dependencies: Analysis and Computational Modeling
Stefan Grünewald
P. Piccirilli
Annemarie Friedrich
54
8
0
16 Mar 2021
Anaphoric Binding: an integrated overview
Anaphoric Binding: an integrated overview
António Branco
23
0
0
11 Mar 2021
CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language
  Representation
CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation
J. Clark
Dan Garrette
Iulia Turc
John Wieting
121
224
0
11 Mar 2021
The Interplay of Variant, Size, and Task Type in Arabic Pre-trained
  Language Models
The Interplay of Variant, Size, and Task Type in Arabic Pre-trained Language Models
Go Inoue
Bashar Alhafni
Nurpeiis Baimukan
Houda Bouamor
Nizar Habash
107
236
0
11 Mar 2021
The Rediscovery Hypothesis: Language Models Need to Meet Linguistics
The Rediscovery Hypothesis: Language Models Need to Meet Linguistics
Vassilina Nikoulina
Maxat Tezekbayev
Nuradil Kozhakhmet
Madina Babazhanova
Matthias Gallé
Z. Assylbekov
65
8
0
02 Mar 2021
Vyākarana: A Colorless Green Benchmark for Syntactic Evaluation in
  Indic Languages
Vyākarana: A Colorless Green Benchmark for Syntactic Evaluation in Indic Languages
Rajaswa Patil
Jasleen Dhillon
Siddhant Mahurkar
Saumitra Kulkarni
M. Malhotra
V. Baths
65
1
0
01 Mar 2021
Investigating the Limitations of Transformers with Simple Arithmetic
  Tasks
Investigating the Limitations of Transformers with Simple Arithmetic Tasks
Rodrigo Nogueira
Zhiying Jiang
Jimmy J. Li
LRM
122
130
0
25 Feb 2021
Using Prior Knowledge to Guide BERT's Attention in Semantic Textual
  Matching Tasks
Using Prior Knowledge to Guide BERT's Attention in Semantic Textual Matching Tasks
Tingyu Xia
Yue Wang
Yuan Tian
Yi-Ju Chang
65
51
0
22 Feb 2021
Evaluating Contextualized Language Models for Hungarian
Evaluating Contextualized Language Models for Hungarian
Judit Ács
Dániel Lévai
D. Nemeskey
András Kornai
27
1
0
22 Feb 2021
Toward Improving Coherence and Diversity of Slogan Generation
Toward Improving Coherence and Diversity of Slogan Generation
Yiping Jin
Akshay Bhatia
Dittaya Wanvarie
Phu T. V. Le
28
5
0
11 Feb 2021
Measuring and Improving Consistency in Pretrained Language Models
Measuring and Improving Consistency in Pretrained Language Models
Yanai Elazar
Nora Kassner
Shauli Ravfogel
Abhilasha Ravichander
Eduard H. Hovy
Hinrich Schütze
Yoav Goldberg
HILM
339
371
0
01 Feb 2021
Multilingual Email Zoning
Multilingual Email Zoning
Bruno Jardim
Ricardo Rei
Mariana S. C. Almeida
21
0
0
31 Jan 2021
A Neural Few-Shot Text Classification Reality Check
A Neural Few-Shot Text Classification Reality Check
Thomas Dopierre
Christophe Gravier
Wilfried Logerais
VLM
61
19
0
28 Jan 2021
Us vs. Them: A Dataset of Populist Attitudes, News Bias and Emotions
Us vs. Them: A Dataset of Populist Attitudes, News Bias and Emotions
Pere-Lluís Huguet Cabot
David Abadi
A. Fischer
Ekaterina Shutova
85
23
0
28 Jan 2021
On the Evolution of Syntactic Information Encoded by BERT's
  Contextualized Representations
On the Evolution of Syntactic Information Encoded by BERT's Contextualized Representations
Laura Pérez-Mayos
Roberto Carlini
Miguel Ballesteros
Leo Wanner
62
7
0
27 Jan 2021
Language Modelling as a Multi-Task Problem
Language Modelling as a Multi-Task Problem
Leon Weber
Jaap Jumelet
Elia Bruni
Dieuwke Hupkes
86
13
0
27 Jan 2021
First Align, then Predict: Understanding the Cross-Lingual Ability of
  Multilingual BERT
First Align, then Predict: Understanding the Cross-Lingual Ability of Multilingual BERT
Benjamin Muller
Yanai Elazar
Benoît Sagot
Djamé Seddah
LRM
67
77
0
26 Jan 2021
Attention Can Reflect Syntactic Structure (If You Let It)
Attention Can Reflect Syntactic Structure (If You Let It)
Vinit Ravishankar
Artur Kulmizev
Mostafa Abdou
Anders Søgaard
Joakim Nivre
66
32
0
26 Jan 2021
Cross-lingual Visual Pre-training for Multimodal Machine Translation
Cross-lingual Visual Pre-training for Multimodal Machine Translation
Ozan Caglayan
Menekse Kuyu
Mustafa Sercan Amac
Pranava Madhyastha
Erkut Erdem
Aykut Erdem
Lucia Specia
VLM
71
46
0
25 Jan 2021
Active Learning for Sequence Tagging with Deep Pre-trained Models and
  Bayesian Uncertainty Estimates
Active Learning for Sequence Tagging with Deep Pre-trained Models and Bayesian Uncertainty Estimates
Artem Shelmanov
Dmitri Puzyrev
L. Kupriyanova
D. Belyakov
Daniil Larionov
Nikita Khromov
Olga Kozlova
Ekaterina Artemova
Dmitry V. Dylov
Alexander Panchenko
BDLUQLMUQCV
90
55
0
20 Jan 2021
Classifying Scientific Publications with BERT -- Is Self-Attention a
  Feature Selection Method?
Classifying Scientific Publications with BERT -- Is Self-Attention a Feature Selection Method?
Andrés García-Silva
José Manuél Gómez-Pérez
43
11
0
20 Jan 2021
Learning to Augment for Data-Scarce Domain BERT Knowledge Distillation
Learning to Augment for Data-Scarce Domain BERT Knowledge Distillation
Lingyun Feng
Minghui Qiu
Yaliang Li
Haitao Zheng
Ying Shen
90
10
0
20 Jan 2021
Of Non-Linearity and Commutativity in BERT
Of Non-Linearity and Commutativity in BERT
Sumu Zhao
Damian Pascual
Gino Brunner
Roger Wattenhofer
105
17
0
12 Jan 2021
Trankit: A Light-Weight Transformer-based Toolkit for Multilingual
  Natural Language Processing
Trankit: A Light-Weight Transformer-based Toolkit for Multilingual Natural Language Processing
Minh Nguyen
Viet Dac Lai
Amir Pouran Ben Veyseh
Thien Huu Nguyen
141
137
0
09 Jan 2021
On Explaining Your Explanations of BERT: An Empirical Study with
  Sequence Classification
On Explaining Your Explanations of BERT: An Empirical Study with Sequence Classification
Zhengxuan Wu
Desmond C. Ong
78
22
0
01 Jan 2021
Coreference Reasoning in Machine Reading Comprehension
Coreference Reasoning in Machine Reading Comprehension
Mingzhu Wu
N. Moosavi
Dan Roth
Iryna Gurevych
LRM
47
8
0
31 Dec 2020
FiD-Ex: Improving Sequence-to-Sequence Models for Extractive Rationale
  Generation
FiD-Ex: Improving Sequence-to-Sequence Models for Extractive Rationale Generation
Kushal Lakhotia
Bhargavi Paranjape
Asish Ghoshal
Wen-tau Yih
Yashar Mehdad
Srini Iyer
63
28
0
31 Dec 2020
Deriving Contextualised Semantic Features from BERT (and Other
  Transformer Model) Embeddings
Deriving Contextualised Semantic Features from BERT (and Other Transformer Model) Embeddings
Jacob Turton
D. Vinson
Robert Smith
44
25
0
30 Dec 2020
ECONET: Effective Continual Pretraining of Language Models for Event
  Temporal Reasoning
ECONET: Effective Continual Pretraining of Language Models for Event Temporal Reasoning
Rujun Han
Xiang Ren
Nanyun Peng
106
57
0
30 Dec 2020
Reservoir Transformers
Reservoir Transformers
Sheng Shen
Alexei Baevski
Ari S. Morcos
Kurt Keutzer
Michael Auli
Douwe Kiela
88
18
0
30 Dec 2020
Transformer Feed-Forward Layers Are Key-Value Memories
Transformer Feed-Forward Layers Are Key-Value Memories
Mor Geva
R. Schuster
Jonathan Berant
Omer Levy
KELM
209
851
0
29 Dec 2020
CascadeBERT: Accelerating Inference of Pre-trained Language Models via
  Calibrated Complete Models Cascade
CascadeBERT: Accelerating Inference of Pre-trained Language Models via Calibrated Complete Models Cascade
Lei Li
Yankai Lin
Deli Chen
Shuhuai Ren
Peng Li
Jie Zhou
Xu Sun
115
52
0
29 Dec 2020
Disentangling semantics in language through VAEs and a certain
  architectural choice
Disentangling semantics in language through VAEs and a certain architectural choice
G. Felhi
Joseph Le Roux
Djamé Seddah
CoGeDRL
48
1
0
24 Dec 2020
Pre-Training a Language Model Without Human Language
Pre-Training a Language Model Without Human Language
Cheng-Han Chiang
Hung-yi Lee
71
13
0
22 Dec 2020
Undivided Attention: Are Intermediate Layers Necessary for BERT?
Undivided Attention: Are Intermediate Layers Necessary for BERT?
S. N. Sridhar
Anthony Sarah
66
15
0
22 Dec 2020
Parameter-Efficient Transfer Learning with Diff Pruning
Parameter-Efficient Transfer Learning with Diff Pruning
Demi Guo
Alexander M. Rush
Yoon Kim
92
406
0
14 Dec 2020
Open Knowledge Graphs Canonicalization using Variational Autoencoders
Open Knowledge Graphs Canonicalization using Variational Autoencoders
Sarthak Dash
Gaetano Rossiello
Nandana Mihindukulasooriya
S. Bagchi
A. Gliozzo
BDL
81
15
0
08 Dec 2020
Neurosymbolic AI for Situated Language Understanding
Neurosymbolic AI for Situated Language Understanding
Nikhil Krishnaswamy
James Pustejovsky
NAI
78
4
0
05 Dec 2020
Self-Explaining Structures Improve NLP Models
Self-Explaining Structures Improve NLP Models
Zijun Sun
Chun Fan
Qinghong Han
Xiaofei Sun
Yuxian Meng
Leilei Gan
Jiwei Li
MILMXAILRMFAtt
117
38
0
03 Dec 2020
Circles are like Ellipses, or Ellipses are like Circles? Measuring the
  Degree of Asymmetry of Static and Contextual Embeddings and the Implications
  to Representation Learning
Circles are like Ellipses, or Ellipses are like Circles? Measuring the Degree of Asymmetry of Static and Contextual Embeddings and the Implications to Representation Learning
Wei Zhang
Murray Campbell
Yang Yu
Yara Rizk
42
0
0
03 Dec 2020
How Can We Know When Language Models Know? On the Calibration of
  Language Models for Question Answering
How Can We Know When Language Models Know? On the Calibration of Language Models for Question Answering
Zhengbao Jiang
Jun Araki
Haibo Ding
Graham Neubig
UQCV
67
439
0
02 Dec 2020
Neural Representations for Modeling Variation in Speech
Neural Representations for Modeling Variation in Speech
Martijn Bartelds
Wietse de Vries
Faraz Sanal
Caitlin Richter
M. Liberman
Martijn B. Wieling
SSLDRL
48
23
0
25 Nov 2020
Enhancing deep neural networks with morphological information
Enhancing deep neural networks with morphological information
Matej Klemen
Luka Krsnik
Marko Robnik-Šikonja
98
12
0
24 Nov 2020
Previous
123...121314151617
Next