ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.05950
  4. Cited By
BERT Rediscovers the Classical NLP Pipeline
v1v2 (latest)

BERT Rediscovers the Classical NLP Pipeline

15 May 2019
Ian Tenney
Dipanjan Das
Ellie Pavlick
    MILMSSeg
ArXiv (abs)PDFHTML

Papers citing "BERT Rediscovers the Classical NLP Pipeline"

50 / 821 papers shown
Title
Effect and Analysis of Large-scale Language Model Rescoring on
  Competitive ASR Systems
Effect and Analysis of Large-scale Language Model Rescoring on Competitive ASR Systems
Takuma Udagawa
Masayuki Suzuki
Gakuto Kurata
N. Itoh
G. Saon
115
24
0
01 Apr 2022
Transformer Feed-Forward Layers Build Predictions by Promoting Concepts
  in the Vocabulary Space
Transformer Feed-Forward Layers Build Predictions by Promoting Concepts in the Vocabulary Space
Mor Geva
Avi Caciularu
Ke Wang
Yoav Goldberg
KELM
146
390
0
28 Mar 2022
Metaphors in Pre-Trained Language Models: Probing and Generalization
  Across Datasets and Languages
Metaphors in Pre-Trained Language Models: Probing and Generalization Across Datasets and Languages
Ehsan Aghazadeh
Mohsen Fayyaz
Yadollah Yaghoobzadeh
63
53
0
26 Mar 2022
Probing for Labeled Dependency Trees
Probing for Labeled Dependency Trees
Max Müller-Eberstein
Rob van der Goot
Barbara Plank
55
8
0
24 Mar 2022
Input-specific Attention Subnetworks for Adversarial Detection
Input-specific Attention Subnetworks for Adversarial Detection
Emil Biju
Anirudh Sriram
Pratyush Kumar
Mitesh M Khapra
AAML
45
5
0
23 Mar 2022
Does DQN really learn? Exploring adversarial training schemes in Pong
Does DQN really learn? Exploring adversarial training schemes in Pong
Bowen He
Sreehari Rammohan
Jessica Zosa Forde
Michael Littman
19
0
0
20 Mar 2022
On the Importance of Data Size in Probing Fine-tuned Models
On the Importance of Data Size in Probing Fine-tuned Models
Houman Mehrafarin
S. Rajaee
Mohammad Taher Pilehvar
63
18
0
17 Mar 2022
Coloring the Blank Slate: Pre-training Imparts a Hierarchical Inductive
  Bias to Sequence-to-sequence Models
Coloring the Blank Slate: Pre-training Imparts a Hierarchical Inductive Bias to Sequence-to-sequence Models
Aaron Mueller
Robert Frank
Tal Linzen
Luheng Wang
Sebastian Schuster
AIMat
97
33
0
17 Mar 2022
Contrastive Visual Semantic Pretraining Magnifies the Semantics of
  Natural Language Representations
Contrastive Visual Semantic Pretraining Magnifies the Semantics of Natural Language Representations
Robert Wolfe
Aylin Caliskan
VLM
67
14
0
14 Mar 2022
VAST: The Valence-Assessing Semantics Test for Contextualizing Language
  Models
VAST: The Valence-Assessing Semantics Test for Contextualizing Language Models
Robert Wolfe
Aylin Caliskan
62
13
0
14 Mar 2022
Grounding Commands for Autonomous Vehicles via Layer Fusion with
  Region-specific Dynamic Layer Attention
Grounding Commands for Autonomous Vehicles via Layer Fusion with Region-specific Dynamic Layer Attention
Hou Pong Chan
M. Guo
Chengguang Xu
74
4
0
14 Mar 2022
On the data requirements of probing
On the data requirements of probing
Zining Zhu
Jixuan Wang
Bai Li
Frank Rudzicz
69
5
0
25 Feb 2022
TrimBERT: Tailoring BERT for Trade-offs
TrimBERT: Tailoring BERT for Trade-offs
S. N. Sridhar
Anthony Sarah
Sairam Sundaresan
MQ
76
4
0
24 Feb 2022
Probing BERT's priors with serial reproduction chains
Probing BERT's priors with serial reproduction chains
Takateru Yamakoshi
Thomas Griffiths
Robert D. Hawkins
105
13
0
24 Feb 2022
BERTVision -- A Parameter-Efficient Approach for Question Answering
BERTVision -- A Parameter-Efficient Approach for Question Answering
Siduo Jiang
Cristopher Benge
Will King
41
1
0
24 Feb 2022
Compositional Generalization Requires Compositional Parsers
Compositional Generalization Requires Compositional Parsers
Pia Weissenhorn
Yuekun Yao
L. Donatelli
Alexander Koller
CoGe
78
3
0
24 Feb 2022
Do Transformers know symbolic rules, and would we know if they did?
Do Transformers know symbolic rules, and would we know if they did?
Tommi Gröndahl
Yu-Wen Guo
Nirmal Asokan
68
0
0
19 Feb 2022
Open-Ended Reinforcement Learning with Neural Reward Functions
Open-Ended Reinforcement Learning with Neural Reward Functions
Robert Meier
Asier Mujika
101
7
0
16 Feb 2022
Interpreting a Machine Learning Model for Detecting Gravitational Waves
Interpreting a Machine Learning Model for Detecting Gravitational Waves
M. Safarzadeh
Asad Khan
Eliu A. Huerta
Martin Wattenberg
103
2
0
15 Feb 2022
What Do They Capture? -- A Structural Analysis of Pre-Trained Language
  Models for Source Code
What Do They Capture? -- A Structural Analysis of Pre-Trained Language Models for Source Code
Yao Wan
Wei Zhao
Hongyu Zhang
Yulei Sui
Guandong Xu
Hairong Jin
103
113
0
14 Feb 2022
Examining Scaling and Transfer of Language Model Architectures for
  Machine Translation
Examining Scaling and Transfer of Language Model Architectures for Machine Translation
Biao Zhang
Behrooz Ghorbani
Ankur Bapna
Yong Cheng
Xavier Garcia
Jonathan Shen
Orhan Firat
84
23
0
01 Feb 2022
Grammatical cues to subjecthood are redundant in a majority of simple
  clauses across languages
Grammatical cues to subjecthood are redundant in a majority of simple clauses across languages
Kyle Mahowald
Evgeniia Diachek
E. Gibson
Evelina Fedorenko
Richard Futrell
104
10
0
30 Jan 2022
Do Transformers Encode a Foundational Ontology? Probing Abstract Classes
  in Natural Language
Do Transformers Encode a Foundational Ontology? Probing Abstract Classes in Natural Language
Mael Jullien
Marco Valentino
André Freitas
64
9
0
25 Jan 2022
An Application of Pseudo-Log-Likelihoods to Natural Language Scoring
An Application of Pseudo-Log-Likelihoods to Natural Language Scoring
Darren Abramson
Ali Emami
66
3
0
23 Jan 2022
TourBERT: A pretrained language model for the tourism industry
TourBERT: A pretrained language model for the tourism industry
Veronika Arefieva
R. Egger
33
4
0
19 Jan 2022
Interpreting Arabic Transformer Models
Ahmed Abdelali
Nadir Durrani
Fahim Dalvi
Hassan Sajjad
70
2
0
19 Jan 2022
Unintended Bias in Language Model-driven Conversational Recommendation
Unintended Bias in Language Model-driven Conversational Recommendation
Tianshu Shen
Jiaru Li
Mohamed Reda Bouadjenek
Zheda Mai
Scott Sanner
68
7
0
17 Jan 2022
Diagnosing BERT with Retrieval Heuristics
Diagnosing BERT with Retrieval Heuristics
A. Câmara
C. Hauff
72
33
0
12 Jan 2022
Does Entity Abstraction Help Generative Transformers Reason?
Does Entity Abstraction Help Generative Transformers Reason?
Nicolas Angelard-Gontier
Siva Reddy
C. Pal
77
5
0
05 Jan 2022
Do Multi-Lingual Pre-trained Language Models Reveal Consistent Token
  Attributions in Different Languages?
Do Multi-Lingual Pre-trained Language Models Reveal Consistent Token Attributions in Different Languages?
Junxiang Wang
Xuchao Zhang
Bo Zong
Yanchi Liu
Wei Cheng
Jingchao Ni
Haifeng Chen
Liang Zhao
AAML
62
0
0
23 Dec 2021
How Should Pre-Trained Language Models Be Fine-Tuned Towards Adversarial
  Robustness?
How Should Pre-Trained Language Models Be Fine-Tuned Towards Adversarial Robustness?
Xinhsuai Dong
Anh Tuan Luu
Min Lin
Shuicheng Yan
Hanwang Zhang
SILMAAML
71
62
0
22 Dec 2021
Linguistic Frameworks Go Toe-to-Toe at Neuro-Symbolic Language Modeling
Linguistic Frameworks Go Toe-to-Toe at Neuro-Symbolic Language Modeling
Jakob Prange
Nathan Schneider
Lingpeng Kong
38
11
0
15 Dec 2021
Large Language Models are not Models of Natural Language: they are
  Corpus Models
Large Language Models are not Models of Natural Language: they are Corpus Models
Csaba Veres
67
20
0
13 Dec 2021
Analysis and Prediction of NLP Models Via Task Embeddings
Analysis and Prediction of NLP Models Via Task Embeddings
Damien Sileo
Marie-Francine Moens
57
4
0
10 Dec 2021
Marvin: an Innovative Omni-Directional Robotic Assistant for Domestic
  Environments
Marvin: an Innovative Omni-Directional Robotic Assistant for Domestic Environments
Andrea Eirale
Mauro Martini
L. Tagliavini
Dario Gandini
Marcello Chiaberge
G. Quaglia
91
28
0
10 Dec 2021
LoNLI: An Extensible Framework for Testing Diverse Logical Reasoning
  Capabilities for NLI
LoNLI: An Extensible Framework for Testing Diverse Logical Reasoning Capabilities for NLI
Ishan Tarunesh
Somak Aditya
Monojit Choudhury
ELMLRM
53
4
0
04 Dec 2021
Siamese BERT-based Model for Web Search Relevance Ranking Evaluated on a
  New Czech Dataset
Siamese BERT-based Model for Web Search Relevance Ranking Evaluated on a New Czech Dataset
M. Kocián
Jakub Náplava
Daniel Stancl
V. Kadlec
44
18
0
03 Dec 2021
Inducing Causal Structure for Interpretable Neural Networks
Inducing Causal Structure for Interpretable Neural Networks
Atticus Geiger
Zhengxuan Wu
Hanson Lu
J. Rozner
Elisa Kreiss
Thomas Icard
Noah D. Goodman
Christopher Potts
CMLOOD
96
76
0
01 Dec 2021
Exploring Low-Cost Transformer Model Compression for Large-Scale
  Commercial Reply Suggestions
Exploring Low-Cost Transformer Model Compression for Large-Scale Commercial Reply Suggestions
Vaishnavi Shrivastava
Radhika Gaonkar
Shashank Gupta
Abhishek Jha
32
0
0
27 Nov 2021
To Augment or Not to Augment? A Comparative Study on Text Augmentation
  Techniques for Low-Resource NLP
To Augment or Not to Augment? A Comparative Study on Text Augmentation Techniques for Low-Resource NLP
Gözde Gül Sahin
68
34
0
18 Nov 2021
LAnoBERT: System Log Anomaly Detection based on BERT Masked Language
  Model
LAnoBERT: System Log Anomaly Detection based on BERT Masked Language Model
Yukyung Lee
Jina Kim
Pilsung Kang
64
84
0
18 Nov 2021
Interpreting Language Models Through Knowledge Graph Extraction
Interpreting Language Models Through Knowledge Graph Extraction
Vinitra Swamy
Angelika Romanou
Martin Jaggi
65
20
0
16 Nov 2021
Discovering Supply Chain Links with Augmented Intelligence
Discovering Supply Chain Links with Augmented Intelligence
Achintya Gopal
Chun-Han Chang
61
3
0
02 Nov 2021
LMdiff: A Visual Diff Tool to Compare Language Models
LMdiff: A Visual Diff Tool to Compare Language Models
Hendrik Strobelt
Benjamin Hoover
Arvind Satyanarayan
Sebastian Gehrmann
VLM
94
19
0
02 Nov 2021
Recent Advances in Natural Language Processing via Large Pre-Trained
  Language Models: A Survey
Recent Advances in Natural Language Processing via Large Pre-Trained Language Models: A Survey
Bonan Min
Hayley L Ross
Elior Sulem
Amir Pouran Ben Veyseh
Thien Huu Nguyen
Oscar Sainz
Eneko Agirre
Ilana Heinz
Dan Roth
LM&MAVLMAI4CE
197
1,100
0
01 Nov 2021
Interpreting Deep Learning Models in Natural Language Processing: A
  Review
Interpreting Deep Learning Models in Natural Language Processing: A Review
Xiaofei Sun
Diyi Yang
Xiaoya Li
Tianwei Zhang
Yuxian Meng
Han Qiu
Guoyin Wang
Eduard H. Hovy
Jiwei Li
97
47
0
20 Oct 2021
Inductive Biases and Variable Creation in Self-Attention Mechanisms
Inductive Biases and Variable Creation in Self-Attention Mechanisms
Benjamin L. Edelman
Surbhi Goel
Sham Kakade
Cyril Zhang
106
125
0
19 Oct 2021
BERMo: What can BERT learn from ELMo?
BERMo: What can BERT learn from ELMo?
Sangamesh Kodge
Kaushik Roy
65
3
0
18 Oct 2021
Quantifying the Task-Specific Information in Text-Based Classifications
Quantifying the Task-Specific Information in Text-Based Classifications
Zining Zhu
Aparna Balagopalan
Marzyeh Ghassemi
Frank Rudzicz
76
4
0
17 Oct 2021
Schrödinger's Tree -- On Syntax and Neural Language Models
Schrödinger's Tree -- On Syntax and Neural Language Models
Artur Kulmizev
Joakim Nivre
75
6
0
17 Oct 2021
Previous
123...91011...151617
Next