ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.05950
  4. Cited By
BERT Rediscovers the Classical NLP Pipeline

BERT Rediscovers the Classical NLP Pipeline

15 May 2019
Ian Tenney
Dipanjan Das
Ellie Pavlick
    MILM
    SSeg
ArXivPDFHTML

Papers citing "BERT Rediscovers the Classical NLP Pipeline"

50 / 296 papers shown
Title
Probing for Labeled Dependency Trees
Probing for Labeled Dependency Trees
Max Müller-Eberstein
Rob van der Goot
Barbara Plank
19
7
0
24 Mar 2022
Coloring the Blank Slate: Pre-training Imparts a Hierarchical Inductive
  Bias to Sequence-to-sequence Models
Coloring the Blank Slate: Pre-training Imparts a Hierarchical Inductive Bias to Sequence-to-sequence Models
Aaron Mueller
Robert Frank
Tal Linzen
Luheng Wang
Sebastian Schuster
AIMat
19
33
0
17 Mar 2022
Contrastive Visual Semantic Pretraining Magnifies the Semantics of
  Natural Language Representations
Contrastive Visual Semantic Pretraining Magnifies the Semantics of Natural Language Representations
Robert Wolfe
Aylin Caliskan
VLM
21
13
0
14 Mar 2022
Grounding Commands for Autonomous Vehicles via Layer Fusion with
  Region-specific Dynamic Layer Attention
Grounding Commands for Autonomous Vehicles via Layer Fusion with Region-specific Dynamic Layer Attention
Hou Pong Chan
M. Guo
Chengguang Xu
30
4
0
14 Mar 2022
TrimBERT: Tailoring BERT for Trade-offs
TrimBERT: Tailoring BERT for Trade-offs
S. N. Sridhar
Anthony Sarah
Sairam Sundaresan
MQ
21
4
0
24 Feb 2022
Probing BERT's priors with serial reproduction chains
Probing BERT's priors with serial reproduction chains
Takateru Yamakoshi
Thomas L. Griffiths
Robert D. Hawkins
29
12
0
24 Feb 2022
Do Transformers know symbolic rules, and would we know if they did?
Do Transformers know symbolic rules, and would we know if they did?
Tommi Gröndahl
Yu-Wen Guo
Nirmal Asokan
27
0
0
19 Feb 2022
Open-Ended Reinforcement Learning with Neural Reward Functions
Open-Ended Reinforcement Learning with Neural Reward Functions
Robert Meier
Asier Mujika
37
7
0
16 Feb 2022
What Do They Capture? -- A Structural Analysis of Pre-Trained Language
  Models for Source Code
What Do They Capture? -- A Structural Analysis of Pre-Trained Language Models for Source Code
Yao Wan
Wei-Ye Zhao
Hongyu Zhang
Yulei Sui
Guandong Xu
Hairong Jin
35
105
0
14 Feb 2022
TourBERT: A pretrained language model for the tourism industry
TourBERT: A pretrained language model for the tourism industry
Veronika Arefieva
R. Egger
14
4
0
19 Jan 2022
Interpreting Arabic Transformer Models
Ahmed Abdelali
Nadir Durrani
Fahim Dalvi
Hassan Sajjad
41
2
0
19 Jan 2022
Does Entity Abstraction Help Generative Transformers Reason?
Does Entity Abstraction Help Generative Transformers Reason?
Nicolas Angelard-Gontier
Siva Reddy
C. Pal
34
5
0
05 Jan 2022
How Should Pre-Trained Language Models Be Fine-Tuned Towards Adversarial
  Robustness?
How Should Pre-Trained Language Models Be Fine-Tuned Towards Adversarial Robustness?
Xinhsuai Dong
Anh Tuan Luu
Min-Bin Lin
Shuicheng Yan
Hanwang Zhang
SILM
AAML
20
55
0
22 Dec 2021
Linguistic Frameworks Go Toe-to-Toe at Neuro-Symbolic Language Modeling
Linguistic Frameworks Go Toe-to-Toe at Neuro-Symbolic Language Modeling
Jakob Prange
Nathan Schneider
Lingpeng Kong
22
10
0
15 Dec 2021
Inducing Causal Structure for Interpretable Neural Networks
Inducing Causal Structure for Interpretable Neural Networks
Atticus Geiger
Zhengxuan Wu
Hanson Lu
J. Rozner
Elisa Kreiss
Thomas Icard
Noah D. Goodman
Christopher Potts
CML
OOD
35
71
0
01 Dec 2021
To Augment or Not to Augment? A Comparative Study on Text Augmentation
  Techniques for Low-Resource NLP
To Augment or Not to Augment? A Comparative Study on Text Augmentation Techniques for Low-Resource NLP
Gözde Gül Sahin
40
33
0
18 Nov 2021
Interpreting Language Models Through Knowledge Graph Extraction
Interpreting Language Models Through Knowledge Graph Extraction
Vinitra Swamy
Angelika Romanou
Martin Jaggi
30
20
0
16 Nov 2021
Discovering Supply Chain Links with Augmented Intelligence
Discovering Supply Chain Links with Augmented Intelligence
Achintya Gopal
Chun-Han Chang
40
3
0
02 Nov 2021
LMdiff: A Visual Diff Tool to Compare Language Models
LMdiff: A Visual Diff Tool to Compare Language Models
Hendrik Strobelt
Benjamin Hoover
Arvind Satyanarayan
Sebastian Gehrmann
VLM
37
19
0
02 Nov 2021
Recent Advances in Natural Language Processing via Large Pre-Trained
  Language Models: A Survey
Recent Advances in Natural Language Processing via Large Pre-Trained Language Models: A Survey
Bonan Min
Hayley L Ross
Elior Sulem
Amir Pouran Ben Veyseh
Thien Huu Nguyen
Oscar Sainz
Eneko Agirre
Ilana Heinz
Dan Roth
LM&MA
VLM
AI4CE
83
1,035
0
01 Nov 2021
Interpreting Deep Learning Models in Natural Language Processing: A
  Review
Interpreting Deep Learning Models in Natural Language Processing: A Review
Xiaofei Sun
Diyi Yang
Xiaoya Li
Tianwei Zhang
Yuxian Meng
Han Qiu
Guoyin Wang
Eduard H. Hovy
Jiwei Li
19
44
0
20 Oct 2021
Inductive Biases and Variable Creation in Self-Attention Mechanisms
Inductive Biases and Variable Creation in Self-Attention Mechanisms
Benjamin L. Edelman
Surbhi Goel
Sham Kakade
Cyril Zhang
27
117
0
19 Oct 2021
BERMo: What can BERT learn from ELMo?
BERMo: What can BERT learn from ELMo?
Sangamesh Kodge
Kaushik Roy
38
3
0
18 Oct 2021
Identifying Introductions in Podcast Episodes from Automatically
  Generated Transcripts
Identifying Introductions in Podcast Episodes from Automatically Generated Transcripts
Elise Jing
K. Schneck
Dennis Egan
Scott A. Waterman
26
2
0
14 Oct 2021
Global Explainability of BERT-Based Evaluation Metrics by Disentangling
  along Linguistic Factors
Global Explainability of BERT-Based Evaluation Metrics by Disentangling along Linguistic Factors
Marvin Kaster
Wei-Ye Zhao
Steffen Eger
33
24
0
08 Oct 2021
Text analysis and deep learning: A network approach
Text analysis and deep learning: A network approach
Ingo Marquart
25
0
0
08 Oct 2021
BadPre: Task-agnostic Backdoor Attacks to Pre-trained NLP Foundation
  Models
BadPre: Task-agnostic Backdoor Attacks to Pre-trained NLP Foundation Models
Kangjie Chen
Yuxian Meng
Xiaofei Sun
Shangwei Guo
Tianwei Zhang
Jiwei Li
Chun Fan
SILM
34
106
0
06 Oct 2021
MoEfication: Transformer Feed-forward Layers are Mixtures of Experts
MoEfication: Transformer Feed-forward Layers are Mixtures of Experts
Zhengyan Zhang
Yankai Lin
Zhiyuan Liu
Peng Li
Maosong Sun
Jie Zhou
MoE
27
117
0
05 Oct 2021
Low Frequency Names Exhibit Bias and Overfitting in Contextualizing
  Language Models
Low Frequency Names Exhibit Bias and Overfitting in Contextualizing Language Models
Robert Wolfe
Aylin Caliskan
87
51
0
01 Oct 2021
SlovakBERT: Slovak Masked Language Model
SlovakBERT: Slovak Masked Language Model
Matúš Pikuliak
Stefan Grivalsky
Martin Konopka
Miroslav Blšták
Martin Tamajka
Viktor Bachratý
Marian Simko
Pavol Balázik
Michal Trnka
Filip Uhlárik
35
26
0
30 Sep 2021
Analysing the Effect of Masking Length Distribution of MLM: An
  Evaluation Framework and Case Study on Chinese MRC Datasets
Analysing the Effect of Masking Length Distribution of MLM: An Evaluation Framework and Case Study on Chinese MRC Datasets
Changchang Zeng
Shaobo Li
24
6
0
29 Sep 2021
Fine-Tuned Transformers Show Clusters of Similar Representations Across
  Layers
Fine-Tuned Transformers Show Clusters of Similar Representations Across Layers
Jason Phang
Haokun Liu
Samuel R. Bowman
30
25
0
17 Sep 2021
Incorporating Residual and Normalization Layers into Analysis of Masked
  Language Models
Incorporating Residual and Normalization Layers into Analysis of Masked Language Models
Goro Kobayashi
Tatsuki Kuribayashi
Sho Yokoi
Kentaro Inui
160
46
0
15 Sep 2021
The Stem Cell Hypothesis: Dilemma behind Multi-Task Learning with
  Transformer Encoders
The Stem Cell Hypothesis: Dilemma behind Multi-Task Learning with Transformer Encoders
Han He
Jinho Choi
51
87
0
14 Sep 2021
ePiC: Employing Proverbs in Context as a Benchmark for Abstract Language
  Understanding
ePiC: Employing Proverbs in Context as a Benchmark for Abstract Language Understanding
Sayan Ghosh
Shashank Srivastava
18
11
0
14 Sep 2021
Can Language Models Encode Perceptual Structure Without Grounding? A
  Case Study in Color
Can Language Models Encode Perceptual Structure Without Grounding? A Case Study in Color
Mostafa Abdou
Artur Kulmizev
Daniel Hershcovich
Stella Frank
Ellie Pavlick
Anders Søgaard
22
114
0
13 Sep 2021
Not All Models Localize Linguistic Knowledge in the Same Place: A
  Layer-wise Probing on BERToids' Representations
Not All Models Localize Linguistic Knowledge in the Same Place: A Layer-wise Probing on BERToids' Representations
Mohsen Fayyaz
Ehsan Aghazadeh
Ali Modarressi
Hosein Mohebbi
Mohammad Taher Pilehvar
18
21
0
13 Sep 2021
Tiered Reasoning for Intuitive Physics: Toward Verifiable Commonsense
  Language Understanding
Tiered Reasoning for Intuitive Physics: Toward Verifiable Commonsense Language Understanding
Shane Storks
Qiaozi Gao
Yichi Zhang
J. Chai
ReLM
LRM
49
22
0
10 Sep 2021
Studying word order through iterative shuffling
Studying word order through iterative shuffling
Nikolay Malkin
Sameera Lanka
Pranav Goel
Nebojsa Jojic
31
14
0
10 Sep 2021
Avoiding Inference Heuristics in Few-shot Prompt-based Finetuning
Avoiding Inference Heuristics in Few-shot Prompt-based Finetuning
Prasetya Ajie Utama
N. Moosavi
Victor Sanh
Iryna Gurevych
AAML
61
35
0
09 Sep 2021
Mixed Attention Transformer for Leveraging Word-Level Knowledge to
  Neural Cross-Lingual Information Retrieval
Mixed Attention Transformer for Leveraging Word-Level Knowledge to Neural Cross-Lingual Information Retrieval
Zhiqi Huang
Hamed Bonab
Sheikh Muhammad Sarwar
Razieh Rahimi
James Allan
40
10
0
07 Sep 2021
BERT might be Overkill: A Tiny but Effective Biomedical Entity Linker
  based on Residual Convolutional Neural Networks
BERT might be Overkill: A Tiny but Effective Biomedical Entity Linker based on Residual Convolutional Neural Networks
T. Lai
Heng Ji
ChengXiang Zhai
30
31
0
06 Sep 2021
How Does Adversarial Fine-Tuning Benefit BERT?
How Does Adversarial Fine-Tuning Benefit BERT?
J. Ebrahimi
Hao Yang
Wei Zhang
AAML
26
4
0
31 Aug 2021
Neuron-level Interpretation of Deep NLP Models: A Survey
Neuron-level Interpretation of Deep NLP Models: A Survey
Hassan Sajjad
Nadir Durrani
Fahim Dalvi
MILM
AI4CE
35
80
0
30 Aug 2021
Rethinking Why Intermediate-Task Fine-Tuning Works
Rethinking Why Intermediate-Task Fine-Tuning Works
Ting-Yun Chang
Chi-Jen Lu
LRM
19
29
0
26 Aug 2021
ComSum: Commit Messages Summarization and Meaning Preservation
ComSum: Commit Messages Summarization and Meaning Preservation
Leshem Choshen
Idan Amit
17
4
0
23 Aug 2021
Representation learning for neural population activity with Neural Data
  Transformers
Representation learning for neural population activity with Neural Data Transformers
Joel Ye
C. Pandarinath
AI4TS
AI4CE
11
52
0
02 Aug 2021
Evaluation of contextual embeddings on less-resourced languages
Evaluation of contextual embeddings on less-resourced languages
Matej Ulvcar
Alevs vZagar
C. S. Armendariz
Andravz Repar
Senja Pollak
Matthew Purver
Marko Robnik-vSikonja
36
11
0
22 Jul 2021
Trusting RoBERTa over BERT: Insights from CheckListing the Natural
  Language Inference Task
Trusting RoBERTa over BERT: Insights from CheckListing the Natural Language Inference Task
Ishan Tarunesh
Somak Aditya
Monojit Choudhury
15
17
0
15 Jul 2021
Hate versus Politics: Detection of Hate against Policy makers in Italian
  tweets
Hate versus Politics: Detection of Hate against Policy makers in Italian tweets
Armend Duzha
Cristiano Casadei
Michael Tosi
Fabio Celli
25
6
0
12 Jul 2021
Previous
123456
Next