Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1905.05950
Cited By
v1
v2 (latest)
BERT Rediscovers the Classical NLP Pipeline
15 May 2019
Ian Tenney
Dipanjan Das
Ellie Pavlick
MILM
SSeg
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"BERT Rediscovers the Classical NLP Pipeline"
50 / 821 papers shown
Title
Exploring the Relationship between Alignment and Cross-lingual Transfer in Multilingual Transformers
Félix Gaschi
Patricio Cerda
Parisa Rastin
Y. Toussaint
88
13
0
05 Jun 2023
A Mathematical Abstraction for Balancing the Trade-off Between Creativity and Reality in Large Language Models
Ritwik Sinha
Zhao Song
Dinesh Manocha
102
25
0
04 Jun 2023
A Simple yet Effective Self-Debiasing Framework for Transformer Models
Xiaoyue Wang
Lijie Wang
Xin Liu
Suhang Wu
Jinsong Su
Huasen Wu
68
4
0
02 Jun 2023
Learning Transformer Programs
Dan Friedman
Alexander Wettig
Danqi Chen
89
36
0
01 Jun 2023
Empirical Sufficiency Lower Bounds for Language Modeling with Locally-Bootstrapped Semantic Structures
Jakob Prange
Emmanuele Chersoni
76
0
0
30 May 2023
A Method for Studying Semantic Construal in Grammatical Constructions with Interpretable Contextual Embedding Spaces
Gabriella Chronis
Kyle Mahowald
K. Erk
62
10
0
29 May 2023
NeuroX Library for Neuron Analysis of Deep NLP Models
Fahim Dalvi
Hassan Sajjad
Nadir Durrani
71
11
0
26 May 2023
Large Language Models Are Partially Primed in Pronoun Interpretation
S. Lam
Qingcheng Zeng
Kexun Zhang
Chenyu You
Rob Voigt
52
4
0
26 May 2023
Not wacky vs. definitely wacky: A study of scalar adverbs in pretrained language models
Isabelle Lorge
J. Pierrehumbert
70
0
0
25 May 2023
Language Models Implement Simple Word2Vec-style Vector Arithmetic
Jack Merullo
Carsten Eickhoff
Ellie Pavlick
KELM
95
66
0
25 May 2023
Towards Adaptive Prefix Tuning for Parameter-Efficient Language Model Fine-tuning
Zhen-Ru Zhang
Chuanqi Tan
Haiyang Xu
Chengyu Wang
Jun Huang
Songfang Huang
73
38
0
24 May 2023
MuLER: Detailed and Scalable Reference-based Evaluation
Taelin Karidi
Leshem Choshen
Gal Patel
Omri Abend
74
0
0
24 May 2023
Measuring the Knowledge Acquisition-Utilization Gap in Pretrained Language Models
Amirhossein Kazemnejad
Mehdi Rezagholizadeh
Prasanna Parthasarathi
Sarath Chandar
ELM
60
2
0
24 May 2023
All Roads Lead to Rome? Exploring the Invariance of Transformers' Representations
Yuxin Ren
Qipeng Guo
Zhijing Jin
Shauli Ravfogel
Mrinmaya Sachan
Bernhard Schölkopf
Ryan Cotterell
77
4
0
23 May 2023
On Robustness of Finetuned Transformer-based NLP Models
Pavan Kalyan Reddy Neerudu
Subba Reddy Oota
Mounika Marreddy
Venkateswara Rao Kagita
Manish Gupta
81
9
0
23 May 2023
Revisiting Acceptability Judgements
Hai Hu
Ziyin Zhang
Wei-Ping Huang
J. Lai
Aini Li
Yi Ma
Jiahui Huang
Peng Zhang
Chien-Jer Charles Lin
Rui Wang
71
2
0
23 May 2023
Can Language Models Understand Physical Concepts?
Lei Li
Jingjing Xu
Qingxiu Dong
Ce Zheng
Qi Liu
Lingpeng Kong
Xu Sun
ALM
61
22
0
23 May 2023
A Trip Towards Fairness: Bias and De-Biasing in Large Language Models
Leonardo Ranaldi
Elena Sofia Ruzzetti
Davide Venditti
Dario Onorati
Fabio Massimo Zanzotto
93
37
0
23 May 2023
Automatic Readability Assessment for Closely Related Languages
Joseph Marvin Imperial
E. Kochmar
57
9
0
22 May 2023
Can LLMs facilitate interpretation of pre-trained language models?
Basel Mousi
Nadir Durrani
Fahim Dalvi
93
13
0
22 May 2023
Should We Attend More or Less? Modulating Attention for Fairness
A. Zayed
Gonçalo Mordido
Samira Shabanian
Sarath Chandar
83
10
0
22 May 2023
Explaining How Transformers Use Context to Build Predictions
Javier Ferrando
Gerard I. Gállego
Ioannis Tsiamas
Marta R. Costa-jussá
65
37
0
21 May 2023
Interactive Learning of Hierarchical Tasks from Dialog with GPT
Lane Lawley
Christopher MacLellan
RALM
15
1
0
17 May 2023
Probing the Role of Positional Information in Vision-Language Models
Philipp J. Rösch
Jindrich Libovický
63
8
0
17 May 2023
Interpretability at Scale: Identifying Causal Mechanisms in Alpaca
Zhengxuan Wu
Atticus Geiger
Thomas Icard
Christopher Potts
Noah D. Goodman
MILM
87
93
0
15 May 2023
Investigating Forgetting in Pre-Trained Representations Through Continual Learning
Yun Luo
Zhen Yang
Xuefeng Bai
Fandong Meng
Jie Zhou
Yue Zhang
CLL
KELM
103
17
0
10 May 2023
Self-supervised Learning for Pre-Training 3D Point Clouds: A Survey
Ben Fei
Weidong Yang
Liwen Liu
Tian-jian Luo
Rui Zhang
Yixuan Li
Ying He
3DPC
113
19
0
08 May 2023
PreCog: Exploring the Relation between Memorization and Performance in Pre-trained Language Models
Leonardo Ranaldi
Elena Sofia Ruzzetti
Fabio Massimo Zanzotto
69
6
0
08 May 2023
Improved Logical Reasoning of Language Models via Differentiable Symbolic Programming
Hanlin Zhang
Jiani Huang
Ziyang Li
Mayur Naik
Eric P. Xing
ReLM
LRM
84
28
0
05 May 2023
AttentionViz: A Global View of Transformer Attention
Catherine Yeh
Yida Chen
Aoyu Wu
Cynthia Chen
Fernanda Viégas
Martin Wattenberg
ViT
79
55
0
04 May 2023
Learning Language-Specific Layers for Multilingual Machine Translation
Telmo Pires
Robin M. Schmidt
Yi-Hsiu Liao
Stephan Peitz
109
19
0
04 May 2023
The System Model and the User Model: Exploring AI Dashboard Design
Fernanda Viégas
Martin Wattenberg
58
6
0
04 May 2023
Approximating CKY with Transformers
Ghazal Khalighinejad
Ollie Liu
Sam Wiseman
109
2
0
03 May 2023
Entity Tracking in Language Models
Najoung Kim
Sebastian Schuster
147
22
0
03 May 2023
Exploring Linguistic Properties of Monolingual BERTs with Typological Classification among Languages
Elena Sofia Ruzzetti
Federico Ranaldi
F. Logozzo
Michele Mastromattei
Leonardo Ranaldi
Fabio Massimo Zanzotto
66
9
0
03 May 2023
Finding Neurons in a Haystack: Case Studies with Sparse Probing
Wes Gurnee
Neel Nanda
Matthew Pauly
Katherine Harvey
Dmitrii Troitskii
Dimitris Bertsimas
MILM
282
218
0
02 May 2023
Evaluating statistical language models as pragmatic reasoners
Benjamin Lipkin
L. Wong
Gabriel Grand
J. Tenenbaum
133
15
0
01 May 2023
Redundancy and Concept Analysis for Code-trained Language Models
Arushi Sharma
Zefu Hu
Christopher Quinn
Ali Jannesari
140
2
0
01 May 2023
How does GPT-2 compute greater-than?: Interpreting mathematical abilities in a pre-trained language model
Michael Hanna
Ollie Liu
Alexandre Variengien
LRM
307
132
0
30 Apr 2023
Dissecting Recall of Factual Associations in Auto-Regressive Language Models
Mor Geva
Jasmijn Bastings
Katja Filippova
Amir Globerson
KELM
287
324
0
28 Apr 2023
The Closeness of In-Context Learning and Weight Shifting for Softmax Regression
Shuai Li
Zhao Song
Yu Xia
Tong Yu
Dinesh Manocha
84
43
0
26 Apr 2023
What does BERT learn about prosody?
Sofoklis Kakouros
Johannah O'Mahony
MILM
55
6
0
25 Apr 2023
Towards Efficient Fine-tuning of Pre-trained Code Models: An Experimental Study and Beyond
Ensheng Shi
Yanlin Wang
Hongyu Zhang
Lun Du
Shi Han
Dongmei Zhang
Hongbin Sun
89
45
0
11 Apr 2023
Low-Shot Learning for Fictional Claim Verification
Viswanath Chadalapaka
Derek Nguyen
Joonwon Choi
Shaunak Joshi
Mohammad Rostami
47
1
0
05 Apr 2023
Larger Probes Tell a Different Story: Extending Psycholinguistic Datasets Via In-Context Learning
Namrata Shivagunde
Vladislav Lialin
Anna Rumshisky
56
1
0
29 Mar 2023
Coupling Artificial Neurons in BERT and Biological Neurons in the Human Brain
Xu Liu
Mengyue Zhou
Gaosheng Shi
Yu Du
Lin Zhao
Zihao Wu
David Liu
Tianming Liu
Xintao Hu
77
11
0
27 Mar 2023
Language Model Behavior: A Comprehensive Survey
Tyler A. Chang
Benjamin Bergen
VLM
LRM
LM&MA
115
109
0
20 Mar 2023
Neural Architecture Search for Effective Teacher-Student Knowledge Transfer in Language Models
Aashka Trivedi
Takuma Udagawa
Michele Merler
Yikang Shen
Yousef El-Kurdi
Bishwaranjan Bhattacharjee
84
7
0
16 Mar 2023
Jump to Conclusions: Short-Cutting Transformers With Linear Transformations
Alexander Yom Din
Taelin Karidi
Leshem Choshen
Mor Geva
61
68
0
16 Mar 2023
Do large language models resemble humans in language use?
Zhenguang G. Cai
Xufeng Duan
David A. Haslett
Shuqi Wang
M. Pickering
ALM
127
41
0
10 Mar 2023
Previous
1
2
3
...
5
6
7
...
15
16
17
Next