ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.12471
  4. Cited By
Neural Network Acceptability Judgments
v1v2v3 (latest)

Neural Network Acceptability Judgments

31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
ArXiv (abs)PDFHTML

Papers citing "Neural Network Acceptability Judgments"

50 / 894 papers shown
Title
Revisiting Token Dropping Strategy in Efficient BERT Pretraining
Revisiting Token Dropping Strategy in Efficient BERT Pretraining
Qihuang Zhong
Liang Ding
Juhua Liu
Xuebo Liu
Min Zhang
Bo Du
Dacheng Tao
VLM
73
10
0
24 May 2023
Inference-Time Policy Adapters (IPA): Tailoring Extreme-Scale LMs
  without Fine-tuning
Inference-Time Policy Adapters (IPA): Tailoring Extreme-Scale LMs without Fine-tuning
Ximing Lu
Faeze Brahman
Peter West
Jaehun Jang
Khyathi Chandu
...
Bill Yuchen Lin
Skyler Hallinan
Xiang Ren
Sean Welleck
Yejin Choi
120
29
0
24 May 2023
Trade-Offs Between Fairness and Privacy in Language Modeling
Trade-Offs Between Fairness and Privacy in Language Modeling
Cleo Matzken
Steffen Eger
Ivan Habernal
SILM
112
6
0
24 May 2023
On Robustness of Finetuned Transformer-based NLP Models
On Robustness of Finetuned Transformer-based NLP Models
Pavan Kalyan Reddy Neerudu
Subba Reddy Oota
Mounika Marreddy
Venkateswara Rao Kagita
Manish Gupta
81
9
0
23 May 2023
Revisiting Acceptability Judgements
Revisiting Acceptability Judgements
Hai Hu
Ziyin Zhang
Wei-Ping Huang
J. Lai
Aini Li
Yi Ma
Jiahui Huang
Peng Zhang
Chien-Jer Charles Lin
Rui Wang
71
2
0
23 May 2023
A Trip Towards Fairness: Bias and De-Biasing in Large Language Models
A Trip Towards Fairness: Bias and De-Biasing in Large Language Models
Leonardo Ranaldi
Elena Sofia Ruzzetti
Davide Venditti
Dario Onorati
Fabio Massimo Zanzotto
93
37
0
23 May 2023
DADA: Dialect Adaptation via Dynamic Aggregation of Linguistic Rules
DADA: Dialect Adaptation via Dynamic Aggregation of Linguistic Rules
Yanchen Liu
William B. Held
Diyi Yang
148
11
0
22 May 2023
Learning Interpretable Style Embeddings via Prompting LLMs
Learning Interpretable Style Embeddings via Prompting LLMs
Ajay Patel
D. Rao
Ansh Kothary
Kathleen McKeown
Chris Callison-Burch
93
26
0
22 May 2023
Task-agnostic Distillation of Encoder-Decoder Language Models
Task-agnostic Distillation of Encoder-Decoder Language Models
Chen Zhang
Yang Yang
Jingang Wang
Dawei Song
62
5
0
21 May 2023
Mitigating Catastrophic Forgetting in Task-Incremental Continual
  Learning with Adaptive Classification Criterion
Mitigating Catastrophic Forgetting in Task-Incremental Continual Learning with Adaptive Classification Criterion
Yun Luo
Xiaotian Lin
Zhen Yang
Fandong Meng
Jie Zhou
Yue Zhang
CLL
68
5
0
20 May 2023
Dynamic Transformers Provide a False Sense of Efficiency
Dynamic Transformers Provide a False Sense of Efficiency
Yiming Chen
Simin Chen
Zexin Li
Wei Yang
Cong Liu
R. Tan
Haizhou Li
AAML
90
12
0
20 May 2023
Lifting the Curse of Capacity Gap in Distilling Language Models
Lifting the Curse of Capacity Gap in Distilling Language Models
Chen Zhang
Yang Yang
Jiahao Liu
Jingang Wang
Yunsen Xian
Benyou Wang
Dawei Song
MoE
69
20
0
20 May 2023
Prompting with Pseudo-Code Instructions
Prompting with Pseudo-Code Instructions
Mayank Mishra
Praveen Venkateswaran
Riyaz Ahmad Bhat
V. Rudramurthy
Danish Contractor
Srikanth G. Tamilselvam
103
14
0
19 May 2023
How does the task complexity of masked pretraining objectives affect
  downstream performance?
How does the task complexity of masked pretraining objectives affect downstream performance?
Atsuki Yamaguchi
Hiroaki Ozaki
Terufumi Morishita
Gaku Morio
Yasuhiro Sogawa
82
2
0
18 May 2023
Towards More Robust NLP System Evaluation: Handling Missing Scores in
  Benchmarks
Towards More Robust NLP System Evaluation: Handling Missing Scores in Benchmarks
Anas Himmi
Ekhine Irurozki
Nathan Noiry
Stephan Clémençon
Pierre Colombo
193
9
0
17 May 2023
AD-KD: Attribution-Driven Knowledge Distillation for Language Model
  Compression
AD-KD: Attribution-Driven Knowledge Distillation for Language Model Compression
Siyue Wu
Hongzhan Chen
Xiaojun Quan
Qifan Wang
Rui Wang
VLM
86
20
0
17 May 2023
On Dataset Transferability in Active Learning for Transformers
On Dataset Transferability in Active Learning for Transformers
Fran Jelenić
Josip Jukić
Nina Drobac
Jan vSnajder
66
2
0
16 May 2023
Weight-Inherited Distillation for Task-Agnostic BERT Compression
Weight-Inherited Distillation for Task-Agnostic BERT Compression
Taiqiang Wu
Cheng-An Hou
Shanshan Lao
Jiayi Li
Ngai Wong
Zhe Zhao
Yujiu Yang
129
10
0
16 May 2023
TESS: Text-to-Text Self-Conditioned Simplex Diffusion
TESS: Text-to-Text Self-Conditioned Simplex Diffusion
Rabeeh Karimi Mahabadi
Hamish Ivison
Jaesung Tae
James Henderson
Iz Beltagy
Matthew E. Peters
Arman Cohan
100
28
0
15 May 2023
Learning Non-linguistic Skills without Sacrificing Linguistic
  Proficiency
Learning Non-linguistic Skills without Sacrificing Linguistic Proficiency
Mandar Sharma
Nikhil Muralidhar
Naren Ramakrishnan
CLL
59
4
0
14 May 2023
Frequency-aware Dimension Selection for Static Word Embedding by Mixed
  Product Distance
Frequency-aware Dimension Selection for Static Word Embedding by Mixed Product Distance
Lingfeng Shen
Haiyun Jiang
Lemao Liu
Ying Chen
50
0
0
13 May 2023
Towards Versatile and Efficient Visual Knowledge Integration into
  Pre-trained Language Models with Cross-Modal Adapters
Towards Versatile and Efficient Visual Knowledge Integration into Pre-trained Language Models with Cross-Modal Adapters
Xinyun Zhang
Haochen Tan
Han Wu
Bei Yu
KELM
36
1
0
12 May 2023
Accelerating Batch Active Learning Using Continual Learning Techniques
Accelerating Batch Active Learning Using Continual Learning Techniques
Arnav M. Das
Gantavya Bhatt
M. Bhalerao
Vianne R. Gao
Rui Yang
J. Bilmes
VLMCLL
68
10
0
10 May 2023
Investigating Forgetting in Pre-Trained Representations Through
  Continual Learning
Investigating Forgetting in Pre-Trained Representations Through Continual Learning
Yun Luo
Zhen Yang
Xuefeng Bai
Fandong Meng
Jie Zhou
Yue Zhang
CLLKELM
103
17
0
10 May 2023
WikiSQE: A Large-Scale Dataset for Sentence Quality Estimation in
  Wikipedia
WikiSQE: A Large-Scale Dataset for Sentence Quality Estimation in Wikipedia
Kenichiro Ando
Satoshi Sekine
Mamoru Komachi
66
2
0
10 May 2023
PreCog: Exploring the Relation between Memorization and Performance in
  Pre-trained Language Models
PreCog: Exploring the Relation between Memorization and Performance in Pre-trained Language Models
Leonardo Ranaldi
Elena Sofia Ruzzetti
Fabio Massimo Zanzotto
69
6
0
08 May 2023
Pre-training Language Model as a Multi-perspective Course Learner
Pre-training Language Model as a Multi-perspective Course Learner
Beiduo Chen
Shaohan Huang
Zi-qiang Zhang
Wu Guo
Zhen-Hua Ling
Haizhen Huang
Furu Wei
Weiwei Deng
Qi Zhang
56
0
0
06 May 2023
Where We Have Arrived in Proving the Emergence of Sparse Symbolic
  Concepts in AI Models
Where We Have Arrived in Proving the Emergence of Sparse Symbolic Concepts in AI Models
Qihan Ren
Maximilian Brunner
Wen Shen
S. Mintchev
91
12
0
03 May 2023
Towards Imperceptible Document Manipulations against Neural Ranking
  Models
Towards Imperceptible Document Manipulations against Neural Ranking Models
Xuanang Chen
Xianpei Han
Zheng Ye
Le Sun
Yingfei Sun
AAML
79
17
0
03 May 2023
Don't Stop Pretraining? Make Prompt-based Fine-tuning Powerful Learner
Don't Stop Pretraining? Make Prompt-based Fine-tuning Powerful Learner
Zhengxiang Shi
Aldo Lipani
VLMCLL
80
22
0
02 May 2023
ISAAC Newton: Input-based Approximate Curvature for Newton's Method
ISAAC Newton: Input-based Approximate Curvature for Newton's Method
Felix Petersen
Tobias Sutter
Christian Borgelt
Dongsung Huh
Hilde Kuehne
Yuekai Sun
Oliver Deussen
ODL
84
5
0
01 May 2023
Empirical Analysis of the Strengths and Weaknesses of PEFT Techniques
  for LLMs
Empirical Analysis of the Strengths and Weaknesses of PEFT Techniques for LLMs
George Pu
Anirudh Jain
Jihan Yin
Russell Kaplan
75
43
0
28 Apr 2023
ChatGPT as an Attack Tool: Stealthy Textual Backdoor Attack via Blackbox
  Generative Model Trigger
ChatGPT as an Attack Tool: Stealthy Textual Backdoor Attack via Blackbox Generative Model Trigger
Jiazhao Li
Yijin Yang
Zhuofeng Wu
V. Vydiswaran
Chaowei Xiao
SILM
167
46
0
27 Apr 2023
ChatLog: Carefully Evaluating the Evolution of ChatGPT Across Time
ChatLog: Carefully Evaluating the Evolution of ChatGPT Across Time
Shangqing Tu
Chunyang Li
Jifan Yu
Xiaozhi Wang
Lei Hou
Juanzi Li
LLMAGAI4MH
158
10
0
27 Apr 2023
Can BERT eat RuCoLA? Topological Data Analysis to Explain
Can BERT eat RuCoLA? Topological Data Analysis to Explain
Irina Proskurina
Irina Piontkovskaya
Ekaterina Artemova
107
4
0
04 Apr 2023
ScandEval: A Benchmark for Scandinavian Natural Language Processing
ScandEval: A Benchmark for Scandinavian Natural Language Processing
Dan Saattrup Nielsen
ELM
80
14
0
03 Apr 2023
TransCODE: Co-design of Transformers and Accelerators for Efficient
  Training and Inference
TransCODE: Co-design of Transformers and Accelerators for Efficient Training and Inference
Shikhar Tuli
N. Jha
103
6
0
27 Mar 2023
Trained on 100 million words and still in shape: BERT meets British
  National Corpus
Trained on 100 million words and still in shape: BERT meets British National Corpus
David Samuel
Andrey Kutuzov
Lilja Øvrelid
Erik Velldal
99
32
0
17 Mar 2023
SmartBERT: A Promotion of Dynamic Early Exiting Mechanism for
  Accelerating BERT Inference
SmartBERT: A Promotion of Dynamic Early Exiting Mechanism for Accelerating BERT Inference
Boren Hu
Yun Zhu
Jiacheng Li
Siliang Tang
58
9
0
16 Mar 2023
Gradient-Free Structured Pruning with Unlabeled Data
Gradient-Free Structured Pruning with Unlabeled Data
Azade Nova
H. Dai
Dale Schuurmans
SyDa
84
22
0
07 Mar 2023
Multitask Prompt Tuning Enables Parameter-Efficient Transfer Learning
Multitask Prompt Tuning Enables Parameter-Efficient Transfer Learning
Zhen Wang
Yikang Shen
Leonid Karlinsky
Rogerio Feris
Huan Sun
Yoon Kim
VLMVPVLM
96
118
0
06 Mar 2023
Revisiting the Noise Model of Stochastic Gradient Descent
Revisiting the Noise Model of Stochastic Gradient Descent
Barak Battash
Ofir Lindenbaum
56
11
0
05 Mar 2023
Calibrating Transformers via Sparse Gaussian Processes
Calibrating Transformers via Sparse Gaussian Processes
Wenlong Chen
Yingzhen Li
UQCV
108
12
0
04 Mar 2023
Bayesian Neural Networks Avoid Encoding Complex and
  Perturbation-Sensitive Concepts
Bayesian Neural Networks Avoid Encoding Complex and Perturbation-Sensitive Concepts
Qihan Ren
Huiqi Deng
Yunuo Chen
Siyu Lou
Quanshi Zhang
BDLAAML
87
10
0
25 Feb 2023
Does a Neural Network Really Encode Symbolic Concepts?
Does a Neural Network Really Encode Symbolic Concepts?
Mingjie Li
Quanshi Zhang
91
31
0
25 Feb 2023
MUX-PLMs: Data Multiplexing for High-throughput Language Models
MUX-PLMs: Data Multiplexing for High-throughput Language Models
Vishvak Murahari
Ameet Deshpande
Carlos E. Jimenez
Izhak Shafran
Mingqiu Wang
Yuan Cao
Karthik Narasimhan
MoE
59
5
0
24 Feb 2023
ChatGPT: Jack of all trades, master of none
ChatGPT: Jack of all trades, master of none
Jan Kocoñ
Igor Cichecki
Oliwier Kaszyca
Mateusz Kochanek
Dominika Szydło
...
Maciej Piasecki
Lukasz Radliñski
Konrad Wojtasik
Stanislaw Wo'zniak
Przemyslaw Kazienko
AI4MH
168
558
0
21 Feb 2023
HomoDistil: Homotopic Task-Agnostic Distillation of Pre-trained
  Transformers
HomoDistil: Homotopic Task-Agnostic Distillation of Pre-trained Transformers
Chen Liang
Haoming Jiang
Zheng Li
Xianfeng Tang
Bin Yin
Tuo Zhao
VLM
126
25
0
19 Feb 2023
Can ChatGPT Understand Too? A Comparative Study on ChatGPT and
  Fine-tuned BERT
Can ChatGPT Understand Too? A Comparative Study on ChatGPT and Fine-tuned BERT
Qihuang Zhong
Liang Ding
Juhua Liu
Bo Du
Dacheng Tao
AI4MH
125
245
0
19 Feb 2023
Learning Language Representations with Logical Inductive Bias
Learning Language Representations with Logical Inductive Bias
Jianshu Chen
NAIAI4CELRM
51
3
0
19 Feb 2023
Previous
123...789...161718
Next