ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.12471
  4. Cited By
Neural Network Acceptability Judgments
v1v2v3 (latest)

Neural Network Acceptability Judgments

31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
ArXiv (abs)PDFHTML

Papers citing "Neural Network Acceptability Judgments"

50 / 894 papers shown
Title
AdapterBias: Parameter-efficient Token-dependent Representation Shift
  for Adapters in NLP Tasks
AdapterBias: Parameter-efficient Token-dependent Representation Shift for Adapters in NLP Tasks
Chin-Lun Fu
Zih-Ching Chen
Yun-Ru Lee
Hung-yi Lee
85
49
0
30 Apr 2022
Tailor: A Prompt-Based Approach to Attribute-Based Controlled Text
  Generation
Tailor: A Prompt-Based Approach to Attribute-Based Controlled Text Generation
Kexin Yang
Dayiheng Liu
Wenqiang Lei
Baosong Yang
Mingfeng Xue
Boxing Chen
Jun Xie
79
29
0
28 Apr 2022
An End-to-End Dialogue Summarization System for Sales Calls
An End-to-End Dialogue Summarization System for Sales Calls
Abedelkadir Asi
Song Wang
Roy Eisenstadt
Dean Geckt
Yarin Kuper
Yi Mao
Royi Ronen
95
16
0
27 Apr 2022
Unsupervised Ranking and Aggregation of Label Descriptions for Zero-Shot
  Classifiers
Unsupervised Ranking and Aggregation of Label Descriptions for Zero-Shot Classifiers
Angelo Basile
Marc Franco-Salvador
Paolo Rosso
VLM
57
2
0
20 Apr 2022
ALBETO and DistilBETO: Lightweight Spanish Language Models
ALBETO and DistilBETO: Lightweight Spanish Language Models
J. Canete
S. Donoso
Felipe Bravo-Marquez
Andrés Carvallo
Vladimir Araujo
74
20
0
19 Apr 2022
DecBERT: Enhancing the Language Understanding of BERT with Causal
  Attention Masks
DecBERT: Enhancing the Language Understanding of BERT with Causal Attention Masks
Ziyang Luo
Yadong Xi
Jing Ma
Zhiwei Yang
Xiaoxi Mao
Changjie Fan
Rongsheng Zhang
35
3
0
19 Apr 2022
Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners
Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners
Shashank Gupta
Subhabrata Mukherjee
K. Subudhi
Eduardo Gonzalez
Damien Jose
Ahmed Hassan Awadallah
Jianfeng Gao
MoE
76
50
0
16 Apr 2022
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided
  Adaptation
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation
Simiao Zuo
Qingru Zhang
Chen Liang
Pengcheng He
T. Zhao
Weizhu Chen
MoE
195
41
0
15 Apr 2022
METRO: Efficient Denoising Pretraining of Large Scale Autoencoding
  Language Models with Model Generated Signals
METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Payal Bajaj
Chenyan Xiong
Guolin Ke
Xiaodong Liu
Di He
Saurabh Tiwary
Tie-Yan Liu
Paul N. Bennett
Xia Song
Jianfeng Gao
116
32
0
13 Apr 2022
CAMERO: Consistency Regularized Ensemble of Perturbed Language Models
  with Weight Sharing
CAMERO: Consistency Regularized Ensemble of Perturbed Language Models with Weight Sharing
Chen Liang
Pengcheng He
Yelong Shen
Weizhu Chen
T. Zhao
FedML
48
6
0
13 Apr 2022
Automatic Multi-Label Prompting: Simple and Interpretable Few-Shot
  Classification
Automatic Multi-Label Prompting: Simple and Interpretable Few-Shot Classification
Han Wang
Canwen Xu
Julian McAuley
VLM
67
12
0
13 Apr 2022
TangoBERT: Reducing Inference Cost by using Cascaded Architecture
TangoBERT: Reducing Inference Cost by using Cascaded Architecture
Jonathan Mamou
Oren Pereg
Moshe Wasserblat
Roy Schwartz
39
12
0
13 Apr 2022
Contextual Representation Learning beyond Masked Language Modeling
Contextual Representation Learning beyond Masked Language Modeling
Zhiyi Fu
Wangchunshu Zhou
Jingjing Xu
Hao Zhou
Lei Li
75
26
0
08 Apr 2022
Pretraining Text Encoders with Adversarial Mixture of Training Signal
  Generators
Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators
Yu Meng
Chenyan Xiong
Payal Bajaj
Saurabh Tiwary
Paul N. Bennett
Jiawei Han
Xia Song
MoE
78
16
0
07 Apr 2022
Fusing finetuned models for better pretraining
Fusing finetuned models for better pretraining
Leshem Choshen
Elad Venezian
Noam Slonim
Yoav Katz
FedMLAI4CEMoMe
130
96
0
06 Apr 2022
VALUE: Understanding Dialect Disparity in NLU
VALUE: Understanding Dialect Disparity in NLU
Caleb Ziems
Jiaao Chen
Camille Harris
J. Anderson
Diyi Yang
ELM
122
44
0
06 Apr 2022
Structured Pruning Learns Compact and Accurate Models
Structured Pruning Learns Compact and Accurate Models
Mengzhou Xia
Zexuan Zhong
Danqi Chen
VLM
109
189
0
01 Apr 2022
Feature Structure Distillation with Centered Kernel Alignment in BERT
  Transferring
Feature Structure Distillation with Centered Kernel Alignment in BERT Transferring
Heeseung Jung
Doyeon Kim
Seung-Hoon Na
Kangil Kim
81
6
0
01 Apr 2022
LinkBERT: Pretraining Language Models with Document Links
LinkBERT: Pretraining Language Models with Document Links
Michihiro Yasunaga
J. Leskovec
Percy Liang
KELM
108
361
0
29 Mar 2022
A Fast Post-Training Pruning Framework for Transformers
A Fast Post-Training Pruning Framework for Transformers
Woosuk Kwon
Sehoon Kim
Michael W. Mahoney
Joseph Hassoun
Kurt Keutzer
A. Gholami
110
156
0
29 Mar 2022
Few-Shot Learning with Siamese Networks and Label Tuning
Few-Shot Learning with Siamese Networks and Label Tuning
Thomas Müller
Guillermo Pérez-Torró
Marc Franco-Salvador
VLM
93
41
0
28 Mar 2022
MKQ-BERT: Quantized BERT with 4-bits Weights and Activations
MKQ-BERT: Quantized BERT with 4-bits Weights and Activations
Hanlin Tang
Xipeng Zhang
Kai Liu
Jianchen Zhu
Zhanhui Kang
VLMMQ
51
15
0
25 Mar 2022
Mix and Match: Learning-free Controllable Text Generation using Energy
  Language Models
Mix and Match: Learning-free Controllable Text Generation using Energy Language Models
Fatemehsadat Mireshghallah
Kartik Goyal
Taylor Berg-Kirkpatrick
71
80
0
24 Mar 2022
minicons: Enabling Flexible Behavioral and Representational Analyses of
  Transformer Language Models
minicons: Enabling Flexible Behavioral and Representational Analyses of Transformer Language Models
Kanishka Misra
84
63
0
24 Mar 2022
GRS: Combining Generation and Revision in Unsupervised Sentence
  Simplification
GRS: Combining Generation and Revision in Unsupervised Sentence Simplification
Mohammad Dehghan
Dhruv Kumar
Lukasz Golab
68
13
0
18 Mar 2022
On the Importance of Data Size in Probing Fine-tuned Models
On the Importance of Data Size in Probing Fine-tuned Models
Houman Mehrafarin
S. Rajaee
Mohammad Taher Pilehvar
63
18
0
17 Mar 2022
RoMe: A Robust Metric for Evaluating Natural Language Generation
RoMe: A Robust Metric for Evaluating Natural Language Generation
Md. Rony
Liubov Kovriguina
Debanjan Chaudhuri
Ricardo Usbeck
Jens Lehmann
75
12
0
17 Mar 2022
Hyperdecoders: Instance-specific decoders for multi-task NLP
Hyperdecoders: Instance-specific decoders for multi-task NLP
Hamish Ivison
Matthew E. Peters
AI4CE
113
22
0
15 Mar 2022
Do BERTs Learn to Use Browser User Interface? Exploring Multi-Step Tasks
  with Unified Vision-and-Language BERTs
Do BERTs Learn to Use Browser User Interface? Exploring Multi-Step Tasks with Unified Vision-and-Language BERTs
Taichi Iki
Akiko Aizawa
LLMAG
59
6
0
15 Mar 2022
VAST: The Valence-Assessing Semantics Test for Contextualizing Language
  Models
VAST: The Valence-Assessing Semantics Test for Contextualizing Language Models
Robert Wolfe
Aylin Caliskan
62
13
0
14 Mar 2022
PERT: Pre-training BERT with Permuted Language Model
PERT: Pre-training BERT with Permuted Language Model
Yiming Cui
Ziqing Yang
Ting Liu
85
37
0
14 Mar 2022
Delta Tuning: A Comprehensive Study of Parameter Efficient Methods for
  Pre-trained Language Models
Delta Tuning: A Comprehensive Study of Parameter Efficient Methods for Pre-trained Language Models
Ning Ding
Yujia Qin
Guang Yang
Fu Wei
Zonghan Yang
...
Jianfei Chen
Yang Liu
Jie Tang
Juan Li
Maosong Sun
108
205
0
14 Mar 2022
Model soups: averaging weights of multiple fine-tuned models improves
  accuracy without increasing inference time
Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time
Mitchell Wortsman
Gabriel Ilharco
S. Gadre
Rebecca Roelofs
Raphael Gontijo-Lopes
...
Hongseok Namkoong
Ali Farhadi
Y. Carmon
Simon Kornblith
Ludwig Schmidt
MoMe
199
1,013
1
10 Mar 2022
Pre-trained Token-replaced Detection Model as Few-shot Learner
Pre-trained Token-replaced Detection Model as Few-shot Learner
Zicheng Li
Shoushan Li
Guodong Zhou
77
9
0
07 Mar 2022
ILDAE: Instance-Level Difficulty Analysis of Evaluation Data
ILDAE: Instance-Level Difficulty Analysis of Evaluation Data
Neeraj Varshney
Swaroop Mishra
Chitta Baral
69
19
0
07 Mar 2022
Do Prompts Solve NLP Tasks Using Natural Language?
Do Prompts Solve NLP Tasks Using Natural Language?
Sen Yang
Yunchen Zhang
Leyang Cui
Yue Zhang
LRM
80
4
0
02 Mar 2022
Neural reality of argument structure constructions
Neural reality of argument structure constructions
Bai Li
Zining Zhu
Guillaume Thomas
Frank Rudzicz
Yang Xu
96
29
0
24 Feb 2022
Probing BERT's priors with serial reproduction chains
Probing BERT's priors with serial reproduction chains
Takateru Yamakoshi
Thomas Griffiths
Robert D. Hawkins
84
13
0
24 Feb 2022
LAMP: Extracting Text from Gradients with Language Model Priors
LAMP: Extracting Text from Gradients with Language Model Priors
Mislav Balunović
Dimitar I. Dimitrov
Nikola Jovanović
Martin Vechev
65
62
0
17 Feb 2022
Revisiting Over-smoothing in BERT from the Perspective of Graph
Revisiting Over-smoothing in BERT from the Perspective of Graph
Han Shi
Jiahui Gao
Hang Xu
Xiaodan Liang
Zhenguo Li
Lingpeng Kong
Stephen M. S. Lee
James T. Kwok
89
76
0
17 Feb 2022
No One Left Behind: Inclusive Federated Learning over Heterogeneous
  Devices
No One Left Behind: Inclusive Federated Learning over Heterogeneous Devices
Ruixuan Liu
Fangzhao Wu
Chuhan Wu
Yanlin Wang
Lingjuan Lyu
Hong Chen
Xing Xie
FedML
89
72
0
16 Feb 2022
Should You Mask 15% in Masked Language Modeling?
Should You Mask 15% in Masked Language Modeling?
Alexander Wettig
Tianyu Gao
Zexuan Zhong
Danqi Chen
CVBM
105
167
0
16 Feb 2022
A Differential Entropy Estimator for Training Neural Networks
A Differential Entropy Estimator for Training Neural Networks
Georg Pichler
Pierre Colombo
Malik Boudiaf
Günther Koliander
Pablo Piantanida
163
23
0
14 Feb 2022
Scaling Laws Under the Microscope: Predicting Transformer Performance
  from Small Scale Experiments
Scaling Laws Under the Microscope: Predicting Transformer Performance from Small Scale Experiments
Maor Ivgi
Y. Carmon
Jonathan Berant
80
17
0
13 Feb 2022
Generating Training Data with Language Models: Towards Zero-Shot
  Language Understanding
Generating Training Data with Language Models: Towards Zero-Shot Language Understanding
Yu Meng
Jiaxin Huang
Yu Zhang
Jiawei Han
SyDa
79
235
0
09 Feb 2022
What are the best systems? New perspectives on NLP Benchmarking
What are the best systems? New perspectives on NLP Benchmarking
Pierre Colombo
Nathan Noiry
Ekhine Irurozki
Stephan Clémençon
201
42
0
08 Feb 2022
Nonparametric Uncertainty Quantification for Single Deterministic Neural
  Network
Nonparametric Uncertainty Quantification for Single Deterministic Neural Network
Nikita Kotelevskii
A. Artemenkov
Kirill Fedyanin
Fedor Noskov
Alexander Fishkov
Artem Shelmanov
Artem Vazhentsev
Aleksandr Petiushko
Maxim Panov
UQCVBDL
106
30
0
07 Feb 2022
No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for
  Training Large Transformer Models
No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models
Chen Liang
Haoming Jiang
Simiao Zuo
Pengcheng He
Xiaodong Liu
Jianfeng Gao
Weizhu Chen
T. Zhao
72
14
0
06 Feb 2022
AutoDistil: Few-shot Task-agnostic Neural Architecture Search for
  Distilling Large Language Models
AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models
Dongkuan Xu
Subhabrata Mukherjee
Xiaodong Liu
Debadeepta Dey
Wenhui Wang
Xiang Zhang
Ahmed Hassan Awadallah
Jianfeng Gao
56
4
0
29 Jan 2022
Describing Differences between Text Distributions with Natural Language
Describing Differences between Text Distributions with Natural Language
Ruiqi Zhong
Charles Burton Snell
Dan Klein
Jacob Steinhardt
VLM
198
44
0
28 Jan 2022
Previous
123...111213...161718
Next