ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.12471
  4. Cited By
Neural Network Acceptability Judgments

Neural Network Acceptability Judgments

31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
ArXivPDFHTML

Papers citing "Neural Network Acceptability Judgments"

50 / 880 papers shown
Title
Improving Pre-trained Language Model Fine-tuning with Noise Stability
  Regularization
Improving Pre-trained Language Model Fine-tuning with Noise Stability Regularization
Hang Hua
Xingjian Li
Dejing Dou
Chengzhong Xu
Jiebo Luo
35
15
0
12 Jun 2022
pFL-Bench: A Comprehensive Benchmark for Personalized Federated Learning
pFL-Bench: A Comprehensive Benchmark for Personalized Federated Learning
Daoyuan Chen
Dawei Gao
Weirui Kuang
Yaliang Li
Bolin Ding
FedML
37
64
0
08 Jun 2022
Exploring Cross-lingual Textual Style Transfer with Large Multilingual
  Language Models
Exploring Cross-lingual Textual Style Transfer with Large Multilingual Language Models
Daniil Moskovskiy
Daryna Dementieva
Alexander Panchenko
26
3
0
05 Jun 2022
ZeroQuant: Efficient and Affordable Post-Training Quantization for
  Large-Scale Transformers
ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers
Z. Yao
Reza Yazdani Aminabadi
Minjia Zhang
Xiaoxia Wu
Conglong Li
Yuxiong He
VLM
MQ
73
444
0
04 Jun 2022
Extreme Compression for Pre-trained Transformers Made Simple and
  Efficient
Extreme Compression for Pre-trained Transformers Made Simple and Efficient
Xiaoxia Wu
Z. Yao
Minjia Zhang
Conglong Li
Yuxiong He
MQ
19
31
0
04 Jun 2022
BayesFormer: Transformer with Uncertainty Estimation
BayesFormer: Transformer with Uncertainty Estimation
Karthik Abinav Sankararaman
Sinong Wang
Han Fang
UQCV
BDL
30
10
0
02 Jun 2022
Task-Specific Expert Pruning for Sparse Mixture-of-Experts
Task-Specific Expert Pruning for Sparse Mixture-of-Experts
Tianyu Chen
Shaohan Huang
Yuan Xie
Binxing Jiao
Daxin Jiang
Haoyi Zhou
Jianxin Li
Furu Wei
MoE
34
39
0
01 Jun 2022
VLUE: A Multi-Task Benchmark for Evaluating Vision-Language Models
VLUE: A Multi-Task Benchmark for Evaluating Vision-Language Models
Wangchunshu Zhou
Yan Zeng
Shizhe Diao
Xinsong Zhang
CoGe
VLM
32
13
0
30 May 2022
E2S2: Encoding-Enhanced Sequence-to-Sequence Pretraining for Language
  Understanding and Generation
E2S2: Encoding-Enhanced Sequence-to-Sequence Pretraining for Language Understanding and Generation
Qihuang Zhong
Liang Ding
Juhua Liu
Bo Du
Dacheng Tao
54
27
0
30 May 2022
MiniDisc: Minimal Distillation Schedule for Language Model Compression
MiniDisc: Minimal Distillation Schedule for Language Model Compression
Chen Zhang
Yang Yang
Qifan Wang
Jiahao Liu
Jingang Wang
Wei Wu
Dawei Song
47
4
0
29 May 2022
StereoKG: Data-Driven Knowledge Graph Construction for Cultural
  Knowledge and Stereotypes
StereoKG: Data-Driven Knowledge Graph Construction for Cultural Knowledge and Stereotypes
Awantee V. Deshpande
Dana Ruiter
Marius Mosbach
Dietrich Klakow
20
11
0
27 May 2022
Federated Split BERT for Heterogeneous Text Classification
Federated Split BERT for Heterogeneous Text Classification
Zhengyang Li
Shijing Si
Jianzong Wang
Jing Xiao
FedML
35
21
0
26 May 2022
Eliciting and Understanding Cross-Task Skills with Task-Level
  Mixture-of-Experts
Eliciting and Understanding Cross-Task Skills with Task-Level Mixture-of-Experts
Qinyuan Ye
Juan Zha
Xiang Ren
MoE
18
12
0
25 May 2022
BITE: Textual Backdoor Attacks with Iterative Trigger Injection
BITE: Textual Backdoor Attacks with Iterative Trigger Injection
Jun Yan
Vansh Gupta
Xiang Ren
SILM
31
47
0
25 May 2022
Train Flat, Then Compress: Sharpness-Aware Minimization Learns More
  Compressible Models
Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models
Clara Na
Sanket Vaibhav Mehta
Emma Strubell
64
19
0
25 May 2022
Gradient-Based Constrained Sampling from Language Models
Gradient-Based Constrained Sampling from Language Models
Sachin Kumar
Biswajit Paria
Yulia Tsvetkov
BDL
32
53
0
25 May 2022
Linear Connectivity Reveals Generalization Strategies
Linear Connectivity Reveals Generalization Strategies
Jeevesh Juneja
Rachit Bansal
Kyunghyun Cho
João Sedoc
Naomi Saphra
244
45
0
24 May 2022
ATTEMPT: Parameter-Efficient Multi-task Tuning via Attentional Mixtures
  of Soft Prompts
ATTEMPT: Parameter-Efficient Multi-task Tuning via Attentional Mixtures of Soft Prompts
Akari Asai
Mohammadreza Salehi
Matthew E. Peters
Hannaneh Hajishirzi
130
100
0
24 May 2022
Diverse Lottery Tickets Boost Ensemble from a Single Pretrained Model
Diverse Lottery Tickets Boost Ensemble from a Single Pretrained Model
Sosuke Kobayashi
Shun Kiyono
Jun Suzuki
Kentaro Inui
MoMe
29
7
0
24 May 2022
On the Role of Bidirectionality in Language Model Pre-Training
On the Role of Bidirectionality in Language Model Pre-Training
Mikel Artetxe
Jingfei Du
Naman Goyal
Luke Zettlemoyer
Ves Stoyanov
30
16
0
24 May 2022
Outliers Dimensions that Disrupt Transformers Are Driven by Frequency
Outliers Dimensions that Disrupt Transformers Are Driven by Frequency
Giovanni Puccetti
Anna Rogers
Aleksandr Drozd
F. Dell’Orletta
81
42
0
23 May 2022
Instruction Induction: From Few Examples to Natural Language Task
  Descriptions
Instruction Induction: From Few Examples to Natural Language Task Descriptions
Or Honovich
Uri Shaham
Samuel R. Bowman
Omer Levy
ELM
LRM
120
137
0
22 May 2022
Calibration of Natural Language Understanding Models with Venn--ABERS
  Predictors
Calibration of Natural Language Understanding Models with Venn--ABERS Predictors
Patrizio Giovannotti
38
6
0
21 May 2022
Acceptability Judgements via Examining the Topology of Attention Maps
Acceptability Judgements via Examining the Topology of Attention Maps
D. Cherniavskii
Eduard Tulchinskii
Vladislav Mikhailov
Irina Proskurina
Laida Kushnareva
Ekaterina Artemova
S. Barannikov
Irina Piontkovskaya
D. Piontkovski
Evgeny Burnaev
778
19
0
19 May 2022
PromptDA: Label-guided Data Augmentation for Prompt-based Few-shot
  Learners
PromptDA: Label-guided Data Augmentation for Prompt-based Few-shot Learners
Canyu Chen
Kai Shu
VLM
31
8
0
18 May 2022
Exploiting Social Media Content for Self-Supervised Style Transfer
Exploiting Social Media Content for Self-Supervised Style Transfer
Dana Ruiter
Thomas Kleinbauer
C. España-Bonet
Josef van Genabith
Dietrich Klakow
36
2
0
18 May 2022
When to Use Multi-Task Learning vs Intermediate Fine-Tuning for
  Pre-Trained Encoder Transfer Learning
When to Use Multi-Task Learning vs Intermediate Fine-Tuning for Pre-Trained Encoder Transfer Learning
Orion Weller
Kevin Seppi
Matt Gardner
22
21
0
17 May 2022
A Precis of Language Models are not Models of Language
A Precis of Language Models are not Models of Language
Csaba Veres
37
3
0
16 May 2022
Assessing the Limits of the Distributional Hypothesis in Semantic
  Spaces: Trait-based Relational Knowledge and the Impact of Co-occurrences
Assessing the Limits of the Distributional Hypothesis in Semantic Spaces: Trait-based Relational Knowledge and the Impact of Co-occurrences
Mark Anderson
Jose Camacho-Collados
35
0
0
16 May 2022
Sibylvariant Transformations for Robust Text Classification
Sibylvariant Transformations for Robust Text Classification
Fabrice Harel-Canada
Muhammad Ali Gulzar
Nanyun Peng
Miryung Kim
AAML
VLM
13
4
0
10 May 2022
So Different Yet So Alike! Constrained Unsupervised Text Style Transfer
So Different Yet So Alike! Constrained Unsupervised Text Style Transfer
Abhinav Ramesh Kashyap
Devamanyu Hazarika
Min-Yen Kan
Roger Zimmermann
Soujanya Poria
GAN
18
13
0
09 May 2022
Mixed-effects transformers for hierarchical adaptation
Mixed-effects transformers for hierarchical adaptation
Julia White
Noah D. Goodman
Robert D. Hawkins
24
2
0
03 May 2022
Adaptable Adapters
Adaptable Adapters
N. Moosavi
Quentin Delfosse
Kristian Kersting
Iryna Gurevych
56
21
0
03 May 2022
AdapterBias: Parameter-efficient Token-dependent Representation Shift
  for Adapters in NLP Tasks
AdapterBias: Parameter-efficient Token-dependent Representation Shift for Adapters in NLP Tasks
Chin-Lun Fu
Zih-Ching Chen
Yun-Ru Lee
Hung-yi Lee
33
44
0
30 Apr 2022
Tailor: A Prompt-Based Approach to Attribute-Based Controlled Text
  Generation
Tailor: A Prompt-Based Approach to Attribute-Based Controlled Text Generation
Kexin Yang
Dayiheng Liu
Wenqiang Lei
Baosong Yang
Mingfeng Xue
Boxing Chen
Jun Xie
38
29
0
28 Apr 2022
An End-to-End Dialogue Summarization System for Sales Calls
An End-to-End Dialogue Summarization System for Sales Calls
Abedelkadir Asi
Song Wang
Roy Eisenstadt
Dean Geckt
Yarin Kuper
Yi Mao
Royi Ronen
30
16
0
27 Apr 2022
Unsupervised Ranking and Aggregation of Label Descriptions for Zero-Shot
  Classifiers
Unsupervised Ranking and Aggregation of Label Descriptions for Zero-Shot Classifiers
Angelo Basile
Marc Franco-Salvador
Paolo Rosso
VLM
12
2
0
20 Apr 2022
ALBETO and DistilBETO: Lightweight Spanish Language Models
ALBETO and DistilBETO: Lightweight Spanish Language Models
J. Canete
S. Donoso
Felipe Bravo-Marquez
Andrés Carvallo
Vladimir Araujo
48
20
0
19 Apr 2022
DecBERT: Enhancing the Language Understanding of BERT with Causal
  Attention Masks
DecBERT: Enhancing the Language Understanding of BERT with Causal Attention Masks
Ziyang Luo
Yadong Xi
Jing Ma
Zhiwei Yang
Xiaoxi Mao
Changjie Fan
Rongsheng Zhang
19
3
0
19 Apr 2022
Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners
Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners
Shashank Gupta
Subhabrata Mukherjee
K. Subudhi
Eduardo Gonzalez
Damien Jose
Ahmed Hassan Awadallah
Jianfeng Gao
MoE
27
49
0
16 Apr 2022
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided
  Adaptation
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation
Simiao Zuo
Qingru Zhang
Chen Liang
Pengcheng He
T. Zhao
Weizhu Chen
MoE
24
38
0
15 Apr 2022
METRO: Efficient Denoising Pretraining of Large Scale Autoencoding
  Language Models with Model Generated Signals
METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Payal Bajaj
Chenyan Xiong
Guolin Ke
Xiaodong Liu
Di He
Saurabh Tiwary
Tie-Yan Liu
Paul N. Bennett
Xia Song
Jianfeng Gao
50
32
0
13 Apr 2022
CAMERO: Consistency Regularized Ensemble of Perturbed Language Models
  with Weight Sharing
CAMERO: Consistency Regularized Ensemble of Perturbed Language Models with Weight Sharing
Chen Liang
Pengcheng He
Yelong Shen
Weizhu Chen
T. Zhao
FedML
17
6
0
13 Apr 2022
Automatic Multi-Label Prompting: Simple and Interpretable Few-Shot
  Classification
Automatic Multi-Label Prompting: Simple and Interpretable Few-Shot Classification
Han Wang
Canwen Xu
Julian McAuley
VLM
26
12
0
13 Apr 2022
TangoBERT: Reducing Inference Cost by using Cascaded Architecture
TangoBERT: Reducing Inference Cost by using Cascaded Architecture
Jonathan Mamou
Oren Pereg
Moshe Wasserblat
Roy Schwartz
9
11
0
13 Apr 2022
Contextual Representation Learning beyond Masked Language Modeling
Contextual Representation Learning beyond Masked Language Modeling
Zhiyi Fu
Wangchunshu Zhou
Jingjing Xu
Hao Zhou
Lei Li
31
25
0
08 Apr 2022
Pretraining Text Encoders with Adversarial Mixture of Training Signal
  Generators
Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators
Yu Meng
Chenyan Xiong
Payal Bajaj
Saurabh Tiwary
Paul N. Bennett
Jiawei Han
Xia Song
MoE
44
16
0
07 Apr 2022
Fusing finetuned models for better pretraining
Fusing finetuned models for better pretraining
Leshem Choshen
Elad Venezian
Noam Slonim
Yoav Katz
FedML
AI4CE
MoMe
54
87
0
06 Apr 2022
VALUE: Understanding Dialect Disparity in NLU
VALUE: Understanding Dialect Disparity in NLU
Caleb Ziems
Jiaao Chen
Camille Harris
J. Anderson
Diyi Yang
ELM
49
42
0
06 Apr 2022
Structured Pruning Learns Compact and Accurate Models
Structured Pruning Learns Compact and Accurate Models
Mengzhou Xia
Zexuan Zhong
Danqi Chen
VLM
11
180
0
01 Apr 2022
Previous
123...101112...161718
Next