ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.12471
  4. Cited By
Neural Network Acceptability Judgments
v1v2v3 (latest)

Neural Network Acceptability Judgments

31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
ArXiv (abs)PDFHTML

Papers citing "Neural Network Acceptability Judgments"

50 / 894 papers shown
Title
Black-box Prompt Learning for Pre-trained Language Models
Black-box Prompt Learning for Pre-trained Language Models
Shizhe Diao
Zhichao Huang
Ruijia Xu
Xuechun Li
Yong Lin
Xiao Zhou
Tong Zhang
VLMAAML
102
71
0
21 Jan 2022
NaijaSenti: A Nigerian Twitter Sentiment Corpus for Multilingual
  Sentiment Analysis
NaijaSenti: A Nigerian Twitter Sentiment Corpus for Multilingual Sentiment Analysis
Shamsuddeen Hassan Muhammad
David Ifeoluwa Adelani
Sebastian Ruder
Ibrahim Said Ahmad
Idris Abdulmumin
...
Chris C. Emezue
Saheed Abdul
Anuoluwapo Aremu
Alipio Jeorge
P. Brazdil
83
100
0
20 Jan 2022
The Dark Side of the Language: Pre-trained Transformers in the DarkNet
The Dark Side of the Language: Pre-trained Transformers in the DarkNet
Leonardo Ranaldi
Aria Nourbakhsh
Arianna Patrizi
Elena Sofia Ruzzetti
Dario Onorati
Francesca Fallucchi
Fabio Massimo Zanzotto
VLM
63
21
0
14 Jan 2022
How Does Data Corruption Affect Natural Language Understanding Models? A
  Study on GLUE datasets
How Does Data Corruption Affect Natural Language Understanding Models? A Study on GLUE datasets
Aarne Talman
Marianna Apidianaki
S. Chatzikyriakidis
Jörg Tiedemann
ELM
57
0
0
12 Jan 2022
Latency Adjustable Transformer Encoder for Language Understanding
Latency Adjustable Transformer Encoder for Language Understanding
Sajjad Kachuee
M. Sharifkhani
84
0
0
10 Jan 2022
Transformer Uncertainty Estimation with Hierarchical Stochastic
  Attention
Transformer Uncertainty Estimation with Hierarchical Stochastic Attention
Jiahuan Pei
Cheng-Yu Wang
Gyuri Szarvas
65
23
0
27 Dec 2021
An Empirical Investigation of the Role of Pre-training in Lifelong
  Learning
An Empirical Investigation of the Role of Pre-training in Lifelong Learning
Sanket Vaibhav Mehta
Darshan Patil
Sarath Chandar
Emma Strubell
CLL
148
145
0
16 Dec 2021
LMTurk: Few-Shot Learners as Crowdsourcing Workers in a
  Language-Model-as-a-Service Framework
LMTurk: Few-Shot Learners as Crowdsourcing Workers in a Language-Model-as-a-Service Framework
Mengjie Zhao
Fei Mi
Yasheng Wang
Minglei Li
Xin Jiang
Qun Liu
Hinrich Schütze
RALM
115
11
0
14 Dec 2021
Pruning Pretrained Encoders with a Multitask Objective
Pruning Pretrained Encoders with a Multitask Objective
Patrick Xia
Richard Shin
65
0
0
10 Dec 2021
FLAVA: A Foundational Language And Vision Alignment Model
FLAVA: A Foundational Language And Vision Alignment Model
Amanpreet Singh
Ronghang Hu
Vedanuj Goswami
Guillaume Couairon
Wojciech Galuba
Marcus Rohrbach
Douwe Kiela
CLIPVLM
133
719
0
08 Dec 2021
ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning
ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning
V. Aribandi
Yi Tay
Tal Schuster
J. Rao
H. Zheng
...
Jianmo Ni
Jai Gupta
Kai Hui
Sebastian Ruder
Donald Metzler
MoE
125
216
0
22 Nov 2021
Can depth-adaptive BERT perform better on binary classification tasks
Can depth-adaptive BERT perform better on binary classification tasks
Jing Fan
Xin Zhang
Sheng Zhang
Yan Pan
Lixiang Guo
MQ
40
0
0
22 Nov 2021
Merging Models with Fisher-Weighted Averaging
Merging Models with Fisher-Weighted Averaging
Michael Matena
Colin Raffel
FedMLMoMe
107
403
0
18 Nov 2021
DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with
  Gradient-Disentangled Embedding Sharing
DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing
Pengcheng He
Jianfeng Gao
Weizhu Chen
231
1,212
0
18 Nov 2021
Few-Shot Self-Rationalization with Natural Language Prompts
Few-Shot Self-Rationalization with Natural Language Prompts
Ana Marasović
Iz Beltagy
Doug Downey
Matthew E. Peters
LRM
91
110
0
16 Nov 2021
Variation and generality in encoding of syntactic anomaly information in
  sentence embeddings
Variation and generality in encoding of syntactic anomaly information in sentence embeddings
Qinxuan Wu
Allyson Ettinger
31
2
0
12 Nov 2021
Defining and Quantifying the Emergence of Sparse Concepts in DNNs
Defining and Quantifying the Emergence of Sparse Concepts in DNNs
Jie Ren
Mingjie Li
Qirui Chen
Huiqi Deng
Quanshi Zhang
120
33
0
11 Nov 2021
A Survey on Green Deep Learning
A Survey on Green Deep Learning
Jingjing Xu
Wangchunshu Zhou
Zhiyi Fu
Hao Zhou
Lei Li
VLM
203
84
0
08 Nov 2021
MetaICL: Learning to Learn In Context
MetaICL: Learning to Learn In Context
Sewon Min
M. Lewis
Luke Zettlemoyer
Hannaneh Hajishirzi
LRM
251
492
0
29 Oct 2021
Alignment Attention by Matching Key and Query Distributions
Alignment Attention by Matching Key and Query Distributions
Shujian Zhang
Xinjie Fan
Huangjie Zheng
Korawat Tanwisuth
Mingyuan Zhou
OOD
117
10
0
25 Oct 2021
Evaluating the Evaluation Metrics for Style Transfer: A Case Study in
  Multilingual Formality Transfer
Evaluating the Evaluation Metrics for Style Transfer: A Case Study in Multilingual Formality Transfer
Eleftheria Briakou
Sweta Agrawal
Joel R. Tetreault
Marine Carpuat
79
31
0
20 Oct 2021
The CoRa Tensor Compiler: Compilation for Ragged Tensors with Minimal
  Padding
The CoRa Tensor Compiler: Compilation for Ragged Tensors with Minimal Padding
Pratik Fegade
Tianqi Chen
Phillip B. Gibbons
T. Mowry
79
29
0
19 Oct 2021
Sparse Progressive Distillation: Resolving Overfitting under
  Pretrain-and-Finetune Paradigm
Sparse Progressive Distillation: Resolving Overfitting under Pretrain-and-Finetune Paradigm
Shaoyi Huang
Dongkuan Xu
Ian En-Hsu Yen
Yijue Wang
Sung-En Chang
...
Shiyang Chen
Mimi Xie
Sanguthevar Rajasekaran
Hang Liu
Caiwen Ding
CLLVLM
76
32
0
15 Oct 2021
SPoT: Better Frozen Model Adaptation through Soft Prompt Transfer
SPoT: Better Frozen Model Adaptation through Soft Prompt Transfer
Tu Vu
Brian Lester
Noah Constant
Rami Al-Rfou
Daniel Cer
VLMLRM
221
290
0
15 Oct 2021
Exploring Universal Intrinsic Task Subspace via Prompt Tuning
Exploring Universal Intrinsic Task Subspace via Prompt Tuning
Yujia Qin
Xiaozhi Wang
Yusheng Su
Yankai Lin
Ning Ding
...
Juanzi Li
Lei Hou
Peng Li
Maosong Sun
Jie Zhou
VLMVPVLM
190
29
0
15 Oct 2021
bert2BERT: Towards Reusable Pretrained Language Models
bert2BERT: Towards Reusable Pretrained Language Models
Cheng Chen
Yichun Yin
Lifeng Shang
Xin Jiang
Yujia Qin
Fengyu Wang
Zhi Wang
Xiao Chen
Zhiyuan Liu
Qun Liu
VLM
85
64
0
14 Oct 2021
A Survey On Neural Word Embeddings
A Survey On Neural Word Embeddings
Erhan Sezerer
Selma Tekir
AI4TS
76
13
0
05 Oct 2021
MoEfication: Transformer Feed-forward Layers are Mixtures of Experts
MoEfication: Transformer Feed-forward Layers are Mixtures of Experts
Zhengyan Zhang
Yankai Lin
Zhiyuan Liu
Peng Li
Maosong Sun
Jie Zhou
MoE
102
129
0
05 Oct 2021
Focused Contrastive Training for Test-based Constituency Analysis
Focused Contrastive Training for Test-based Constituency Analysis
Benjamin Roth
Erion cCano
31
0
0
30 Sep 2021
Shaking Syntactic Trees on the Sesame Street: Multilingual Probing with
  Controllable Perturbations
Shaking Syntactic Trees on the Sesame Street: Multilingual Probing with Controllable Perturbations
Ekaterina Taktasheva
Vladislav Mikhailov
Ekaterina Artemova
91
13
0
28 Sep 2021
Monolingual and Cross-Lingual Acceptability Judgments with the Italian
  CoLA corpus
Monolingual and Cross-Lingual Acceptability Judgments with the Italian CoLA corpus
Daniela Trotta
R. Guarasci
Elisa Leonardelli
Sara Tonelli
103
31
0
24 Sep 2021
Revisiting the Uniform Information Density Hypothesis
Revisiting the Uniform Information Density Hypothesis
Clara Meister
Tiago Pimentel
Patrick Haller
Lena Jäger
Ryan Cotterell
R. Levy
106
77
0
23 Sep 2021
Dynamic Knowledge Distillation for Pre-trained Language Models
Dynamic Knowledge Distillation for Pre-trained Language Models
Lei Li
Yankai Lin
Shuhuai Ren
Peng Li
Jie Zhou
Xu Sun
91
49
0
23 Sep 2021
Survey: Transformer based Video-Language Pre-training
Survey: Transformer based Video-Language Pre-training
Ludan Ruan
Qin Jin
VLMViT
125
45
0
21 Sep 2021
Training Dynamic based data filtering may not work for NLP datasets
Training Dynamic based data filtering may not work for NLP datasets
Arka Talukdar
Monika Dagar
Prachi Gupta
Varun G. Menon
NoLa
74
3
0
19 Sep 2021
Preventing Author Profiling through Zero-Shot Multilingual
  Back-Translation
Preventing Author Profiling through Zero-Shot Multilingual Back-Translation
David Ifeoluwa Adelani
Miaoran Zhang
Xiaoyu Shen
A. Davody
Thomas Kleinbauer
Dietrich Klakow
57
5
0
19 Sep 2021
Text Detoxification using Large Pre-trained Neural Models
Text Detoxification using Large Pre-trained Neural Models
David Dale
Anton Voronov
Daryna Dementieva
V. Logacheva
Olga Kozlova
Nikita Semenov
Alexander Panchenko
124
74
0
18 Sep 2021
Fine-Tuned Transformers Show Clusters of Similar Representations Across
  Layers
Fine-Tuned Transformers Show Clusters of Similar Representations Across Layers
Jason Phang
Haokun Liu
Samuel R. Bowman
86
29
0
17 Sep 2021
SupCL-Seq: Supervised Contrastive Learning for Downstream Optimized
  Sequence Representations
SupCL-Seq: Supervised Contrastive Learning for Downstream Optimized Sequence Representations
Hooman Sedghamiz
Shivam Raval
Enrico Santus
Tuka Alhanai
M. Ghassemi
SSL
34
17
0
15 Sep 2021
EfficientBERT: Progressively Searching Multilayer Perceptron via Warm-up
  Knowledge Distillation
EfficientBERT: Progressively Searching Multilayer Perceptron via Warm-up Knowledge Distillation
Chenhe Dong
Guangrun Wang
Hang Xu
Jiefeng Peng
Xiaozhe Ren
Xiaodan Liang
77
28
0
15 Sep 2021
ARCH: Efficient Adversarial Regularized Training with Caching
ARCH: Efficient Adversarial Regularized Training with Caching
Simiao Zuo
Chen Liang
Haoming Jiang
Pengcheng He
Xiaodong Liu
Jianfeng Gao
Weizhu Chen
T. Zhao
AAML
69
3
0
15 Sep 2021
Summarize-then-Answer: Generating Concise Explanations for Multi-hop
  Reading Comprehension
Summarize-then-Answer: Generating Concise Explanations for Multi-hop Reading Comprehension
Naoya Inoue
H. Trivedi
Steven K. Sinha
Niranjan Balasubramanian
Kentaro Inui
78
16
0
14 Sep 2021
LM-Critic: Language Models for Unsupervised Grammatical Error Correction
LM-Critic: Language Models for Unsupervised Grammatical Error Correction
Michihiro Yasunaga
J. Leskovec
Percy Liang
84
50
0
14 Sep 2021
Not All Models Localize Linguistic Knowledge in the Same Place: A
  Layer-wise Probing on BERToids' Representations
Not All Models Localize Linguistic Knowledge in the Same Place: A Layer-wise Probing on BERToids' Representations
Mohsen Fayyaz
Ehsan Aghazadeh
Ali Modarressi
Hosein Mohebbi
Mohammad Taher Pilehvar
35
21
0
13 Sep 2021
Learning to Teach with Student Feedback
Learning to Teach with Student Feedback
Yitao Liu
Tianxiang Sun
Xipeng Qiu
Xuanjing Huang
VLM
67
6
0
10 Sep 2021
Automated Robustness with Adversarial Training as a Post-Processing Step
Automated Robustness with Adversarial Training as a Post-Processing Step
Ambrish Rawat
M. Sinn
Beat Buesser
AAML
27
0
0
06 Sep 2021
ALLWAS: Active Learning on Language models in WASserstein space
ALLWAS: Active Learning on Language models in WASserstein space
Anson Bastos
Manohar Kaul
MedIm
55
1
0
03 Sep 2021
Finetuned Language Models Are Zero-Shot Learners
Finetuned Language Models Are Zero-Shot Learners
Jason W. Wei
Maarten Bosma
Vincent Zhao
Kelvin Guu
Adams Wei Yu
Brian Lester
Nan Du
Andrew M. Dai
Quoc V. Le
ALMUQCV
320
3,805
0
03 Sep 2021
Does Knowledge Help General NLU? An Empirical Study
Does Knowledge Help General NLU? An Empirical Study
Ruochen Xu
Yuwei Fang
Chenguang Zhu
Michael Zeng
ELM
70
9
0
01 Sep 2021
Shatter: An Efficient Transformer Encoder with Single-Headed
  Self-Attention and Relative Sequence Partitioning
Shatter: An Efficient Transformer Encoder with Single-Headed Self-Attention and Relative Sequence Partitioning
Ran Tian
Joshua Maynez
Ankur P. Parikh
ViT
56
2
0
30 Aug 2021
Previous
123...121314...161718
Next