ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.12471
  4. Cited By
Neural Network Acceptability Judgments

Neural Network Acceptability Judgments

31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
ArXivPDFHTML

Papers citing "Neural Network Acceptability Judgments"

50 / 880 papers shown
Title
Adversarial Regularization as Stackelberg Game: An Unrolled Optimization
  Approach
Adversarial Regularization as Stackelberg Game: An Unrolled Optimization Approach
Simiao Zuo
Chen Liang
Haoming Jiang
Xiaodong Liu
Pengcheng He
Jianfeng Gao
Weizhu Chen
T. Zhao
55
9
0
11 Apr 2021
Adapting Language Models for Zero-shot Learning by Meta-tuning on
  Dataset and Prompt Collections
Adapting Language Models for Zero-shot Learning by Meta-tuning on Dataset and Prompt Collections
Ruiqi Zhong
Kristy Lee
Zheng-Wei Zhang
Dan Klein
33
166
0
10 Apr 2021
EXPATS: A Toolkit for Explainable Automated Text Scoring
EXPATS: A Toolkit for Explainable Automated Text Scoring
Hitoshi Manabe
Masato Hagiwara
8
4
0
07 Apr 2021
Exploring the Role of BERT Token Representations to Explain Sentence
  Probing Results
Exploring the Role of BERT Token Representations to Explain Sentence Probing Results
Hosein Mohebbi
Ali Modarressi
Mohammad Taher Pilehvar
MILM
27
23
0
03 Apr 2021
Evaluating the Morphosyntactic Well-formedness of Generated Texts
Evaluating the Morphosyntactic Well-formedness of Generated Texts
Adithya Pratapa
Antonios Anastasopoulos
Shruti Rijhwani
Aditi Chaudhary
David R. Mortensen
Graham Neubig
Yulia Tsvetkov
33
8
0
30 Mar 2021
Retraining DistilBERT for a Voice Shopping Assistant by Using Universal
  Dependencies
Retraining DistilBERT for a Voice Shopping Assistant by Using Universal Dependencies
P. Jayarao
Arpit Sharma
16
2
0
29 Mar 2021
A Practical Survey on Faster and Lighter Transformers
A Practical Survey on Faster and Lighter Transformers
Quentin Fournier
G. Caron
Daniel Aloise
14
93
0
26 Mar 2021
Approximating Instance-Dependent Noise via Instance-Confidence Embedding
Approximating Instance-Dependent Noise via Instance-Confidence Embedding
Yivan Zhang
Masashi Sugiyama
39
8
0
25 Mar 2021
UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New
  Multitask Benchmark
UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New Multitask Benchmark
Nicholas Lourie
Ronan Le Bras
Chandra Bhagavatula
Yejin Choi
LRM
30
137
0
24 Mar 2021
The NLP Cookbook: Modern Recipes for Transformer based Deep Learning
  Architectures
The NLP Cookbook: Modern Recipes for Transformer based Deep Learning Architectures
Sushant Singh
A. Mahmood
AI4TS
60
92
0
23 Mar 2021
Unsupervised Contextual Paraphrase Generation using Lexical Control and
  Reinforcement Learning
Unsupervised Contextual Paraphrase Generation using Lexical Control and Reinforcement Learning
Sonal Garg
Sumanth Prabhu
Hemant Misra
G. Srinivasaraghavan
22
14
0
23 Mar 2021
TAG: Gradient Attack on Transformer-based Language Models
TAG: Gradient Attack on Transformer-based Language Models
Jieren Deng
Yijue Wang
Ji Li
Chao Shang
Hang Liu
Sanguthevar Rajasekaran
Caiwen Ding
FedML
PILM
8
73
0
11 Mar 2021
FairFil: Contrastive Neural Debiasing Method for Pretrained Text
  Encoders
FairFil: Contrastive Neural Debiasing Method for Pretrained Text Encoders
Pengyu Cheng
Weituo Hao
Siyang Yuan
Shijing Si
Lawrence Carin
33
100
0
11 Mar 2021
Split Computing and Early Exiting for Deep Learning Applications: Survey
  and Research Challenges
Split Computing and Early Exiting for Deep Learning Applications: Survey and Research Challenges
Yoshitomo Matsubara
Marco Levorato
Francesco Restuccia
33
199
0
08 Mar 2021
Rissanen Data Analysis: Examining Dataset Characteristics via
  Description Length
Rissanen Data Analysis: Examining Dataset Characteristics via Description Length
Ethan Perez
Douwe Kiela
Kyunghyun Cho
22
24
0
05 Mar 2021
Token-Modification Adversarial Attacks for Natural Language Processing:
  A Survey
Token-Modification Adversarial Attacks for Natural Language Processing: A Survey
Tom Roth
Yansong Gao
A. Abuadbba
Surya Nepal
Wei Liu
AAML
31
12
0
01 Mar 2021
SparseBERT: Rethinking the Importance Analysis in Self-attention
SparseBERT: Rethinking the Importance Analysis in Self-attention
Han Shi
Jiahui Gao
Xiaozhe Ren
Hang Xu
Xiaodan Liang
Zhenguo Li
James T. Kwok
23
54
0
25 Feb 2021
Analyzing Curriculum Learning for Sentiment Analysis along Task
  Difficulty, Pacing and Visualization Axes
Analyzing Curriculum Learning for Sentiment Analysis along Task Difficulty, Pacing and Visualization Axes
Anvesh Rao Vijjini
Kaveri Anuranjana
R. Mamidi
38
2
0
19 Feb 2021
COCO-LM: Correcting and Contrasting Text Sequences for Language Model
  Pretraining
COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
Yu Meng
Chenyan Xiong
Payal Bajaj
Saurabh Tiwary
Paul N. Bennett
Jiawei Han
Xia Song
125
203
0
16 Feb 2021
AutoFreeze: Automatically Freezing Model Blocks to Accelerate
  Fine-tuning
AutoFreeze: Automatically Freezing Model Blocks to Accelerate Fine-tuning
Yuhan Liu
Saurabh Agarwal
Shivaram Venkataraman
OffRL
16
53
0
02 Feb 2021
Explaining Natural Language Processing Classifiers with Occlusion and
  Language Modeling
Explaining Natural Language Processing Classifiers with Occlusion and Language Modeling
David Harbecke
AAML
27
2
0
28 Jan 2021
CLiMP: A Benchmark for Chinese Language Model Evaluation
CLiMP: A Benchmark for Chinese Language Model Evaluation
Beilei Xiang
Changbing Yang
Yu Li
Alex Warstadt
Katharina Kann
ALM
25
38
0
26 Jan 2021
Muppet: Massive Multi-task Representations with Pre-Finetuning
Muppet: Massive Multi-task Representations with Pre-Finetuning
Armen Aghajanyan
Anchit Gupta
Akshat Shrivastava
Xilun Chen
Luke Zettlemoyer
Sonal Gupta
33
266
0
26 Jan 2021
WER-BERT: Automatic WER Estimation with BERT in a Balanced Ordinal
  Classification Paradigm
WER-BERT: Automatic WER Estimation with BERT in a Balanced Ordinal Classification Paradigm
Akshay Krishna Sheshadri
Anvesh Rao Vijjini
S. Kharbanda
16
8
0
14 Jan 2021
I-BERT: Integer-only BERT Quantization
I-BERT: Integer-only BERT Quantization
Sehoon Kim
A. Gholami
Z. Yao
Michael W. Mahoney
Kurt Keutzer
MQ
105
341
0
05 Jan 2021
WARP: Word-level Adversarial ReProgramming
WARP: Word-level Adversarial ReProgramming
Karen Hambardzumyan
Hrant Khachatrian
Jonathan May
AAML
254
342
0
01 Jan 2021
MiniLMv2: Multi-Head Self-Attention Relation Distillation for
  Compressing Pretrained Transformers
MiniLMv2: Multi-Head Self-Attention Relation Distillation for Compressing Pretrained Transformers
Wenhui Wang
Hangbo Bao
Shaohan Huang
Li Dong
Furu Wei
MQ
24
257
0
31 Dec 2020
Making Pre-trained Language Models Better Few-shot Learners
Making Pre-trained Language Models Better Few-shot Learners
Tianyu Gao
Adam Fisch
Danqi Chen
243
1,924
0
31 Dec 2020
Out of Order: How Important Is The Sequential Order of Words in a
  Sentence in Natural Language Understanding Tasks?
Out of Order: How Important Is The Sequential Order of Words in a Sentence in Natural Language Understanding Tasks?
Thang M. Pham
Trung Bui
Long Mai
Anh Totti Nguyen
217
122
0
30 Dec 2020
Accurate Word Representations with Universal Visual Guidance
Accurate Word Representations with Universal Visual Guidance
Zhuosheng Zhang
Haojie Yu
Hai Zhao
Rui-cang Wang
Masao Utiyama
22
0
0
30 Dec 2020
BURT: BERT-inspired Universal Representation from Learning Meaningful
  Segment
BURT: BERT-inspired Universal Representation from Learning Meaningful Segment
Yian Li
Hai Zhao
SSL
13
0
0
28 Dec 2020
ALP-KD: Attention-Based Layer Projection for Knowledge Distillation
ALP-KD: Attention-Based Layer Projection for Knowledge Distillation
Peyman Passban
Yimeng Wu
Mehdi Rezagholizadeh
Qun Liu
13
122
0
27 Dec 2020
Pre-Training Transformers as Energy-Based Cloze Models
Pre-Training Transformers as Energy-Based Cloze Models
Kevin Clark
Minh-Thang Luong
Quoc V. Le
Christopher D. Manning
15
78
0
15 Dec 2020
LRC-BERT: Latent-representation Contrastive Knowledge Distillation for
  Natural Language Understanding
LRC-BERT: Latent-representation Contrastive Knowledge Distillation for Natural Language Understanding
Hao Fu
Shaojun Zhou
Qihong Yang
Junjie Tang
Guiquan Liu
Kaikui Liu
Xiaolong Li
40
57
0
14 Dec 2020
Infusing Finetuning with Semantic Dependencies
Infusing Finetuning with Semantic Dependencies
Zhaofeng Wu
Hao Peng
Noah A. Smith
25
36
0
10 Dec 2020
Mapping the Space of Chemical Reactions Using Attention-Based Neural
  Networks
Mapping the Space of Chemical Reactions Using Attention-Based Neural Networks
P. Schwaller
Daniel Probst
Alain C. Vaucher
Vishnu H. Nair
D. Kreutter
Teodoro Laino
J. Reymond
147
225
0
09 Dec 2020
Towards Coinductive Models for Natural Language Understanding. Bringing
  together Deep Learning and Deep Semantics
Towards Coinductive Models for Natural Language Understanding. Bringing together Deep Learning and Deep Semantics
Wlodek Zadrozny
AI4CE
17
1
0
09 Dec 2020
An Investigation of Language Model Interpretability via Sentence Editing
An Investigation of Language Model Interpretability via Sentence Editing
Samuel Stevens
Yu-Chuan Su
LRM
13
8
0
28 Nov 2020
CoRe: An Efficient Coarse-refined Training Framework for BERT
CoRe: An Efficient Coarse-refined Training Framework for BERT
Cheng Yang
Shengnan Wang
Yuechuan Li
Chao Yang
Ming Yan
Jingqiao Zhang
Fangquan Lin
14
0
0
27 Nov 2020
AGenT Zero: Zero-shot Automatic Multiple-Choice Question Generation for
  Skill Assessments
AGenT Zero: Zero-shot Automatic Multiple-Choice Question Generation for Skill Assessments
Eric Li
Jingyi Su
Hao Sheng
Lawrence Wai
14
2
0
25 Nov 2020
Data-Informed Global Sparseness in Attention Mechanisms for Deep Neural
  Networks
Data-Informed Global Sparseness in Attention Mechanisms for Deep Neural Networks
Ileana Rugina
Rumen Dangovski
L. Jing
Preslav Nakov
Marin Soljacic
26
0
0
20 Nov 2020
Mixing ADAM and SGD: a Combined Optimization Method
Mixing ADAM and SGD: a Combined Optimization Method
Nicola Landro
I. Gallo
Riccardo La Grassa
ODL
8
24
0
16 Nov 2020
Know What You Don't Need: Single-Shot Meta-Pruning for Attention Heads
Know What You Don't Need: Single-Shot Meta-Pruning for Attention Heads
Zhengyan Zhang
Fanchao Qi
Zhiyuan Liu
Qun Liu
Maosong Sun
VLM
41
30
0
07 Nov 2020
CharBERT: Character-aware Pre-trained Language Model
CharBERT: Character-aware Pre-trained Language Model
Wentao Ma
Yiming Cui
Chenglei Si
Ting Liu
Shijin Wang
Guoping Hu
31
104
0
03 Nov 2020
Word Frequency Does Not Predict Grammatical Knowledge in Language Models
Word Frequency Does Not Predict Grammatical Knowledge in Language Models
Charles Yu
Ryan Sie
Nicolas Tedeschi
Leon Bergen
17
3
0
26 Oct 2020
ANLIzing the Adversarial Natural Language Inference Dataset
ANLIzing the Adversarial Natural Language Inference Dataset
Adina Williams
Tristan Thrush
Douwe Kiela
AAML
179
46
0
24 Oct 2020
Improving Classification through Weak Supervision in Context-specific
  Conversational Agent Development for Teacher Education
Improving Classification through Weak Supervision in Context-specific Conversational Agent Development for Teacher Education
Debajyoti Datta
Maria Phillips
Jennifer L. Chiu
G. Watson
J. Bywater
Laura E. Barnes
Donald E. Brown
23
7
0
23 Oct 2020
ERNIE-Gram: Pre-Training with Explicitly N-Gram Masked Language Modeling
  for Natural Language Understanding
ERNIE-Gram: Pre-Training with Explicitly N-Gram Masked Language Modeling for Natural Language Understanding
Dongling Xiao
Yukun Li
Han Zhang
Yu Sun
Hao Tian
Hua Wu
Haifeng Wang
27
38
0
23 Oct 2020
Detecting and Exorcising Statistical Demons from Language Models with
  Anti-Models of Negative Data
Detecting and Exorcising Statistical Demons from Language Models with Anti-Models of Negative Data
Michael L. Wick
Kate Silverstein
Jean-Baptiste Tristan
Adam Craig Pocock
Mark Johnson
17
3
0
22 Oct 2020
Towards Fully Bilingual Deep Language Modeling
Towards Fully Bilingual Deep Language Modeling
Li-Hsin Chang
S. Pyysalo
Jenna Kanerva
Filip Ginter
34
3
0
22 Oct 2020
Previous
123...1415161718
Next