ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.12471
  4. Cited By
Neural Network Acceptability Judgments

Neural Network Acceptability Judgments

31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
ArXivPDFHTML

Papers citing "Neural Network Acceptability Judgments"

50 / 880 papers shown
Title
Evaluating Large Language Models with fmeval
Evaluating Large Language Models with fmeval
Pola Schwöbel
Luca Franceschi
Muhammad Bilal Zafar
Keerthan Vasist
Aman Malhotra
Tomer Shenhar
Pinal Tailor
Pinar Yilmaz
Michael Diamond
Michele Donini
LM&MA
ELM
27
2
0
15 Jul 2024
SHERL: Synthesizing High Accuracy and Efficient Memory for
  Resource-Limited Transfer Learning
SHERL: Synthesizing High Accuracy and Efficient Memory for Resource-Limited Transfer Learning
Haiwen Diao
Bo Wan
Xu Jia
Yunzhi Zhuge
Ying Zhang
Huchuan Lu
Long Chen
VLM
50
4
0
10 Jul 2024
Testing learning hypotheses using neural networks by manipulating
  learning data
Testing learning hypotheses using neural networks by manipulating learning data
Cara Su-Yi Leong
Tal Linzen
31
4
0
05 Jul 2024
Efficient Training of Language Models with Compact and Consistent Next
  Token Distributions
Efficient Training of Language Models with Compact and Consistent Next Token Distributions
Ashutosh Sathe
Sunita Sarawagi
40
0
0
03 Jul 2024
MLKD-BERT: Multi-level Knowledge Distillation for Pre-trained Language
  Models
MLKD-BERT: Multi-level Knowledge Distillation for Pre-trained Language Models
Ying Zhang
Ziheng Yang
Shufan Ji
KELM
24
1
0
03 Jul 2024
Survey on Knowledge Distillation for Large Language Models: Methods,
  Evaluation, and Application
Survey on Knowledge Distillation for Large Language Models: Methods, Evaluation, and Application
Chuanpeng Yang
Wang Lu
Yao Zhu
Yidong Wang
Qian Chen
Chenlong Gao
Bingjie Yan
Yiqiang Chen
ALM
KELM
44
23
0
02 Jul 2024
CPT: Consistent Proxy Tuning for Black-box Optimization
CPT: Consistent Proxy Tuning for Black-box Optimization
Yuanyang He
Zitong Huang
Xinxing Xu
Rick Siow Mong Goh
Salman Khan
W. Zuo
Yong Liu
Chun-Mei Feng
45
0
0
01 Jul 2024
Exploring Advanced Large Language Models with LLMsuite
Exploring Advanced Large Language Models with LLMsuite
Giorgio Roffo
LLMAG
24
0
0
01 Jul 2024
Locate&Edit: Energy-based Text Editing for Efficient, Flexible, and
  Faithful Controlled Text Generation
Locate&Edit: Energy-based Text Editing for Efficient, Flexible, and Faithful Controlled Text Generation
Hye Ryung Son
Jay-Yoon Lee
46
0
0
30 Jun 2024
IDT: Dual-Task Adversarial Attacks for Privacy Protection
IDT: Dual-Task Adversarial Attacks for Privacy Protection
Pedro Faustini
Shakila Mahjabin Tonni
Annabelle McIver
Qiongkai Xu
Mark Dras
SILM
AAML
52
0
0
28 Jun 2024
Solving Token Gradient Conflict in Mixture-of-Experts for Large Vision-Language Model
Solving Token Gradient Conflict in Mixture-of-Experts for Large Vision-Language Model
Longrong Yang
Dong Shen
Chaoxiang Cai
Fan Yang
Size Li
Di Zhang
Xi Li
MoE
56
2
0
28 Jun 2024
LLMs instead of Human Judges? A Large Scale Empirical Study across 20
  NLP Evaluation Tasks
LLMs instead of Human Judges? A Large Scale Empirical Study across 20 NLP Evaluation Tasks
A. Bavaresco
Raffaella Bernardi
Leonardo Bertolazzi
Desmond Elliott
Raquel Fernández
...
David Schlangen
Alessandro Suglia
Aditya K Surikuchi
Ece Takmaz
A. Testoni
ALM
ELM
54
62
0
26 Jun 2024
Grass: Compute Efficient Low-Memory LLM Training with Structured Sparse
  Gradients
Grass: Compute Efficient Low-Memory LLM Training with Structured Sparse Gradients
Aashiq Muhamed
Oscar Li
David Woodruff
Mona Diab
Virginia Smith
61
7
0
25 Jun 2024
TinyStyler: Efficient Few-Shot Text Style Transfer with Authorship
  Embeddings
TinyStyler: Efficient Few-Shot Text Style Transfer with Authorship Embeddings
Zachary Horvitz
Ajay Patel
Kanishk Singh
Chris Callison-Burch
Kathleen McKeown
Zhou Yu
36
4
0
21 Jun 2024
Information Guided Regularization for Fine-tuning Language Models
Information Guided Regularization for Fine-tuning Language Models
Mandar Sharma
Nikhil Muralidhar
Shengzhe Xu
Raquib Bin Yousuf
Naren Ramakrishnan
38
0
0
20 Jun 2024
Open Generative Large Language Models for Galician
Open Generative Large Language Models for Galician
Pablo Gamallo
Pablo Rodríguez
Iria de-Dios-Flores
Susana Sotelo
Silvia Paniagua
Daniel Bardanca
José Ramom Pichel
Marcos Garcia
48
3
0
19 Jun 2024
UBENCH: Benchmarking Uncertainty in Large Language Models with Multiple
  Choice Questions
UBENCH: Benchmarking Uncertainty in Large Language Models with Multiple Choice Questions
Xunzhi Wang
Zhuowei Zhang
Qiongyu Li
Gaonan Chen
Mengting Hu
Zhiyu li
Bitong Luo
Hang Gao
Zhixin Han
Haotian Wang
ELM
48
3
0
18 Jun 2024
Fighting Randomness with Randomness: Mitigating Optimisation Instability
  of Fine-Tuning using Delayed Ensemble and Noisy Interpolation
Fighting Randomness with Randomness: Mitigating Optimisation Instability of Fine-Tuning using Delayed Ensemble and Noisy Interpolation
Branislav Pecher
Ján Cegin
Róbert Belanec
Jakub Simko
Ivan Srba
Maria Bielikova
47
1
0
18 Jun 2024
GW-MoE: Resolving Uncertainty in MoE Router with Global Workspace Theory
GW-MoE: Resolving Uncertainty in MoE Router with Global Workspace Theory
Haoze Wu
Zihan Qiu
Zili Wang
Hang Zhao
Jie Fu
MoE
51
3
0
18 Jun 2024
Knowledge Fusion By Evolving Weights of Language Models
Knowledge Fusion By Evolving Weights of Language Models
Guodong Du
Jing Li
Hanting Liu
Runhua Jiang
Shuyang Yu
Yifei Guo
S. Goh
Ho-Kin Tang
MoMe
44
8
0
18 Jun 2024
Style Transfer with Multi-iteration Preference Optimization
Style Transfer with Multi-iteration Preference Optimization
Shuai Liu
Jonathan May
45
4
0
17 Jun 2024
FamiCom: Further Demystifying Prompts for Language Models with
  Task-Agnostic Performance Estimation
FamiCom: Further Demystifying Prompts for Language Models with Task-Agnostic Performance Estimation
Bangzheng Li
Ben Zhou
Xingyu Fu
Fei Wang
Dan Roth
Muhao Chen
28
3
0
17 Jun 2024
Symmetric Dot-Product Attention for Efficient Training of BERT Language
  Models
Symmetric Dot-Product Attention for Efficient Training of BERT Language Models
Martin Courtois
Malte Ostendorff
Leonhard Hennig
Georg Rehm
39
2
0
10 Jun 2024
SuperPos-Prompt: Enhancing Soft Prompt Tuning of Language Models with
  Superposition of Multi Token Embeddings
SuperPos-Prompt: Enhancing Soft Prompt Tuning of Language Models with Superposition of Multi Token Embeddings
MohammadAli SadraeiJavaeri
Ehsaneddin Asgari
A. Mchardy
Hamid R. Rabiee
VLM
AAML
26
0
0
07 Jun 2024
VTrans: Accelerating Transformer Compression with Variational
  Information Bottleneck based Pruning
VTrans: Accelerating Transformer Compression with Variational Information Bottleneck based Pruning
Oshin Dutta
Ritvik Gupta
Sumeet Agarwal
44
1
0
07 Jun 2024
What Makes Language Models Good-enough?
What Makes Language Models Good-enough?
Daiki Asami
Saku Sugawara
37
1
0
06 Jun 2024
Seeing the Forest through the Trees: Data Leakage from Partial
  Transformer Gradients
Seeing the Forest through the Trees: Data Leakage from Partial Transformer Gradients
Weijun Li
Qiongkai Xu
Mark Dras
PILM
32
1
0
03 Jun 2024
Unveiling the Lexical Sensitivity of LLMs: Combinatorial Optimization
  for Prompt Enhancement
Unveiling the Lexical Sensitivity of LLMs: Combinatorial Optimization for Prompt Enhancement
Pengwei Zhan
Zhen Xu
Qian Tan
Jie Song
Ru Xie
51
7
0
31 May 2024
DoRA: Enhancing Parameter-Efficient Fine-Tuning with Dynamic Rank
  Distribution
DoRA: Enhancing Parameter-Efficient Fine-Tuning with Dynamic Rank Distribution
Yulong Mao
Kaiyu Huang
Changhao Guan
Ganglin Bao
Fengran Mo
Jinan Xu
37
11
0
27 May 2024
Unifying Demonstration Selection and Compression for In-Context Learning
Unifying Demonstration Selection and Compression for In-Context Learning
Jun Gao
Ziqiang Cao
Wenjie Li
43
3
0
27 May 2024
Understanding Linear Probing then Fine-tuning Language Models from NTK
  Perspective
Understanding Linear Probing then Fine-tuning Language Models from NTK Perspective
Akiyoshi Tomihari
Issei Sato
30
4
0
27 May 2024
Learning to Reason via Program Generation, Emulation, and Search
Learning to Reason via Program Generation, Emulation, and Search
Nathaniel Weir
Muhammad Khalifa
Linlu Qiu
Orion Weller
Peter Clark
SyDa
ReLM
LRM
90
5
0
25 May 2024
Filtered Corpus Training (FiCT) Shows that Language Models can
  Generalize from Indirect Evidence
Filtered Corpus Training (FiCT) Shows that Language Models can Generalize from Indirect Evidence
Abhinav Patil
Jaap Jumelet
Yu Ying Chiu
Andy Lapastora
Peter Shen
Lexie Wang
Clevis Willrich
Shane Steinert-Threlkeld
35
13
0
24 May 2024
DAGER: Exact Gradient Inversion for Large Language Models
DAGER: Exact Gradient Inversion for Large Language Models
Ivo Petrov
Dimitar I. Dimitrov
Maximilian Baader
Mark Niklas Muller
Martin Vechev
FedML
63
3
0
24 May 2024
Leakage-Resilient and Carbon-Neutral Aggregation Featuring the Federated
  AI-enabled Critical Infrastructure
Leakage-Resilient and Carbon-Neutral Aggregation Featuring the Federated AI-enabled Critical Infrastructure
Zehang Deng
Ruoxi Sun
Minhui Xue
Sheng Wen
S. Çamtepe
Surya Nepal
Yang Xiang
45
1
0
24 May 2024
VB-LoRA: Extreme Parameter Efficient Fine-Tuning with Vector Banks
VB-LoRA: Extreme Parameter Efficient Fine-Tuning with Vector Banks
Yang Li
Shaobo Han
Shihao Ji
MoE
27
11
0
24 May 2024
EMR-Merging: Tuning-Free High-Performance Model Merging
EMR-Merging: Tuning-Free High-Performance Model Merging
Chenyu Huang
Peng Ye
Tao Chen
Tong He
Xiangyu Yue
Wanli Ouyang
MoMe
46
29
0
23 May 2024
Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models
Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models
Yongxin Guo
Zhenglin Cheng
Xiaoying Tang
Tao R. Lin
Tao Lin
MoE
66
7
0
23 May 2024
Dynamic Context Adaptation and Information Flow Control in Transformers:
  Introducing the Evaluator Adjuster Unit and Gated Residual Connections
Dynamic Context Adaptation and Information Flow Control in Transformers: Introducing the Evaluator Adjuster Unit and Gated Residual Connections
Sahil Rajesh Dhayalkar
24
1
0
22 May 2024
Comparative Analysis of Different Efficient Fine Tuning Methods of Large
  Language Models (LLMs) in Low-Resource Setting
Comparative Analysis of Different Efficient Fine Tuning Methods of Large Language Models (LLMs) in Low-Resource Setting
Krishna Prasad Varadarajan Srinivasan
Prasanth Gumpena
Madhusudhana Yattapu
Vishal H. Brahmbhatt
24
1
0
21 May 2024
Effective In-Context Example Selection through Data Compression
Effective In-Context Example Selection through Data Compression
Zhongxiang Sun
Kepu Zhang
Haoyu Wang
Xiao Zhang
Jun Xu
38
2
0
19 May 2024
Efficient Prompt Tuning by Multi-Space Projection and Prompt Fusion
Efficient Prompt Tuning by Multi-Space Projection and Prompt Fusion
Pengxiang Lan
Enneng Yang
Yuting Liu
Guibing Guo
Linying Jiang
Jianzhe Zhao
Xingwei Wang
VLM
AAML
38
1
0
19 May 2024
Keep It Private: Unsupervised Privatization of Online Text
Keep It Private: Unsupervised Privatization of Online Text
Calvin Bao
Marine Carpuat
DeLMO
37
3
0
16 May 2024
DEPTH: Discourse Education through Pre-Training Hierarchically
DEPTH: Discourse Education through Pre-Training Hierarchically
Zachary Bamberger
Ofek Glick
Chaim Baskin
Yonatan Belinkov
67
0
0
13 May 2024
From Human Judgements to Predictive Models: Unravelling Acceptability in Code-Mixed Sentences
From Human Judgements to Predictive Models: Unravelling Acceptability in Code-Mixed Sentences
Prashant Kodali
Anmol Goel
Likhith Asapu
Vamshi Krishna Bonagiri
Anirudh Govil
Monojit Choudhury
Manish Shrivastava
Ponnurangam Kumaraguru
44
0
0
09 May 2024
SPAFIT: Stratified Progressive Adaptation Fine-tuning for Pre-trained
  Large Language Models
SPAFIT: Stratified Progressive Adaptation Fine-tuning for Pre-trained Large Language Models
Samir Arora
Liangliang Wang
24
0
0
30 Apr 2024
Knowledge Distillation vs. Pretraining from Scratch under a Fixed
  (Computation) Budget
Knowledge Distillation vs. Pretraining from Scratch under a Fixed (Computation) Budget
Minh Duc Bui
Fabian David Schmidt
Goran Glavaš
K. Wense
28
0
0
30 Apr 2024
More Room for Language: Investigating the Effect of Retrieval on
  Language Models
More Room for Language: Investigating the Effect of Retrieval on Language Models
David Samuel
Lucas Georges Gabriel Charpentier
Sondre Wold
LRM
RALM
KELM
30
1
0
16 Apr 2024
Variance-reduced Zeroth-Order Methods for Fine-Tuning Language Models
Variance-reduced Zeroth-Order Methods for Fine-Tuning Language Models
Tanmay Gautam
Youngsuk Park
Hao Zhou
Parameswaran Raman
Wooseok Ha
43
11
0
11 Apr 2024
PORTULAN ExtraGLUE Datasets and Models: Kick-starting a Benchmark for
  the Neural Processing of Portuguese
PORTULAN ExtraGLUE Datasets and Models: Kick-starting a Benchmark for the Neural Processing of Portuguese
T. Osório
Bernardo Leite
Henrique Lopes Cardoso
Luís Gomes
João Rodrigues
Rodrigo Santos
António Branco
32
3
0
08 Apr 2024
Previous
123456...161718
Next