ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.10044
  4. Cited By
BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions

BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions

24 May 2019
Christopher Clark
Kenton Lee
Ming-Wei Chang
Tom Kwiatkowski
Michael Collins
Kristina Toutanova
ArXiv (abs)PDFHTML

Papers citing "BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions"

50 / 1,143 papers shown
Title
Compressing Sine-Activated Low-Rank Adapters through Post-Training Quantization
Compressing Sine-Activated Low-Rank Adapters through Post-Training Quantization
Cameron Gordon
Yiping Ji
Hemanth Saratchandran
Paul Albert
Simon Lucey
MQ
61
0
0
28 May 2025
Highly Efficient and Effective LLMs with Multi-Boolean Architectures
Highly Efficient and Effective LLMs with Multi-Boolean Architectures
Ba-Hien Tran
Van Minh Nguyen
MQ
42
0
0
28 May 2025
SlimLLM: Accurate Structured Pruning for Large Language Models
SlimLLM: Accurate Structured Pruning for Large Language Models
Jialong Guo
Xinghao Chen
Yehui Tang
Yunhe Wang
13
0
0
28 May 2025
FireQ: Fast INT4-FP8 Kernel and RoPE-aware Quantization for LLM Inference Acceleration
FireQ: Fast INT4-FP8 Kernel and RoPE-aware Quantization for LLM Inference Acceleration
Daehyeon Baek
Jieun Choi
Jimyoung Son
Kyungmin Bin
Seungbeom Choi
Kihyo Moon
Minsung Jang
Hyojung Lee
MQ
21
0
0
27 May 2025
Efficient Large Language Model Inference with Neural Block Linearization
Efficient Large Language Model Inference with Neural Block Linearization
Mete Erdogan
F. Tonin
Volkan Cevher
66
0
0
27 May 2025
DenseLoRA: Dense Low-Rank Adaptation of Large Language Models
DenseLoRA: Dense Low-Rank Adaptation of Large Language Models
Lin Mu
Xiaoyu Wang
Li Ni
Yang Li
Zhize Wu
Peiquan Jin
Yiwen Zhang
ALMAI4CE
24
0
0
27 May 2025
LayerIF: Estimating Layer Quality for Large Language Models using Influence Functions
LayerIF: Estimating Layer Quality for Large Language Models using Influence Functions
Hadi Askari
Shivanshu Gupta
Fei Wang
Anshuman Chhabra
Muhao Chen
TDI
41
0
0
27 May 2025
Towards Objective Fine-tuning: How LLMs' Prior Knowledge Causes Potential Poor Calibration?
Towards Objective Fine-tuning: How LLMs' Prior Knowledge Causes Potential Poor Calibration?
Ziming Wang
Zeyu Shi
Haoyi Zhou
Shiqi Gao
Qingyun Sun
Jianxin Li
21
0
0
27 May 2025
QwT-v2: Practical, Effective and Efficient Post-Training Quantization
QwT-v2: Practical, Effective and Efficient Post-Training Quantization
Ningyuan Tang
Minghao Fu
Hao Yu
Jianxin Wu
MQ
79
0
0
27 May 2025
DLP: Dynamic Layerwise Pruning in Large Language Models
DLP: Dynamic Layerwise Pruning in Large Language Models
Yuli Chen
B. Cheng
Jiale Han
Yingying Zhang
Yingting Li
Shuhao Zhang
33
0
0
27 May 2025
BASE-Q: Bias and Asymmetric Scaling Enhanced Rotational Quantization for Large Language Models
Liulu He
Shenli Zhen
Karwei Sun
Yijiang Liu
Yufei Zhao
Chongkang Tan
Huanrui Yang
Yuan Du
Li Du
MQ
12
0
0
26 May 2025
Dynamically Learned Test-Time Model Routing in Language Model Zoos with Service Level Guarantees
Dynamically Learned Test-Time Model Routing in Language Model Zoos with Service Level Guarantees
Herbert Woisetschläger
Ryan Zhang
Shiqiang Wang
Hans-Arno Jacobsen
39
0
0
26 May 2025
Learning to Select In-Context Demonstration Preferred by Large Language Model
Learning to Select In-Context Demonstration Preferred by Large Language Model
Zheng Zhang
Shaocheng Lan
Lei Song
Jiang Bian
Yexin Li
Kan Ren
24
0
0
26 May 2025
Parameter-Efficient Fine-Tuning with Column Space Projection
Parameter-Efficient Fine-Tuning with Column Space Projection
Junseo Hwang
Wonguk Cho
Taesup Kim
48
0
0
26 May 2025
GraLoRA: Granular Low-Rank Adaptation for Parameter-Efficient Fine-Tuning
GraLoRA: Granular Low-Rank Adaptation for Parameter-Efficient Fine-Tuning
Yeonjoon Jung
Daehyun Ahn
Hyungjun Kim
Taesu Kim
Eunhyeok Park
51
0
0
26 May 2025
HD-PiSSA: High-Rank Distributed Orthogonal Adaptation
HD-PiSSA: High-Rank Distributed Orthogonal Adaptation
Yiding Wang
Fauxu meng
Xuefeng Zhang
Fan Jiang
Pingzhi Tang
Muhan Zhang
32
0
0
24 May 2025
BTC-LLM: Efficient Sub-1-Bit LLM Quantization via Learnable Transformation and Binary Codebook
BTC-LLM: Efficient Sub-1-Bit LLM Quantization via Learnable Transformation and Binary Codebook
Hao Gu
Lujun Li
Zheyu Wang
B. Liu
Qiyuan Zhu
Sirui Han
Yike Guo
MQ
12
0
0
24 May 2025
KerZOO: Kernel Function Informed Zeroth-Order Optimization for Accurate and Accelerated LLM Fine-Tuning
KerZOO: Kernel Function Informed Zeroth-Order Optimization for Accurate and Accelerated LLM Fine-Tuning
Zhendong Mi
Qitao Tan
Xiaodong Yu
Zining Zhu
Geng Yuan
Shaoyi Huang
206
0
0
24 May 2025
How Does Sequence Modeling Architecture Influence Base Capabilities of Pre-trained Language Models? Exploring Key Architecture Design Principles to Avoid Base Capabilities Degradation
How Does Sequence Modeling Architecture Influence Base Capabilities of Pre-trained Language Models? Exploring Key Architecture Design Principles to Avoid Base Capabilities Degradation
Xin Lu
Yanyan Zhao
Si Wei
Shijin Wang
Bing Qin
Ting Liu
43
0
0
24 May 2025
The Quest for Efficient Reasoning: A Data-Centric Benchmark to CoT Distillation
The Quest for Efficient Reasoning: A Data-Centric Benchmark to CoT Distillation
Ruichen Zhang
Rana Muhammad Shahroz Khan
Zhen Tan
Dawei Li
Song Wang
Tianlong Chen
LRM
45
0
0
24 May 2025
GIM: Improved Interpretability for Large Language Models
GIM: Improved Interpretability for Large Language Models
Joakim Edin
Róbert Csordás
Tuukka Ruotsalo
Zhengxuan Wu
Maria Maistro
Jing-ling Huang
Lars Maaløe
124
0
0
23 May 2025
How Can I Publish My LLM Benchmark Without Giving the True Answers Away?
How Can I Publish My LLM Benchmark Without Giving the True Answers Away?
Takashi Ishida
Thanawat Lodkaew
Ikko Yamane
203
0
0
23 May 2025
NeuroTrails: Training with Dynamic Sparse Heads as the Key to Effective Ensembling
NeuroTrails: Training with Dynamic Sparse Heads as the Key to Effective Ensembling
Bram Grooten
Farid Hasanov
Chenxiang Zhang
Q. Xiao
Boqian Wu
...
Shiwei Liu
L. Yin
Elena Mocanu
Mykola Pechenizkiy
Decebal Constantin Mocanu
52
0
0
23 May 2025
CoMoE: Contrastive Representation for Mixture-of-Experts in Parameter-Efficient Fine-tuning
Jinyuan Feng
Chaopeng Wei
Tenghai Qiu
Tianyi Hu
Zhiqiang Pu
MoE
86
0
0
23 May 2025
TRIM: Achieving Extreme Sparsity with Targeted Row-wise Iterative Metric-driven Pruning
TRIM: Achieving Extreme Sparsity with Targeted Row-wise Iterative Metric-driven Pruning
Florentin Beck
William Rudman
Carsten Eickhoff
51
0
0
22 May 2025
Mitigating Fine-tuning Risks in LLMs via Safety-Aware Probing Optimization
Mitigating Fine-tuning Risks in LLMs via Safety-Aware Probing Optimization
Chengcan Wu
Zhixin Zhang
Zeming Wei
Yihao Zhang
Meng Sun
AAML
59
1
0
22 May 2025
Hierarchical Safety Realignment: Lightweight Restoration of Safety in Pruned Large Vision-Language Models
Hierarchical Safety Realignment: Lightweight Restoration of Safety in Pruned Large Vision-Language Models
Yue Li
Xin Yi
Dongsheng Shi
Gerard de Melo
Xiaoling Wang
Linlin Wang
68
0
0
22 May 2025
HOFT: Householder Orthogonal Fine-tuning
HOFT: Householder Orthogonal Fine-tuning
Alejandro Moreno Arcas
Albert Sanchis
Jorge Civera
Alfons Juan
63
0
0
22 May 2025
ScholarBench: A Bilingual Benchmark for Abstraction, Comprehension, and Reasoning Evaluation in Academic Contexts
ScholarBench: A Bilingual Benchmark for Abstraction, Comprehension, and Reasoning Evaluation in Academic Contexts
Dongwon Noh
Donghyeok Koh
Junghun Yuk
Gyuwan Kim
Jaeyong Lee
Kyungtae Lim
Cheoneum Park
ELM
68
0
0
22 May 2025
Amplify Adjacent Token Differences: Enhancing Long Chain-of-Thought Reasoning with Shift-FFN
Amplify Adjacent Token Differences: Enhancing Long Chain-of-Thought Reasoning with Shift-FFN
Yao Xu
Mingyu Xu
Fangyu Lei
Wangtao Sun
Xiangrong Zeng
Bingning Wang
Guang Liu
Shizhu He
Jun Zhao
Kang Liu
LRM
77
1
0
22 May 2025
Tools in the Loop: Quantifying Uncertainty of LLM Question Answering Systems That Use Tools
Tools in the Loop: Quantifying Uncertainty of LLM Question Answering Systems That Use Tools
Panagiotis Lymperopoulos
Vasanth Sarathy
104
0
0
22 May 2025
LFTF: Locating First and Then Fine-Tuning for Mitigating Gender Bias in Large Language Models
LFTF: Locating First and Then Fine-Tuning for Mitigating Gender Bias in Large Language Models
Zhanyue Qin
Yue Ding
Deyuan Liu
Qingbin Liu
Junxian Cai
Xi Chen
Zhiying Tu
Dianhui Chu
Cuiyun Gao
Dianbo Sui
76
0
0
21 May 2025
Revealing Language Model Trajectories via Kullback-Leibler Divergence
Revealing Language Model Trajectories via Kullback-Leibler Divergence
Ryo Kishino
Yusuke Takase
Momose Oyama
Hiroaki Yamagiwa
Hidetoshi Shimodaira
82
0
0
21 May 2025
Social Bias in Popular Question-Answering Benchmarks
Social Bias in Popular Question-Answering Benchmarks
Angelie Kraft
Judith Simon
Sonja Schimmler
110
0
0
21 May 2025
On the Generalization vs Fidelity Paradox in Knowledge Distillation
On the Generalization vs Fidelity Paradox in Knowledge Distillation
Suhas Kamasetty Ramesh
Ayan Sengupta
Tanmoy Chakraborty
31
0
0
21 May 2025
OSoRA: Output-Dimension and Singular-Value Initialized Low-Rank Adaptation
OSoRA: Output-Dimension and Singular-Value Initialized Low-Rank Adaptation
Jialong Han
Si Zhang
Ke Zhang
37
0
0
20 May 2025
Void in Language Models
Void in Language Models
Mani Shemiranifar
62
0
0
20 May 2025
Dual Precision Quantization for Efficient and Accurate Deep Neural Networks Inference
Dual Precision Quantization for Efficient and Accurate Deep Neural Networks Inference
Tomer Gafni
Asaf Karnieli
Yair Hanani
MQ
74
0
0
20 May 2025
Breaking Bad Tokens: Detoxification of LLMs Using Sparse Autoencoders
Breaking Bad Tokens: Detoxification of LLMs Using Sparse Autoencoders
Agam Goyal
Vedant Rathi
William Yeh
Yian Wang
Yuen Chen
Hari Sundaram
100
0
0
20 May 2025
ABBA: Highly Expressive Hadamard Product Adaptation for Large Language Models
ABBA: Highly Expressive Hadamard Product Adaptation for Large Language Models
Raghav Singhal
Kaustubh Ponkshe
Rohit Vartak
Praneeth Vepakomma
118
0
0
20 May 2025
YESciEval: Robust LLM-as-a-Judge for Scientific Question Answering
YESciEval: Robust LLM-as-a-Judge for Scientific Question Answering
Jennifer D'Souza
Hamed Babaei Giglou
Quentin Münch
ELM
109
0
0
20 May 2025
Dual Decomposition of Weights and Singular Value Low Rank Adaptation
Dual Decomposition of Weights and Singular Value Low Rank Adaptation
Jialong Han
Si Zhang
Ke Zhang
43
0
0
20 May 2025
Fine-tuning Quantized Neural Networks with Zeroth-order Optimization
Fine-tuning Quantized Neural Networks with Zeroth-order Optimization
Sifeng Shang
Jiayi Zhou
Chenyu Lin
Minxian Li
Kaiyang Zhou
MQ
60
0
0
19 May 2025
Safety Alignment Can Be Not Superficial With Explicit Safety Signals
Safety Alignment Can Be Not Superficial With Explicit Safety Signals
Jianwei Li
Jung-Eng Kim
AAML
185
1
0
19 May 2025
Occult: Optimizing Collaborative Communication across Experts for Accelerated Parallel MoE Training and Inference
Occult: Optimizing Collaborative Communication across Experts for Accelerated Parallel MoE Training and Inference
Shuqing Luo
Pingzhi Li
Jie Peng
Hanrui Wang
Yang
Zhao
Yu Cheng
Tianlong Chen
MoE
70
0
0
19 May 2025
LoRASuite: Efficient LoRA Adaptation Across Large Language Model Upgrades
LoRASuite: Efficient LoRA Adaptation Across Large Language Model Upgrades
Yanan Li
Fanxu Meng
Muhan Zhang
Shiai Zhu
Shangguang Wang
Mengwei Xu
MoMe
73
0
0
17 May 2025
A Systematic Analysis of Base Model Choice for Reward Modeling
A Systematic Analysis of Base Model Choice for Reward Modeling
Kian Ahrabian
Pegah Jandaghi
Negar Mokhberian
Sai Praneeth Karimireddy
Jay Pujara
129
0
0
16 May 2025
GenKnowSub: Improving Modularity and Reusability of LLMs through General Knowledge Subtraction
GenKnowSub: Improving Modularity and Reusability of LLMs through General Knowledge Subtraction
Mohammadtaha Bagherifard
Sahar Rajabi
Ali Edalat
Yadollah Yaghoobzadeh
KELM
69
0
0
16 May 2025
ZeroTuning: Unlocking the Initial Token's Power to Enhance Large Language Models Without Training
ZeroTuning: Unlocking the Initial Token's Power to Enhance Large Language Models Without Training
Feijiang Han
Xiaodong Yu
Jianheng Tang
Lyle Ungar
97
0
0
16 May 2025
Memory-Efficient Orthogonal Fine-Tuning with Principal Subspace Adaptation
Memory-Efficient Orthogonal Fine-Tuning with Principal Subspace Adaptation
Fei Wu
Jia Hu
Geyong Min
Shiqiang Wang
97
0
0
16 May 2025
Previous
12345...212223
Next