Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2306.04751
Cited By
How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources
7 June 2023
Yizhong Wang
Hamish Ivison
Pradeep Dasigi
Jack Hessel
Tushar Khot
Khyathi Raghavi Chandu
David Wadden
Kelsey MacMillan
Noah A. Smith
Iz Beltagy
Hannaneh Hajishirzi
ALM
ELM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources"
50 / 298 papers shown
Title
WorldView-Bench: A Benchmark for Evaluating Global Cultural Perspectives in Large Language Models
Abdullah Mushtaq
Imran Taj
Rafay Naeem
Ibrahim Ghaznavi
Junaid Qadir
28
0
0
14 May 2025
Large Language Models for Computer-Aided Design: A Survey
Licheng Zhang
Bach Le
Naveed Akhtar
Siew-Kei Lam
Tuan Ngo
3DV
AI4CE
48
0
0
13 May 2025
Incentivizing Inclusive Contributions in Model Sharing Markets
Enpei Zhang
Jingyi Chai
Guangyi Liu
Yanfeng Wang
Siheng Chen
TDI
FedML
230
0
0
05 May 2025
What do Language Model Probabilities Represent? From Distribution Estimation to Response Prediction
Eitan Wagner
Omri Abend
43
0
0
04 May 2025
Can We Enhance Bug Report Quality Using LLMs?: An Empirical Study of LLM-Based Bug Report Generation
Jagrit Acharya
Gouri Ginde
56
0
0
26 Apr 2025
Honey, I Shrunk the Language Model: Impact of Knowledge Distillation Methods on Performance and Explainability
Daniel Hendriks
Philipp Spitzer
Niklas Kühl
G. Satzger
27
2
0
22 Apr 2025
Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
Yang Yue
Zhiqi Chen
Rui Lu
Andrew Zhao
Zhaokai Wang
Yang Yue
Shiji Song
Gao Huang
ReLM
LRM
63
23
0
18 Apr 2025
LazyReview A Dataset for Uncovering Lazy Thinking in NLP Peer Reviews
Sukannya Purkayastha
Zhuang Li
Anne Lauscher
Lizhen Qu
Iryna Gurevych
37
0
0
15 Apr 2025
NorEval: A Norwegian Language Understanding and Generation Evaluation Benchmark
Vladislav Mikhailov
Tita Ranveig Enstad
David Samuel
Hans Christian Farsethås
Andrey Kutuzov
Erik Velldal
Lilja Øvrelid
ELM
45
0
0
10 Apr 2025
Understanding Learner-LLM Chatbot Interactions and the Impact of Prompting Guidelines
Cansu Koyuturk
Emily Theophilou
Sabrina Patania
Gregor Donabauer
Andrea Martinenghi
...
Udo Kruschwitz
D. Taibi
Simona Amenta
Martin Ruskov
Dimitri Ognibene
50
0
0
10 Apr 2025
Task-Specific Data Selection for Instruction Tuning via Monosemantic Neuronal Activations
Da Ma
Gonghu Shang
Zhi Chen
L. Qin
Yijie Luo
Lei Pan
Shuai Fan
Lu Chen
Kai Yu
46
0
0
19 Mar 2025
reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs
Zhaofeng Wu
Michihiro Yasunaga
Andrew Cohen
Yoon Kim
Asli Celikyilmaz
Marjan Ghazvininejad
48
2
0
14 Mar 2025
Take Off the Training Wheels Progressive In-Context Learning for Effective Alignment
Zhenyu Liu
Dongfang Li
Xinshuo Hu
X. Zhao
Yibin Chen
Baotian Hu
Min-Ling Zhang
56
1
0
13 Mar 2025
Large-Scale Data Selection for Instruction Tuning
Hamish Ivison
Muru Zhang
Faeze Brahman
Pang Wei Koh
Pradeep Dasigi
ALM
75
1
0
03 Mar 2025
CrowdSelect: Synthetic Instruction Data Selection with Multi-LLM Wisdom
Yisen Li
Lingfeng Yang
Wenxuan Shen
Pan Zhou
Yao Wan
Weiwei Lin
Danny Chen
77
0
0
03 Mar 2025
Stackelberg Game Preference Optimization for Data-Efficient Alignment of Language Models
Xu Chu
Zhixin Zhang
Tianyu Jia
Yujie Jin
82
0
0
25 Feb 2025
PiCO: Peer Review in LLMs based on the Consistency Optimization
Kun-Peng Ning
Shuo Yang
Yu-Yang Liu
Jia-Yu Yao
Zhen-Hui Liu
Yu Wang
Ming Pang
Li Yuan
ALM
71
8
0
24 Feb 2025
RewardDS: Privacy-Preserving Fine-Tuning for Large Language Models via Reward Driven Data Synthesis
Jianwei Wang
Junyao Yang
Haoran Li
Huiping Zhuang
Cen Chen
Ziqian Zeng
SyDa
49
0
0
23 Feb 2025
RIDE: Enhancing Large Language Model Alignment through Restyled In-Context Learning Demonstration Exemplars
Yuncheng Hua
Lizhen Qu
Zhuang Li
Hao Xue
Flora D. Salim
Gholamreza Haffari
ALM
132
0
0
17 Feb 2025
Ensembles of Low-Rank Expert Adapters
Yinghao Li
Vianne Gao
Chao Zhang
MohamadAli Torkamani
80
0
0
31 Jan 2025
Improving Influence-based Instruction Tuning Data Selection for Balanced Learning of Diverse Capabilities
Qirun Dai
Dylan Zhang
Jiaqi W. Ma
Hao Peng
TDI
63
1
0
21 Jan 2025
CDS: Data Synthesis Method Guided by Cognitive Diagnosis Theory
Haokun Zhao
Jinyi Han
Jiaqing Liang
Yanghua Xiao
59
0
0
13 Jan 2025
Codebook LLMs: Evaluating LLMs as Measurement Tools for Political Science Concepts
Andrew Halterman
Katherine A. Keith
58
2
0
10 Jan 2025
Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback
Lester James V. Miranda
Yizhong Wang
Yanai Elazar
Sachin Kumar
Valentina Pyatkin
Faeze Brahman
Noah A. Smith
Hannaneh Hajishirzi
Pradeep Dasigi
55
8
0
08 Jan 2025
PRD: Peer Rank and Discussion Improve Large Language Model based Evaluations
Ruosen Li
Teerth Patel
Xinya Du
LLMAG
ALM
73
97
0
03 Jan 2025
Extending LLMs to New Languages: A Case Study of Llama and Persian Adaptation
Samin Mahdizadeh Sani
Pouya Sadeghi
Thuy-Trang Vu
Yadollah Yaghoobzadeh
Gholamreza Haffari
81
2
0
17 Dec 2024
Unveiling the Secret Recipe: A Guide For Supervised Fine-Tuning Small LLMs
Aldo Pareja
Nikhil Shivakumar Nayak
Hao Wang
Krishnateja Killamsetty
Shivchander Sudalairaj
...
Guangxuan Xu
Kai Xu
Ligong Han
Luke Inglis
Akash Srivastava
98
6
0
17 Dec 2024
CoinMath: Harnessing the Power of Coding Instruction for Math LLMs
Chengwei Wei
Bin Wang
Jung-jae Kim
Guimei Liu
Nancy F. Chen
LRM
87
1
0
16 Dec 2024
ROSE: A Reward-Oriented Data Selection Framework for LLM Task-Specific Instruction Tuning
Yang Wu
Huayi Zhang
Yizheng Jiao
Lin Ma
Xiaozhong Liu
Jinhong Yu
Dongyu Zhang
Dezhi Yu
Wei Xu
90
1
0
01 Dec 2024
What Really is Commonsense Knowledge?
Quyet V. Do
Junze Li
Tung-Duong Vuong
Zhaowei Wang
Yangqiu Song
Xiaojuan Ma
28
1
0
06 Nov 2024
A Bayesian Approach to Data Point Selection
Xinnuo Xu
Minyoung Kim
Royson Lee
Brais Martínez
Timothy M. Hospedales
35
0
0
06 Nov 2024
On the Loss of Context-awareness in General Instruction Fine-tuning
Yihan Wang
Andrew Bai
Nanyun Peng
Cho-Jui Hsieh
177
1
0
05 Nov 2024
VPO: Leveraging the Number of Votes in Preference Optimization
Jae Hyeon Cho
Minkyung Park
Byung-Jun Lee
27
1
0
30 Oct 2024
Prompting and Fine-Tuning of Small LLMs for Length-Controllable Telephone Call Summarization
David Thulke
Yingbo Gao
Rricha Jalota
Christian Dugast
Hermann Ney
29
3
0
24 Oct 2024
Influential Language Data Selection via Gradient Trajectory Pursuit
Zhiwei Deng
Tao Li
Yang Li
36
1
0
22 Oct 2024
DEAN: Deactivating the Coupled Neurons to Mitigate Fairness-Privacy Conflicts in Large Language Models
Chen Qian
Dongrui Liu
Jie Zhang
Yong Liu
Jing Shao
40
1
0
22 Oct 2024
The effect of fine-tuning on language model toxicity
Will Hawkins
Brent Mittelstadt
Chris Russell
33
5
0
21 Oct 2024
Compute-Constrained Data Selection
Junjie Oscar Yin
Alexander M. Rush
42
0
0
21 Oct 2024
MCQG-SRefine: Multiple Choice Question Generation and Evaluation with Iterative Self-Critique, Correction, and Comparison Feedback
Zonghai Yao
Aditya Parashar
Huixue Zhou
Won Seok Jang
Feiyun Ouyang
Zhichao Yang
Hong-ye Yu
ELM
53
2
0
17 Oct 2024
Merge to Learn: Efficiently Adding Skills to Language Models with Model Merging
Jacob Morrison
Noah A. Smith
Hannaneh Hajishirzi
Pang Wei Koh
Jesse Dodge
Pradeep Dasigi
KELM
MoMe
CLL
50
1
0
16 Oct 2024
QSpec: Speculative Decoding with Complementary Quantization Schemes
Juntao Zhao
Wenhao Lu
Sheng Wang
Lingpeng Kong
Chuan Wu
MQ
74
5
0
15 Oct 2024
Semantic Image Inversion and Editing using Rectified Stochastic Differential Equations
Litu Rout
Yujia Chen
Nataniel Ruiz
C. Caramanis
Sanjay Shakkottai
Wen-Sheng Chu
DiffM
64
23
0
14 Oct 2024
SensorBench: Benchmarking LLMs in Coding-Based Sensor Processing
Pengrui Quan
Xiaomin Ouyang
J. Jeyakumar
Ziqi Wang
Yang Xing
Mani B. Srivastava
34
2
0
14 Oct 2024
3DS: Decomposed Difficulty Data Selection's Case Study on LLM Medical Domain Adaptation
Hongxin Ding
Yue Fang
Runchuan Zhu
Xinke Jiang
Jinyang Zhang
Yongxin Xu
Xu Chu
Junfeng Zhao
Yasha Wang
35
0
0
13 Oct 2024
Rethinking Data Selection at Scale: Random Selection is Almost All You Need
Tingyu Xia
Bowen Yu
K. Dang
An Yang
Yuan Wu
Yuan Tian
Yi-Ju Chang
Junyang Lin
ALM
54
5
0
12 Oct 2024
Simultaneous Reward Distillation and Preference Learning: Get You a Language Model Who Can Do Both
Abhijnan Nath
Changsoo Jung
Ethan Seefried
Nikhil Krishnaswamy
224
1
0
11 Oct 2024
Packing Analysis: Packing Is More Appropriate for Large Models or Datasets in Supervised Fine-tuning
Shuhe Wang
Guoyin Wang
Yucheng Wang
Jiwei Li
Eduard H. Hovy
Chen Guo
37
4
0
10 Oct 2024
Reward-Augmented Data Enhances Direct Preference Alignment of LLMs
Shenao Zhang
Zhihan Liu
Boyi Liu
Wenjie Qu
Yingxiang Yang
Yunxing Liu
Liyu Chen
Tao Sun
Ziyi Wang
101
3
0
10 Oct 2024
TOWER: Tree Organized Weighting for Evaluating Complex Instructions
Noah Ziems
Zhihan Zhang
Meng Jiang
ALM
32
0
0
08 Oct 2024
KnowledgeSG: Privacy-Preserving Synthetic Text Generation with Knowledge Distillation from Server
Wenhao Wang
Xiaoyu Liang
Rui Ye
Jingyi Chai
Siheng Chen
Yanfeng Wang
SyDa
36
4
0
08 Oct 2024
1
2
3
4
5
6
Next