ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2407.18921
  4. Cited By
Mobile Edge Intelligence for Large Language Models: A Contemporary Survey
v1v2 (latest)

Mobile Edge Intelligence for Large Language Models: A Contemporary Survey

9 July 2024
Guanqiao Qu
Qiyuan Chen
Wei Wei
Zheng Lin
Xianhao Chen
Kaibin Huang
ArXiv (abs)PDFHTML

Papers citing "Mobile Edge Intelligence for Large Language Models: A Contemporary Survey"

50 / 211 papers shown
Title
Unified Language-Vision Pretraining in LLM with Dynamic Discrete Visual
  Tokenization
Unified Language-Vision Pretraining in LLM with Dynamic Discrete Visual Tokenization
Yang Jin
Kun Xu
Kun Xu
Liwei Chen
Chao Liao
...
Xiaoqiang Lei
Di Zhang
Wenwu Ou
Kun Gai
Yadong Mu
MLLMVLM
79
50
0
09 Sep 2023
LLMCad: Fast and Scalable On-device Large Language Model Inference
LLMCad: Fast and Scalable On-device Large Language Model Inference
Daliang Xu
Wangsong Yin
Xin Jin
Yanzhe Zhang
Shiyun Wei
Mengwei Xu
Xuanzhe Liu
69
50
0
08 Sep 2023
Benchmarking Large Language Models in Retrieval-Augmented Generation
Benchmarking Large Language Models in Retrieval-Augmented Generation
Jiawei Chen
Hongyu Lin
Xianpei Han
Le Sun
3DVRALM
119
312
0
04 Sep 2023
A Study on the Implementation of Generative AI Services Using an
  Enterprise Data-Based LLM Application Architecture
A Study on the Implementation of Generative AI Services Using an Enterprise Data-Based LLM Application Architecture
CheonSu Jeong
72
49
0
03 Sep 2023
FederatedScope-LLM: A Comprehensive Package for Fine-tuning Large
  Language Models in Federated Learning
FederatedScope-LLM: A Comprehensive Package for Fine-tuning Large Language Models in Federated Learning
Weirui Kuang
Bingchen Qian
Zitao Li
Daoyuan Chen
Dawei Gao
Xuchen Pan
Yuexiang Xie
Yaliang Li
Bolin Ding
Jingren Zhou
FedML
124
136
0
01 Sep 2023
Efficient Model Personalization in Federated Learning via
  Client-Specific Prompt Generation
Efficient Model Personalization in Federated Learning via Client-Specific Prompt Generation
Fu-En Yang
Chien-Yi Wang
Yu-Chiang Frank Wang
VLMFedML
112
69
0
29 Aug 2023
OmniQuant: Omnidirectionally Calibrated Quantization for Large Language
  Models
OmniQuant: Omnidirectionally Calibrated Quantization for Large Language Models
Wenqi Shao
Mengzhao Chen
Zhaoyang Zhang
Peng Xu
Lirui Zhao
Zhiqiang Li
Kaipeng Zhang
Peng Gao
Yu Qiao
Ping Luo
MQ
127
206
0
25 Aug 2023
IncreLoRA: Incremental Parameter Allocation Method for
  Parameter-Efficient Fine-tuning
IncreLoRA: Incremental Parameter Allocation Method for Parameter-Efficient Fine-tuning
Feiyu F. Zhang
Liangzhi Li
Jun-Cheng Chen
Zhouqian Jiang
Bowen Wang
Yiming Qian
95
37
0
23 Aug 2023
Towards an On-device Agent for Text Rewriting
Towards an On-device Agent for Text Rewriting
Yun Zhu
Yinxiao Liu
Felix Stahlberg
Shankar Kumar
Yu-hui Chen
Liangchen Luo
Lei Shu
Renjie Liu
Jindong Chen
Lei Meng
LLMAG
59
7
0
22 Aug 2023
Instruction Tuning for Large Language Models: A Survey
Instruction Tuning for Large Language Models: A Survey
Shengyu Zhang
Linfeng Dong
Xiaoya Li
Sen Zhang
Xiaofei Sun
...
Jiwei Li
Runyi Hu
Tianwei Zhang
Leilei Gan
Guoyin Wang
LM&MA
110
610
0
21 Aug 2023
Optimal Resource Allocation for U-Shaped Parallel Split Learning
Optimal Resource Allocation for U-Shaped Parallel Split Learning
Song Lyu
Zhengyi Lin
Guanqiao Qu
Xianhao Chen
Xiaoxia Huang
P. Li
89
29
0
17 Aug 2023
A Survey on Model Compression for Large Language Models
A Survey on Model Compression for Large Language Models
Xunyu Zhu
Jian Li
Yong Liu
Can Ma
Weiping Wang
139
233
0
15 Aug 2023
SLoRA: Federated Parameter Efficient Fine-Tuning of Language Models
SLoRA: Federated Parameter Efficient Fine-Tuning of Language Models
Sara Babakniya
A. Elkordy
Yahya H. Ezzeldin
Qingfeng Liu
Kee-Bong Song
Mostafa El-Khamy
Salman Avestimehr
76
72
0
12 Aug 2023
Efficient Multiuser AI Downloading via Reusable Knowledge Broadcasting
Efficient Multiuser AI Downloading via Reusable Knowledge Broadcasting
Hai Wu
Qunsong Zeng
Kaibin Huang
104
8
0
28 Jul 2023
Backdoor Attacks for In-Context Learning with Language Models
Backdoor Attacks for In-Context Learning with Language Models
Nikhil Kandpal
Matthew Jagielski
Florian Tramèr
Nicholas Carlini
SILMAAML
118
84
0
27 Jul 2023
Low-Parameter Federated Learning with Large Language Models
Low-Parameter Federated Learning with Large Language Models
Jing Jiang
Xiangyang Liu
Chenyou Fan
61
26
0
26 Jul 2023
Llama 2: Open Foundation and Fine-Tuned Chat Models
Llama 2: Open Foundation and Fine-Tuned Chat Models
Hugo Touvron
Louis Martin
Kevin R. Stone
Peter Albert
Amjad Almahairi
...
Sharan Narang
Aurelien Rodriguez
Robert Stojnic
Sergey Edunov
Thomas Scialom
AI4MHALM
516
12,128
0
18 Jul 2023
Large Language Models
Large Language Models
Michael R Douglas
LLMAGLM&MA
177
645
0
11 Jul 2023
A Survey on Evaluation of Large Language Models
A Survey on Evaluation of Large Language Models
Yu-Chu Chang
Xu Wang
Jindong Wang
Yuanyi Wu
Linyi Yang
...
Yue Zhang
Yi-Ju Chang
Philip S. Yu
Qian Yang
Xingxu Xie
ELMLM&MAALM
223
1,764
0
06 Jul 2023
Large Language Models Empowered Autonomous Edge AI for Connected
  Intelligence
Large Language Models Empowered Autonomous Edge AI for Connected Intelligence
Yifei Shen
Jiawei Shao
Xinjie Zhang
Zehong Lin
Hao Pan
Dongsheng Li
Jun Zhang
Khaled B. Letaief
142
90
0
06 Jul 2023
H$_2$O: Heavy-Hitter Oracle for Efficient Generative Inference of Large
  Language Models
H2_22​O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models
Zhenyu Zhang
Ying Sheng
Dinesh Manocha
Tianlong Chen
Lianmin Zheng
...
Yuandong Tian
Christopher Ré
Clark W. Barrett
Zhangyang Wang
Beidi Chen
VLM
180
314
0
24 Jun 2023
Beyond Scale: The Diversity Coefficient as a Data Quality Metric for Variability in Natural Language Data
Beyond Scale: The Diversity Coefficient as a Data Quality Metric for Variability in Natural Language Data
Alycia Lee
Brando Miranda
Sudharsan Sundar
Allison Casasola
Rylan Schaeffer
Elyas Obbad
Sanmi Koyejo
131
17
0
24 Jun 2023
Split Learning in 6G Edge Networks
Split Learning in 6G Edge Networks
Zhengyi Lin
Guanqiao Qu
Xianhao Chen
Kaibin Huang
116
76
0
21 Jun 2023
SqueezeLLM: Dense-and-Sparse Quantization
SqueezeLLM: Dense-and-Sparse Quantization
Sehoon Kim
Coleman Hooper
A. Gholami
Zhen Dong
Xiuyu Li
Sheng Shen
Michael W. Mahoney
Kurt Keutzer
MQ
150
198
0
13 Jun 2023
SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight
  Compression
SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression
Tim Dettmers
Ruslan Svirschevski
Vage Egiazarian
Denis Kuznedelev
Elias Frantar
Saleh Ashkboos
Alexander Borzunov
Torsten Hoefler
Dan Alistarh
MQ
78
257
0
05 Jun 2023
An Overview on Generative AI at Scale with Edge-Cloud Computing
An Overview on Generative AI at Scale with Edge-Cloud Computing
Yun Cheng Wang
Jintang Xue
Chengwei Wei
C.-C. Jay Kuo
69
35
0
02 Jun 2023
Reducing Communication for Split Learning by Randomized Top-k
  Sparsification
Reducing Communication for Split Learning by Randomized Top-k Sparsification
Fei Zheng
Chaochao Chen
Lingjuan Lyu
Binhui Yao
FedML
79
11
0
29 May 2023
LLM-QAT: Data-Free Quantization Aware Training for Large Language Models
LLM-QAT: Data-Free Quantization Aware Training for Large Language Models
Zechun Liu
Barlas Oğuz
Changsheng Zhao
Ernie Chang
Pierre Stock
Yashar Mehdad
Yangyang Shi
Raghuraman Krishnamoorthi
Vikas Chandra
MQ
128
208
0
29 May 2023
PuMer: Pruning and Merging Tokens for Efficient Vision Language Models
PuMer: Pruning and Merging Tokens for Efficient Vision Language Models
Qingqing Cao
Bhargavi Paranjape
Hannaneh Hajishirzi
MLLMVLM
75
27
0
27 May 2023
Fine-Tuning Language Models with Just Forward Passes
Fine-Tuning Language Models with Just Forward Passes
Sadhika Malladi
Tianyu Gao
Eshaan Nichani
Alexandru Damian
Jason D. Lee
Danqi Chen
Sanjeev Arora
160
205
0
27 May 2023
Scissorhands: Exploiting the Persistence of Importance Hypothesis for
  LLM KV Cache Compression at Test Time
Scissorhands: Exploiting the Persistence of Importance Hypothesis for LLM KV Cache Compression at Test Time
Zichang Liu
Aditya Desai
Fangshuo Liao
Weitao Wang
Victor Xie
Zhaozhuo Xu
Anastasios Kyrillidis
Anshumali Shrivastava
84
237
0
26 May 2023
QLoRA: Efficient Finetuning of Quantized LLMs
QLoRA: Efficient Finetuning of Quantized LLMs
Tim Dettmers
Artidoro Pagnoni
Ari Holtzman
Luke Zettlemoyer
ALM
163
2,641
0
23 May 2023
LLM-Pruner: On the Structural Pruning of Large Language Models
LLM-Pruner: On the Structural Pruning of Large Language Models
Xinyin Ma
Gongfan Fang
Xinchao Wang
173
445
0
19 May 2023
Towards Building the Federated GPT: Federated Instruction Tuning
Towards Building the Federated GPT: Federated Instruction Tuning
Jianyi Zhang
Saeed Vahidian
Martin Kuo
Chunyuan Li
Ruiyi Zhang
Tong Yu
Yufan Zhou
Guoyin Wang
Yiran Chen
ALMFedML
85
132
0
09 May 2023
LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale
  Instructions
LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructions
Minghao Wu
Abdul Waheed
Chiyu Zhang
Muhammad Abdul-Mageed
Alham Fikri Aji
ALM
206
128
0
27 Apr 2023
Vehicle as a Service (VaaS): Leverage Vehicles to Build Service Networks
  and Capabilities for Smart Cities
Vehicle as a Service (VaaS): Leverage Vehicles to Build Service Networks and Capabilities for Smart Cities
Xianhao Chen
Yiqin Deng
Haichuan Ding
Guanqiao Qu
Haixia Zhang
P. Li
Yuguang Fang
GNN
92
44
0
22 Apr 2023
Instruction Tuning with GPT-4
Instruction Tuning with GPT-4
Baolin Peng
Chunyuan Li
Pengcheng He
Michel Galley
Jianfeng Gao
SyDaALMLM&MA
236
625
0
06 Apr 2023
Efficient Parallel Split Learning over Resource-constrained Wireless
  Edge Networks
Efficient Parallel Split Learning over Resource-constrained Wireless Edge Networks
Zhengyi Lin
Guangyu Zhu
Yiqin Deng
Xianhao Chen
Yue Gao
Kaibin Huang
Yuguang Fang
94
112
0
26 Mar 2023
GPT-4 Technical Report
GPT-4 Technical Report
OpenAI OpenAI
OpenAI Josh Achiam
Steven Adler
Sandhini Agarwal
Lama Ahmad
...
Shengjia Zhao
Tianhao Zheng
Juntang Zhuang
William Zhuk
Barret Zoph
LLMAGMLLM
1.6K
14,832
0
15 Mar 2023
Dynamic Stashing Quantization for Efficient Transformer Training
Dynamic Stashing Quantization for Efficient Transformer Training
Guofu Yang
Daniel Lo
Robert D. Mullins
Yiren Zhao
MQ
79
8
0
09 Mar 2023
Joint Task and Data Oriented Semantic Communications: A Deep Separate
  Source-channel Coding Scheme
Joint Task and Data Oriented Semantic Communications: A Deep Separate Source-channel Coding Scheme
Jianhao Huang
Dongxu Li
Chenyu Huang
Xiaoqi Qin
Wei Zhang
93
32
0
27 Feb 2023
DepGraph: Towards Any Structural Pruning
DepGraph: Towards Any Structural Pruning
Gongfan Fang
Xinyin Ma
Mingli Song
Michael Bi Mi
Xinchao Wang
GNN
172
274
0
30 Jan 2023
Prompt Federated Learning for Weather Forecasting: Toward Foundation
  Models on Meteorological Data
Prompt Federated Learning for Weather Forecasting: Toward Foundation Models on Meteorological Data
Shen Chen
Guodong Long
Tao Shen
Jing Jiang
AI4TS
115
39
0
22 Jan 2023
ChatGPT is not all you need. A State of the Art Review of large
  Generative AI models
ChatGPT is not all you need. A State of the Art Review of large Generative AI models
Roberto Gozalo-Brizuela
E.C. Garrido-Merchán
91
267
0
11 Jan 2023
SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot
SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot
Elias Frantar
Dan Alistarh
VLM
133
737
0
02 Jan 2023
Robust Split Federated Learning for U-shaped Medical Image Networks
Robust Split Federated Learning for U-shaped Medical Image Networks
Ziyuan Yang
Yingyu Chen
Huijie Huangfu
Maosong Ran
Hui Wang
Xiaoxiao Li
Yi Zhang
OODFedML
80
11
0
13 Dec 2022
Fast Inference from Transformers via Speculative Decoding
Fast Inference from Transformers via Speculative Decoding
Yaniv Leviathan
Matan Kalman
Yossi Matias
LRM
155
738
0
30 Nov 2022
On the Effectiveness of Parameter-Efficient Fine-Tuning
On the Effectiveness of Parameter-Efficient Fine-Tuning
Z. Fu
Haoran Yang
Anthony Man-Cho So
Wai Lam
Lidong Bing
Nigel Collier
76
162
0
28 Nov 2022
SmoothQuant: Accurate and Efficient Post-Training Quantization for Large
  Language Models
SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Guangxuan Xiao
Ji Lin
Mickael Seznec
Hao Wu
Julien Demouth
Song Han
MQ
255
844
0
18 Nov 2022
GPTQ: Accurate Post-Training Quantization for Generative Pre-trained
  Transformers
GPTQ: Accurate Post-Training Quantization for Generative Pre-trained Transformers
Elias Frantar
Saleh Ashkboos
Torsten Hoefler
Dan Alistarh
MQ
194
1,013
0
31 Oct 2022
Previous
12345
Next