Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2306.14893
Cited By
LongCoder: A Long-Range Pre-trained Language Model for Code Completion
26 June 2023
Daya Guo
Canwen Xu
Nan Duan
Jian Yin
Julian McAuley
Re-assign community
ArXiv (abs)
PDF
HTML
Github (2519★)
Papers citing
"LongCoder: A Long-Range Pre-trained Language Model for Code Completion"
49 / 49 papers shown
Title
HPCTransCompile: An AI Compiler Generated Dataset for High-Performance CUDA Transpilation and LLM Preliminary Exploration
Jiaqi Lv
Xufeng He
Yanchen Liu
Xu Dai
Aocheng Shen
Yinghao Li
Jiachen Hao
Jianrong Ding
Yang Hu
Shouyi Yin
103
0
0
12 Jun 2025
SwiftSpec: Ultra-Low Latency LLM Decoding by Scaling Asynchronous Speculative Decoding
Ziyi Zhang
Ziheng Jiang
Chengquan Jiang
Menghan Yu
Size Zheng
H. Lin
Henry Hoffmann
Xin Liu
25
0
0
12 Jun 2025
Code Researcher: Deep Research Agent for Large Systems Code and Commit History
Ramneet Singh
Sathvik Joel
Abhav Mehrotra
Nalin Wadhwa
Ramakrishna Bairi
Aditya Kanade
Nagarajan Natarajan
LLMAG
49
0
0
27 May 2025
Alignment-Augmented Speculative Decoding with Alignment Sampling and Conditional Verification
Jikai Wang
Zhenxu Tian
Jilong Li
Qingrong Xia
Xinyu Duan
Zhefeng Wang
Baoxing Huai
Min Zhang
61
0
0
19 May 2025
Sparse Attention Remapping with Clustering for Efficient LLM Decoding on PIM
Zehao Fan
Garrett Gagnon
Zhenyu Liu
Liu Liu
56
0
0
09 May 2025
Can LLMs handle WebShell detection? Overcoming Detection Challenges with Behavioral Function-Aware Framework
Feijiang Han
Jiaming Zhang
Chuyi Deng
Jianheng Tang
Yunhuai Liu
67
1
0
14 Apr 2025
Towards an Understanding of Context Utilization in Code Intelligence
Yanlin Wang
Kefeng Duan
Dewu Zheng
Ensheng Shi
F. Zhang
...
Xilin Liu
Yuchi Ma
Hongyu Zhang
Qianxiang Wang
Zibin Zheng
92
0
0
11 Apr 2025
Safe Screening Rules for Group OWL Models
Runxue Bao
Quanchao Lu
Yanfu Zhang
110
0
0
04 Apr 2025
GPU-Accelerated Motion Planning of an Underactuated Forestry Crane in Cluttered Environments
M. Vu
Gerald Ebmer
Alexander Watcher
Marc-Philip Ecker
Giang Nguyen
Tobias Glueck
133
3
0
18 Mar 2025
A Survey on Transformer Context Extension: Approaches and Evaluation
Yijun Liu
Jinzheng Yu
Yang Xu
Zhongyang Li
Qingfu Zhu
LLMAG
128
3
0
17 Mar 2025
Code to Think, Think to Code: A Survey on Code-Enhanced Reasoning and Reasoning-Driven Code Intelligence in LLMs
Dayu Yang
Tianyang Liu
Daoan Zhang
Antoine Simoulin
Xiaoyi Liu
...
Zhaopu Teng
Xin Qian
Grey Yang
Jiebo Luo
Julian McAuley
ReLM
OffRL
LRM
153
12
0
26 Feb 2025
DBudgetKV: Dynamic Budget in KV Cache Compression for Ensuring Optimal Performance
Xuanfan Ni
Liyan Xu
Chenyang Lyu
Longyue Wang
Mo Yu
Lemao Liu
Fandong Meng
Jie Zhou
Piji Li
108
0
0
24 Feb 2025
LongSpec: Long-Context Lossless Speculative Decoding with Efficient Drafting and Verification
Penghui Yang
Cunxiao Du
Fengzhuo Zhang
Haonan Wang
Tianyu Pang
Chao Du
Bo An
RALM
101
2
0
24 Feb 2025
Neural Attention Search
Difan Deng
Marius Lindauer
144
0
0
21 Feb 2025
Twilight: Adaptive Attention Sparsity with Hierarchical Top-
p
p
p
Pruning
C. Lin
Jiaming Tang
Shuo Yang
Hanshuo Wang
Tian Tang
Boyu Tian
Ion Stoica
Enze Xie
Mingyu Gao
171
5
0
04 Feb 2025
Raiders of the Lost Dependency: Fixing Dependency Conflicts in Python using LLMs
Antony Bartlett
Cynthia C. S. Liem
Annibale Panichella
39
0
0
28 Jan 2025
Ada-KV: Optimizing KV Cache Eviction by Adaptive Budget Allocation for Efficient LLM Inference
Yuan Feng
Junlin Lv
Yukun Cao
Xike Xie
S. K. Zhou
VLM
150
44
0
28 Jan 2025
Efficient Prompt Compression with Evaluator Heads for Long-Context Transformer Inference
Weizhi Fei
Xueyan Niu
Guoqing Xie
Yingqing Liu
Bo Bai
Wei Han
130
1
0
22 Jan 2025
AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Janghwan Lee
Jiwoong Park
Jinseok Kim
Yongjik Kim
Jungju Oh
Jinwook Oh
Jungwook Choi
73
2
0
15 Nov 2024
LLM-Inference-Bench: Inference Benchmarking of Large Language Models on AI Accelerators
Krishna Teja Chitty-Venkata
Siddhisanket Raskar
B. Kale
Farah Ferdaus
Aditya Tanikanti
Ken Raffenetti
Valerie Taylor
M. Emani
V. Vishwanath
150
12
0
31 Oct 2024
WAFFLE: Finetuning Multi-Modal Model for Automated Front-End Development
Shanchao Liang
Nan Jiang
Shangshu Qian
Lin Tan
87
1
0
24 Oct 2024
Fine-grained Attention I/O Complexity: Comprehensive Analysis for Backward Passes
Xiaoyu Li
Yingyu Liang
Zhenmei Shi
Zhao Song
Yufa Zhou
94
18
0
12 Oct 2024
Rodimus*: Breaking the Accuracy-Efficiency Trade-Off with Efficient Attentions
Zhihao He
Hang Yu
Zi Gong
Shizhan Liu
Jia-Nan Li
Weiyao Lin
VLM
104
2
0
09 Oct 2024
Unlocking Structured Thinking in Language Models with Cognitive Prompting
Oliver Kramer
Jill Baumann
ReLM
LRM
77
4
0
03 Oct 2024
Retrieval Or Holistic Understanding? Dolce: Differentiate Our Long Context Evaluation Tasks
Zi Yang
44
2
0
10 Sep 2024
DataSculpt: Crafting Data Landscapes for Long-Context LLMs through Multi-Objective Partitioning
Keer Lu
Xiaonan Nie
Zhuoran Zhang
Zheng Liang
Da Pan
...
Weipeng Chen
Guosheng Dong
Bin Cui
Bin Cui
Wentao Zhang
109
0
0
02 Sep 2024
Self-evolving Agents with reflective and memory-augmented abilities
Xuechen Liang
Yangfan He
Yinghui Xia
Xinyuan Song
Jianhui Wang
...
Keqin Li
Jiaqi Chen
Jinsong Yang
Siyuan Chen
Tianyu Shi
LLMAG
KELM
CLL
147
4
0
01 Sep 2024
Palu: Compressing KV-Cache with Low-Rank Projection
Chi-Chih Chang
Wei-Cheng Lin
Chien-Yu Lin
Chong-Yan Chen
Yu-Fang Hu
Pei-Shuo Wang
N. Huang
Luis Ceze
Kai-Chiang Wu
110
2
0
30 Jul 2024
Scaling Granite Code Models to 128K Context
Matt Stallone
Vaibhav Saxena
Leonid Karlinsky
Bridget McGinn
Tim Bula
...
Rogerio Feris
Nirmit Desai
David D. Cox
Ruchir Puri
Yikang Shen
69
4
0
18 Jul 2024
On Evaluating The Performance of Watermarked Machine-Generated Texts Under Adversarial Attacks
Zesen Liu
Tianshuo Cong
Xinlei He
Qi Li
AAML
WaLM
115
1
0
05 Jul 2024
Is It Really Long Context if All You Need Is Retrieval? Towards Genuinely Difficult Long Context NLP
Omer Goldman
Alon Jacovi
Aviv Slobodkin
Aviya Maimon
Ido Dagan
Reut Tsarfaty
116
11
0
29 Jun 2024
Sparser is Faster and Less is More: Efficient Sparse Attention for Long-Range Transformers
Chao Lou
Zixia Jia
Zilong Zheng
Kewei Tu
ODL
85
26
0
24 Jun 2024
An LLM-Assisted Easy-to-Trigger Backdoor Attack on Code Completion Models: Injecting Disguised Vulnerabilities against Strong Detection
Shenao Yan
Shen Wang
Yue Duan
Hanbin Hong
Kiho Lee
Doowon Kim
Yuan Hong
AAML
SILM
76
26
0
10 Jun 2024
Chain of Agents: Large Language Models Collaborating on Long-Context Tasks
Yusen Zhang
Ruoxi Sun
Yanfei Chen
Tomas Pfister
Rui Zhang
Sercan O. Arik
RALM
AI4CE
LLMAG
133
57
0
04 Jun 2024
PyramidKV: Dynamic KV Cache Compression based on Pyramidal Information Funneling
Zefan Cai
Yichi Zhang
Bofei Gao
Yuliang Liu
Yongqian Li
...
Wayne Xiong
Yue Dong
Baobao Chang
Junjie Hu
Wen Xiao
196
107
0
04 Jun 2024
A Survey on Large Language Models for Code Generation
Juyong Jiang
Fan Wang
Jiasi Shen
Sungju Kim
Sunghun Kim
145
204
0
01 Jun 2024
Long Context is Not Long at All: A Prospector of Long-Dependency Data for Large Language Models
Longze Chen
Ziqiang Liu
Wanwei He
Yunshui Li
Run Luo
Min Yang
75
12
0
28 May 2024
Compressing Lengthy Context With UltraGist
Peitian Zhang
Zheng Liu
Shitao Xiao
Ninglu Shao
Qiwei Ye
Zhicheng Dou
46
4
0
26 May 2024
Are Long-LLMs A Necessity For Long-Context Tasks?
Hongjin Qian
Zheng Liu
Peitian Zhang
Kelong Mao
Yujia Zhou
Xu Chen
Zhicheng Dou
71
13
0
24 May 2024
Performance-Aligned LLMs for Generating Fast Code
Daniel Nichols
Pranav Polasam
Harshitha Menon
Aniruddha Marathe
T. Gamblin
A. Bhatele
81
10
0
29 Apr 2024
CORM: Cache Optimization with Recent Message for Large Language Model Inference
Jincheng Dai
Zhuowei Huang
Haiyun Jiang
Chen Chen
Deng Cai
Wei Bi
Shuming Shi
109
3
0
24 Apr 2024
Identify Critical Nodes in Complex Network with Large Language Models
Jinzhu Mao
Dongyun Zou
Li Sheng
Siyi Liu
Chen Gao
Yue Wang
Yong Li
62
4
0
01 Mar 2024
WKVQuant: Quantizing Weight and Key/Value Cache for Large Language Models Gains More
Yuxuan Yue
Zhihang Yuan
Haojie Duanmu
Sifan Zhou
Jianlong Wu
Liqiang Nie
MQ
93
49
0
19 Feb 2024
Nova: Generative Language Models for Assembly Code with Hierarchical Attention and Contrastive Learning
Nan Jiang
Chengxiao Wang
Kevin Liu
Xiangzhe Xu
Lin Tan
Xiangyu Zhang
OffRL
85
8
0
22 Nov 2023
Prompt Cache: Modular Attention Reuse for Low-Latency Inference
In Gim
Guojun Chen
Seung-seob Lee
Nikhil Sarda
Anurag Khandelwal
Lin Zhong
121
88
0
07 Nov 2023
Attention Alignment and Flexible Positional Embeddings Improve Transformer Length Extrapolation
Ta-Chung Chi
Ting-Han Fan
Alexander I. Rudnicky
39
4
0
01 Nov 2023
LLM for SoC Security: A Paradigm Shift
Dipayan Saha
Shams Tarek
Katayoon Yahyaei
S. Saha
Jingbo Zhou
M. Tehranipoor
Farimah Farahmandi
172
54
0
09 Oct 2023
LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding
Yushi Bai
Xin Lv
Jiajie Zhang
Hong Lyu
Jiankai Tang
...
Aohan Zeng
Lei Hou
Yuxiao Dong
Jie Tang
Juanzi Li
LLMAG
RALM
119
604
0
28 Aug 2023
Code Llama: Open Foundation Models for Code
Baptiste Rozière
Jonas Gehring
Fabian Gloeckle
Sten Sootla
Itai Gat
...
Hugo Touvron
Louis Martin
Nicolas Usunier
Thomas Scialom
Gabriel Synnaeve
ELM
ALM
143
2,112
0
24 Aug 2023
1