ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.16782
  4. Cited By
CipherPrune: Efficient and Scalable Private Transformer Inference

CipherPrune: Efficient and Scalable Private Transformer Inference

24 February 2025
Yancheng Zhang
Jinbao Xue
Mengxin Zheng
Mimi Xie
Mingzhe Zhang
Lei Jiang
Qian Lou
ArXivPDFHTML

Papers citing "CipherPrune: Efficient and Scalable Private Transformer Inference"

11 / 11 papers shown
Title
TFHE-Coder: Evaluating LLM-agentic Fully Homomorphic Encryption Code Generation
TFHE-Coder: Evaluating LLM-agentic Fully Homomorphic Encryption Code Generation
Mayank Kumar
Jinbao Xue
Mengxin Zheng
Qian Lou
87
3
0
15 Mar 2025
CryptoTrain: Fast Secure Training on Encrypted Dataset
CryptoTrain: Fast Secure Training on Encrypted Dataset
Jiaqi Xue
Yancheng Zhang
YanShan Wang
Xueqiang Wang
Hao Zheng
Qian Lou
47
3
0
25 Sep 2024
SnapKV: LLM Knows What You are Looking for Before Generation
SnapKV: LLM Knows What You are Looking for Before Generation
Yuhong Li
Yingbing Huang
Bowen Yang
Bharat Venkitesh
Acyr Locatelli
Hanchen Ye
Tianle Cai
Patrick Lewis
Deming Chen
VLM
100
180
0
22 Apr 2024
TrojLLM: A Black-box Trojan Prompt Attack on Large Language Models
TrojLLM: A Black-box Trojan Prompt Attack on Large Language Models
Jiaqi Xue
Mengxin Zheng
Ting Hua
Yilin Shen
Ye Liu
Ladislau Bölöni
Qian Lou
53
31
0
12 Jun 2023
MPCFormer: fast, performant and private Transformer inference with MPC
MPCFormer: fast, performant and private Transformer inference with MPC
Dacheng Li
Rulin Shao
Hongyi Wang
Han Guo
Eric P. Xing
Haotong Zhang
47
81
0
02 Nov 2022
Learned Token Pruning for Transformers
Learned Token Pruning for Transformers
Sehoon Kim
Sheng Shen
D. Thorsley
A. Gholami
Woosuk Kwon
Joseph Hassoun
Kurt Keutzer
38
152
0
02 Jul 2021
I-BERT: Integer-only BERT Quantization
I-BERT: Integer-only BERT Quantization
Sehoon Kim
A. Gholami
Z. Yao
Michael W. Mahoney
Kurt Keutzer
MQ
135
348
0
05 Jan 2021
SpAtten: Efficient Sparse Attention Architecture with Cascade Token and
  Head Pruning
SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning
Hanrui Wang
Zhekai Zhang
Song Han
99
384
0
17 Dec 2020
CrypTFlow2: Practical 2-Party Secure Inference
CrypTFlow2: Practical 2-Party Secure Inference
Deevashwer Rathee
Mayank Rathee
Nishant Kumar
Nishanth Chandran
Divya Gupta
Aseem Rastogi
Rahul Sharma
100
308
0
13 Oct 2020
BERT: Pre-training of Deep Bidirectional Transformers for Language
  Understanding
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Jacob Devlin
Ming-Wei Chang
Kenton Lee
Kristina Toutanova
VLM
SSL
SSeg
1.2K
93,936
0
11 Oct 2018
Attention Is All You Need
Attention Is All You Need
Ashish Vaswani
Noam M. Shazeer
Niki Parmar
Jakob Uszkoreit
Llion Jones
Aidan Gomez
Lukasz Kaiser
Illia Polosukhin
3DV
521
129,831
0
12 Jun 2017
1