ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2201.06618
  4. Cited By
VAQF: Fully Automatic Software-Hardware Co-Design Framework for Low-Bit
  Vision Transformer

VAQF: Fully Automatic Software-Hardware Co-Design Framework for Low-Bit Vision Transformer

17 January 2022
Mengshu Sun
Haoyu Ma
Guoliang Kang
Yi Ding
Tianlong Chen
Xiaolong Ma
Zhangyang Wang
Yanzhi Wang
    ViT
ArXivPDFHTML

Papers citing "VAQF: Fully Automatic Software-Hardware Co-Design Framework for Low-Bit Vision Transformer"

29 / 29 papers shown
Title
COBRA: Algorithm-Architecture Co-optimized Binary Transformer Accelerator for Edge Inference
COBRA: Algorithm-Architecture Co-optimized Binary Transformer Accelerator for Edge Inference
Ye Qiao
Zhiheng Cheng
Yian Wang
Yifan Zhang
Yunzhe Deng
Sitao Huang
79
0
0
22 Apr 2025
HG-PIPE: Vision Transformer Acceleration with Hybrid-Grained Pipeline
HG-PIPE: Vision Transformer Acceleration with Hybrid-Grained Pipeline
Qingyu Guo
Jiayong Wan
Songqiang Xu
Meng Li
Yuan Wang
36
1
0
25 Jul 2024
CHOSEN: Compilation to Hardware Optimization Stack for Efficient Vision
  Transformer Inference
CHOSEN: Compilation to Hardware Optimization Stack for Efficient Vision Transformer Inference
Mohammad Erfan Sadeghi
A. Fayyazi
Suhas Somashekar
Massoud Pedram
34
2
0
17 Jul 2024
ViT-1.58b: Mobile Vision Transformers in the 1-bit Era
ViT-1.58b: Mobile Vision Transformers in the 1-bit Era
Zhengqing Yuan
Rong-Er Zhou
Hongyi Wang
Lifang He
Yanfang Ye
Lichao Sun
MQ
27
8
0
26 Jun 2024
P$^2$-ViT: Power-of-Two Post-Training Quantization and Acceleration for
  Fully Quantized Vision Transformer
P2^22-ViT: Power-of-Two Post-Training Quantization and Acceleration for Fully Quantized Vision Transformer
Huihong Shi
Xin Cheng
Wendong Mao
Zhongfeng Wang
MQ
45
3
0
30 May 2024
From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of
  Deep Neural Networks
From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of Deep Neural Networks
Xue Geng
Zhe Wang
Chunyun Chen
Qing Xu
Kaixin Xu
...
Zhenghua Chen
M. Aly
Jie Lin
Min-man Wu
Xiaoli Li
33
1
0
09 May 2024
Trio-ViT: Post-Training Quantization and Acceleration for Softmax-Free
  Efficient Vision Transformer
Trio-ViT: Post-Training Quantization and Acceleration for Softmax-Free Efficient Vision Transformer
Huihong Shi
Haikuo Shao
Wendong Mao
Zhongfeng Wang
ViT
MQ
41
3
0
06 May 2024
Model Quantization and Hardware Acceleration for Vision Transformers: A
  Comprehensive Survey
Model Quantization and Hardware Acceleration for Vision Transformers: A Comprehensive Survey
Dayou Du
Gu Gong
Xiaowen Chu
MQ
38
7
0
01 May 2024
Lightweight Deep Learning for Resource-Constrained Environments: A
  Survey
Lightweight Deep Learning for Resource-Constrained Environments: A Survey
Hou-I Liu
Marco Galindo
Hongxia Xie
Lai-Kuan Wong
Hong-Han Shuai
Yung-Hui Li
Wen-Huang Cheng
55
48
0
08 Apr 2024
An FPGA-Based Reconfigurable Accelerator for Convolution-Transformer
  Hybrid EfficientViT
An FPGA-Based Reconfigurable Accelerator for Convolution-Transformer Hybrid EfficientViT
Haikuo Shao
Huihong Shi
Wendong Mao
Zhongfeng Wang
21
2
0
29 Mar 2024
Faster Inference of Integer SWIN Transformer by Removing the GELU
  Activation
Faster Inference of Integer SWIN Transformer by Removing the GELU Activation
Mohammadreza Tayaranian
S. H. Mozafari
James J. Clark
Brett H. Meyer
Warren Gross
28
1
0
02 Feb 2024
BETA: Binarized Energy-Efficient Transformer Accelerator at the Edge
BETA: Binarized Energy-Efficient Transformer Accelerator at the Edge
Yuhao Ji
Chao Fang
Zhongfeng Wang
30
3
0
22 Jan 2024
Accelerating Neural Networks for Large Language Models and Graph
  Processing with Silicon Photonics
Accelerating Neural Networks for Large Language Models and Graph Processing with Silicon Photonics
Salma Afifi
Febin P. Sunny
Mahdi Nikdast
S. Pasricha
GNN
29
2
0
12 Jan 2024
A Cost-Efficient FPGA Implementation of Tiny Transformer Model using
  Neural ODE
A Cost-Efficient FPGA Implementation of Tiny Transformer Model using Neural ODE
Ikumi Okubo
Keisuke Sugiura
Hiroki Matsutani
36
2
0
05 Jan 2024
Improved TokenPose with Sparsity
Improved TokenPose with Sparsity
Anning Li
ViT
34
0
0
16 Nov 2023
Quantized Feature Distillation for Network Quantization
Quantized Feature Distillation for Network Quantization
Kevin Zhu
Yin He
Jianxin Wu
MQ
29
9
0
20 Jul 2023
Edge-MoE: Memory-Efficient Multi-Task Vision Transformer Architecture
  with Task-level Sparsity via Mixture-of-Experts
Edge-MoE: Memory-Efficient Multi-Task Vision Transformer Architecture with Task-level Sparsity via Mixture-of-Experts
Rishov Sarkar
Hanxue Liang
Zhiwen Fan
Zhangyang Wang
Cong Hao
MoE
27
17
0
30 May 2023
Distilling Token-Pruned Pose Transformer for 2D Human Pose Estimation
Distilling Token-Pruned Pose Transformer for 2D Human Pose Estimation
Feixiang Ren
ViT
19
2
0
12 Apr 2023
TRON: Transformer Neural Network Acceleration with Non-Coherent Silicon
  Photonics
TRON: Transformer Neural Network Acceleration with Non-Coherent Silicon Photonics
S. Afifi
Febin P. Sunny
Mahdi Nikdast
S. Pasricha
GNN
32
20
0
22 Mar 2023
ViTA: A Vision Transformer Inference Accelerator for Edge Applications
ViTA: A Vision Transformer Inference Accelerator for Edge Applications
Shashank Nag
Gourav Datta
Souvik Kundu
N. Chandrachoodan
P. Beerel
ViT
20
25
0
17 Feb 2023
M$^3$ViT: Mixture-of-Experts Vision Transformer for Efficient Multi-task
  Learning with Model-Accelerator Co-design
M3^33ViT: Mixture-of-Experts Vision Transformer for Efficient Multi-task Learning with Model-Accelerator Co-design
Hanxue Liang
Zhiwen Fan
Rishov Sarkar
Ziyu Jiang
Tianlong Chen
Kai Zou
Yu Cheng
Cong Hao
Zhangyang Wang
MoE
31
81
0
26 Oct 2022
ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design
ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design
Haoran You
Zhanyi Sun
Huihong Shi
Zhongzhi Yu
Yang Katie Zhao
Yongan Zhang
Chaojian Li
Baopu Li
Yingyan Lin
ViT
22
76
0
18 Oct 2022
PPT: token-Pruned Pose Transformer for monocular and multi-view human
  pose estimation
PPT: token-Pruned Pose Transformer for monocular and multi-view human pose estimation
Haoyu Ma
Zhe Wang
Yifei Chen
Deying Kong
Liangjian Chen
Xingwei Liu
Xiangyi Yan
Hao Tang
Xiaohui Xie
ViT
35
47
0
16 Sep 2022
MobileViT: Light-weight, General-purpose, and Mobile-friendly Vision
  Transformer
MobileViT: Light-weight, General-purpose, and Mobile-friendly Vision Transformer
Sachin Mehta
Mohammad Rastegari
ViT
218
1,213
0
05 Oct 2021
Transformer in Transformer
Transformer in Transformer
Kai Han
An Xiao
Enhua Wu
Jianyuan Guo
Chunjing Xu
Yunhe Wang
ViT
284
1,524
0
27 Feb 2021
Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction
  without Convolutions
Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions
Wenhai Wang
Enze Xie
Xiang Li
Deng-Ping Fan
Kaitao Song
Ding Liang
Tong Lu
Ping Luo
Ling Shao
ViT
277
3,623
0
24 Feb 2021
BinaryBERT: Pushing the Limit of BERT Quantization
BinaryBERT: Pushing the Limit of BERT Quantization
Haoli Bai
Wei Zhang
Lu Hou
Lifeng Shang
Jing Jin
Xin Jiang
Qun Liu
Michael Lyu
Irwin King
MQ
142
221
0
31 Dec 2020
The Lottery Ticket Hypothesis for Pre-trained BERT Networks
The Lottery Ticket Hypothesis for Pre-trained BERT Networks
Tianlong Chen
Jonathan Frankle
Shiyu Chang
Sijia Liu
Yang Zhang
Zhangyang Wang
Michael Carbin
153
345
0
23 Jul 2020
Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT
Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT
Sheng Shen
Zhen Dong
Jiayu Ye
Linjian Ma
Z. Yao
A. Gholami
Michael W. Mahoney
Kurt Keutzer
MQ
230
575
0
12 Sep 2019
1