10
0

DPUV4E: High-Throughput DPU Architecture Design for CNN on Versal ACAP

Main:8 Pages
10 Figures
Bibliography:2 Pages
4 Tables
Abstract

Convolutional Neural Networks (CNNs) remain prevalent in computer vision applications, and FPGAs, known for their flexibility and energy efficiency, have become essential components in heterogeneous acceleration systems. However, traditional FPGAs face challenges in balancing performance and versatility due to limited on-chip resources. AMD's Versal ACAP architecture, tailored for AI applications, incorporates AI Engines (AIEs) to deliver high computational power. Nevertheless, the platform suffers from insufficient memory bandwidth, hindering the full utilization of the AIEs' theoretical performance. In this paper, we present DPUV4E for the Versal architecture, providing configurations ranging from 2PE (32.632.6 TOPS) to 8PE (131.0131.0 TOPS). We design two computation units, Conv PE and DWC PE, to support different computational patterns. Each computation unit's data flow efficiently utilizes the data reuse opportunities to mitigate bandwidth bottlenecks. Additionally, we extend the functionality of each PE to utilize AIEs for non-convolutional operations, reducing resource overhead. Experiments on over 50 models show that compared to previous designs, our design provides 8.6×8.6\times the TOPS/W of traditional FPGA-based DPU designs, while reducing DSP usage by 95.8%95.8\%, LUT usage by 44.7%44.7\%, and latency to 68.5%68.5\% under single-batch conditions. For end-to-end inference, our design improving throughput by up to 2.2×2.2\times for depth-wise convolution models and up to 1.3×1.3\times for standard models.

View on arXiv
@article{li2025_2506.11441,
  title={ DPUV4E: High-Throughput DPU Architecture Design for CNN on Versal ACAP },
  author={ Guoyu Li and Pengbo Zheng and Jian Weng and Enshan Yang },
  journal={arXiv preprint arXiv:2506.11441},
  year={ 2025 }
}
Comments on this paper