ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2510.16075
12
0

Optimization of the quantization of dense neural networks from an exact QUBO formulation

17 October 2025
Sergio Muñiz Subiñas
Manuel L. González
Jorge Ruiz Gómez
Alejandro Mata Ali
Jorge Martínez Martín
Miguel Franco Hernando
Ángel Miguel García-Vico
    MQ
ArXiv (abs)PDFHTML
Main:10 Pages
2 Figures
Bibliography:3 Pages
2 Tables
Appendix:2 Pages
Abstract

This work introduces a post-training quantization (PTQ) method for dense neural networks via a novel ADAROUND-based QUBO formulation. Using the Frobenius distance between the theoretical output and the dequantized output (before the activation function) as the objective, an explicit QUBO whose binary variables represent the rounding choice for each weight and bias is obtained. Additionally, by exploiting the structure of the coefficient QUBO matrix, the global problem can be exactly decomposed into nnn independent subproblems of size f+1f+1f+1, which can be efficiently solved using some heuristics such as simulated annealing. The approach is evaluated on MNIST, Fashion-MNIST, EMNIST, and CIFAR-10 across integer precisions from int8 to int1 and compared with a round-to-nearest traditional quantization methodology.

View on arXiv
Comments on this paper