ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2010.08512
37
1

An Approximation Algorithm for Optimal Subarchitecture Extraction

16 October 2020
Adrian de Wynter
ArXivPDFHTML
Abstract

We consider the problem of finding the set of architectural parameters for a chosen deep neural network which is optimal under three metrics: parameter size, inference speed, and error rate. In this paper we state the problem formally, and present an approximation algorithm that, for a large subset of instances behaves like an FPTAS with an approximation error of ρ≤∣1−ϵ∣\rho \leq |{1- \epsilon}|ρ≤∣1−ϵ∣, and that runs in O(∣Ξ∣+∣WT∗∣(1+∣Θ∣∣B∣∣Ξ∣/(ϵ s3/2)))O(|{\Xi}| + |{W^*_T}|(1 + |{\Theta}||{B}||{\Xi}|/({\epsilon\, s^{3/2})}))O(∣Ξ∣+∣WT∗​∣(1+∣Θ∣∣B∣∣Ξ∣/(ϵs3/2))) steps, where ϵ\epsilonϵ and sss are input parameters; ∣B∣|{B}|∣B∣ is the batch size; ∣WT∗∣|{W^*_T}|∣WT∗​∣ denotes the cardinality of the largest weight set assignment; and ∣Ξ∣|{\Xi}|∣Ξ∣ and ∣Θ∣|{\Theta}|∣Θ∣ are the cardinalities of the candidate architecture and hyperparameter spaces, respectively.

View on arXiv
Comments on this paper