ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.05641
37
0

Projectable Models: One-Shot Generation of Small Specialized Transformers from Large Ones

6 June 2025
A. Zhmoginov
Jihwan Lee
Mark Sandler
ArXiv (abs)PDFHTML
Main:4 Pages
4 Figures
Bibliography:2 Pages
4 Tables
Appendix:5 Pages
Abstract

Modern Foundation Models (FMs) are typically trained on corpora spanning a wide range of different data modalities, topics and downstream tasks. Utilizing these models can be very computationally expensive and is out of reach for most consumer devices. Furthermore, most of the broad FM knowledge may actually be irrelevant for a specific task at hand. Here we explore a technique for mapping parameters of a large Transformer to parameters of a smaller specialized model. By making this transformation task-specific, we aim to capture a narrower scope of the knowledge needed for performing a specific task by a smaller model. We study our method on image modeling tasks, showing that performance of generated models exceeds that of universal conditional models.

View on arXiv
@article{zhmoginov2025_2506.05641,
  title={ Projectable Models: One-Shot Generation of Small Specialized Transformers from Large Ones },
  author={ Andrey Zhmoginov and Jihwan Lee and Mark Sandler },
  journal={arXiv preprint arXiv:2506.05641},
  year={ 2025 }
}
Comments on this paper