ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2411.18092
109
0
v1v2 (latest)

Training Noise Token Pruning

27 November 2024
Mingxing Rao
Bohan Jiang
Daniel Moyer
    ViT
ArXiv (abs)PDFHTML
Abstract

In the present work we present Training Noise Token (TNT) Pruning for vision transformers. Our method relaxes the discrete token dropping condition to continuous additive noise, providing smooth optimization in training, while retaining discrete dropping computational gains in deployment settings. We provide theoretical connections to Rate-Distortion literature, and empirical evaluations on the ImageNet dataset using ViT and DeiT architectures demonstrating TNT's advantages over previous pruning methods.

View on arXiv
@article{rao2025_2411.18092,
  title={ Training Noise Token Pruning },
  author={ Mingxing Rao and Bohan Jiang and Daniel Moyer },
  journal={arXiv preprint arXiv:2411.18092},
  year={ 2025 }
}
Comments on this paper