ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2104.10972
  4. Cited By
ImageNet-21K Pretraining for the Masses

ImageNet-21K Pretraining for the Masses

22 April 2021
T. Ridnik
Emanuel Ben-Baruch
Asaf Noy
Lihi Zelnik-Manor
    SSeg
    VLM
    CLIP
ArXivPDFHTML

Papers citing "ImageNet-21K Pretraining for the Masses"

50 / 151 papers shown
Title
Leveraging Self-Supervised Training for Unintentional Action Recognition
Leveraging Self-Supervised Training for Unintentional Action Recognition
Enea Duka
Anna Kukleva
Bernt Schiele
33
1
0
23 Sep 2022
Synthetic Latent Fingerprint Generator
Synthetic Latent Fingerprint Generator
André Brasil Vieira Wyzykowski
A.K. Jain
32
13
0
29 Aug 2022
An Impartial Take to the CNN vs Transformer Robustness Contest
An Impartial Take to the CNN vs Transformer Robustness Contest
Francesco Pinto
Philip Torr
P. Dokania
UQCV
AAML
33
49
0
22 Jul 2022
On the Generalizability and Predictability of Recommender Systems
On the Generalizability and Predictability of Recommender Systems
Duncan C. McElfresh
Sujay Khandagale
Jonathan Valverde
John P. Dickerson
Colin White
41
10
0
23 Jun 2022
EdgeNeXt: Efficiently Amalgamated CNN-Transformer Architecture for
  Mobile Vision Applications
EdgeNeXt: Efficiently Amalgamated CNN-Transformer Architecture for Mobile Vision Applications
Muhammad Maaz
Abdelrahman M. Shaker
Hisham Cholakkal
Salman Khan
Syed Waqas Zamir
Rao Muhammad Anwer
Fahad Shahbaz Khan
ViT
29
184
0
21 Jun 2022
Unified-IO: A Unified Model for Vision, Language, and Multi-Modal Tasks
Unified-IO: A Unified Model for Vision, Language, and Multi-Modal Tasks
Jiasen Lu
Christopher Clark
Rowan Zellers
Roozbeh Mottaghi
Aniruddha Kembhavi
ObjD
VLM
MLLM
56
393
0
17 Jun 2022
Efficient Adaptive Ensembling for Image Classification
Efficient Adaptive Ensembling for Image Classification
A. Bruno
Davide Moroni
M. Martinelli
34
18
0
15 Jun 2022
Differentiable Top-k Classification Learning
Differentiable Top-k Classification Learning
Felix Petersen
Hilde Kuehne
Christian Borgelt
Oliver Deussen
59
28
0
15 Jun 2022
Separable Self-attention for Mobile Vision Transformers
Separable Self-attention for Mobile Vision Transformers
Sachin Mehta
Mohammad Rastegari
ViT
MQ
26
252
0
06 Jun 2022
Which models are innately best at uncertainty estimation?
Which models are innately best at uncertainty estimation?
Ido Galil
Mohammed Dabbah
Ran El-Yaniv
UQCV
34
5
0
05 Jun 2022
Self-Supervised Pre-training of Vision Transformers for Dense Prediction
  Tasks
Self-Supervised Pre-training of Vision Transformers for Dense Prediction Tasks
Jaonary Rabarisoa
Velentin Belissen
Florian Chabot
Q. C. Pham
VLM
ViT
SSL
MDE
23
2
0
30 May 2022
A Closer Look at Self-Supervised Lightweight Vision Transformers
A Closer Look at Self-Supervised Lightweight Vision Transformers
Shaoru Wang
Jin Gao
Zeming Li
Jian Sun
Weiming Hu
ViT
67
41
0
28 May 2022
Vision Transformers in 2022: An Update on Tiny ImageNet
Vision Transformers in 2022: An Update on Tiny ImageNet
Ethan Huynh
ViT
31
11
0
21 May 2022
A Unified and Biologically-Plausible Relational Graph Representation of
  Vision Transformers
A Unified and Biologically-Plausible Relational Graph Representation of Vision Transformers
Yuzhong Chen
Yu Du
Zhe Xiao
Lin Zhao
Lu Zhang
...
Dajiang Zhu
Tuo Zhang
Xintao Hu
Tianming Liu
Xi Jiang
ViT
27
5
0
20 May 2022
Identical Image Retrieval using Deep Learning
Identical Image Retrieval using Deep Learning
Sayan Nath
Nikhil Nayak
VLM
34
1
0
10 May 2022
Leaner and Faster: Two-Stage Model Compression for Lightweight
  Text-Image Retrieval
Leaner and Faster: Two-Stage Model Compression for Lightweight Text-Image Retrieval
Siyu Ren
Kenny Q. Zhu
VLM
30
7
0
29 Apr 2022
Where in the World is this Image? Transformer-based Geo-localization in
  the Wild
Where in the World is this Image? Transformer-based Geo-localization in the Wild
Shraman Pramanick
E. Nowara
Joshua Gleason
Carlos D. Castillo
Rama Chellappa
ViT
21
30
0
29 Apr 2022
The MeVer DeepFake Detection Service: Lessons Learnt from Developing and
  Deploying in the Wild
The MeVer DeepFake Detection Service: Lessons Learnt from Developing and Deploying in the Wild
Spyridon Baxevanakis
Giorgos Kordopatis-Zilos
Panagiotis Galopoulos
Lazaros Apostolidis
Killian Levacher
Ipek B. Schlicht
Denis Teyssou
I. Kompatsiaris
Symeon Papadopoulos
42
8
0
27 Apr 2022
VISTA: Vision Transformer enhanced by U-Net and Image Colorfulness Frame
  Filtration for Automatic Retail Checkout
VISTA: Vision Transformer enhanced by U-Net and Image Colorfulness Frame Filtration for Automatic Retail Checkout
Md. Istiak Hossain Shihab
Nazia Tasnim
H. Zunair
L. Rupty
Nabeel Mohammed
33
8
0
23 Apr 2022
Learning to Scaffold: Optimizing Model Explanations for Teaching
Learning to Scaffold: Optimizing Model Explanations for Teaching
Patrick Fernandes
Marcos Vinícius Treviso
Danish Pruthi
André F. T. Martins
Graham Neubig
FAtt
25
22
0
22 Apr 2022
DualPrompt: Complementary Prompting for Rehearsal-free Continual
  Learning
DualPrompt: Complementary Prompting for Rehearsal-free Continual Learning
Zifeng Wang
Zizhao Zhang
Sayna Ebrahimi
Ruoxi Sun
Han Zhang
...
Xiaoqi Ren
Guolong Su
Vincent Perot
Jennifer Dy
Tomas Pfister
CLL
VLM
VPVLM
36
460
0
10 Apr 2022
Solving ImageNet: a Unified Scheme for Training any Backbone to Top
  Results
Solving ImageNet: a Unified Scheme for Training any Backbone to Top Results
T. Ridnik
Hussam Lawen
Emanuel Ben-Baruch
Asaf Noy
38
11
0
07 Apr 2022
How stable are Transferability Metrics evaluations?
How stable are Transferability Metrics evaluations?
A. Agostinelli
Michal Pándy
J. Uijlings
Thomas Mensink
V. Ferrari
35
22
0
04 Apr 2022
CLIP-Mesh: Generating textured meshes from text using pretrained
  image-text models
CLIP-Mesh: Generating textured meshes from text using pretrained image-text models
N. Khalid
Tianhao Xie
Eugene Belilovsky
Tiberiu Popa
CLIP
13
291
0
24 Mar 2022
X-Learner: Learning Cross Sources and Tasks for Universal Visual
  Representation
X-Learner: Learning Cross Sources and Tasks for Universal Visual Representation
Yinan He
Gengshi Huang
Siyu Chen
Jianing Teng
Wang Kun
Zhen-fei Yin
Lu Sheng
Ziwei Liu
Yu Qiao
Jing Shao
VLM
SSL
ViT
43
7
0
16 Mar 2022
Knowledge Distillation as Efficient Pre-training: Faster Convergence,
  Higher Data-efficiency, and Better Transferability
Knowledge Distillation as Efficient Pre-training: Faster Convergence, Higher Data-efficiency, and Better Transferability
Ruifei He
Shuyang Sun
Jihan Yang
Song Bai
Xiaojuan Qi
34
36
0
10 Mar 2022
RankSeg: Adaptive Pixel Classification with Image Category Ranking for
  Segmentation
RankSeg: Adaptive Pixel Classification with Image Category Ranking for Segmentation
Hao He
Yuhui Yuan
Xiangyu Yue
Han Hu
VOS
VLM
27
13
0
08 Mar 2022
MetaFormer: A Unified Meta Framework for Fine-Grained Recognition
MetaFormer: A Unified Meta Framework for Fine-Grained Recognition
Qishuai Diao
Yi-Xin Jiang
Bin Wen
Jianxiang Sun
Zehuan Yuan
39
60
0
05 Mar 2022
Navigating Neural Space: Revisiting Concept Activation Vectors to Overcome Directional Divergence
Navigating Neural Space: Revisiting Concept Activation Vectors to Overcome Directional Divergence
Frederik Pahde
Maximilian Dreyer
Leander Weber
Moritz Weckbecker
Christopher J. Anders
Thomas Wiegand
Wojciech Samek
Sebastian Lapuschkin
63
7
0
07 Feb 2022
It's All in the Head: Representation Knowledge Distillation through
  Classifier Sharing
It's All in the Head: Representation Knowledge Distillation through Classifier Sharing
Emanuel Ben-Baruch
M. Karklinsky
Yossi Biton
Avi Ben-Cohen
Hussam Lawen
Nadav Zamir
24
11
0
18 Jan 2022
Video Transformers: A Survey
Video Transformers: A Survey
Javier Selva
A. S. Johansen
Sergio Escalera
Kamal Nasrollahi
T. Moeslund
Albert Clapés
ViT
22
103
0
16 Jan 2022
Towards General and Efficient Active Learning
Towards General and Efficient Active Learning
Yichen Xie
Masayoshi Tomizuka
Wei Zhan
VLM
35
10
0
15 Dec 2021
Transformaly -- Two (Feature Spaces) Are Better Than One
Transformaly -- Two (Feature Spaces) Are Better Than One
M. Cohen
S. Avidan
ViT
13
27
0
08 Dec 2021
Learning to Detect Every Thing in an Open World
Learning to Detect Every Thing in an Open World
Kuniaki Saito
Ping Hu
Trevor Darrell
Kate Saenko
ObjD
VLM
19
45
0
03 Dec 2021
DenseCLIP: Language-Guided Dense Prediction with Context-Aware Prompting
DenseCLIP: Language-Guided Dense Prediction with Context-Aware Prompting
Yongming Rao
Wenliang Zhao
Guangyi Chen
Yansong Tang
Zheng Zhu
Guan Huang
Jie Zhou
Jiwen Lu
VLM
CLIP
94
551
0
02 Dec 2021
ML-Decoder: Scalable and Versatile Classification Head
ML-Decoder: Scalable and Versatile Classification Head
T. Ridnik
Gilad Sharir
Avi Ben-Cohen
Emanuel Ben-Baruch
Asaf Noy
VLM
32
102
0
25 Nov 2021
Multi-label Classification with Partial Annotations using Class-aware
  Selective Loss
Multi-label Classification with Partial Annotations using Class-aware Selective Loss
Emanuel Ben-Baruch
T. Ridnik
Itamar Friedman
Avi Ben-Cohen
Nadav Zamir
Asaf Noy
Lihi Zelnik-Manor
32
38
0
21 Oct 2021
Inferring Offensiveness In Images From Natural Language Supervision
Inferring Offensiveness In Images From Natural Language Supervision
P. Schramowski
Kristian Kersting
32
2
0
08 Oct 2021
LIGAR: Lightweight General-purpose Action Recognition
LIGAR: Lightweight General-purpose Action Recognition
Evgeny Izutov
15
3
0
30 Aug 2021
WikiChurches: A Fine-Grained Dataset of Architectural Styles with
  Real-World Challenges
WikiChurches: A Fine-Grained Dataset of Architectural Styles with Real-World Challenges
Björn Barz
Joachim Denzler
36
7
0
16 Aug 2021
Query2Label: A Simple Transformer Way to Multi-Label Classification
Query2Label: A Simple Transformer Way to Multi-Label Classification
Shilong Liu
Lei Zhang
Xiao Yang
Hang Su
Jun Zhu
24
187
0
22 Jul 2021
ResViT: Residual vision transformers for multi-modal medical image
  synthesis
ResViT: Residual vision transformers for multi-modal medical image synthesis
Onat Dalmaz
Mahmut Yurt
Tolga Çukur
ViT
MedIm
32
338
0
30 Jun 2021
Early Convolutions Help Transformers See Better
Early Convolutions Help Transformers See Better
Tete Xiao
Mannat Singh
Eric Mintun
Trevor Darrell
Piotr Dollár
Ross B. Girshick
20
753
0
28 Jun 2021
How to train your ViT? Data, Augmentation, and Regularization in Vision
  Transformers
How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers
Andreas Steiner
Alexander Kolesnikov
Xiaohua Zhai
Ross Wightman
Jakob Uszkoreit
Lucas Beyer
ViT
66
616
0
18 Jun 2021
Person Re-Identification with a Locally Aware Transformer
Person Re-Identification with a Locally Aware Transformer
Charu Sharma
S. R. Kapil
David Chapman
ViT
48
45
0
07 Jun 2021
MLP-Mixer: An all-MLP Architecture for Vision
MLP-Mixer: An all-MLP Architecture for Vision
Ilya O. Tolstikhin
N. Houlsby
Alexander Kolesnikov
Lucas Beyer
Xiaohua Zhai
...
Andreas Steiner
Daniel Keysers
Jakob Uszkoreit
Mario Lucic
Alexey Dosovitskiy
277
2,606
0
04 May 2021
Video Transformer Network
Video Transformer Network
Daniel Neimark
Omri Bar
Maya Zohar
Dotan Asselmann
ViT
204
422
0
01 Feb 2021
Re-labeling ImageNet: from Single to Multi-Labels, from Global to
  Localized Labels
Re-labeling ImageNet: from Single to Multi-Labels, from Global to Localized Labels
Sangdoo Yun
Seong Joon Oh
Byeongho Heo
Dongyoon Han
Junsuk Choe
Sanghyuk Chun
414
142
0
13 Jan 2021
Scaling Out-of-Distribution Detection for Real-World Settings
Scaling Out-of-Distribution Detection for Real-World Settings
Dan Hendrycks
Steven Basart
Mantas Mazeika
Andy Zou
Joe Kwon
Mohammadreza Mostajabi
Jacob Steinhardt
D. Song
OODD
17
455
0
25 Nov 2019
A disciplined approach to neural network hyper-parameters: Part 1 --
  learning rate, batch size, momentum, and weight decay
A disciplined approach to neural network hyper-parameters: Part 1 -- learning rate, batch size, momentum, and weight decay
L. Smith
208
1,020
0
26 Mar 2018
Previous
1234
Next