Skip to content

[FEATURE] ImageNet1k weights for ViT Huge? #2163

Discussion options

You must be logged in to vote

There are some ImageNet-22k H/14 weights without a valid head from the original google ViT weights. Not very good though.

Best non CLIP option on ImageNet-22k/1-k is the Deit-III H/14. Although it was pretty clearly running out of steam vs the L. I don't feel ImageNet-22k is enough to train H optimally so best to stick with big dataset pretrain w/ image-text or semi/self-supervised learning.

The SO400M ViT would probably be a better bet but don't think anyone has done a supervised ImageNet-22k - 1k fine-tune with that...

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by NightMachinery
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
2 participants
Converted from issue

This discussion was converted from issue #2161 on April 30, 2024 00:35.