Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

vision-transformer problem report #1184

Open
ChenDaiwei-99 opened this issue Aug 24, 2023 · 4 comments
Open

vision-transformer problem report #1184

ChenDaiwei-99 opened this issue Aug 24, 2023 · 4 comments

Comments

@ChenDaiwei-99
Copy link

image
image
image

@msaroufim
Copy link
Member

If you wanna contribute some fixes would be happy to merge

@ChenDaiwei-99
Copy link
Author

Sure, will do it in a few days :)

@colinosterman
Copy link

I can confirm this example is badly broken. I added some code to compare individual labels to predictions and discovered the forward pass of ViT always returns the same tensor. No matter the input. The tensor it returns is different each time I run it, even if I load the same weights from the save file and don't do any training. It's no wonder it can't do better than 2.3. Always giving the same prediction should accidentally hit on the correct label about as often as random guessing.

@fanqieguo
Copy link

I added the printout of accuracy in the code, but during the training process, the accuracy does not improve, and the loss does not converge, even after training for many epochs. Does this model really work? I think this model has serious problems. I hope to get an answer, this is very important to me

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants