-
Notifications
You must be signed in to change notification settings - Fork 147
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
problem about the result #29
Comments
Network: Inception V2 All these together is enough to make good performance with contrasstive loss. |
Thanks very much ! |
Wow, i used the contrastive loss in Cub200 and then got rank1 0.68(but highest baseline in your chart is 0.66). When you train the model , have you use the data augmentation 'rand_crop' which is in your codes? |
yes |
Found in my experiment, why BNInception in your code If there is no freeze BN-layer's result will be about two points lower than after freeze. Why is there no such phenomenon in the InceptionBN network in other people's papers? Looking forward to your reply. |
Hello, thanks for your good jobs!
When I use your codes, i find almost all types of loss can reach the result that rank1 is over 0.60. But when i read the paper, if the method is over 0.60(rank1), it can be seen as a good method. Even the contrastive loss can achieve rank1 over 0.60. Why? Thanks
The text was updated successfully, but these errors were encountered: