Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Do different pre-trained models have a big impact #38

Open
marongxing opened this issue Aug 7, 2019 · 4 comments
Open

Do different pre-trained models have a big impact #38

marongxing opened this issue Aug 7, 2019 · 4 comments

Comments

@marongxing
Copy link

I tried to replace the BN-inception with resnet50 pretrained on imagenet.
it seems that the performence droped a lot.
With your code, I can reproduce the result on cub-200, about 0.65.
And I inplemented a gluon(mxnet) version of the binomial loss.
But i can't find the pretrained model on mxnet.
Therefor, I tried different pretrained models. And the performences varies from 0.50 - 0.60 on cub-200.
Thus, I wonder know if different models have a big impact on the permformece?
Thank you!

@bnu-wangxun
Copy link
Owner

Different network need different hyperparamter: we also use resnet-50 and the performance is comparable with BN-inception on cars-196.

@marongxing
Copy link
Author

Maybe I have found the reason.
the BN-inception you used is different from what I used.
params of your BN-inception is about 22M.(pytorch)
while I used is about 11M. best performence: 60%.(gluon)
I just want to implement it on gluon(mxnet). But I cannot find a right pretrained model.
Thank you!

@ZhangHZ9
Copy link
Collaborator

Cub-200 is very small and the results may be affected by many details. I recommend you compare the performance with and without XBM under the same conditions except for the learning rate. Experiments on SOP or In-shop will be more stable.

@fengshi-cherish
Copy link

Is there any suggestion about the hyperparamters in training resnet50. When i used ms-loss with resnet50 in CUB-200, the performance was dropped a lot, about 0.52 in rank1.I'm not sure what make such difference, hyperparameters or my code?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants