Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training loss doesn't decrease #2

Open
Uirseita opened this issue Apr 6, 2019 · 5 comments
Open

Training loss doesn't decrease #2

Uirseita opened this issue Apr 6, 2019 · 5 comments

Comments

@Uirseita
Copy link

Uirseita commented Apr 6, 2019

Hi, I tried to run your notebooks. For notebook 2, my training loss is still around 0.19 after 200 epochs so the resulting model is still pretty bad. The only thing I changed is "use_multiprocessing" to False because it keeps getting an error when it's True.

Do you know how to fix this issue?

@liuxuankai
Copy link

me too ,tho loss doesn't decrease!!

@Golbstein
Copy link
Owner

Use notebook 2.1
The reason it doesn't work for notebook 1 and 2 is because we train it on a very small model compared to what we really want to train on.
It's just an example of how to fine-tune your pre-trained model on your internal dataset

@Golbstein
Copy link
Owner

I've made few changes now. Check it out

@Uirseita
Copy link
Author

I've made few changes now. Check it out

Thanks for the response. I was able to run notebook 2.1 on my own pc with GTX1080Ti, but it always crashes because out of graphics memory on the computation center of my university which has Tesla V100 32GB. Do you know the reason for this by any chance?

@Golbstein
Copy link
Owner

How can I know if you didn't provide any details, not even the error message...
Usually it's due to the fact that your GPU is already "full" or you didn't reset it. Try to reboot the server..

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants