Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to finetuning with lower memory fp16 version for p100 GPUs? #39

Open
xurongqiang opened this issue Oct 9, 2019 · 5 comments
Open

Comments

@xurongqiang
Copy link

xurongqiang commented Oct 9, 2019

For finetuning with lower memory fp16 version(for fp32 version , OOMs occur. ), How should I modify the training.py script?

@xurongqiang xurongqiang changed the title How to finetuning with lower memory fp16 version ? How to finetuning with lower memory fp16 version for p100 GPUs? Oct 10, 2019
@keskarnitish
Copy link
Contributor

This is quite delicate and doesn't quite seem to work out of the box. I'm going to need more time to look into this.

@xurongqiang
Copy link
Author

This is quite delicate and doesn't quite seem to work out of the box. I'm going to need more time to look into this.

The background to this problem is that we have a large number of p100 machines, but they cannot run on the fp32 version. Thank you for your improvement.

@v1nc3nt27
Copy link

Is there some update on this matter? We have the same problem, unfortunately.

@itsuncheng
Copy link

itsuncheng commented Feb 11, 2020

Yes same here for us. Both huggingface and this repo seem to have the same OOM error when running on Google Colab free GPU like p100. Any fix or workaround yet?

@nastyatrvl
Copy link

The problem still persists, unfortunately. Fine-tuning doesn't really work with collab resources..

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants