Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Memory requirements for a new model #55

Open
DanrunFR opened this issue Feb 6, 2023 · 3 comments
Open

Memory requirements for a new model #55

DanrunFR opened this issue Feb 6, 2023 · 3 comments

Comments

@DanrunFR
Copy link

DanrunFR commented Feb 6, 2023

Hello,
I am trying to train a new ACE model on a 3.5m word corpus. I want to use the same embedding configuration of the CoNLL 2003 model. I have tried a few times, but I kept running into memory problems. I had about 250G CPU memory but it didn't even get through 3 embedding models. I was wondering how much memory would you recommend using to train my model ?

@DanrunFR
Copy link
Author

DanrunFR commented Feb 6, 2023

And also how much time would the training take ? I'm asking because I'm using our lab's computation facility, and I need to reserve the resources I need.

Thank you in advance for your response.

@wangxinyu0922
Copy link
Member

Hi, to run conll 2003 dataset, it takes about 40~60G CPU memory. It seems that a 3.5m word corpus is about 10 times larger than the conll 2003 dataset. So it takes about 400G cpu memory to run the code. I suggest you may remove the flair embeddings (especially the multilingual embeddings) to reduce the feature size.

@wangxinyu0922
Copy link
Member

For training, it takes about 2~4 days to train conll 2003 dataset on a single Tesla V100 GPU. You may reduce the training epoch to reduce the training time.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants