Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Multi-GPU training & batchsize problem #102

Open
X-Drunker opened this issue Nov 14, 2023 · 0 comments
Open

Multi-GPU training & batchsize problem #102

X-Drunker opened this issue Nov 14, 2023 · 0 comments

Comments

@X-Drunker
Copy link

X-Drunker commented Nov 14, 2023

Hi, I really appreciate your work and now I'm going to train the model on this pipeline. My issues are as follows:

  1. I note that you have adapted the code to multi-GPU versions with DDP, but I cant figure out how to train with multi-GPU. Maybe I should set self.use_ddp = True here ?
  2. In the paper you mentioned that you trained DiffSinger on 1 NVIDIA V100 GPU with 48 batch size. However, I can't find any customizable variable related to batch size. Is it necessary to set batch size to match the number of GPU, if I want to train with multi-GPU?
    Any suggestion is welcome.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant