Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Increasing Memory Size Problem #20

Open
dmadhitama opened this issue Sep 29, 2020 · 1 comment
Open

Increasing Memory Size Problem #20

dmadhitama opened this issue Sep 29, 2020 · 1 comment

Comments

@dmadhitama
Copy link

dmadhitama commented Sep 29, 2020

Hello there,

I was using this repo, trying to resume training your pretrained model using my own dataset. I already prepared the dataset as the readme tutorial guided. The problem now is; I ran the train_rnn.py script. Yes, I successfully run and freeze at:

20-09-29 15:51:21 [train_rnn.py:66 - INFO ] Building the model of Dual-Path-RNN
20-09-29 15:51:21 [train_rnn.py:69 - INFO ] Building the optimizer of Dual-Path-RNN
20-09-29 15:51:21 [train_rnn.py:72 - INFO ] Building the dataloader of Dual-Path-RNN

It's been an hour and I saw the htop my RAM increasing above 200G. Is this normal or the script is running the dataset on the fly?

Addition: I am using your default .yml config, batch size also 1. Is adjusting larger batch size help this problem?
FYI, for each speaker my dataset size s 60-ish GB. So it means, total dataset for speaker 1 & 2, and mix are about 180 GB.

The tasks also increasing, is it normal?
Screenshot from 2020-09-29 16-56-20

@JusperLee
Copy link
Owner

If it is a large data set, I recommend you to use this script's Dataset. Because my script will read data into memory, which is not friendly for large data sets.
https://github.com/JusperLee/Conv-TasNet/blob/master/Conv-TasNet_lightning/Datasets.py

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants