Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About average_checkpoints in neural_sp/bin/eval_utils.py. #218

Open
lfgogogo opened this issue Dec 22, 2020 · 3 comments
Open

About average_checkpoints in neural_sp/bin/eval_utils.py. #218

lfgogogo opened this issue Dec 22, 2020 · 3 comments
Labels

Comments

@lfgogogo
Copy link

I have reproduced the result of aishell,but when i check the weight files,i find model-avg10 is about one fifth size of the others.
I only see computing average in the average_checkpoints,how can this process decreases the size of the model,just like quantization.
I also use the distiller to make comparison,the model-avg10 is almost the same size as quantization model.

@hirofumi0810
Copy link
Owner

@lfgogogo That is because parameters in the optimizer were removed after model averaging.

@lfgogogo
Copy link
Author

@hirofumi0810 Why do model average computing?For speeding up?But i test both,it doesn't seem to improve much,on my test dataset which consists of 30 thousands speeches,the original model.epoch-25 cost 15+ hours,while the averaged model cost 12 huors.Do you have some suggestions for speeding up the process,hiro?Thank you very much.

@hirofumi0810
Copy link
Owner

@lfgogogo If you have a lot of training data, checkpoint averaging might be ineffective. Please try to change n_average in score.sh.
But averaging is not related to speed performance at all.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants