Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is this support for hyperparameters tuning? #5

Open
zy455761544 opened this issue Aug 2, 2020 · 3 comments
Open

Is this support for hyperparameters tuning? #5

zy455761544 opened this issue Aug 2, 2020 · 3 comments
Assignees

Comments

@zy455761544
Copy link

I am looking for a multi-task learning model and found this repo is a good start. Thanks for your contribution! I am wondering if there is any easy way to implement hyperparameters tuning in your codes?

@saransh-mehta
Copy link
Contributor

saransh-mehta commented Aug 3, 2020

Hi @zy455761544
You can provide values for following hyper-parameters as arguments to the training script - train.py

epochs
train_batch_size
eval_batch_size
learning_rate
epsilon
grad_accumulation_step
num_of_warmup_steps
grad_clip_value

To know about all the arguments present you can try $ train.py -h from your command line.
In case, you're looking for any specific hyper-parameters, you can let me know.
Thanks!

@saransh-mehta saransh-mehta self-assigned this Aug 3, 2020
@zy455761544
Copy link
Author

Thank you. I plan to experiment on auto hyper-parameters tuning, using library such as gridserachCV (sklearn) instead of changing those hyper-parameters manually. Maybe I need to modify partial codes inside the train.py

@saransh-mehta
Copy link
Contributor

For auto hyper-parameters tuning, you can find the optimizer/scheduler/ related codes in models/model.py under multiTaskModel class. Libraries like Optuna might also be useful for hyper-parameters tuning.
If you get good results with adding auto-tuning into the codes and wish to contribute, you can raise a pull request and we can add it as a feature too.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants