Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Conditions to stop training when target return is reached #331

Open
excellentiam-prog opened this issue Sep 12, 2023 · 0 comments
Open

Conditions to stop training when target return is reached #331

excellentiam-prog opened this issue Sep 12, 2023 · 0 comments

Comments

@excellentiam-prog
Copy link

Hello, I am testing elegantRL for different environments, but I could not find any examples of completed training. I have been running the LunarLander example for a few days, using the original parameters (attached). How can I tell when the model is fully trained? Is there a minimum number of episodes that need to exceed the target return? If so, where can I specify that? Or is the training indefinite and I have to stop it manually?

While the original example is still running, I tried changing some parameters to see if the training would finish, or if there were any statistics besides the ones shown in the log. I set the target return to 2 and the eval times also 2, and the model surpassed the target return, but it kept training even after several episodes with avgR above 2.

Could you please explain how the stop condition works and what are the best practices for evaluating the model performance? Thank you very much.

LearningCurve

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant