Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The learning rate of the inverse KKT method in the code is inconsistent with that in the paper. #3

Open
beijiguang94 opened this issue Oct 19, 2023 · 2 comments

Comments

@beijiguang94
Copy link

hi @wanxinjin. I noticed that in the paper the learning rate for PDP, inverse KKT, and neural policy cloning methods in imitation learning was set to $\eta=10^{-4}$. But in scripts like "cartpole_inverseKKT.py", the parameter lr equals 1e-7. Why so?

@wanxinjin
Copy link
Owner

Hi @beijiguang94, thanks for your interest.
I highly likely have changed the codes afterward. Please use a learning rate that is stable.

@beijiguang94
Copy link
Author

thanks. When useing different learning rates, I suppose it would be better to compare the imitation loss of these methods by changing the X label from 'iterations' to 'consumed time'.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants