Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Saved the trained agent and hold the thetas unchanged for simulation on new dataset #18

Open
hanchenresearch opened this issue Mar 22, 2020 · 0 comments

Comments

@hanchenresearch
Copy link

Dear Robin,

This is not a bug report but more like a new feature request.

We know that the theta is updated after the agents' every interaction with the bandit. What I want to ask is that is it possible to save the "trained" agent with the theta for later use on another dataset. The logic behind this is that the trained agent acts as an oracle/ground truth of the environment, then I want to add a benchmark full information model based on this oracle.In this way, I can look at what is the maximum reward I can theoretically get if I initiate my offline evaluation with this oracle, without knowing the ground truth until the ends of my simulation.

Basically, to achieve this goal, I need to save the trained agents with the thetas, and break the thata updating chain and hold the thetas unchanged when used for another dataset.

Thank you so much for your help!

Best,
Han

@hanchenresearch hanchenresearch changed the title Saved the trained agent and break the updating when applying to new data Saved the trained agent and hold the thetas unchanged when applying to new dataset Mar 22, 2020
@hanchenresearch hanchenresearch changed the title Saved the trained agent and hold the thetas unchanged when applying to new dataset Saved the trained agent and hold the thetas unchanged for simulation on new dataset Mar 22, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants