Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Observationspace/Infostate #16

Open
ILikePoker opened this issue Nov 26, 2020 · 1 comment
Open

Observationspace/Infostate #16

ILikePoker opened this issue Nov 26, 2020 · 1 comment

Comments

@ILikePoker
Copy link

Hey Eric,

thanks for making this public, haven't found a good env so far that implements Multiplayer NL. Am I understanding the code right that the observationspace isn't actually perfect information? E.g. it is only the last couple actions? Do you have any research on how this affects convergence?
I had a bit of trouble understanding the code, so I apologize if I just didn't read it right.

@EricSteinberger
Copy link
Owner

Heyo,

You're welcome! I've written a wrapper that tracks the action history for limit games. For no-limit games, this piece is slightly less obvious, so I would default to the recurrent option that tracks the history of public observations, which is also sound. So, perfect recall is still supported for NL through recurrent NNs but you did correctly spot that action history is only tracked through that and not explicitly.

If you wish to track it explicitly, you could just track it manually in the training code or write an appropriate "Wrapper" for the NL environment. However, this will require you to adjust the NN observation space accordingly. TL;DR: Recurrent is cleaner and more scalable.

Cheers,
Eric

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants