Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implementaion of transfomer into model #31

Open
kimbring2 opened this issue May 22, 2019 · 1 comment
Open

Implementaion of transfomer into model #31

kimbring2 opened this issue May 22, 2019 · 1 comment

Comments

@kimbring2
Copy link

kimbring2 commented May 22, 2019

Hello, thank you for sharing good code.

I am trying to solve a DefeatRoaches minigame by using a Relational Network.

I found a example code of Transformer for MNIST classification and modified a fully_conv.py file for it. Unlike original code, I only use a screen feature without a minimap feature. But, result is still not good.

Would you like give me recommendation how to modify it for reaching performance of DeepMind?

Thank you.
From Dohyeong

@inoryy
Copy link
Owner

inoryy commented May 22, 2019

Hello Dohyeong,

I assume you're trying to replicate the AlphaStar architecture?

First, note that spatial information is still processed by a normal conv model and transformer body is applied only to a flattened list of unit information (see architecture figure below).

Second, the transformer would require a full view of the game map as opposed to only the camera viewport we currently get from PySC2. There is some WIP patch to add support for it, but no ETA on when it will be released.

Third, from your results figures I see that mean rewards are stuck around 14 which as far as I remember is about the score you would get from random actions. I'm afraid this implies your model hasn't learned much beyond the initial jump, so improving from there might be quite difficult.

What I would recommend is to try to simplify the model as much as possible - reduce state / action spaces to the bare minimum, remove non-spatial and minimap info blocks. After verifying that it works try to add transformer body as a separate layer only on player_relative feature and merge it with the final spatial state block before the final policy / value layers.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants