Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How is Elo calculated with agent self-play? #270

Open
dbsxdbsx opened this issue May 9, 2022 · 0 comments
Open

How is Elo calculated with agent self-play? #270

dbsxdbsx opened this issue May 9, 2022 · 0 comments

Comments

@dbsxdbsx
Copy link

dbsxdbsx commented May 9, 2022

I came up this question when reading paper of Alpha(Go)Zero, and I know formula to calculate Elo.

But when it comes to self-play, for example, player A and B (which is the same agent) starts with score 1400 for game chess. and after 1 round, A score is 1500 and B score is 1300.

Then... no matter whether the agent is trained or not, it starts to beat itself again, but what is the basic score used to calculate Elo for agent after 2nd round? The 1500, 1300 or some other way? I didn't come up a right way to do so.

In a word, I don't know how data of Elo trend graph is generated with trained steps as x axis and Elo as yaxis, when game is a 2 side zero-sum game and trained with only self-play.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant