Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

PyTorchRL: Modular and scalable reinforcement learning in pytorch reaching new high in obstacle tower challenge #124

Open
giadefa opened this issue Jul 29, 2020 · 3 comments
Assignees

Comments

@giadefa
Copy link

giadefa commented Jul 29, 2020

Hi,
don't know if it is the right place, but it is probably relevant for anyone interested in obstacle-tower-env.

https://arxiv.org/abs/2007.02622

We have scaled up our second position approach in the official challenge. We now reach a max of 23.6 floors on the test seeds and consistently above 20. Good news is that there is more work to do, as it does seem to plateau with the current method. The entire code and trained policies will be available soon.

[If there is a better place to put this let me and delete it]

@MarcoMeter
Copy link
Contributor

Hi!

I could imagine a section on the readme.md to list such publications.

@awjuliani awjuliani self-assigned this Jul 29, 2020
@awjuliani
Copy link
Contributor

Hi @giadefa

This is great work to see, and I have no problem keeping this issue open so it can be discovered. The work is relevant to helping others who might want to get started doing research with Obstacle Tower.

@giadefa giadefa changed the title NAPPO: Modular and scalable reinforcement learning in pytorch reaching new high in obstacle tower challenge PyTorchRL: Modular and scalable reinforcement learning in pytorch reaching new high in obstacle tower challenge Feb 11, 2021
@giadefa
Copy link
Author

giadefa commented Feb 11, 2021

HI,
We have just release the final version of PyTorchRL:
https://github.com/PyTorchRL/pytorchrl
Main points are:

  • It is scalable on multi GPUs and multi nodes and
  • It provides a modular approach which helps in developping multiple algorithms, both on-policy and off-policy, We provide several.

and the final paper:
https://arxiv.org/abs/2007.02622

The docs:
https://pytorchrl.readthedocs.io/en/latest/

I think that this is relevant for people who would like to try multiple algorithms or scale it up. We would be happy to work with people who are keen on this.

Best,
Gianni

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants