Skip to content

jbakams/slimebot-volleyball

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

slimebot-volleyball:

Slimebot-volleyball is a 3D version of the slime volleyball game. The game is implemented on top of slimevolleygym and uses Webots as a simulator for visualization. The major difference with the slimevolleygym environment is the additional z-axis, which makes the training more challenging for RL agents such as PPO.


Evaluation of a team of 2 independent PPO agents trained with decentralized policies using Incremental RL and Self-Play Learning. The training script is given here.


Requirements

1. Webots R2021b:

Webots is open-source and multi-platform software used to simulate robots. It is friendly and doesn't have strong device requirements. Install it if you are interested in running simulations (Agents can be trained without a simulator). We recommand version R2021b to avoid bugs with recent versions.

2. Python 3.X:

Defaultly, Webots uses the operating system Python. However, we recommend using a virtual environment. We used Python 3.7 for its compatibility with Stablebaselines. If you prefer a Webots version different from R2021b, please refer to the Webots User Guide to set up Python. Otherwise, use the following steps to easily set up Python in Webots R2021b.

3. Gym

Slimebot Volleyball is a gym-like environment. So one needs to install a gym

The Environment

Although Webots has a built-in physics system, we used the same physics as with the slimevolleygym game. This allows running the environment in a CMD without Webots. The training side is yellow and the blue side is the opponent.

The environment can be dynamic along the z-axis (the depth). It can take an initial value between 0 and 24 and can change during the training. See this post for more details. The objects in the 3D scene were created with Blender (for flexibility) and imported into Webots.

1. Observation space

Although agents have cameras showing their respective views of the environment, We trained agents using state observation. The pixels version is not yet set up. At each timestep, the agent reads the location and speed XYZ-coordinates of the ball and each agent present in the scene. Depending on the scenario, the observation_space is $(n+1)\times 6$, with $n$ number of agents in the game.

  • 1 vs 1: The input shape is 18
  • 2 vs 2: The input shape is 30

2. Action Space:

The are 6 basic actions represented with the following 1 hot encoding:

  • forward : [1,0,0,0,0]
  • backward : [0,1,0,0,0]
  • up : [0,0,1,0,0]
  • right : [0,0,0,1,0]
  • left : [0,0,0,0,1]
  • stay still : [0,0,0,0,0]

3 actions maximum can be combined as long as they don't contradict each other(combining left and right doesn't make sense). Which gives a total of 18 possible actions.

Scenarios

We have two scenarios at the moment:

The yellow player represents the agent we aim to train to defeat the blue agent. Cameras at each bottom corner show views for respective agents.

The yellow agents represent the team we aim to train to defeat the blue agents. Teammates can bump into each other. The training is more challenging compared to the 1 vs 1 case.

Training Overview

For now, we have been able to train a PPO agent to play the game using self-play and incremental learning. Depending on the initialization, the agent can start playing in the full 3D space and last the maximum episode length before 10 Million timesteps of training. see the script and this post for more details.

Though performing well, the pretrained agent didn't reach the same perfection in 3D as it did in the 2D version of the game. The actual champion is the best we got after trying different training settings and seedings. It can be replaced by any other model which beats it during evaluation.

Coming Next

Webots built-in sensors

The use of state observation shows weakness in the input dimension when we need to move an agent trained in a single environment to a team play environment (which will increase the observation size). This is why using sensors such as a Camera would be a better approach. The input dimension will never change in any environment in which the agent would be placed.

  • Camera: A camera placed on the bot can allow the agent to use pixels as an observation based on its personal view of the 3D environment.
  • Distance sensor: The built-in distance sensor may be used to generate observations as the agent can detect objects by distance.
  • Combine sensors: Refers to Human abilities to use different senses at the same time. Combining different sensors can bring better training or in the worse case brings confusion to the agent during training.

Citing the project

@misc{slimebot-volleyball,
  author = {Bakambana, Jeremie},
  title = {Slimebot Volleyball: A 3D multi-agents gym environment for the slime volleyball game},
  year = {2022},
  publisher = {GitHub},
  journal = {GitHub repository},
  howpublished = {\url{https://github.com/jbakams/slimebot-volleyball}},
}

Conclusion

The code might miss some elegance in its structure and syntax. Any contribution to improve the project is welcome.

We only used Incremental Learning to solve the problem in both cases. We are expecting to discover other novel methods to solve the game.

About

3D gym environments to train RL agents to play the Slime Volleyball game in 3 dimensions using Webots as simulator.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages