Skip to content

Tensorflow implementation of asyncronous 1-step Q learning in "Asynchronous Methods for Deep Reinforcement Learning" with improvement on weight update process (use minibatch) to speed up training.

Notifications You must be signed in to change notification settings

snowkylin/async_rl

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Play Atari Games with TensorFlow and Asynchronous RL

This is a Tensorflow implementation of asyncronous 1-step Q learning with improvement on weight update process (use minibatch) to speed up training. Algorithm can be fount at Asynchronous Methods for Deep Reinforcement Learning

Demo

Play Flappy Bird with TensorFlow

Dependencies

  • Python
  • TensorFlow
  • gym (with atari environment)
  • OpenCV-Python

Usage

Run play.py to play atari game (default is Breakout-v0) by trained network.

Run train.py to train the network on your computer.

You will get a comparatively good result (40+ score) when t is larger than 2000000. On my computer (i5-4590/16GB/GTX 1060 6GB), the training process need at least 2-3 hours.

Evaluation

You can find the eval at https://gym.openai.com/evaluations/eval_03aUUz45Sc6TBg0vifljwA , which takes 40 hours to train the network.

Credit

About

Tensorflow implementation of asyncronous 1-step Q learning in "Asynchronous Methods for Deep Reinforcement Learning" with improvement on weight update process (use minibatch) to speed up training.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages