ctguxp / Replicating-DeepMind

Reproducing the results of "Playing Atari with Deep Reinforcement Learning" by DeepMind

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Replicating-DeepMind

Reproducing the results of "Playing Atari with Deep Reinforcement Learning" by DeepMind. All the information is in our Wiki.

Progress: System is up and running on a GPU cluster with cuda-convnet2. It can learn to play better than random but not much better yet :) It is rather fast but still about 2x slower than DeepMind's original system. It does not have RMSprop implemented at the moment which is our next goal.

Note 1: You can also check out a popular science article we wrote about the system to Robohub.

Note 2: Nathan Sprague has a implementation based on Theano. It can do fairly well. See his github for more details.

About

Reproducing the results of "Playing Atari with Deep Reinforcement Learning" by DeepMind

License:GNU General Public License v3.0


Languages

Language:C++ 52.9%Language:Cuda 17.9%Language:Prolog 14.0%Language:Python 6.1%Language:TeX 5.0%Language:C 3.5%Language:Makefile 0.5%Language:Shell 0.1%Language:MATLAB 0.0%