muupan / dqn-in-the-caffe

An implementation of Deep Q-Network using Caffe

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Parameters used to learn Pong?

mhauskn opened this issue · comments

What parameters were used to learn the pong player showed in the video (https://www.youtube.com/watch?v=p88R2_3yWPA)? Specifically what is the gamma/memory size/iterations needed to train dqn? I attempted to retrain the agent using the default parameters but it shows no progress after 2-million iterations.

Thanks in advance, and very cool repo!

Thanks for forking my repo! Default parameters in dqn_solver.prototxt was outdated, so I've just updated it and now it's the one used in the Pong demo. The size of replay memory was 500,000.

Thanks, it seems to be working much better now :)