[ExI] Breakout success for reinforcement learning

Alejandro Dubrovsky alito at organicrobot.com
Sat Feb 28 13:08:07 UTC 2015


On 28/02/15 22:35, Anders Sandberg wrote:
> Google DeepMind's paper on reinforcement learning playing Atari games is
> now out in Nature: Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A.,
> Antonoglou, I., Wierstra, D., & Riedmiller, M. (2013). Playing atari
> with deep reinforcement learning. arXiv preprint arXiv:1312.5602.
>
> http://googleresearch.blogspot.co.uk/2015/02/from-pixels-to-actions-human-level.html
> http://www.nature.com/nature/journal/v518/n7540/full/nature14236.html
> https://www.youtube.com/watch?v=iqXKQf2BOSE
> http://arxiv.org/abs/1312.5602
>

Yep, DeepMind are killing it at the moment. They've also released their 
source code with their Nature paper. You can get it from here: 
https://sites.google.com/a/deepmind.com/dqn/

There's also code independently developed by Nathan Sprague based on the 
2013 version of the paper here: https://github.com/spragunr/deep_q_rl. 
The 2013 version used a narrower and shallower network, and it also 
didn't use their new two-network system that they adapted in the Nature 
version to avoid oscillatory behaviour. At the moment the only advantage 
of that code would be the more liberal licencing since DeepMind's code 
can only be used for research purposes. It shouldn't take long for it to 
catch up with the newer version, and I suspect it will become the go-to 
version because of the licencing advantage.

BTW, if you are going to play with the code, make sure to get the 
beastliest Nvidia GPU that you can afford.



More information about the extropy-chat mailing list