标签:.com img nts image dom put 分享 who lap
This gives us the final deep Q-learning algorithm with experience replay:
There are many more tricks that DeepMind used to actually make it work – like target network, error clipping, reward clipping etc, but these are out of scope for this introduction.
The most amazing part of this algorithm is that it learns anything at all. Just think about it – because our Q-function is initialized randomly, it initially outputs complete garbage. And we are using this garbage (the maximum Q-value of the next state) as targets for the network, only occasionally folding in a tiny reward. That sounds insane, how could it learn anything meaningful at all? The fact is, that it does.
Learning Notes: Morvan - Reinforcement Learning, Part 4: Deep Q Network
标签:.com img nts image dom put 分享 who lap
原文地址:http://www.cnblogs.com/casperwin/p/6295404.html