Machine Learning with Phil dives into Deep Q Learning with Tensorflow 2 and Keras.

Dueling Deep Q Learning is easier than ever with Tensorflow 2 and Keras. In this tutorial for deep reinforcement learning beginners we’ll code up the dueling deep q network and agent from scratch, with no prior experience needed. We’ll train an agent to land a spacecraft on the surface of the moon, using the lunar lander environment from the OpenAI Gym.

The dueling network can be applied to both regular and double q learning, as it’s just a new network architecture. It doesn’t require any change to the q learning or double q learning algorithms. We simply have to change up our feed forward to accommodate the new value and advantage streams, and combine them in a way that makes sense.

Machine Learning with Phil has a great tutorial on how to do Deep Q Learning in PyTorch.

The PyTorch deep learning framework makes coding a deep q learning agent in python easier than ever. We’re going to code up the simplest possible deep Q learning agent, and show that we only need a replay memory to get some serious results in the Lunar Lander environment from the Open AI Gym. We don’t really need the target network, though it has been known to help the deep Q learning algorithm with convergence.

Machine Learning with Phil has got another interesting look at Deep Q Learning as part of a preview of his course.

The two biggest innovations in deep Q learning were the introduction of the target network and the replay memory. One would think that simply bolting a deep neural network to the Q learning algorithm would be enough for a robust deep Q learning agent, but that isn’t the case. In this video I’ll show you how this naive implementation of the deep q learning agent fails, and spectacularly at that.

This is an excerpt from my new course, Deep Q Learning From Paper to Code which you can get on sale with this link

https://www.udemy.com/course/deep-q-learning-from-paper-to-code/?couponCode=CYBERMONDAY19