Search icon CANCEL
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Hands-On Q-Learning with Python

You're reading from   Hands-On Q-Learning with Python Practical Q-learning with OpenAI Gym, Keras, and TensorFlow

Arrow left icon
Product type Paperback
Published in Apr 2019
Publisher Packt
ISBN-13 9781789345803
Length 212 pages
Edition 1st Edition
Languages
Tools
Arrow right icon
Author (1):
Arrow left icon
Nazia Habib Nazia Habib
Author Profile Icon Nazia Habib
Nazia Habib
Arrow right icon
View More author details
Toc

Table of Contents (14) Chapters Close

Preface 1. Section 1: Q-Learning: A Roadmap FREE CHAPTER
2. Brushing Up on Reinforcement Learning Concepts 3. Getting Started with the Q-Learning Algorithm 4. Setting Up Your First Environment with OpenAI Gym 5. Teaching a Smartcab to Drive Using Q-Learning 6. Section 2: Building and Optimizing Q-Learning Agents
7. Building Q-Networks with TensorFlow 8. Digging Deeper into Deep Q-Networks with Keras and TensorFlow 9. Section 3: Advanced Q-Learning Challenges with Keras, TensorFlow, and OpenAI Gym
10. Decoupling Exploration and Exploitation in Multi-Armed Bandits 11. Further Q-Learning Research and Future Projects 12. Assessments 13. Other Books You May Enjoy

Questions

  1. Why do we choose to use the words state and observation interchangeably? When would be a more appropriate time to use the word state?
  2. How do we know when the Q-function has converged?
  3. What happens to the Q-table when the Q-function has converged?
  4. When do we know the agent has found the optimal path to the goal? Describe in terms of the previous two questions.
  5. What does numpy.argmax() return?
  6. What does numpy.max() return?
  7. Why does the randomly-acting agent take thousands of time steps to reach the goal? How does the Q-learning agent perform better?
  8. Describe one benefit of decaying alpha.
  9. What is overfitting and how does it apply in the context of an RL model?
  10. By what order of magnitude does the number of time steps needed to reach the goal reduce when the number of training episodes is multiplied by 10? Give a general response to this; there may be multiple valid...
lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime