Sign In Start Free Trial
Account

Add to playlist

Create a Playlist

Modal Close icon
You need to login to use this feature.
  • Book Overview & Buying Hands-On Q-Learning with Python
  • Table Of Contents Toc
  • Feedback & Rating feedback
Hands-On Q-Learning with Python

Hands-On Q-Learning with Python

By : Nazia Habib
2.3 (3)
close
close
Hands-On Q-Learning with Python

Hands-On Q-Learning with Python

2.3 (3)
By: Nazia Habib

Overview of this book

Q-learning is a machine learning algorithm used to solve optimization problems in artificial intelligence (AI). It is one of the most popular fields of study among AI researchers. This book starts off by introducing you to reinforcement learning and Q-learning, in addition to helping you become familiar with OpenAI Gym as well as libraries such as Keras and TensorFlow. A few chapters into the book, you will gain insights into model-free Q-learning and use deep Q-networks and double deep Q-networks to solve complex problems. This book will guide you in exploring use cases such as self-driving vehicles and OpenAI Gym’s CartPole problem. You will also learn how to tune and optimize Q-networks and their hyperparameters. As you progress, you will understand the reinforcement learning approach to solving real-world problems. You will also explore how to use Q-learning and related algorithms in scientific research. Toward the end, you’ll gain insight into what’s in store for reinforcement learning. By the end of this book, you will be equipped with the skills you need to solve reinforcement learning problems using Q-learning algorithms with OpenAI Gym, Keras, and TensorFlow.
Table of Contents (14 chapters)
close
close
Free Chapter
1
Section 1: Q-Learning: A Roadmap
6
Section 2: Building and Optimizing Q-Learning Agents
9
Section 3: Advanced Q-Learning Challenges with Keras, TensorFlow, and OpenAI Gym

Chapter 2, Getting Started with the Q-Learning Algorithm

  1. Generally speaking, a control process is designed to optimize a value or a set of values within a set of limitations.
  2. A Markov chain does not incorporate actions or rewards; it only has states and events that will lead from one state to the next.
  3. The Markov property is the certainty that knowledge of a system's future states does not depend on knowledge of past states, but only on the current state.
  4. The Taxi-v2 environment has 500 states based on the values the state variables can take. State variables are the location of the taxi, the location of the destination, and the location of the passenger.
  5. We include these states for simplicity in enumerating the state space. They are unreachable in the task because, in some cases, the environment reaches a terminal state before they can be reached. For example, when the taxi...

Unlock full access

Continue reading for free

A Packt free trial gives you instant online access to our library of over 7000 practical eBooks and videos, constantly updated with the latest in tech

Create a Note

Modal Close icon
You need to login to use this feature.
notes
bookmark search playlist download font-size

Change the font size

margin-width

Change margin width

day-mode

Change background colour

Close icon Search
Country selected

Close icon Your notes and bookmarks

Delete Bookmark

Modal Close icon
Are you sure you want to delete it?
Cancel
Yes, Delete

Delete Note

Modal Close icon
Are you sure you want to delete it?
Cancel
Yes, Delete

Edit Note

Modal Close icon
Write a note (max 255 characters)
Cancel
Update Note

Confirmation

Modal Close icon
claim successful

Buy this book with your credits?

Modal Close icon
Are you sure you want to buy this book with one of your credits?
Close
YES, BUY