万本电子书0元读

万本电子书0元读

顶部广告

TensorFlow Reinforcement Learning Quick Start Guide电子书

售       价:¥

1人正在读 | 0人评论 9.8

作       者:Kaushik Balakrishnan

出  版  社:Packt Publishing

出版时间:2019-03-30

字       数:21.5万

所属分类: 进口书 > 外文原版书 > 电脑/网络

温馨提示:数字商品不支持退换货,不提供源文件,不支持导出打印

为你推荐

  • 读书简介
  • 目录
  • 累计评论(0条)
  • 读书简介
  • 目录
  • 累计评论(0条)
Leverage the power of Tensorflow to Create powerful software agents that can self-learn to perform real-world tasks Key Features * Explore efficient Reinforcement Learning algorithms and code them using TensorFlow and Python * Train Reinforcement Learning agents for problems, ranging from computer games to autonomous driving. * Formulate and devise selective algorithms and techniques in your applications in no time. Book Description Advances in reinforcement learning algorithms have made it possible to use them for optimal control in several different industrial applications. With this book, you will apply Reinforcement Learning to a range of problems, from computer games to autonomous driving. The book starts by introducing you to essential Reinforcement Learning concepts such as agents, environments, rewards, and advantage functions. You will also master the distinctions between on-policy and off-policy algorithms, as well as model-free and model-based algorithms. You will also learn about several Reinforcement Learning algorithms, such as SARSA, Deep Q-Networks (DQN), Deep Deterministic Policy Gradients (DDPG), Asynchronous Advantage Actor-Critic (A3C), Trust Region Policy Optimization (TRPO), and Proximal Policy Optimization (PPO). The book will also show you how to code these algorithms in TensorFlow and Python and apply them to solve computer games from OpenAI Gym. Finally, you will also learn how to train a car to drive autonomously in the Torcs racing car simulator. By the end of the book, you will be able to design, build, train, and evaluate feed-forward neural networks and convolutional neural networks. You will also have mastered coding state-of-the-art algorithms and also training agents for various control problems. What you will learn * Understand the theory and concepts behind modern Reinforcement Learning algorithms * Code state-of-the-art Reinforcement Learning algorithms with discrete or continuous actions * Develop Reinforcement Learning algorithms and apply them to training agents to play computer games * Explore DQN, DDQN, and Dueling architectures to play Atari's Breakout using TensorFlow * Use A3C to play CartPole and LunarLander * Train an agent to drive a car autonomously in a simulator Who this book is for Data scientists and AI developers who wish to quickly get started with training effective reinforcement learning models in TensorFlow will find this book very useful. Prior knowledge of machine learning and deep learning concepts (as well as exposure to Python programming) will be useful.
目录展开

Dedication

About Packt

Why subscribe?

Packt.com

Contributors

About the author

About the reviewer

Packt is searching for authors like you

Preface

Who this book is for

What this book covers

To get the most out of this book

Download the example code files

Download the color images

Conventions used

Get in touch

Reviews

Up and Running with Reinforcement Learning

Why RL?

Formulating the RL problem

The relationship between an agent and its environment

Defining the states of the agent

Defining the actions of the agent

Understanding policy, value, and advantage functions

Identifying episodes

Identifying reward functions and the concept of discounted rewards

Rewards

Learning the Markov decision process

Defining the Bellman equation

On-policy versus off-policy learning

On-policy method

Off-policy method

Model-free and model-based training

Algorithms covered in this book

Summary

Questions

Further reading

Temporal Difference, SARSA, and Q-Learning

Technical requirements

Understanding TD learning

Relation between the value functions and state

Understanding SARSA and Q-Learning

Learning SARSA

Understanding Q-learning

Cliff walking and grid world problems

Cliff walking with SARSA

Cliff walking with Q-learning

Grid world with SARSA

Summary

Further reading

Deep Q-Network

Technical requirements

Learning the theory behind a DQN

Understanding target networks

Learning about replay buffer

Getting introduced to the Atari environment

Summary of Atari games

Pong

Breakout

Space Invaders

LunarLander

The Arcade Learning Environment

Coding a DQN in TensorFlow

Using the model.py file

Using the funcs.py file

Using the dqn.py file

Evaluating the performance of the DQN on Atari Breakout

Summary

Questions

Further reading

Double DQN, Dueling Architectures, and Rainbow

Technical requirements

Understanding Double DQN

Updating the Bellman equation

Coding DDQN and training to play Atari Breakout

Evaluating the performance of DDQN on Atari Breakout

Understanding dueling network architectures

Coding dueling network architecture and training it to play Atari Breakout

Combining V and A to obtain Q

Evaluating the performance of dueling architectures on Atari Breakout

Understanding Rainbow networks

DQN improvements

Prioritized experience replay

Multi-step learning

Distributional RL

Noisy nets

Running a Rainbow network on Dopamine

Rainbow using Dopamine

Summary

Questions

Further reading

Deep Deterministic Policy Gradient

Technical requirements

Actor-Critic algorithms and policy gradients

Policy gradient

Deep Deterministic Policy Gradient

Coding ddpg.py

Coding AandC.py

Coding TrainOrTest.py

Coding replay_buffer.py

Training and testing the DDPG on Pendulum-v0

Summary

Questions

Further reading

Asynchronous Methods - A3C and A2C

Technical requirements

The A3C algorithm

Loss functions

CartPole and LunarLander

CartPole

LunarLander

The A3C algorithm applied to CartPole

Coding cartpole.py

Coding a3c.py

The AC class

The Worker() class

Coding utils.py

Training on CartPole

The A3C algorithm applied to LunarLander

Coding lunar.py

Training on LunarLander

The A2C algorithm

Summary

Questions

Further reading

Trust Region Policy Optimization and Proximal Policy Optimization

Technical requirements

Learning TRPO

TRPO equations

Learning PPO

PPO loss functions

Using PPO to solve the MountainCar problem

Coding the class_ppo.py file

Coding train_test.py file

Evaluating the performance

Full throttle

Random throttle

Summary

Questions

Further reading

Deep RL Applied to Autonomous Driving

Technical requirements

Car driving simulators

Learning to use TORCS

State space

Support files

Training a DDPG agent to learn to drive

Coding ddpg.py

Coding AandC.py

Coding TrainOrTest.py

Training a PPO agent

Summary

Questions

Further reading

Assessment

Chapter 1

Chapter 3

Chapter 4

Chapter 5

Chapter 6

Chapter 7

Chapter 8

Other Books You May Enjoy

Leave a review - let other readers know what you think

累计评论(0条) 0个书友正在讨论这本书 发表评论

发表评论

发表评论,分享你的想法吧!

买过这本书的人还买过

读了这本书的人还在读

回顶部