万本电子书0元读

万本电子书0元读

顶部广告

Artificial Intelligence By Example电子书

售       价:¥

12人正在读 | 0人评论 6.2

作       者:Denis Rothman

出  版  社:Packt Publishing

出版时间:2018-05-31

字       数:56.9万

所属分类: 进口书 > 外文原版书 > 电脑/网络

温馨提示:数字商品不支持退换货,不提供源文件,不支持导出打印

为你推荐

  • 读书简介
  • 目录
  • 累计评论(0条)
  • 读书简介
  • 目录
  • 累计评论(0条)
Be an adaptive thinker that leads the way to Artificial Intelligence About This Book ? AI-based examples to guide you in designing and implementing machine intelligence ? Develop your own method for future AI solutions ? Acquire advanced AI, machine learning, and deep learning design skills Who This Book Is For Artificial Intelligence by Example is a simple, explanatory, and descriptive guide for junior developers, experienced developers, technology consultants, and those interested in AI who want to understand the fundamentals of Artificial Intelligence and implement it practically by devising smart solutions. Prior experience with Python and statistical knowledge is essential to make the most out of this book. What You Will Learn ? Use adaptive thinking to solve real-life AI case studies ? Rise beyond being a modern-day factory code worker ? Acquire advanced AI, machine learning, and deep learning designing skills ? Learn about cognitive NLP chatbots, quantum computing, and IoT and blockchain technology ? Understand future AI solutions and adapt quickly to them ? Develop out-of-the-box thinking to face any challenge the market presents In Detail Artificial Intelligence has the potential to replicate humans in every field. This book serves as a starting point for you to understand how AI is built, with the help of intriguing examples and case studies. Artificial Intelligence By Example will make you an adaptive thinker and help you apply concepts to real-life scenarios. Using some of the most interesting AI examples, right from a simple chess engine to a cognitive chatbot, you will learn how to tackle the machine you are competing with. You will study some of the most advanced machine learning models, understand how to apply AI to blockchain and IoT, and develop emotional quotient in chatbots using neural networks. You will move on to designing AI solutions in a simple manner rather than get confused by complex architectures and techniques. This comprehensive guide will be a starter kit for you to develop AI applications on your own. By the end of this book, will have understood the fundamentals of AI and worked through a number of case studies that will help you develop business vision. Style and approach An easy-to-follow step by step guide which will help you get to grips with real world application of Artificial Intelligence
目录展开

Title Page

Copyright and Credits

Artificial Intelligence By Example

Dedication

Packt Upsell

Why subscribe?

PacktPub.com

Contributors

About the author

About the reviewers

Packt is searching for authors like you

Preface

Who this book is for

What this book covers

To get the most out of this book

Download the example code files

Download the color images

Code in Action

Conventions used

Get in touch

Reviews

Become an Adaptive Thinker

Technical requirements

How to be an adaptive thinker

Addressing real-life issues before coding a solution

Step 1 – MDP in natural language

Step 2 – the mathematical representation of the Bellman equation and MDP

From MDP to the Bellman equation

Step 3 – implementing the solution in Python

The lessons of reinforcement learning

How to use the outputs

Machine learning versus traditional applications

Summary

Questions

Further reading

Think like a Machine

Technical requirements

Designing datasets – where the dream stops and the hard work begins

Designing datasets in natural language meetings

Using the McCulloch-Pitts neuron

The McCulloch-Pitts neuron

The architecture of Python TensorFlow

Logistic activation functions and classifiers

Overall architecture

Logistic classifier

Logistic function

Softmax

Summary

Questions

Further reading

Apply Machine Thinking to a Human Problem

Technical requirements

Determining what and how to measure

Convergence

Implicit convergence

Numerical – controlled convergence

Applying machine thinking to a human problem

Evaluating a position in a chess game

Applying the evaluation and convergence process to a business problem

Using supervised learning to evaluate result quality

Summary

Questions

Further reading

Become an Unconventional Innovator

Technical requirements

The XOR limit of the original perceptron

XOR and linearly separable models

Linearly separable models

The XOR limit of a linear model, such as the original perceptron

Building a feedforward neural network from scratch

Step 1 – Defining a feedforward neural network

Step 2 – how two children solve the XOR problem every day

Implementing a vintage XOR solution in Python with an FNN and backpropagation

A simplified version of a cost function and gradient descent

Linear separability was achieved

Applying the FNN XOR solution to a case study to optimize subsets of data

Summary

Questions

Further reading

Manage the Power of Machine Learning and Deep Learning

Technical requirements

Building the architecture of an FNN with TensorFlow

Writing code using the data flow graph as an architectural roadmap

A data flow graph translated into source code

The input data layer

The hidden layer

The output layer

The cost or loss function

Gradient descent and backpropagation

Running the session

Checking linear separability

Using TensorBoard to design the architecture of your machine learning and deep learning solutions

Designing the architecture of the data flow graph

Displaying the data flow graph in TensorBoard

The final source code with TensorFlow and TensorBoard

Using TensorBoard in a corporate environment

Using TensorBoard to explain the concept of classifying customer products to a CEO

Will your views on the project survive this meeting?

Summary

Questions

Further reading

References

Don't Get Lost in Techniques – Focus on Optimizing Your Solutions

Technical requirements

Dataset optimization and control

Designing a dataset and choosing an ML/DL model

Approval of the design matrix

Agreeing on the format of the design matrix

Dimensionality reduction

The volume of a training dataset

Implementing a k-means clustering solution

The vision

The data

Conditioning management

The strategy

The k-means clustering program

The mathematical definition of k-means clustering

Lloyd's algorithm

The goal of k-means clustering in this case study

The Python program

1 – The training dataset

2 – Hyperparameters

3 – The k-means clustering algorithm

4 – Defining the result labels

5 – Displaying the results – data points and clusters

Test dataset and prediction

Analyzing and presenting the results

AGV virtual clusters as a solution

Summary

Questions

Further reading

When and How to Use Artificial Intelligence

Technical requirements

Checking whether AI can be avoided

Data volume and applying k-means clustering

Proving your point

NP-hard – the meaning of P

NP-hard – The meaning of non-deterministic

The meaning of hard

Random sampling

The law of large numbers – LLN

The central limit theorem

Using a Monte Carlo estimator

Random sampling applications

Cloud solutions – AWS

Preparing your baseline model

Training the full sample training dataset

Training a random sample of the training dataset

Shuffling as an alternative to random sampling

AWS – data management

Buckets

Uploading files

Access to output results

SageMaker notebook

Creating a job

Running a job

Reading the results

Recommended strategy

Summary

Questions

Further reading

Revolutions Designed for Some Corporations and Disruptive Innovations for Small to Large Companies

Technical requirements

Is AI disruptive?

What is new and what isn't in AI

AI is based on mathematical theories that are not new

Neural networks are not new

Cloud server power, data volumes, and web sharing of the early 21st century started to make AI disruptive

Public awareness contributed to making AI disruptive

Inventions versus innovations

Revolutionary versus disruptive solutions

Where to start?

Discover a world of opportunities with Google Translate

Getting started

The program

The header

Implementing Google's translation service

Google Translate from a linguist's perspective

Playing with the tool

Linguistic assessment of Google Translate

Lexical field theory

Jargon

Translating is not just translating but interpreting

How to check a translation

AI as a new frontier

Lexical field and polysemy

Exploring the frontier – the program

k-nearest neighbor algorithm

The KNN algorithm

The knn_polysemy.py program

Implementing the KNN compressed function in Google_Translate_Customized.py

Conclusions on the Google Translate customized experiment

The disruptive revolutionary loop

Summary

Questions

Further reading

Getting Your Neurons to Work

Technical requirements

Defining a CNN

Defining a CNN

Initializing the CNN

Adding a 2D convolution

Kernel

Intuitive approach

Developers' approach

Mathematical approach

Shape

ReLu

Pooling

Next convolution and pooling layer

Flattening

Dense layers

Dense activation functions

Training a CNN model

The goal

Compiling the model

Loss function

Quadratic loss function

Binary cross-entropy

Adam optimizer

Metrics

Training dataset

Data augmentation

Loading the data

Testing dataset

Data augmentation

Loading the data

Training with the classifier

Saving the model

Next steps

Summary

Questions

Further reading and references

Applying Biomimicking to Artificial Intelligence

Technical requirements

Human biomimicking

TensorFlow, an open source machine learning framework

Does deep learning represent our brain or our mind?

A TensorBoard representation of our mind

Input data

Layer 1 – managing the inputs to the network

Weights, biases, and preactivation

Displaying the details of the activation function through the preactivation process

The activation function of Layer 1

Dropout and Layer 2

Layer 2

Measuring the precision of prediction of a network through accuracy values

Correct prediction

accuracy

Cross-entropy

Training

Optimizing speed with Google's Tensor Processing Unit

Summary

Questions

Further reading

Conceptual Representation Learning

Technical requirements

Generate profit with transfer learning

The motivation of transfer learning

Inductive thinking

Inductive abstraction

The problem AI needs to solve

The Γ gap concept

Loading the Keras model after training

Loading the model to optimize training

Loading the model to use it

Using transfer learning to be profitable or see a project stopped

Defining the strategy

Applying the model

Making the model profitable by using it for another problem

Where transfer learning ends and domain learning begins

Domain learning

How to use the programs

The trained models used in this section

The training model program

GAP – loaded or unloaded

GAP – jammed or open lanes

The gap dataset

Generalizing the Γ(gap conceptual dataset)

Generative adversarial networks

Generating conceptual representations

The use of autoencoders

The motivation of conceptual representation learning meta-models

The curse of dimensionality

The blessing of dimensionality

Scheduling and blockchains

Chatbots

Self-driving cars

Summary

Questions

Further reading

Automated Planning and Scheduling

Technical requirements

Planning and scheduling today and tomorrow

A real-time manufacturing process

Amazon must expand its services to face competition

A real-time manufacturing revolution

CRLMM applied to an automated apparel manufacturing process

An apparel manufacturing process

Training the CRLMM

Generalizing the unit-training dataset

Food conveyor belt processing – positive pγ and negative nγ gaps

Apparel conveyor belt processing – undetermined gaps

The beginning of an abstract notion of gaps

Modifying the hyperparameters

Running a prediction program

Building the DQN-CRLMM

A circular process

Implementing a CNN-CRLMM to detect gaps and optimize

Q-Learning – MDP

MDP inputs and outputs

The input is a neutral reward matrix

The standard output of the MDP function

A graph interpretation of the MDP output matrix

The optimizer

The optimizer as a regulator

Implementing Z – squashing the MDP result matrix

Implementing Z – squashing the vertex weights vector

Finding the main target for the MDP function

Circular DQN-CRLMM – a stream-like system that never starts nor ends

Summary

Questions

Further reading

AI and the Internet of Things (IoT)

Technical requirements

The Iotham City project

Setting up the DQN-CRLMM model

Training the CRLMM

The dataset

Training and testing the model

Classifying the parking lots

Adding an SVM function

Motivation – using an SVM to increase safety levels

Definition of a support vector machine

Python function

Running the CRLMM

Finding a parking space

Deciding how to get to the parking lot

Support vector machine

The itinerary graph

The weight vector

Summary

Questions

Further reading

References

Optimizing Blockchains with AI

Technical requirements

Blockchain technology background

Mining bitcoins

Using cryptocurrency

Using blockchains

Using blockchains in the A-F network

Creating a block

Exploring the blocks

Using naive Bayes in a blockchain process

A naive Bayes example

The blockchain anticipation novelty

The goal

Step 1 the dataset

Step 2 frequency

Step 3 likelihood

Step 4 naive Bayes equation

Implementation

Gaussian naive Bayes

The Python program

Implementing your ideas

Summary

Questions

Further reading

Cognitive NLP Chatbots

Technical requirements

IBM Watson

Intents

Testing the subsets

Entities

Dialog flow

Scripting and building up the model

Adding services to a chatbot

A cognitive chatbot service

The case study

A cognitive dataset

Cognitive natural language processing

Activating an image + word cognitive chat

Solving the problem

Implementation

Summary

Questions

Further reading

Improve the Emotional Intelligence Deficiencies of Chatbots

Technical requirements

Building a mind

How to read this chapter

The profiling scenario

Restricted Boltzmann Machines

The connections between visible and hidden units

Energy-based models

Gibbs random sampling

Running the epochs and analyzing the results

Sentiment analysis

Parsing the datasets

Conceptual representation learning meta-models

Profiling with images

RNN for data augmentation

RNNs and LSTMs

RNN, LSTM, and vanishing gradients

Prediction as data augmentation

Step1 – providing an input file

Step 2 – running an RNN

Step 3 – producing data augmentation

Word embedding

The Word2vec model

Principal component analysis

Intuitive explanation

Mathematical explanation

Variance

Covariance

Eigenvalues and eigenvectors

Creating the feature vector

Deriving the dataset

Summing it up

TensorBoard Projector

Using Jacobian matrices

Summary

Questions

Further reading

Quantum Computers That Think

Technical requirements

The rising power of quantum computers

Quantum computer speed

Defining a qubit

Representing a qubit

The position of a qubit

Radians, degrees, and rotations

Bloch sphere

Composing a quantum score

Quantum gates with Quirk

A quantum computer score with Quirk

A quantum computer score with IBM Q

A thinking quantum computer

Representing our mind's concepts

Expanding MindX's conceptual representations

Concepts in the mind-dataset of MindX

Positive thinking

Negative thinking

Gaps

Distances

The embedding program

The MindX experiment

Preparing the data

Transformation Functions – the situation function

Transformation functions – the quantum function

Creating and running the score

Using the output

IBM Watson and scripts

Summary

Questions

Further reading

Answers to the Questions

Chapter 1 – Become an Adaptive Thinker

Chapter 2 – Think like a Machine

Chapter 3 – Apply Machine Thinking to a Human Problem

Chapter 4 – Become an Unconventional Innovator

Chapter 5 – Manage the Power of Machine Learning and Deep Learning

Chapter 6 – Don't Get Lost in Techniques, Focus on Optimizing Your Solutions

Chapter 7 – When and How to Use Artificial Intelligence

Chapter 8 – Revolutions Designed for Some Corporations and Disruptive Innovations for Small to Large Companies

Chapter 9 – Getting Your Neurons to Work

Chapter 10 – Applying Biomimicking to AI

Chapter 11 – Conceptual Representation Learning

Chapter 12 – Automated Planning and Scheduling

Chapter 13 – AI and the Internet of Things

Chapter 14 – Optimizing Blockchains with AI

Chapter 15 – Cognitive NLP Chatbots

Chapter 16 – Improve the Emotional Intelligence Deficiencies of Chatbots

Chapter 17 – Quantum Computers That Think

累计评论(0条) 0个书友正在讨论这本书 发表评论

发表评论

发表评论,分享你的想法吧!

买过这本书的人还买过

读了这本书的人还在读

回顶部