万本电子书0元读

万本电子书0元读

顶部广告

Deep Learning with R for Beginners电子书

售       价:¥

3人正在读 | 0人评论 9.8

作       者:Mark Hodnett

出  版  社:Packt Publishing

出版时间:2019-05-20

字       数:74.6万

所属分类: 进口书 > 外文原版书 > 电脑/网络

温馨提示:数字商品不支持退换货,不提供源文件,不支持导出打印

为你推荐

  • 读书简介
  • 目录
  • 累计评论(0条)
  • 读书简介
  • 目录
  • 累计评论(0条)
Explore the world of neural networks by building powerful deep learning models using the R ecosystem Key Features * Get to grips with the fundamentals of deep learning and neural networks * Use R 3.5 and its libraries and APIs to build deep learning models for computer vision and text processing * Implement effective deep learning systems in R with the help of end-to-end projects Book Description Deep learning finds practical applications in several domains, while R is the preferred language for designing and deploying deep learning models. This Learning Path introduces you to the basics of deep learning and even teaches you to build a neural network model from scratch. As you make your way through the chapters, you’ll explore deep learning libraries and understand how to create deep learning models for a variety of challenges, right from anomaly detection to recommendation systems. The book will then help you cover advanced topics, such as generative adversarial networks (GANs), transfer learning, and large-scale deep learning in the cloud, in addition to model optimization, overfitting, and data augmentation. Through real-world projects, you’ll also get up to speed with training convolutional neural networks (CNNs), recurrent neural networks (RNNs), and long short-term memory networks (LSTMs) in R. By the end of this Learning Path, you’ll be well versed with deep learning and have the skills you need to implement a number of deep learning concepts in your research work or projects. This Learning Path includes content from the following Packt products: * R Deep Learning Essentials - Second Edition by F. Wiley and Mark Hodnett * R Deep Learning Projects by Yuxi (Hayden) Liu and Pablo Maldonado What you will learn * Implement credit card fraud detection with autoencoders * Train neural networks to perform handwritten digit recognition using MXNet * Reconstruct images using variational autoencoders * Explore the applications of autoencoder neural networks in clustering and dimensionality reduction * Create natural language processing (NLP) models using Keras and TensorFlow in R * Prevent models from overfitting the data to improve generalizability * Build shallow neural network prediction models Who this book is for This Learning Path is for aspiring data scientists, data analysts, machine learning developers, and deep learning enthusiasts who are well versed in machine learning concepts and are looking to explore the deep learning paradigm using R. A fundamental understanding of R programming and familiarity with the basic concepts of deep learning are necessary to get the most out of this Learning Path.
目录展开

About Packt

Why subscribe?

Packt.com

Contributors

About the authors

Packt is searching for authors like you

Preface

Who this book is for

What this book covers

To get the most out of this book

Download the example code files

Download the color images of this book

Conventions used

Get in touch

Reviews

Getting Started with Deep Learning

What is deep learning?

A conceptual overview of neural networks

Neural networks as an extension of linear regression

Neural networks as a network of memory cells

Deep neural networks

Some common myths about deep learning

Setting up your R environment

Deep learning frameworks for R

MXNet

Keras

Do I need a GPU (and what is it, anyway)?

Setting up reproducible results

Summary

Training a Prediction Model

Neural networks in R

Building neural network models

Generating predictions from a neural network

The problem of overfitting data – the consequences explained

Use case – building and applying a neural network

Summary

Deep Learning Fundamentals

Building neural networks from scratch in R

Neural network web application

Neural network code

Back to deep learning

The symbol, X, y, and ctx parameters

The num.round and begin.round parameters

The optimizer parameter

The initializer parameter

The eval.metric and eval.data parameters

The epoch.end.callback parameter

The array.batch.size parameter

Using regularization to overcome overfitting

L1 penalty

L1 penalty in action

L2 penalty

L2 penalty in action

Weight decay (L2 penalty in neural networks)

Ensembles and model-averaging

Use case – improving out-of-sample model performance using dropout

Summary

Training Deep Prediction Models

Getting started with deep feedforward neural networks

Activation functions

Introduction to the MXNet deep learning library

Deep learning layers

Building a deep learning model

Use case – using MXNet for classification and regression

Data download and exploration

Preparing the data for our models

The binary classification model

The regression model

Improving the binary classification model

The unreasonable effectiveness of data

Summary

Image Classification Using Convolutional Neural Networks

CNNs

Convolutional layers

Pooling layers

Dropout

Flatten layers, dense layers, and softmax

Image classification using the MXNet library

Base model (no convolutional layers)

LeNet

Classification using the fashion MNIST dataset

References/further reading

Summary

Tuning and Optimizing Models

Evaluation metrics and evaluating performance

Types of evaluation metric

Evaluating performance

Data preparation

Different data distributions

Data partition between training, test, and validation sets

Standardization

Data leakage

Data augmentation

Using data augmentation to increase the training data

Test time augmentation

Using data augmentation in deep learning libraries

Tuning hyperparameters

Grid search

Random search

Use case—using LIME for interpretability

Model interpretability with LIME

Summary

Natural Language Processing Using Deep Learning

Document classification

The Reuters dataset

Traditional text classification

Deep learning text classification

Word vectors

Comparing traditional text classification and deep learning

Advanced deep learning text classification

1D convolutional neural network model

Recurrent neural network model

Long short term memory model

Gated Recurrent Units model

Bidirectional LSTM model

Stacked bidirectional model

Bidirectional with 1D convolutional neural network model

Comparing the deep learning NLP architectures

Summary

Deep Learning Models Using TensorFlow in R

Introduction to the TensorFlow library

Using TensorBoard to visualize deep learning networks

TensorFlow models

Linear regression using TensorFlow

Convolutional neural networks using TensorFlow

TensorFlow estimators and TensorFlow runs packages

TensorFlow estimators

TensorFlow runs package

Summary

Anomaly Detection and Recommendation Systems

What is unsupervised learning?

How do auto-encoders work?

Regularized auto-encoders

Penalized auto-encoders

Denoising auto-encoders

Training an auto-encoder in R

Accessing the features of the auto-encoder model

Using auto-encoders for anomaly detection

Use case – collaborative filtering

Preparing the data

Building a collaborative filtering model

Building a deep learning collaborative filtering model

Applying the deep learning model to a business problem

Summary

Running Deep Learning Models in the Cloud

Setting up a local computer for deep learning

How do I know if my model is training on a GPU?

Using AWS for deep learning

A brief introduction to AWS

Creating a deep learning GPU instance in AWS

Creating a deep learning AMI in AWS

Using Azure for deep learning

Using Google Cloud for deep learning

Using Paperspace for deep learning

Summary

The Next Level in Deep Learning

Image classification models

Building a complete image classification solution

Creating the image data

Building the deep learning model

Using the saved deep learning model

The ImageNet dataset

Loading an existing model

Transfer learning

Deploying TensorFlow models

Other deep learning topics

Generative adversarial networks

Reinforcement learning

Summary

Handwritten Digit Recognition using Convolutional Neural Networks

What is deep learning and why do we need it?

What makes deep learning special?

What are the applications of deep learning?

Handwritten digit recognition using CNNs

Get started with exploring MNIST

First attempt – logistic regression

Going from logistic regression to single-layer neural networks

Adding more hidden layers to the networks

Extracting richer representation with CNNs

Summary

Traffic Signs Recognition for Intelligent Vehicles

How is deep learning applied in self-driving cars?

How does deep learning become a state-of-the-art solution?

Traffic sign recognition using CNN

Getting started with exploring GTSRB

First solution – convolutional neural networks using MXNet

Trying something new – CNNs using Keras with TensorFlow

Reducing overfitting with dropout

Dealing with a small training set – data augmentation

Reviewing methods to prevent overfitting in CNNs

Summary

Fraud Detection with Autoencoders

Getting ready

Installing Keras and TensorFlow for R

Installing H2O

Our first examples

A simple 2D example

Autoencoders and MNIST

Outlier detection in MNIST

Credit card fraud detection with autoencoders

Exploratory data analysis

The autoencoder approach – Keras

Fraud detection with H2O

Exercises

Variational Autoencoders

Image reconstruction using VAEs

Outlier detection in MNIST

Text fraud detection

From unstructured text data to a matrix

From text to matrix representation — the Enron dataset

Autoencoder on the matrix representation

Exercises

Summary

Text Generation using Recurrent Neural Networks

What is so exciting about recurrent neural networks?

But what is a recurrent neural network, really?

LSTM and GRU networks

LSTM

GRU

RNNs from scratch in R

Classes in R with R6

Perceptron as an R6 class

Logistic regression

Multi-layer perceptron

Implementing a RNN

Implementation as an R6 class

Implementation without R6

RNN without derivatives — the cross-entropy method

RNN using Keras

A simple benchmark implementation

Generating new text from old

Exercises

Summary

Sentiment Analysis with Word Embedding

Warm-up – data exploration

Working with tidy text

The more, the merrier – calculating n-grams instead of single words

Bag of words benchmark

Preparing the data

Implementing a benchmark – logistic regression

Exercises

Word embeddings

word2vec

GloVe

Sentiment analysis from movie reviews

Data preprocessing

From words to vectors

Sentiment extraction

The importance of data cleansing

Vector embeddings and neural networks

Bi-directional LSTM networks

Other LSTM architectures

Exercises

Mining sentiment from Twitter

Connecting to the Twitter API

Building our model

Exploratory data analysis

Using a trained model

Summary

Other Books You May Enjoy

Leave a review - let other readers know what you think

累计评论(0条) 0个书友正在讨论这本书 发表评论

发表评论

发表评论,分享你的想法吧!

买过这本书的人还买过

读了这本书的人还在读

回顶部