万本电子书0元读

万本电子书0元读

顶部广告

Hands-On Ensemble Learning with R电子书

售       价:¥

2人正在读 | 0人评论 9.8

作       者:Prabhanjan Narayanachar Tattar

出  版  社:Packt Publishing

出版时间:2018-07-27

字       数:196.2万

所属分类: 进口书 > 外文原版书 > 电脑/网络

温馨提示:数字商品不支持退换货,不提供源文件,不支持导出打印

为你推荐

  • 读书简介
  • 目录
  • 累计评论(0条)
  • 读书简介
  • 目录
  • 累计评论(0条)
Explore powerful R packages to create predictive models using ensemble methods Key Features *Implement machine learning algorithms to build ensemble-efficient models *Explore powerful R packages to create predictive models using ensemble methods *Learn to build ensemble models on large datasets using a practical approach Book Description Ensemble techniques are used for combining two or more similar or dissimilar machine learning algorithms to create a stronger model. Such a model delivers superior prediction power and can give your datasets a boost in accuracy. Hands-On Ensemble Learning with R begins with the important statistical resampling methods. You will then walk through the central trilogy of ensemble techniques – bagging, random forest, and boosting – then you'll learn how they can be used to provide greater accuracy on large datasets using popular R packages. You will learn how to combine model predictions using different machine learning algorithms to build ensemble models. In addition to this, you will explore how to improve the performance of your ensemble models. By the end of this book, you will have learned how machine learning algorithms can be combined to reduce common problems and build simple efficient ensemble models with the help of real-world examples. What you will learn *Carry out an essential review of re-sampling methods, bootstrap, and jackknife *Explore the key ensemble methods: bagging, random forests, and boosting *Use multiple algorithms to make strong predictive models *Enjoy a comprehensive treatment of boosting methods *Supplement methods with statistical tests, such as ROC *Walk through data structures in classification, regression, survival, and time series data *Use the supplied R code to implement ensemble methods *Learn stacking method to combine heterogeneous machine learning models Who this book is for This book is for you if you are a data scientist or machine learning developer who wants to implement machine learning techniques by building ensemble models with the power of R. You will learn how to combine different machine learning algorithms to perform efficient data processing. Basic knowledge of machine learning techniques and programming knowledge of R would be an added advantage.
目录展开

Hands-On Ensemble Learning with R

Why subscribe?

PacktPub.com

Contributors

About the author

About the reviewer

Packt is Searching for Authors Like You

Preface

Who this book is for

What this book covers

To get the most out of this book

Download the example code files

Download the color images

Conventions used

Get in touch

Reviews

Chapter 1. Introduction to Ensemble Techniques

Datasets

Hypothyroid

Waveform

German Credit

Iris

Pima Indians Diabetes

US Crime

Overseas visitors

Primary Biliary Cirrhosis

Multishapes

Board Stiffness

Statistical/machine learning models

Logistic regression model

Neural networks

Naïve Bayes classifier

Decision tree

Support vector machines

The right model dilemma!

An ensemble purview

Complementary statistical tests

Permutation test

Chi-square and McNemar test

ROC test

Summary

Chapter 2. Bootstrapping

Technical requirements

The jackknife technique

The jackknife method for mean and variance

Pseudovalues method for survival data

Bootstrap – a statistical method

The standard error of correlation coefficient

The parametric bootstrap

Eigen values

The boot package

Bootstrap and testing hypotheses

Bootstrapping regression models

Bootstrapping survival models*

Bootstrapping time series models*

Summary

Chapter 3. Bagging

Technical requirements

Classification trees and pruning

Bagging

k-NN classifier

Analyzing waveform data

k-NN bagging

Summary

Chapter 4. Random Forests

Technical requirements

Random Forests

Variable importance

Proximity plots

Random Forest nuances

Comparisons with bagging

Missing data imputation

Clustering with Random Forest

Summary

Chapter 5. The Bare Bones Boosting Algorithms

Technical requirements

The general boosting algorithm

Adaptive boosting

Gradient boosting

Using the adabag and gbm packages

Variable importance

Comparing bagging, random forests, and boosting

Summary

Chapter 6. Boosting Refinements

Technical requirements

Why does boosting work?

The gbm package

Boosting for count data

Boosting for survival data

The xgboost package

The h2o package

Summary

Chapter 7. The General Ensemble Technique

Technical requirements

Why does ensembling work?

Ensembling by voting

Majority voting

Weighted voting

Ensembling by averaging

Simple averaging

Weight averaging

Stack ensembling

Summary

Chapter 8. Ensemble Diagnostics

Technical requirements

What is ensemble diagnostics?

Ensemble diversity

Numeric prediction

Pairwise measure

Disagreement measure

Yule's or Q-statistic

Correlation coefficient measure

Cohen's statistic

Double-fault measure

Interrating agreement

Entropy measure

Kohavi-Wolpert measure

Disagreement measure for ensemble

Measurement of interrater agreement

Summary

Chapter 9. Ensembling Regression Models

Technical requirements

Pre-processing the housing data

Visualization and variable reduction

Variable clustering

Regression models

Linear regression model

Neural networks

Regression tree

Prediction for regression models

Bagging and Random Forests

Boosting regression models

Stacking methods for regression models

Summary

Chapter 10. Ensembling Survival Models

Core concepts of survival analysis

Nonparametric inference

Regression models – parametric and Cox proportional hazards models

Survival tree

Ensemble survival models

Summary

Chapter 11. Ensembling Time Series Models

Technical requirements

Time series datasets

Time series visualization

Core concepts and metrics

Essential time series models

Bagging and time series

Ensemble time series models

Summary

Chapter 12. What's Next?

Appendix A. Bibliography

References

R package references

Index

A

B

C

D

E

F

G

H

I

J

K

L

M

N

O

P

Q

R

S

T

U

V

W

X

Y

累计评论(0条) 0个书友正在讨论这本书 发表评论

发表评论

发表评论,分享你的想法吧!

买过这本书的人还买过

读了这本书的人还在读

回顶部