万本电子书0元读

万本电子书0元读

顶部广告

Mastering Machine Learning with scikit-learn - Second Edition电子书

售       价:¥

12人正在读 | 0人评论 9.8

作       者:Gavin Hackeling

出  版  社:Packt Publishing

出版时间:2017-07-24

字       数:30.1万

所属分类: 进口书 > 外文原版书 > 电脑/网络

温馨提示:数字商品不支持退换货,不提供源文件,不支持导出打印

为你推荐

  • 读书简介
  • 目录
  • 累计评论(0条)
  • 读书简介
  • 目录
  • 累计评论(0条)
Use scikit-learn to apply machine learning to real-world problems About This Book ? Master popular machine learning models including k-nearest neighbors, random forests, logistic regression, k-means, naive Bayes, and artificial neural networks ? Learn how to build and evaluate performance of efficient models using scikit-learn ? Practical guide to master your basics and learn from real life applications of machine learning Who This Book Is For This book is intended for software engineers who want to understand how common machine learning algorithms work and develop an intuition for how to use them, and for data scientists who want to learn about the scikit-learn API. Familiarity with machine learning fundamentals and Python are helpful, but not required. What You Will Learn ? Review fundamental concepts such as bias and variance ? Extract features from categorical variables, text, and images ? Predict the values of continuous variables using linear regression and K Nearest Neighbors ? Classify documents and images using logistic regression and support vector machines ? Create ensembles of estimators using bagging and boosting techniques ? Discover hidden structures in data using K-Means clustering ? Evaluate the performance of machine learning systems in common tasks In Detail Machine learning is the buzzword bringing computer science and statistics together to build smart and efficient models. Using powerful algorithms and techniques offered by machine learning you can automate any analytical model. This book examines a variety of machine learning models including popular machine learning algorithms such as k-nearest neighbors, logistic regression, naive Bayes, k-means, decision trees, and artificial neural networks. It discusses data preprocessing, hyperparameter optimization, and ensemble methods. You will build systems that classify documents, recognize images, detect ads, and more. You will learn to use scikit-learn’s API to extract features from categorical variables, text and images; evaluate model performance, and develop an intuition for how to improve your model’s performance. By the end of this book, you will master all required concepts of scikit-learn to build efficient models at work to carry out advanced tasks with the practical approach. Style and approach This book is motivated by the belief that you do not understand something until you can describe it simply. Work through toy problems to develop your understanding of the learning algorithms and models, then apply your learnings to real-life problems.
目录展开

Title Page

Copyright

Mastering Machine Learning with scikit-learn

Second Edition

Credits

About the Author

About the Reviewer

www.PacktPub.com

Why subscribe?

Customer Feedback

Preface

What this book covers

What you need for this book

Who this book is for

Conventions

Reader feedback

Customer support

Downloading the example code

Errata

Piracy

Questions

The Fundamentals of Machine Learning

Defining machine learning

Learning from experience

Machine learning tasks

Training data, testing data, and validation data

Bias and variance

An introduction to scikit-learn

Installing scikit-learn

Installing using pip

Installing on Windows

Installing on Ubuntu 16.04

Installing on Mac OS

Installing Anaconda

Verifying the installation

Installing pandas, Pillow, NLTK, and matplotlib

Summary

Simple Linear Regression

Simple linear regression

Evaluating the fitness of the model with a cost function

Solving OLS for simple linear regression

Evaluating the model

Summary

Classification and Regression with k-Nearest Neighbors

K-Nearest Neighbors

Lazy learning and non-parametric models

Classification with KNN

Regression with KNN

Scaling features

Summary

Feature Extraction

Extracting features from categorical variables

Standardizing features

Extracting features from text

The bag-of-words model

Stop word filtering

Stemming and lemmatization

Extending bag-of-words with tf-idf weights

Space-efficient feature vectorizing with the hashing trick

Word embeddings

Extracting features from images

Extracting features from pixel intensities

Using convolutional neural network activations as features

Summary

From Simple Linear Regression to Multiple Linear Regression

Multiple linear regression

Polynomial regression

Regularization

Applying linear regression

Exploring the data

Fitting and evaluating the model

Gradient descent

Summary

From Linear Regression to Logistic Regression

Binary classification with logistic regression

Spam filtering

Binary classification performance metrics

Accuracy

Precision and recall

Calculating the F1 measure

ROC AUC

Tuning models with grid search

Multi-class classification

Multi-class classification performance metrics

Multi-label classification and problem transformation

Multi-label classification performance metrics

Summary

Naive Bayes

Bayes' theorem

Generative and discriminative models

Naive Bayes

Assumptions of Naive Bayes

Naive Bayes with scikit-learn

Summary

Nonlinear Classification and Regression with Decision Trees

Decision trees

Training decision trees

Selecting the questions

Information gain

Gini impurity

Decision trees with scikit-learn

Advantages and disadvantages of decision trees

Summary

From Decision Trees to Random Forests and Other Ensemble Methods

Bagging

Boosting

Stacking

Summary

The Perceptron

The perceptron

Activation functions

The perceptron learning algorithm

Binary classification with the perceptron

Document classification with the perceptron

Limitations of the perceptron

Summary

From the Perceptron to Support Vector Machines

Kernels and the kernel trick

Maximum margin classification and support vectors

Classifying characters in scikit-learn

Classifying handwritten digits

Classifying characters in natural images

Summary

From the Perceptron to Artificial Neural Networks

Nonlinear decision boundaries

Feed-forward and feedback ANNs

Multi-layer perceptrons

Training multi-layer perceptrons

Backpropagation

Training a multi-layer perceptron to approximate XOR

Training a multi-layer perceptron to classify handwritten digits

Summary

K-means

Clustering

K-means

Local optima

Selecting K with the elbow method

Evaluating clusters

Image quantization

Clustering to learn features

Summary

Dimensionality Reduction with Principal Component Analysis

Principal component analysis

Variance, covariance, and covariance matrices

Eigenvectors and eigenvalues

Performing PCA

Visualizing high-dimensional data with PCA

Face recognition with PCA

Summary

累计评论(0条) 0个书友正在讨论这本书 发表评论

发表评论

发表评论,分享你的想法吧!

买过这本书的人还买过

读了这本书的人还在读

回顶部