Employing Ensemble Methods with scikit-learn
This course covers the theoretical and practical aspects of building ensemble learning solutions in scikit-learn; from random forests built using bagging and pasting to adaptive and gradient boosting and model stacking and hyperparameter tuning.
What you'll learn
Even as the number of machine learning frameworks and libraries increases on a daily basis, scikit-learn is retaining its popularity with ease. In particular, scikit-learn features extremely comprehensive support for ensemble learning, an important technique to mitigate overfitting. In this course, Employing Ensemble Methods with scikit-learn, you will gain the ability to construct several important types of ensemble learning models. First, you will learn decision trees and random forests are ideal building blocks for ensemble learning, and how hard voting and soft voting can be used in an ensemble model. Next, you will discover how bagging and pasting can be used to control the manner in which individual learners in the ensemble are trained. Finally, you will round out your knowledge by utilizing model stacking to combine the output of individual learners. When you’re finished with this course, you will have the skills and knowledge to design and implement sophisticated ensemble learning techniques using the support provided by the scikit-learn framework.
Table of contents
- Version Check 0m
- Module Overview 1m
- Prerequisites and Course Outline 2m
- A Quick Overview of Ensemble Learning 6m
- Averaging and Boosting, Voting and Stacking 7m
- Decision Trees in Ensemble Learning 3m
- Understanding Decision Trees 3m
- Overfitted Models and Ensemble Learning 5m
- Getting Started and Exploring the Environment 2m
- Exploring the Classification Dataset 7m
- Hard Voting 5m
- Soft Voting 4m
- Module Summary 1m
- Module Overview 2m
- Bagging and Pasting 5m
- Random Subspaces and Random Patches 3m
- Extra Trees 3m
- Averaging vs. Boosting 2m
- Exploring the Regression Dataset 4m
- Regression Using Bagging and Pasting 5m
- Regression Using Random Subspaces 2m
- Classification Using Bagging and Pasting 4m
- Classification Using Random Patches 2m
- Regression Using Random Forest 5m
- Regression Using Extra Trees 2m
- Classification Using Random Forest and Extra Trees 3m
- Module Summary 1m
- Module Overview 1m
- Adaptive Boosting (AdaBoost) 3m
- Regression Using AdaBoost 5m
- Classification Using AdaBoost 3m
- Gradient Boosting 3m
- Regression Using Gradient Boosting 6m
- Hyperparameter Tuning of the Gradient Boosting Regressor Using Grid Search 4m
- Hyperparameter Tuning Using Warm Start and Early Stopping 4m
- Module Summary 1m