Employing Ensemble Methods with scikit-learn

by Janani Ravi

This course covers the theoretical and practical aspects of building ensemble learning solutions in scikit-learn; from random forests built using bagging and pasting to adaptive and gradient boosting and model stacking and hyperparameter tuning.

What you'll learn

Even as the number of machine learning frameworks and libraries increases on a daily basis, scikit-learn is retaining its popularity with ease. In particular, scikit-learn features extremely comprehensive support for ensemble learning, an important technique to mitigate overfitting. In this course, Employing Ensemble Methods with scikit-learn, you will gain the ability to construct several important types of ensemble learning models. First, you will learn decision trees and random forests are ideal building blocks for ensemble learning, and how hard voting and soft voting can be used in an ensemble model. Next, you will discover how bagging and pasting can be used to control the manner in which individual learners in the ensemble are trained. Finally, you will round out your knowledge by utilizing model stacking to combine the output of individual learners. When you’re finished with this course, you will have the skills and knowledge to design and implement sophisticated ensemble learning techniques using the support provided by the scikit-learn framework.

Table of contents

Course Overview

About the author

A problem solver at heart, Janani has a Masters degree from Stanford and worked for 7+ years at Google. She was one of the original engineers on Google Docs and holds 4 patents for its real-time collaborative editing framework.

Ready to upskill? Get started