Employing Ensemble Methods with scikit-learn
MP4 | Video: AVC 1280×720 | Audio: AAC 44KHz 2ch | Duration: 2 Hours 14M | 203 MB
Genre: eLearning | Language: English
This course covers the theoretical and practical aspects of building ensemble learning solutions in scikit-learn; from random forests built using bagging and pasting to adaptive and gradient boosting and model stacking and hyperparameter tuning.
Even as the number of machine learning frameworks and libraries increases on a daily basis, scikit-learn is retaining its popularity with ease. In particular, scikit-learn features extremely comprehensive support for ensemble learning, an important technique to mitigate overfitting. In this course, Employing Ensemble Methods with scikit-learn, you will gain the ability to construct several important types of ensemble learning models. First, you will learn decision trees and random forests are ideal building blocks for ensemble learning, and how hard voting and soft voting can be used in an ensemble model. Next, you will discover how bagging and pasting can be used to control the manner in which individual learners in the ensemble are trained. Finally, you will round out your knowledge by utilizing model stacking to combine the output of individual learners. When you’re finished with this course, you will have the skills and knowledge to design and implement sophisticated ensemble learning techniques using the support provided by the scikit-learn framework.
(Buy premium account for maximum speed and resuming ability)