Machine Learning with Scikit-learn

Instructor: Safdar Munir

5.0

Course Duration:40 Hours

Course level:Intermediate

What I will learn?

About Course

Machine learning is the field that teaches machines and computers to learn from existing data to make predictions on new data: Will a tumor be benign or malignant? Which of your customers will take their business elsewhere? Is a particular email spam?

In this course, you’ll learn how to use Python to perform supervised learning, an essential component of machine learning. You’ll learn how to build predictive models, tune their parameters, and determine how well they will perform with unseen data—all while using real world datasets.

You’ll be using scikit-learn, one of the most popular and user-friendly machine learning libraries for Python.

Course Curriculum

In this chapter, you will be introduced to classification problems and learn how to solve them using supervised learning techniques. And you’ll apply what you learn to a different dataset.

  1. Supervised learning
  2. Which of these is a classification problem?
  3. Exploratory data analysis
  4. Numerical EDA
  5. Visual EDA
  6. The classification challenge
  7. k-Nearest Neighbors: Fit
  8. k-Nearest Neighbors: Predict
  9. Measuring model performance
  10. The digits recognition dataset
  11. Train/Test Split + Fit/Predict/Accuracy
  12. Overfitting and underfitting

In the previous chapter, you used image and different datasets to predict binary and multiclass outcomes. But what if your problem requires a continuous outcome? Regression is best suited to solving such problems. You will learn about fundamental concepts in regression and apply them to predict the life expectancy

  1. Introduction to regression
  2. Which of the following is a regression problem?
  3. Importing data for supervised learning
  4. Exploring the data
  5. The basics of linear regression
  6. Fit & predict for regression
  7. Train/test split for regression
  8. Cross-validation
  9. 5-fold cross-validation
  10. K-Fold CV comparison
  11. Regularized regression
  12. Regularization I: Lasso
  13. Regularization II: Ridge
Having trained your model, your next task is to evaluate its performance. In this chapter, you will learn about some of the other metrics available in scikit-learn that will allow you to assess your model’s performance in a more nuanced manner. Next, learn to optimize your classification and regression models using hyperparameter tuning.
 
  1. How good is your model?
  2. Metrics for classification
  3. Logistic regression and the ROC curve
  4. Building a logistic regression model
  5. Plotting an ROC curve
  6. Precision-recall Curve
  7. Area under the ROC curve
  8. AUC computation
  9. Hyperparameter tuning
  10. Hyperparameter tuning with GridSearchCV
  11. Hyperparameter tuning with RandomizedSearchCV
  12. Hold-out set for final evaluation
  13. Hold-out set reasoning
  14. Hold-out set in practice I: Classification
  15. Hold-out set in practice II: Regression

This chapter introduces pipelines, and how scikit-learn allows for transformers and estimators to be chained together and used as a single unit. Preprocessing techniques will be introduced as a way to enhance model performance, and pipelines will tie together concepts from previous chapters.

  1. Preprocessing data
  2. Exploring categorical features
  3. Creating dummy variables
  4. Regression with categorical features
  5. Handling missing data
  6. Dropping missing data
  7. Imputing missing data in a ML Pipeline I
  8. Imputing missing data in a ML Pipeline II
  9. Centering and scaling
  10. Centering and scaling your data
  11. Centering and scaling in a pipeline
  12. Bringing it all together I: Pipeline for classification
  13. Bringing it all together II: Pipeline for regression
  14. Final thoughts

Requirements

Material Includes

Play Video

15000 PKR

Benefits Obtained :

SHARE :

Course related

Beginner

Register Now

Fill out the form below to register in your desired course, and we will be in touch shortly.