Главная
Study mode:
on
1
Intro
2
Uses of Machine Learning
3
Machine Learning Models
4
Data Format and Quality
5
Challenges of Feature Engineering
6
Open-source for Feature engineering
7
Why Open-source
8
Missing Data Imputation Techniques
9
Categorical Variables
10
Categorical Encoding Techniques
11
Encoding Techniques: Rare labels
12
Distributions
13
Mathematical transformations
14
Discretisation
15
Outliers
16
Feature Combination
17
Variable Magnitude
18
Feature scaling methods
19
Datetime Variables
20
Transactions and Time Series
21
Pipeline
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Explore data preprocessing and feature engineering techniques for machine learning in this 40-minute conference talk from the Data Science Festival Summer School 2021. Delve into the crucial steps data scientists take to clean and prepare raw data for model training. Learn about various feature engineering processes, including handling missing values, encoding categorical variables, mathematical transformations, and creating new variables. Discover when and why to use specific techniques, their advantages, assumptions, and limitations, as well as their suitability for different algorithms. Compare implementations of these techniques in open-source Python libraries as presented by Soledad Galli, Lead Data Scientist at Train in Data. Gain insights into topics such as missing data imputation, categorical encoding, handling rare labels, distribution transformations, discretization, outlier treatment, feature combination, variable magnitude, scaling methods, and working with datetime variables, transactions, and time series data. Understand the importance of building efficient data preprocessing pipelines for machine learning projects. Read more

Data Preprocessing and Feature Engineering for Machine Learning

Data Science Festival
Add to list