Feature Selection for Machine Learning

preview_player
Показать описание
#datascience #datasciencefestival #machinelearning

When we build machine learning models with the aim to use them in production, we probably don’t want to use all the variables available in the data. Sure, adding more variables rarely makes a model less accurate, but there are certain disadvantages to including an excess of features. To select the most predictive variables, we can use several feature selection algorithms. They are typically grouped in three categories; filter, wrapper and embedded methods, and those algorithms that do not fit in these categories are sort of hybrid methods. In this video Soledad Galli, Lead Data Scientist at Train in Data, discusses the importance of feature selection, walks us through the categories of feature selection methods and describes the most popular algorithms of each. She also compares the implementation of these feature selection algorithms in open source Python libraries.

#datasciencefestival
#DIsummerschool
#machine learning
Рекомендации по теме