Impute missing values using KNNImputer or IterativeImputer

preview_player
Показать описание
Need something better than SimpleImputer for missing value imputation?
Try KNNImputer or IterativeImputer (inspired by R's MICE package). Both are multivariate approaches (they take other features into account!)

👉 New tips every TUESDAY and THURSDAY! 👈

=== WANT TO GET BETTER AT MACHINE LEARNING? ===

3) LET'S CONNECT!
Рекомендации по теме
Комментарии
Автор

Thanks for watching! 🙌 Let me know if you have any questions about imputation and I'm happy to answer them! 👇

dataschool
Автор

I really love your videos, they are just right, concise and informative, no unnecessary fluff. Thank you so much for these.

levon
Автор

Kevin, you just expanded my column transformation vocabulary. Thank you.

lovejazzbass
Автор

thank you.
love the clarity in your explanation!

zfitvwe
Автор

Thank you, this is exactly what I need. Plus you've explained it very well!

dogsever
Автор

You are awesome man!! Saved me a lot of time yet again!!!!

seansantiagox
Автор

Awesome video, couldn't be clearer. Thanks

ilducedimas
Автор

Fantastic video !! 👏🏼👏🏼👏🏼 … thank you for spreading the knowledge

mapa
Автор

God bless you man such valuable content you are producing!

atiqrehman
Автор

Awesome video! I was wondering if you can share how the process works behind the scenes for cases where we have rows with multiple columns that are null, with respect to the iterative imputer that builds a model behind the scenes. I understand the logic when we only have a single column with null values but can't wrap my head around what will be assigned as training and test data if we have multiple columns with null values. Looking forward to your response. Thanks

fobaogunkeye
Автор

Thanks for posting this. For features where there are missing values, should I be passing in the whole df to impute the missing values, or should I only include features that are correlated with the dependent variable I'm trying to impute?

dizetoot
Автор

Question: If we impute values of a feature based on other features, wouldn't that increase the likelihood of multicollinearity?

-o-
Автор

We can do this for numerical data but what in the case of categoical data?
Can you mention any method for that?

rajnishadhikari
Автор

This imputation return an array as the OHE want a dataframe. How can we solve this if we want to put both inside a pipepline?

joxa
Автор

Hi, I tried encoding my categorical variables (boolean value column) and then running the data through a KNNImputer but instead of getting 1's and 0's I got values inbetween those values, for example 0.4, 0.9 etc. Is there anything I am missing, or is there any way to improve the prediction of this imputer ?

mooncake
Автор

very nice video, however i want to ask, is the knn-imputer can use for data object (string )?

primaezy
Автор

Super helpful, as always. Is IterativeImputer the sklearn version of MICE?

ericsims
Автор

Do you have a recommended tool/package for doing imputation with categorical variables?

dariomelconian
Автор

can iterative imputer and knn imputer works with only numerical values ? Or can it also impute string/alphanumeric values as well?

soumyabanerjee
Автор

Hello ! Thank you very much for your interesting video ! Do you know where I can find a video like this one to know how many neighbors choose ?
Thank you very much

evarondeau