Random Forest in Machine Learning: Easy Explanation for Data Science Interviews

preview_player
Показать описание
Random Forest is one of the most useful pragmatic algorithms for fast, simple, flexible predictive modeling. In this video, I dive into how Random Forest works, how you can use it to reduce variance, what makes it “random,” and the most common pros and cons associated with using this method.

🟢Get all my free data science interview resources

// Comment
Got any questions? Something to add?
Write a comment below to chat.

// Let's connect on LinkedIn:

====================
Contents of this video:
====================
00:00 Introduction
01:09 What Is Random Forest?
02:10 How Random Forest Works
03:53 Why Is Random Forest Random?
04:20 Random Forest vs. Bagging
04:57 Hyperparameters
06:18 Variance Reduction
09:04 Pros and Cons of Random Forest
Рекомендации по теме
Комментарии
Автор

Very clear explaination! Thank you so much!

南南東-sv
Автор

Great Video! Thanks for making it. One minor comment is that at 6:56, sigma^2/k is actually not from CLT, essentially it's just from the basic property of variance.

yuegao
Автор

Keep up the awesome work!, Emma I watched your video one year ago and I got a data science job. Now I start to forget some ML models that I don't use often, it is a very good way to refresh my memory on them!!!

alanzhu
Автор

Looking forward to the notes!! Thanks for sharing, Emma!!!

evag
Автор

Would you mind sharing the notion page with us? Would really appreciate it. :)

Doctor_monk
Автор

Nice lecture, can we get the resource you used. It will be very helpful.

raghu_teja
Автор

A very helpful video on RF. Hi Emma, would you mind actually making a video on how to go about mastering new ML concepts from zero to hero?

ayuumi
Автор

Hi Emma. Thanks for the video. Have a question. I am not sure about how this statement is true. "random forest constructs a large number of trees with random bootstrap samples from the training data". If sample size = replacement, we have all observations in every bootstrap sample. Then, it's not random bootstrap samples. Can you please elaborate what that line is saying?

shawnkim
Автор

Can you clarify how the random feature subset selection happens "without replacement"? Is it that e.g. we have 20 features and tree 1 takes 10 features, tree 2 takes the remaining10 features and now tree 3 can take 10 from the original 20?

davidskarbrevik
Автор

What happens if RF n_estimators(individual decision trees) have conflicting outcome as in 50% of them voted/predicted class A while the other 50% voted/predicted class B.
In this situation, what would be the final outcome??

imran
Автор

can we say if interview ask which algorithm can be used here, and we don't know the Ans we can surely apply random forest here.🤔😜

shubhamkaushik