Reinforcement Learning with Stable Baselines 3 - Introduction (P.1)

preview_player
Показать описание
Welcome to a tutorial series covering how to do reinforcement learning with the Stable Baselines 3 (SB3) package. The objective of the SB3 library is to be for reinforcement learning like what sklearn is for general machine learning.

Рекомендации по теме
Комментарии
Автор

By the way, when you were comparing the models you were still using
Which is why they were almost the same and didn’t look like they were learning

AakashKumar-gtip
Автор

I have been using stable baselines 2 for last year or so for my work and it's super convenient, the docs are great, great examples for custom env etc. It's a great library.

ashu-
Автор

This is very useful. I'm working on an RL video series myself (the theory side, so no overlap here) and I was just looking for prebuilt RL algos. Stable baseline's 3 is by far the most complete/well tested suite I've come across. This really makes a big differences - thanks!

Also, it's nice to see super technical coverage like can yield a 1M+ followers. Awesome.

Mutual_Information
Автор

SentDex youre a legend, brother. The thought of implementing these using deep learning libraries alone, instant grey hair! Thank you

vernonvkayhypetuttz
Автор

Your videos always inspire me to continue working on my own projects!!!

hendrixkid
Автор

Even without watching it . Thanks for your good work and content sentdex

thetiesenvy
Автор

If you're following along using a Conda environment and the Lunar Lander environment gives you an error (namely "module 'gym.envs.box2d' has no attribute 'LunarLander'") then I found that you need to also install two other packages; swig and box2d-py:

conda install -c conda-forge swig box2d-py

OhItsAnthony
Автор

LETS THIS IS EXACTLY WHAT I WANTED THANK YOU SO MUCH

pfrivik
Автор

Honestly loving this series, i hope you make a indepth tutorial series on this. Thanks

amogh
Автор

I had so much fun learning with you.... can't wait to follow you again after completing my web project

alishbakhan
Автор

wow, great video! really can't wait for the rest to come out and learn more.
Thanks for all the info you provide us!

enriquesnetwork
Автор

Thanks for introducing the Stable Baseline 3,
and yeah sometime we forget to use model!

VaibhavSingh-lfps
Автор

I think you were still getting random results because you still had the .sample method call in the rendered tests for A2C and PPO. They learned, but you did not use the trained model for testing.

OTheIDaveO
Автор

Happy New Year SEndex, was learning machine learning during the lockdown & I had no idea in the Field . U teach so well

tytobieola
Автор

Thank you, these video tutorial will be of big help to my thesis. I going to support you.
I have many doubts I hope this can resolved them.

arthurflores
Автор

awesome video, learned a lot, keep up the good work

ahmarhussain
Автор

Thank you for this tutorial. I am just getting into AI. It is over my head immediately, but your overview of the parts such as observation and agent were helpful for the bigger picture.

KennTollens
Автор

Great series as always...needs the next step, developing asynchronous (multiprocessing) models, eg: PPO into Asynchronous-PPO (APPO) on custom environments...Thx

markd
Автор

Can you please talk about how we use the RL to model and optimize satellite networks and HAP( high altitude platforms)??

How we control the direction and angle of a projector embedded into HAP or UAV so that it directs its light beams towards an special area of interest on the Earth??

AIdreamer_AIdreamer
Автор

This is really interesting and new to me! You mentioned going over creating custom environments in future videos which sounds like exactly what I am eager to know next so I’m really looking forward to that video! Is there anything I should educate myself on in the meantime?

DaZMan