Markov Decision Processes - Computerphile

preview_player
Показать описание
Deterministic route finding isn't enough for the real world - Nick Hawes of the Oxford Robotics Institute takes us through some problems featuring probabilities.

This video was previously called "Robot Decision Making"


This video was filmed and edited by Sean Riley.


Рекомендации по теме
Комментарии
Автор

This guy was my lecturer about 10 years ago. He was very down to earth and explained the concepts in a really friendly way. Glad to see he's still doing it.

Deathhead
Автор

This channel makes me appreciate the human brain more. We do all that automatically with barely a moment's thought.

CalvinHikes
Автор

OMG as a Robotics student, I'm amazed how well explained that is. Love it <3

mateuszdziezok
Автор

Just took a RL course. Bellman equation and Markovian assumptions are so familiar. Btw, for those who are interested, the algorithm to solve discrete MDP (or model based RL problems in general) are Value Iterations and Policy Iterations, which are all based on Bellman equation.

tlxyxl
Автор

I made these decisions for my real commute. The train was fastest, but occasionally much longer. The car was fast, but the cost of parking equalled 2 hours of work, so was effectively slowest. The latest I could leave and be sure of being on time was walking.

gasdive
Автор

Where the formal definitions for concepts like MDP can get overwhelming, it really helps to have these easy to understand explanations

SachinVerma-lxbx
Автор

Nice one, I met Professor Nick at Pembroke College Oxford. It was an honour.

engineeringmadeasy
Автор

This was a fantastic simple explanation, very enlightening.

tobiaswegener
Автор

There is a 3% chance that, somewhere along the route, there's a half-duplex roadblock because they're fixing the overhead wires or something. There's a 0.1% chance that a power line or tree fell across the road, forcing you to take an extremely long detour, but half of the time this happens, you could get past it on a bike.

pierreabbat
Автор

I heared a lot about MDP and policy functions in the context of reinforcement learning. But this is the best explanation I ever heared.

Ceelvain
Автор

I'd like an autonomous taxi system that would decide it's all too hard to take me to the office, and would just take me back home, or, indeed, just refuse to take me to the office.
"Sorry, I"m working from home today because the car refused to drive itself."

cerealport
Автор

I rarely put a like on a video, but this one deserves it.
I definitely want to hear more about the algorithms to solve MDP problems.

Ceelvain
Автор

This is such a fascinating breakdown of Markov decision making. I love the mathematics that underpins Markov, but the creativity and imagination applied to the example and its host of solutions are delicious brain food.

tristanlouthrobins
Автор

MDP is the topic of my bachelorthesis and the example really helped understanding everything a lot better and I think I'll be using it throughout the thesis to understand the theory I have to write about. It's a lot easier to understand than some state a, b and c and action 1, 2, 3 :D

phil
Автор

the best explanation of this I've ever heard. many thanks.

elwood.downey
Автор

I literally had my final year project use a kalman filter to solve this problem. That's awesome!

Edit: spelling

asfandiyar
Автор

You can read passion in every word he is pronouncing. Very good explanation.

yvesamevoin
Автор

great video! Really well explained and interesting

BobWaist
Автор

Fascination look into decision-making.

lucrainville
Автор

So is there a way to compute the solutions? Like I assume some matrices show up. One for probabilities and one for the sum of times. Then you can multiply it and get different time distributions for every strategy?

Veptis