136 understanding deep learning parameters batch size

preview_player
Показать описание
Рекомендации по теме
Комментарии
Автор

So far the most clear and concise explaination

ai-video-stories
Автор

Thanks for the great video. It is useful to see how batch size affects the model.

justinnails
Автор

Grate Video! Thanks for the amazing playlist!
One comment about the batch size analysis:
usually we increase the learning rate with the same rate we increase the batch size! This seems to mitigate the convergency issue shown in your analysis.

Raulvic
Автор

This is a great video - I am happy i found your channel. It is amazing.!

Mohan-social
Автор

Thank you for the great content! As always, very helpful and interesting to watch

microcosmos
Автор

I am a little confused on how the parameters (weights) are updated after the batch has been processed. If two different observations in the training set go through the same nodes in the network, it would seem that the contribution the first observation made to changes in the weights would be lost when the second observation pass through the weights since the weights are not changed until the batch is competed. I am obviously missing something, could someone point me in the right direction.

jameshawkes
Автор

Amazing explanation.... and Amazing demonstration...

saikrishnavadali
Автор

Very informative and great video. I am able to learn it first after watching these videos. While explaining the batch size you mentioned that in 1 epoch the model covers all the samples in 94 iterations. I understand that in each batch operation the weight and biases are updated for those samples and then moved forward for next batch. If by 94 iterations all the samples are already visited then what is the use of 5000 epochs? Could you please explain that too. If someone knows the answer please welcome. Thanks once again for such wonderful videos. I am an Msc student and happily learning from this source.

satinathdebnath
Автор

I have 100milion rows dataset, I want to do preprocessing for NLP (like tokenization, rearranging, label encoding etc..) how should I approach this problem .. pls help me

sarabhian
Автор

Great explanation sir, Thanks for sharing knowledge :)

hritikdj
Автор

thank you so much for the explanation and the striking demonstration!

luciendaunas
Автор

Sir, I am still confuse in this. if we have 500 images and we want to set batch size=20 so 500/20= 25 samples in each batch and epoch size is 5 so each epoch 25 sample will be given to model as forward pass and update weights right ?. my question is after given 25 samples and what about next epoch same 25 samples are given or other 25 samples from dataset which were not shown to model ? please answer my question.

ShahidAli-bkqg
Автор

It seems to me that the optimal batch size is a function of how large the training dataset is. Using your example, you've chosen 32 as batch size with a dataset of 3000 rows. That means each batch is approximately .011% of the dataset. If your dataset was much larger, (for example: 1, 000, 000 rows), wouldn't that imply that you should choose a batch size of 11, 000? that assumes that 11, 000 rows fits within the system ram and gpu ram? Am I on the right track here? (great video!)

lakeguy
Автор

Thank you. Good tutorial. Good topic, well prepared and excellently explained.

frankgkilima
Автор

Thank you very very very much, this is deliciously useful

keeprun-a
Автор

I have seen this very often, that the batch size is 2 to some power (4, 16, 32, 64 etc). Any reason behind that? If you have say 3000 samples, why not use a divisible batch size, such as 50?

MrMikael
Автор

Outstanding explanation !!! I want to know why we need 200 epoch as in each epoch all 1000 data is passing through the model.
Why only one epoch is not enough as each epoch use hole dataset ?

a.k.zraselrahman
Автор

Can use this code with GAN ???
And xtrain what choice? Real or fake image?

asraajalilsaeed
Автор

Why is your batch size the number two to the power of n?

SimonZimmermann
join shbcf.ru