Handling Exploding Gradients in Machine Learning

preview_player
Показать описание

This segment addresses the issue of exploding gradients in neural network training, a phenomenon where gradient values increase exponentially, potentially destabilizing the learning process. It presents practical solutions such as gradient clipping to impose thresholds, implementing batch normalization to maintain stable gradient scales, and adjusting network architecture.

Want more machine learning content?

ABOUT US:
Did you enjoy this interview question and answer? Want to land your dream career? Exponent is an online community, course, and coaching platform to help you ace your upcoming interview. Exponent has helped people land their dream careers at companies like Google, Microsoft, Amazon, and high-growth startups. Exponent is currently licensed by Stanford, Yale, UW, and others.

Our courses include interview lessons, questions, and complete answers with video walkthroughs. Access hours of real interview videos, where we analyze what went right or wrong, and our 1000+ community of expert coaches and industry professionals, to help you get your dream job and more!
Рекомендации по теме
Комментарии
Автор

I like his delivery - slow and clear - almost perfect answering machine.

shyama
Автор

Hey!
As per my understanding, skip connections help in vanishing gradient problem and not in exploding gradient problem. Since during backprop, strong gradient gets bypassed through the residual connection and it gets added to the regular weak gradient coming through the hidden layers. For exploding gradients, skip connection doesn’t help, since you add both the gradients in the backprop. So you actually add exploded gradient and skipped gradient which increases the gradient. So let me know if this is wrong.

bhushanghadge
welcome to shbcf.ru