Sparse Activation- Game-changer for the Future of Deep Learning. Devansh Machine Learning Techniques

preview_player
Показать описание
Devansh Machine Learning Made Simple explains why Sparse Activation is such a game-changer in the field of Deep Learning, Data Science, and AI Research. It is recently being given attention because of its use in Google's new Large Language Model (Transformer) called Pathways Language Model (PaLM), “a 540-billion parameter, dense decoder-only Transformer model trained with the Pathways system”.

Adding new parameters to Machine Learning models allows us to add more scale and functionality to our Machine Learning Models.

Adding more params→exponentially increasing runtime. Because the inputs go through network in forward and backward passes in Neural Networks

What if instead of running the entire network for every input/task, we ran the network for a small part of the network. This would allow us to train our model for lots of capabilities, but would make running the model very cheap. Sounds like a dream right?

This is one of the ideas behind Google’s new Large Language Model (Transformer) called Pathways Language Model (PaLM), “a 540-billion parameter, dense decoder-only Transformer model trained with the Pathways system”. To learn more about it, and the system that allowed it to have such great results, read the article in the comments, “Google AI sparks a revolution in Machine Learning.”

Like the video, and follow to stay updated. Will do a more detailed video as I read through the papers.

This is part of my Machine Learning Techniques Playlist. Machine Learning Techniques focuses on how different research teams and problem solvers use various Machine Learning Tools and Techniques to solve various problems. Seeing their innovative approaches should show you the various ways we can combine and use AI and ML tricks and tools to break down very complex problems. It will interest you and hopefully inspire you

Like the video and Follow across other social media to learn more and stay updated about AI.

For monetary support, check out the links below. Any amount helps a lot and is greatly appreciated. Donating provides lots of benefits like special sessions, annotated papers, and discounted consulting rates.

Reach out to me:

#MachineLearning #DeepLearning #Devansh #MachineLearningMadeSimple #Math #Data #NeuralNetworks #AI #ArtificialIntelligence #SoftwareDevelopment
Рекомендации по теме
Комментарии
Автор

How do you decide between sparse activation vs multiple smaller networks? Also, is the entire network designed to solve one task/ problem and what would an example of such a problem be?

daanishdan