Model Quantization in Deep Neural Network (Post Training)

preview_player
Показать описание
#machinelearning #neuralnetwork #quantization

In this video we talk about post training model quantization that allows for reduced precision representations of weights and possibly activation. This helps reducing the storage footprint of model as well as computational need. There are many advantages of quantization especially when models are expected to be deployed in low compute or low storage like mobile or on edge of IOT devices

Below are some article where you can read more on this topic

There are also options of quantization aware training that you can read below

Рекомендации по теме
Комментарии
Автор

Thanks for the helpful video! Very clear~

kyoungsub
Автор

I hope this channel gains popularity between the AI community.

RamkrishanYT
Автор

I am reviewing it again and just hoping that the voice quality can be better next time!
One question is that in 1:18, you mention that the quantization can increase the performance of the model.
May I know in what perspective are you talking about?

kyoungsub
Автор

is it true that accuracy of the model reduces because of quantization?

prudhvirajboddu
Автор

Great Video Sir!
Are You In Twitter loves to follow you!

HeyFaheem