filmov
tv
How to statically quantize a PyTorch model (Eager mode)
![preview_player](https://i.ytimg.com/vi/hlcPz7wuf3M/maxresdefault.jpg)
Показать описание
In this video, we go over the theory of how to statically quantize a PyTorch model in Eager mode. We present:
00:00 Intro
03:05 Required Architecture Changes (QuantStubs/ DeQuantStubs/ FloatFunctionals)
08:54 Fusing modules
12:18 Assignment of QConfigs (recipe for quantization for each module),
15:26 Preparing the model for quantization (i.e. making the model fake-quantizable),
20:25 Converting the model to a "true" quantized int8 model.
23:06 Conclusion
Some links to PyTorch documentation:
00:00 Intro
03:05 Required Architecture Changes (QuantStubs/ DeQuantStubs/ FloatFunctionals)
08:54 Fusing modules
12:18 Assignment of QConfigs (recipe for quantization for each module),
15:26 Preparing the model for quantization (i.e. making the model fake-quantizable),
20:25 Converting the model to a "true" quantized int8 model.
23:06 Conclusion
Some links to PyTorch documentation: