How to statically quantize a PyTorch model (Eager mode)

preview_player
Показать описание
In this video, we go over the theory of how to statically quantize a PyTorch model in Eager mode. We present:
00:00 Intro
03:05 Required Architecture Changes (QuantStubs/ DeQuantStubs/ FloatFunctionals)
08:54 Fusing modules
12:18 Assignment of QConfigs (recipe for quantization for each module),
15:26 Preparing the model for quantization (i.e. making the model fake-quantizable),
20:25 Converting the model to a "true" quantized int8 model.
23:06 Conclusion

Some links to PyTorch documentation:
Рекомендации по теме