How to Choose AI Model Quantization Techniques | AI Model Optimization with Intel® Neural Compressor

preview_player
Показать описание
Learn the fundamentals of AI model quantization. Your application and project have unique requirements, so there are a variety of quantization techniques. See an overview of each technique, tradeoffs, and recommended applications.

AI model quantization is one of the most popular ways to optimize models for deployment. Reducing word lengths of weights and activations reduces the model size and can speed up inference. However, there are a variety of techniques to choose from.

Learn the first principles of what is required to quantize floating-point models to integer formats. This is followed by an overview of each of the main model quantization approaches, covering the amount of effort required and the benefits of each. Subsequent videos in this series will cover each technique and how to use them in Intel Neural Compressor.

About the AI Model Optimization with Intel® Neural Compressor Series:
Learn how to choose and get started with AI model optimization techniques. Get started with examples using Intel® Neural Compressor, which works within PyTorch*, TensorFlow*, and ONNX* Runtime

About Intel Software:
Intel® Developer Zone is committed to empowering and assisting software developers in creating applications for Intel hardware and software products. The Intel Software YouTube channel is an excellent resource for those seeking to enhance their knowledge. Our channel provides the latest news, helpful tips, and engaging product demos from Intel and our numerous industry partners. Our videos cover various topics; you can explore them further by following the links.

Connect with Intel Software:

Powered by oneAPI

#intelsoftware #ai #oneapi

How to Choose AI Model Quantization Techniques | AI Model Optimization with Intel® Neural Compressor
Рекомендации по теме
Комментарии
Автор

How can i define my own quantization methods?

AnthonyKostalvazque