filmov
tv
Named Tensors, Model Quantization, and the Latest PyTorch Features - Part 1

Показать описание
PyTorch, the popular open-source ML framework, has continued to evolve rapidly since the introduction of PyTorch 1.0, which brought an accelerated workflow from research to production. We'll deep dive on some of the most important new advances, including the ability to name tensors, support for quantization-aware training and post-training quantization, improved distributed training on GPUs, and streamlined mobile deployment. We'll also cover new developer tools and domain-specific frameworks including Captum for model interpretability, Detectron2 for computer vision, and speech extensions for Fairseq.
Named Tensors, Model Quantization, and the Latest PyTorch Features - Part 1
Quantization in deep learning | Deep Learning Tutorial 49 (Tensorflow, Keras & Python)
Deep Dive on PyTorch Quantization - Chris Gottbrath
Quantization in Deep Learning (LLMs)
Quantization explained with PyTorch - Post-Training Quantization, Quantization-Aware Training
Tensors for Neural Networks, Clearly Explained!!!
Quantization vs Pruning vs Distillation: Optimizing NNs for Inference
What’s new in PyTorch 1.3 - Lin Qiao
GTC 2021: Systematic Neural Network Quantization
Tutorial (TVMCon 2021) - Neural Network Quantization with Brevitas
Understanding 4bit Quantization: QLoRA explained (w/ Colab)
Hessian AWare Quantization V3: Dyadic Neural Network Quantization
Understanding Quantization for Deep Learning
Model Quantization for Edge Devices with AIMET
Quantization in PyTorch 2.0 Export at PyTorch Conference 2022
INT8 Inference of Quantization-Aware trained models using ONNX-TensorRT
Tensor Processing Units: History and hardware
Neural Network Compression – Dmitri Puzyrev
Quantization - Dmytro Dzhulgakov
What’s new in TensorFlow 2.11
Inside TensorFlow: TF Model Optimization Toolkit (Quantization and Pruning)
TensorFlow model optimization: Quantization and pruning (TF World '19)
How to Change Data types of Tensors - Tensorflow Basics
Multi-Dimensional Pruning: A Unified Framework for Model Compression
Комментарии