Azure Cognitive Service deployment: AI inference with NVIDIA Triton Server | BRKFP04

preview_player
Показать описание
Join us to see how Azure Cognitive Services utilize NVIDIA Triton Inference Server for inference at scale. We highlight two use cases: deploying first-ever Mixture of Expert model for document translation and acoustic model for Microsoft Teams Live Captioning. Tune in to learn about serving models with NVIDIA Triton, ONNX Runtime and custom backends.

Additional Resource:

Recommended Next Step:

Microsoft Build 2022
Рекомендации по теме