Scalable ML acceleration with ONNX Runtime

preview_player
Показать описание
Technical session delivered by Manash Goswami, Principal Program Manager at Microsoft.

ONNX Runtime is an open source inference engine for ONNX Models. ONNX Runtime Execution Providers (EPs) enables the execution of any ONNX model using a single set of inference APIs that provide access to the best hardware acceleration available. In simple terms, developers no longer need to worry about the nuances of hardware specific custom libraries to accelerate their machine learning models. This technical session demonstrates that by enabling the same code to run on different hardware platforms using their respective AI acceleration libraries for optimized execution of the ONNX model.

Stay connected with Arm:
Рекомендации по теме