Deep Learning hardware acceleration with AWS Inferentia

preview_player
Показать описание
In this video, I show you how to get started with AWS Inferentia, and Amazon EC2 Inf1 instances.

Follow me on:
Рекомендации по теме
Комментарии
Автор

Hi Julian, hope you are fine.
I have made a REST API with API Gateway, Lambda and sagemaker endpoint, the time that model take for inference is hardly 1sec, but the API on average takes about 5-9 secs, and sometimes can also take more than 30 secs.
SO, how to make it response fast.
Kindly, guide me to any relevant blog, or video.

Thanks a lot...

MuhammadAli-migg
Автор

Can we use Inferentia, for deploying Pytorch endpoints as REST APIs.Like by integrating with Lambda, and API Gateway..?
Thanks..

MuhammadAli-migg
Автор

Julien, tu ne fais plus de video/talk in Fr ? only in Eng ? Did you move ?

ericd.
Автор

Hi Julian:
Kindly help me with the following error:

UnexpectedStatusException: Error for Compilation job Failed. Reason: ClientError: CompilationError: Unable to compile model for ml_inf1:', 'Please implement aten::alias in native_ops/aten.py')

Thanks for any help...

MuhammadAli-migg