Deploying A Custom Pytorch Model to SageMaker using Terraform, Docker, FastAPI and Pytorch

preview_player
Показать описание
In this video, we show how to deploy a custom Machine Learning Pytorch model to an AWS Sagemaker endpoint, using Infrastructure as Code (IaaC) Terraform, Docker for containerisation and Pytorch for model training and FastAPI for inference.

-------

Рекомендации по теме
Комментарии
Автор

thank you, please continue producing videos

aminasgharisooreh
Автор

Great content! However, it would be even more helpful if you could provide detailed explanations of the steps involved in configuring AWS IAM for Terraform..

huytube
Автор

Hi, Thank you so much for making this tutorial. I have a question, I see that you have defined a "/invocations" endpoint, is it possible to have multiple post/get endpoints and use it? And I have a pretrained CLIP model from transformers library Im using. Essentially my endpoint would take in a video, read frame by frame and insert these embeddings into a vector database. Currently I have it wrapped in a fastapi application deployed to a CPU instance on EC2. My application would benefit a ton from having a GPU and I was wondering if there are any other ways to deploy this on a GPU instance or any recommendations at all :)) I would really appreciate your input!

Tetrax
Автор

How to make the final request via postman?

samsantechsamsan