Azure ML: deploy Hugging Face models in minutes!

preview_player
Показать описание
In this video, I show you how to deploy Hugging Face models in one click on Azure, thanks to the model catalog in Azure ML Studio. Then, I run a small Python example to predict with the model.

⭐️⭐️⭐️ Don't forget to subscribe to be notified of future videos ⭐️⭐️⭐️

To get started, you simply need to navigate to the Azure ML Studio website and open the model catalog. Then, you can click on a model to select it. This will initiate the setup process, which takes care of all the required infrastructure for you. Once the setup is complete, Azure ML Studio provides a sample program and you can start testing the model immediately!

If you want to experiment with the latest state-of-the-art models, Azure ML Studio provides a hassle-free way to do so. Try it out and explore the possibilities of cutting-edge AI models with just one click!

Рекомендации по теме
Комментарии
Автор

Hey, did the whole interface change? I don't see the same things you do in "Azure AI Studio". It only has nav items too: Home / Explore / Build / Manage. If you go to "Explore -> Hugging Face" you can only select from a predefined list of models.

daniellasek
Автор

Hey, can you also show how to deploy a custom model not on the model cards?

joejay
Автор

H, nice vid. I got it working. Maybe a bit specific question: Now i'm trying to use langchain with it, but all langchain chatmodel classes (openAI, mistral, olllama etc) use hardcoded endpoint ".../v1/chat/completions" while Azure generates endpoint ".../score". Also when using a local model via LM Studio the prebuilt server also already points to this /v1/chat/completions endpoint. Should i make a custom class extending baseLLM to point it to my custom azure endpoint? or is there already something built for this i can use out of the box?

buksa
Автор

Hi Julien, I'm a product analyst working with my manager to propose a new AI based product for internal use. We are not a tech company.

Can you advice how to know how big an instance would I need to run LLMs with 7B to 14B paramterers. Just ballpark ideas will do for me.

My intention is to give a ballpark cost estimate before we commission a feasibility study with a software vendor who'd give us more accurate numbers

markthompson