Mistral Large with Function Calling - Review and Code

preview_player
Показать описание

🕵️ Interested in building LLM Agents? Fill out the form below

👨‍💻Github:

⏱️Time Stamps:
00:00 Intro
00:10 Mistral AI Company
01:58 Mistral Large
04:12 Benchmark
07:30 Code: Mistral Large
12:36 Code: Mistral Medium
15:19 Code: Function Calling with Mistral Large
Рекомендации по теме
Комментарии
Автор

Nice - good to see a less censored model _with_ function calling 🎉Will hopefully pressure others to follow likewise as anyone who’s used ChaGPT will run across ridiculous “refusals” which can often be overcome by persisting…

RiczWest
Автор

So far I’ve found this to be the best LLM as a code assistant in producing Python code. Was also good seeing it have less censorship and the issues that brings.

Tech_and_AI
Автор

These AI are definitely gonna rule the world someday.

stevensilvaquevedo
Автор

This is interesting. Interested in what you think of my opinion here. I think we should give the bot the broad level instructions with how the flow can and should go, for example the way that it responds to the user etc, in the system prompt. This is all UI stuff after all. The tool itself should have interfaces that are more programatic. For example the input should be "Date in ISO 8601" etc and the output should be "complete" or "done" or a data structure with the response. The LLM should (and in my experience can) then understand these input requirements and output messages, and it should be the one that generates natural language, as apposed to the tool returning natural language. This means, for example that we can change the style or language of the bot without changing the backend tool code.

Cool demo though, thanks.

mikegchambers
Автор

Awesome 😎 !!
And it's cool to see a "small" french company compete with big players !

Can't wait to see how Gemini will add some news to the Functions calling. I mean, I hope they will do a bit differently, more flexible

Wouldn't it be awesome to have a LLM able to set up its own tools ?

joffreylemery
Автор

@Sam - Any chance of doing a video about Autogen Studio 2? I think that your style of video could do some justice to explaining it and extend on the idea of using mistral for function calling or "skills" as Autogen studio calls them.

thesilentcitadel
Автор

Can you create a video of Gemma with function calling?

apexefficiency
Автор

Great video. It would be interesting to know how it'd do if some of the required parameters are not given. Does it ask for them or will it fill them out arbitrarily? Cause this is a problem I've seen with OpenAI models, where they sometimes ask for the missing parameters and other times they fill them with arbitrary values, even when I make parameters required and tell it to ask for missing ones in the system prompt.

jcksn
Автор

This is an interesting model, though it's in kind of an awkward space where if I wanted to do something with it it's a bit impractical without having baremetal access to the model, so I'd generally just use OpenAI, honestly. If I want customization, I'd rather fine tune a model and run it myself, and if I want a big corporate model behind a wall I would just use OpenAI.

I think it might be kind of interesting if they allowed various compute providers (Groq AI, etc) to provide it at a lower cost (and pay some sort of royalty to Mistral) or at higher throughputs so that people could do really custom, super high bandwidth solutions (like scaling with test-time compute) that can require thousands of responses to a single request to pick the most valid solution, as doing that is a bit impractical with OpenAI at the moment as I see it.

novantha
Автор

Is it possible/useful to add a system prompt with specific rules to follow by the model, before starting the proper conversation with the restaurant customer? Or are function calling and system prompt mutually exclusive?

MaximoPower
Автор

i want to build locally the zephyr-7b-beta.. is it possible on an intel mac 16 gb ram?

samfights
Автор

What does native function calling mean in this context? nvm i found the answer in the video

holthuizenoemoet
Автор

Number of tokens is way too little compared to other models in top 5. But I am very happy that there is a non American solution available.

KeyhanHadjari
Автор

Still seemed pretty censored in my testing, any tips?

SloanMosley
Автор

Why there isn't a library inside LangChain which could automatically take care of OpenAI or Mistral function calling.

maxpaynestory
Автор

But this is not free right? We have to pay for it if we want to use it, just like OpenAI GPT-4 model.

limjuroy
Автор

can we take this model from huggingface

uliukbu
Автор

Just wait. It will go haywire just like Gemini. I'm buying puts on microsoft. Ask Google what happens with there stolen code for that model. Just wait.

williamb
Автор

It's a shame this place is full of thieves and shady people.

williamb
Автор

I have actually better result with GPT3-Turbo than GPT-4 when it comes to coding. Don't know why you guys find GPT-4 better.

KeyhanHadjari