LLMs as Tool Makers [LATM] - GPT-4 *UPGRADES* lower AI Models.

preview_player
Показать описание
🔥 Get my A.I. + Business Newsletter (free):

#ai #openai #gpt4
Рекомендации по теме
Комментарии
Автор

Wow, the quality of this video completely blew me away, you did an amazing job reading, digesting and summarizing this paper, I'm really surprised.

The majority of AI channels in YouTube are just hype makers, they just talk about new AI software that looks fancy in the moment, but you dove deep in this important paper and explained its impact on the field. I'm far more interested in this kind of AI content than on the new tools itself.

Keep up doing this, the effort definitely pays off, as very few people talk about Artificial Intelligence Papers.

There's another paper that came out this week about the Orca model that Microsoft created and it's planning to open source. That paper was as groundbreaking as this one in the video, Microsoft Research showed that you can create a model as capable as GPT4 with just 13B parameters, only using chain of thought questions and answers from GPT4 as the training data. This could change everything in the field and eventually leading to models that can run on smartphones, as they get smaller and smaller.

I'd love to hear you talking about it.

GabrielLima-ghwe
Автор

this is the only video on LATM on youtube. I think its the best AI has to offer atm.

scottlacey
Автор

I think the next leap forward with LLMs will be more about the tools than about the LLMs themselves. Some of these tools, like Wolfram Alpha, will just be "found" and the AI only needs to be good at understanding what it can do and how to format queries. The Minecraft interface is another example of that. Other tools will be made by the AI, but they don't need to be deleted when a session ends. They can be stored in a queriable tool chest and reused. The fun part will be in tool verification, and you might think this would lead to circular dependencies, but it doesn't have to. It might take extra training or even maybe a specialized LLM, but the question "does this tool reliably and efficiently do what it's supposed to?" to be something that the fancy LLMs could handle if you give them good tool-testing tools. The LLM should be prompted to design a test battery to verify that the tool leads to correct outputs, to be reflective about what means it has to make sure it knows what results are right for the test problems independently of the tool, and to be reflective about potential edge cases and what "correct behavior" should look like in those cases. In this process it could query SOTA LLMs, but honestly it should even be querying humans for the hardest questions. Expert human trainers could exist in an addressable pool, and get pinged when the language model needs to verify something or is otherwise stuck. An important feature of such an AI would be its skill in formulating clear and telling questions for the humans while concisely and accurately describing the context of its confusion.

davidhoracek
Автор

I'm loving how engineering LLM applications to get the most out of these models are a big focus in the papers being released this year, i wonder how "LLM-ops" will be shaped by these approaches

aaronward
Автор

So a future model for an AGI can just be a vast network of state of the art models making tools and delegating tasks to faster personal models

jossefyoucef
Автор

You just gave me a idea to try and LangChain.

No, just with the LATM, but combining a whole entire governance system that goes beyond simple tool conceiver and to user but also a governance framework that watches over these two agents.

What we may not immediately realize with any of this is that even with a lot of the new papers that are coming out, people can immediately use already existing tools within the LLM, universe, two prototype any, and all of this, without waiting for any of the large fang stock companies to even get involved. Their entire research that they release can immediately within a week if not sooner be implemented into an idea.

_________
Автор

We need to start talking about the ethics of Tool Making.

So far most of the responsibility seems to be on the user, but I hope this progress of insights on tool making makes OBVIOUS that the makers of the tool making system are pushing this forward, so they should be also reviewing the ethics side.
We cannot continue to ignore it in these conversations, or any conversation really.

LorettaBangBang
Автор

wow, I just asked Bard to summarize the paper, and it hallucinated a whole thing about Bayesian Optimization with Neural Architecture Priors (BONAP) 🤣 had me going for a sec, did a ctrl + f of the doc and none of what it wrote about was there

GNARGNARHEAD
Автор

Is it just me or is this looking more and more like actual cultural evolution among AI? I'm joking ofc, but only partially, especially because of the wording they chose, "tool making". There are some funny similarities with historical tool making and subsequent advancements. This area of technology is truly in its infancy still.

etunimenisukunimeni
Автор

Very good video. Ur channel seems smart. The smart model starting and the other cheaper effitient onces completing is very smart and should be looked into. Absolutely also super effitient tool LLMs maybe too Where the smart model tells what to do that is out of the cheap models capabulity but does as little of the other stuff as possible even if it has to correct the other output and it might be more effitient/cheaper like that. Like a baby programmer that programms and the genious just checks it isntead of letting the genious write boring repetitive easy funcitons he focuses his high paying working hours for more important stuff (opportunitycosts)

hanskraut
Автор

So basically, hey GPT can you please make a game! Sure, let me generate the tools to do this... its like AutoGPT but more specific to tool making. Amazing.

skippersthepenguin
Автор

"Keep that one alive". We laugh now...

neilo
Автор

Oh the overinflated mysterization of llms. I've been using gpt-4 to write code and 3.5 turbo to execute work for me, automatically, for two months now. It's an entirely logical conclusion.

pizzaiq
Автор

For credit and for my benefit. Can you provide the link to the paper, please?

Press return, , click comment

ilozeet
Автор

gpt3.5 is like an infant and gpt 4 is like a young adult. Bard is like a baby

LG-bbzq
Автор

it's hard for me to not see that we are creating a tail-risk of extinction!

Learna_Hydralis
Автор

Yeah the llms are using the tools themselfs and not he humans are thinkering. Sounds like a very grounded opinion 😂 based on real world experience with llms and a deep understand how reliably they work.

davidw
Автор

Except Python language was built by humans for the usage of humans.
Asking AI to use Python is like asking an elephant to play piano.
Why not allow AI to design their own perfect tool (language) with Assembly from the ground up?

kyber.octopus
Автор

yo get those clickbaity, unrelated thumbnails of steve jobs off this post

montagistreel