New $10m Open-Source Foundational LLM Is AMAZING! (DBRX by Databricks)

preview_player
Показать описание
Databricks just dropped DBRX, a brand new Mixture of Experts foundational model with 132b parameters. It's really impressive, let's test it!

Join My Newsletter for Regular AI Updates 👇🏼

Need AI Consulting? ✅

My Links 🔗

Rent a GPU (MassedCompute) 🚀
USE CODE "MatthewBerman" for 50% discount

Media/Sponsorship Inquiries 📈

Links:
Рекомендации по теме
Комментарии
Автор

You need to make a table at the end of the video with all LLMs and the results! I don't remember anymore all tests you made!! lol

inteligenciamilgrau
Автор

Databricks is a database and analytics company, created by the founders of spark. Pretty famous in the space for scaling analytics for gigantic datasets for very very large companies. They do AI better than most in the space as well though, they have the right architecture.

PraxeoIogy
Автор

I appreciate the value that your channel provides. Yours is the only one I see who actually shows how to use a lot of this stuff and for that we salute you.

NakedSageAstrology
Автор

I’m a data engineer that uses Databricks regularly with clients, incredible platform and absolutely love the company. Currently using DLT for near real time streaming pipelines, really cutting edge stuff

josephjoestar
Автор

Yaay! Another tutorial. This man is just unstoppable.

japneetsingh
Автор

"DBRX has 16 experts and chooses 4"

MarkDurbin
Автор

Thanks so much Matt for putting this up. I have a program generation piece in one of my upcoming Hackathons. I think I will try this out. You rock! Mike

QuantumAI_LLC
Автор

Quick Clarification: With Mixture of Experts, the routing is decided on a PER TOKEN basis, not on the entire input. This is important and slightly counter intuitive from the name "expert". The tokens are split between two experts out of the total number of experts in the system.

Love the content, keep up the great work!

JohnLewis-old
Автор

Really appreciate you and your channel. You are now my favourite tech journalist. Not just because of your objectivity and integrity; but also because of your natural teaching ability, excellent voice, and stage presence. I'm especially excited for you to expand into interviews. I would love to hear about all the people you look up to in the tech industry. Thank you.

darkhydrastar
Автор

Claude 3 opus is the most accurate coder I’ve experienced personally. I have also been going to groq, Claude 3, and Gemini advanced when a model hits a brick wall and doesn’t know how to fix the code. Off topic, you can get Claude to stop being so uptight and get it to open up if you ask it to. Maybe it’s the way I came about it but Claude 3 opus is closest to agi than any model I’ve tried thus far. I even began creating a new spoken language with it and created rules and began going through English and making words up. It’s using the structure we came up with.

maxlightning
Автор

9:40 the model was spot on, you just didn’t read the entire answer

anticom
Автор

Databricks trained on the snake game to prepare for your test.

Chris-senc
Автор

1) There is a safety filter in the env, if you try to duplicate the space you will be able to configure it there

2) I think it be mad funny to say yippie like drake everytime a model passes a test with flyin colors

BarrelOfLube-clqq
Автор

Thanks for sharing these tests. Very insightful.

fabriai
Автор

Always enjoy your videos. I've found that some LLMs can get the correct answers for "how many words" and "10 sentences ending in xx" by framing the prompt with an actor and a checker.

JonathanStory
Автор

Hey Matthew. For MoE, you keep saying that it's decides which experts to use for a prompt, so it's not using all the experts for a prompt. This isn't right. It's choosing which experts to use for each TOKEN output. So a single output sentence could be using all the models.

johnflux
Автор

Consider that the "hole" test is too broad. A 1' deep hole @ 50' feet wide is vastly different with valid solutions when inverted. I think your premise is too vague and you should be more specific with width and depth to get to a test that more repeatable.

Love the rubrics and I'm totally supportive of even more complicated versions. Thanks Matthew!

briandinello
Автор

The double-exponential feels off as they STILL cannot write like a first grade student (I taught many who could write a sentence ending in apple), and text IS THEIR BEST DOMAIN....Overhype or am I impatient?

devlogicg
Автор

It looks like mixtral next is better that DBRX.

giahuyhoang
Автор

This group of experts models looks like a great way to go and being open source others can optimise it to run on lower performance hardware. One day on my Pi5?

babbagebrassworks