The CRAZIEST LLM Fine-Tuning I've seen, And It WORKS!!!

preview_player
Показать описание
Mistral AI Hackathon winners fine-tuned Mistral 7B to play doom.

Imo, it's the most innovative and craziest LLM fine-tuning I've ever seen.

This video dives into the building of Mistral dooM!

🔗 Links 🔗

Ref 3 -

❤️ If you want to support the channel ❤️
Support here:

🧭 Follow me on 🧭
Рекомендации по теме
Комментарии
Автор

It is a meme: We reached AGI, but does it run Doom?

bourdainedepiment
Автор

"If you are very old at this point." I take offense to this statement. I'm 45...I didn't think I was old...until now.

king
Автор

Hmm. Looks like a fantastic test case for Groq’s increased token output.

pstefan
Автор

So cool! I had the idea to do it with Cataclysm: dark days ahead !!! I'm sure many people had this idea but it is the first time I see someone do it.

Imagine replacing NPCs in Cataclysm or Dwarf Fortress by LLMs :) That's what I would like to see next...

sinitarium
Автор

This means we can make every LLM multimodal.

thepabli
Автор

we are truly living in the most interesting of times indeed XD
getting a model to play and survive in Dwarf Fortress is obvs the next logical step, especially considering the base game was only ASCII

CYIERPUNK
Автор

Yea, that is pretty deep. Thanks for the video.

marcfruchtman
Автор

Damn. I was going to write a paper on this, and I had started setting up. Didn't know if it would actually work.

thedoctor
Автор

"if you are ... very old..." ...

bstbuddy
Автор

I remember getting “Doom” confused with “Dune” when I was growing up.

lun
Автор

I think the ASCII representation is an overhead map state. not the frame itself. that way it's easy for the LLM to move because it understands where it is in the level space at every turn.

s_the_first
Автор

I made a data set that trains the LLM the think before speaking and works amazing with mistral.02. I made another one inspired by this where the ai is encouraged to make a “mental image” or ascii art for each prompt

spencerfunk
Автор

So, this method calls for some ideas :))

muhammadrezahaghiri
Автор

It's like giving someone sight through electrodes on their tongue!

soccerkenshin
Автор

Man... Where do people get these ideas? 😅

faaz
Автор

ok legit question, How the hell do you deal with tokenizers

BhaswataChoudhury
Автор

It's not news, people have been generating images with llm since gpt2. And model has at least seen ascii and base64 images in the internet scrapped training data

timmygilbert
Автор

I wonder if it simply maps out the layout based on all the limited options and also comprehends like AI image enhancers.

ThankYouESM
Автор

If it exist Doom will be run on it or now by it, I guess. Things have become very strange.

SiCSpiT
Автор

Why do this work ? Because LLM are predictive model, they finetune it on 2048 characters paired with an action. LLama see a text and predict an action.
Now the challenge would be to augment the resolution and types of characters used without exploding the context window.
I remember reading somewhere that LLM can handle compressed text with a dictionary.

vincentvoillot
welcome to shbcf.ru