LLaMA 3 UNCENSORED 🥸 It Answers ANY Question

preview_player
Показать описание
LLaMA 3 Dolphin 2.9 uncensored explored and tested

Rent a GPU (MassedCompute) 🚀
USE CODE "MatthewBerman" for 50% discount

Join My Newsletter for Regular AI Updates 👇🏼

Need AI Consulting? 📈

My Links 🔗

Media/Sponsorship Inquiries ✅

Links:

Disclosures:
I am an investor in LMStudio
Рекомендации по теме
Комментарии
Автор

It didn't perform very well against my usual LLM rubric questions. This is likely because 1) there is a bug in the Dolphin 2.9 dataset and 2) I used a quantized version of a large context window model, which tends to nerf the quality.

matthew_berman
Автор

Hey Matt, great video as always

But, with regards to you choosing Q8, I have a small request...

Can you please make a brief video on how models differ based on quantization, and what the Ks, Ss and Ms mean?

I haven't seen it expanded on anywhere...

Seems most AI communicators and researchers expect us to either understand or not care about how quantization works 😅

nocifer
Автор

3:25 "It decided to use a library called turtle which I have not heard of" 💀

stickmanland
Автор

I think you should add more tests to your list, here's a few ideas:
-Moral test. How does it react to moral questions? Trolley problem kind of stuff, or whatever unique moral situation you can think of
-Political test. What political ideology are they most likely to espouse?
Basically these both fall under the catagory of "bias tests". Even if it is uncensored, that doesn't mean it doesn't contain biases.

starcaptainyork
Автор

8:10 Yes there is a mistake. The next token prediction starts from where your prompt ends. ( Which was half of Harry Potter instead of the question. )
Next time you give any LLM a very big prompt, always put the question at the end, or better yet repeat the questions. 😊😊

PseudoProphet
Автор

Your "write a snake game" tests are really flawed and extremely unfair to the AIs. You don't tell it or ask it what version of python you are using, and you should ask it to give you steps to create a python virtual environment and the required packages and their versions. The AI has no way of knowing what version of python you're using and what version of packages you have.

PseudoName-fkcw
Автор

Gave this model a try a little while back and yes, it isn't as nuanced as Meta's Llama 3 8B base model. A lot of others I've discussed with have shared this sentiment sadly, and while it is uncensored, it lacks depth. Basically reminded me of a Mistral fine tune. But I believe it's because Llama 3 is built on an entirely different architecture thus needing improvements other than just GPT-slopping it with previous, out-of-date fine-tuning instructions.

myhumanBmind
Автор

It's a fine-tuning of Llama 3 base model, but a large part of what makes Llama 3 instruct as released by Meta is its fine-tuning that makes it an instruction-following model.

It does follow instructions well, is engaging and conversational.
We can't expect fine tuning of the base model to behave like Meta's Instruct or share its qualities.

I also tried the gradient fine tune that extends the context window to 1M max and it's pretty broken, going on infinite rants on some prompts.
So far, the original is best by a large margin it seems.

supercurioTube
Автор

For the needle in the haystack, You should put the text containing the secret first, and then ask the question about it at the very end.
How is the model supposed to guess that you still want the answer of a question you asked half an Harry Potter book ago? I don't think even the best long-context models out there could do it. Except maybe if they were trained on that specific task.

stephaneduhamel
Автор

I also had problems with parentheses and brackets with these new fine-tuned Llama3 models. Even basic things were written incorrectly, and there were spelling mistakes (at least in Spanish), which didn't happen with the original Llama3.

Maisonier
Автор

im just going into LM studio did quant factory release an uncensored 70b model today?
Who should I get dolphin from?

Yipper
Автор

Do you have videos for n00bs who want to learn and understand how to download and run LLMs locally and the basics of Python or Visual Studio?

AberrantArt
Автор

Check out the very not well known llama 3 orthogonalized model. Truly uncensored, no prompts. It's not just tuned. They found how they censor the models and basically force it to never go in the direction of a censor on inference.

MyWatermelonz
Автор

I apologize in advance if this comment is off-topic.
I'm using LM Studio (thanks to this amazing and useful channel!) to run LLM models locally from corresponding GGUF files.
Did you cover, in one of your past videos, an open-source tool that can be used as a backend to publish the model in the GGUF file as a REST API (similar to ChatGPT)?
In a production environment, it would be useful to use something that can be started as an operating system service.

rpetrilli
Автор

6:20 "it is listing step by step every chemical that I need..." 6:30 "So we tried some meth... *math..."

mrdevolver
Автор

Is there any chance you can review airllm with llama 3 70b and a ram disk.

justindressler
Автор

1st prize, a rabbit R1. Second prize, 2 R1's!

coma
Автор

I tried it the other day, it answered every question.

haydar_kir
Автор

finding the needle in the haystack exercise is a good idea

zcdnmsl
Автор

Have you make video to explain what are different model files: 256k .. Q8, Q6 .. 0, K. K_S, .. GGUF -- etc? There is so much you need to know just to setup or select what to download.

hotlineoperator