How-To Run Llama 3 LOCALLY with RAG!!! (GPT4ALL Tutorial)

preview_player
Показать описание
In this Llama 3 Tutorial, You'll learn how to run Llama 3 locally. Unlike most other local tutorials, This tutorial also covers Local RAG with llama 3. we'll use a tool called GPT4All for the same.

GPT4ALL is free, open source and hence get started with Llama 3 Local.

This tutorial also takes a fully no-code approach to run llama 3 without command prompt or terminal.

in short, it's a one-click llama 3 installer tutorial!

Chapters:

00:00 Intro
00:19 Download GPT4ALL
00:52 Download Llama 3 Instruct Model
01:42 Chat with local Llama 3
03:30 Download the embedding model for local data ingestion
03:58 Add Documents Folder to Collection
04:48 Chat with PDFs using Llama 3 Locally
05:54 Advanced GPT4ALl Settings (model parameters, prompt template, system prompt)

🔗 Links 🔗

❤️ If you want to support the channel ❤️
Support here:

🧭 Follow me on 🧭
Рекомендации по теме
Комментарии
Автор

Awesome, waiting for such tool since long, for this tutorial thank you so much

om_ashish_soni
Автор

Dunno if i'm more thankful for this gpt4all or this workflow either way great video, thank you!!

build.aiagents
Автор

This is great, GPT4ALL is rarely mentioned by other AI channels, so I didn't know how much it advanced. Buit in RAG is nice.

PS Eventually I forsee all local LLM software integrating a combination of RAG, code interpreter, and agentic frameworks. Even though the latter two can already be used through API

TomM-po
Автор

Amazing, that's the best AI tool I've tried so far ! The only problem it has is that the exe file isn't in an obvious location. Some people could have problems with that but otherwise its great.

Raphy_Afk
Автор

great video, can you please make a video about lmstudio or jan local api ? how to setup and use in local application/ script?

robins.storey
Автор

Have you discovered any good techniques in the system instructions for grounding llama3 on the datasource better?

prestonmccauley
Автор

Thank you for the Video. This is really cool. I work for an organization. For a prototype, i cannot go this approach. Is there any open source code base so that i can setup UI in a server and run RAG ? Can you please suggest on this?

connectingdots-
Автор

why its so fast in GPT4all, for me too. LM studio is too slow for me (1 token per second) but on my same computer GPT4all works fast (5 token per second). same model

nikhilmish
Автор

I can't find SBert or the other embeeding models

santiagomartinez
Автор

Please make a video on how we can do function calling in ollama open web ui or how we can enable internet access to it

mernik
Автор

1) Can you use custom tuned models downloaded from other sources? 2) If I have a good GPU, can it use it to work faster?

KlimovArtem
Автор

Where is the embedded data stored? Will the model auto pick up new files in the same folder? Or do I need to do the embedding process again if I want to add more files

rosemaryng
Автор

Nice video!!! Do you know how to use python trigger this function?

yi-chuan
Автор

Using llama3, which app do you consider is better for RAG? This or ollama web ui or LLM studio?

shopyx
Автор

I want to ask a question about the Streamlit-pyodide video to make sure I understand it correctly. Since the website I created with Streamlit and pyodide uses the processor and gpu of the client side, it will not create an extra load on the server of the website, right?

evrengullupunar
Автор

I tried to run it and it keeps crashing after loading the model, if anyone has any idea, the error log seems to be a .dll file.

mimotron
Автор

Can it be used for files other than pdf?

rahuldinesh
Автор

I have Jan and LM studio, RAG cant be achieved with those ?

bilalalam
Автор

Can you please provide information about How we can run 70B parameter models locally using RTX 4090.
Or we should choose other GPU which can handle same even faster and GPU costing similar price range.

I want to run 70B parameter ai models locally. Could you please help me to build PC within budget of 4500 $, by providing best configuration for PC.

parthwagh
Автор

Is there a way to use llama3 locally via the program?

searchhello