How Does Rag Work? - Vector Database and LLMs
Vložit
- čas přidán 6. 12. 2023
- Join this channel to get access to perks:
/ @python-programming
If you enjoy this video, please subscribe.
✅Be my Patron: / wjbmattingly
✅PayPal: www.paypal.com/cgi-bin/webscr...
If there's a specific video you would like to see or a tutorial series, let me know in the comments and I will try and make it.
If you liked this video, check out www.PythonHumanities.com, where I have Coding Exercises, Lessons, on-site Python shells where you can experiment with code, and a text version of the material discussed here.
You can follow me at:
/ wjb_mattingly - Věda a technologie
Wow I got more in this 60 seconds explanation than all the lengthy confusing videos I watched on RAG.
Simply put. LLMs can't access your company's information directly. RAG lets you share relevant documents with the LLM, so it can answer your questions using your own data.
precisely!
But is it safe to connect our database to llms?
Can the information get leaked?
@@RamandeepSingh_04 It is if you use open-source LLMs that can be hosted locally.
@@python-programming okay so it means downloading and installing the llm on my device and then using it?And can all the open source llms be downloaded?
@@RamandeepSingh_04yes, but Small Language Models will typically run smoothly on newer phones or PCs with lower-end GPUs, Large LMs for more VRAM capable systems, or likely the soon-to-come ai chipsets.
Personally my M1 Mac really struggles to run a smaller Dolphin-Llama model, and I’d need to upgrade that one to M3 silicon or newer I think.
I have 16GB VRAM in my PC GPU, so it might run better on there.
The Vector database is updated accordingly I guess.
Nice explanation ❤
Awesome. This 60secs costs a lot!! Thank you for sharing!
So it's a fancy way of implementing long term memory, cool
Or Real Academia Gallega. An ancient national linguistic institution of Spain.
Lol! Love it
Great explanation
Let's always do alot of good ❤
Does it serve open domain question answering system?
Please in next video explain how to create dataset for RAG
Thanks so much
I think you have not explained it completly in this short!, Generative modles provided based on vdb(fixed data set), but when you combaune with Retrieval augmented, now it will provide you with latest data as it communicates and give you the updated and right answer!
Yeah this is one of the major benefits.
Its RAG the same as langchain?
You can build a RAG system with langchain but they are two different things. RAG is a workflow while LangChain is a framework.
I don’t need how it works I need the demonstration
I didn't understand.
Basically llms are trained on dataset and this dataset may or might not contain relevant information or may contain outdated information. Here in RAG system we store any information in vectorstore and fetch the relevant information when needed. We make sure the vectorstore is updated with updated datasets and allows the model to work without the content in the dataset that the llm was trained on.
@@anuragtekam4588 ohhh