How Does Rag Work? - Vector Database and LLMs

Sdílet
Vložit
  • čas přidán 6. 12. 2023
  • Join this channel to get access to perks:
    / @python-programming
    If you enjoy this video, please subscribe.
    ✅Be my Patron: / wjbmattingly
    ✅PayPal: www.paypal.com/cgi-bin/webscr...
    If there's a specific video you would like to see or a tutorial series, let me know in the comments and I will try and make it.
    If you liked this video, check out www.PythonHumanities.com, where I have Coding Exercises, Lessons, on-site Python shells where you can experiment with code, and a text version of the material discussed here.
    You can follow me at:
    / wjb_mattingly
  • Věda a technologie

Komentáře • 26

  • @dcpotomac20850
    @dcpotomac20850 Před měsícem +25

    Wow I got more in this 60 seconds explanation than all the lengthy confusing videos I watched on RAG.

  • @24035709
    @24035709 Před 2 měsíci +44

    Simply put. LLMs can't access your company's information directly. RAG lets you share relevant documents with the LLM, so it can answer your questions using your own data.

    • @python-programming
      @python-programming  Před 2 měsíci +5

      precisely!

    • @RamandeepSingh_04
      @RamandeepSingh_04 Před měsícem +1

      But is it safe to connect our database to llms?
      Can the information get leaked?

    • @python-programming
      @python-programming  Před měsícem

      @@RamandeepSingh_04 It is if you use open-source LLMs that can be hosted locally.

    • @RamandeepSingh_04
      @RamandeepSingh_04 Před měsícem

      @@python-programming okay so it means downloading and installing the llm on my device and then using it?And can all the open source llms be downloaded?

    • @stateportSound_wav
      @stateportSound_wav Před měsícem

      @@RamandeepSingh_04yes, but Small Language Models will typically run smoothly on newer phones or PCs with lower-end GPUs, Large LMs for more VRAM capable systems, or likely the soon-to-come ai chipsets.
      Personally my M1 Mac really struggles to run a smaller Dolphin-Llama model, and I’d need to upgrade that one to M3 silicon or newer I think.
      I have 16GB VRAM in my PC GPU, so it might run better on there.

  • @kunallahiri
    @kunallahiri Před 3 měsíci +5

    The Vector database is updated accordingly I guess.
    Nice explanation ❤

  • @lazzybirdflying3225
    @lazzybirdflying3225 Před 13 dny

    Awesome. This 60secs costs a lot!! Thank you for sharing!

  • @alextasarov1341
    @alextasarov1341 Před 2 měsíci +2

    So it's a fancy way of implementing long term memory, cool

  • @Antonio-mn5pq
    @Antonio-mn5pq Před 8 měsíci +7

    Or Real Academia Gallega. An ancient national linguistic institution of Spain.

  • @CodeMaster-w5g
    @CodeMaster-w5g Před měsícem +1

    Great explanation

  • @Mari_Selalu_Berbuat_Kebaikan

    Let's always do alot of good ❤

  • @visualsbysri
    @visualsbysri Před měsícem +1

    Does it serve open domain question answering system?

  • @Rudra0x01
    @Rudra0x01 Před 2 měsíci +1

    Please in next video explain how to create dataset for RAG

  • @jatinnandwani6678
    @jatinnandwani6678 Před 5 měsíci +2

    Thanks so much

  • @tejasaimaheshchennamsetti3247
    @tejasaimaheshchennamsetti3247 Před 3 měsíci +6

    I think you have not explained it completly in this short!, Generative modles provided based on vdb(fixed data set), but when you combaune with Retrieval augmented, now it will provide you with latest data as it communicates and give you the updated and right answer!

  • @willyjauregui6541
    @willyjauregui6541 Před měsícem +2

    Its RAG the same as langchain?

    • @python-programming
      @python-programming  Před měsícem

      You can build a RAG system with langchain but they are two different things. RAG is a workflow while LangChain is a framework.

  • @AlexandreMarr-uq8pw
    @AlexandreMarr-uq8pw Před 2 měsíci +2

    I don’t need how it works I need the demonstration

  • @excusemenoexcusemeno1671
    @excusemenoexcusemeno1671 Před 2 měsíci +1

    I didn't understand.

    • @anuragtekam4588
      @anuragtekam4588 Před 17 dny

      Basically llms are trained on dataset and this dataset may or might not contain relevant information or may contain outdated information. Here in RAG system we store any information in vectorstore and fetch the relevant information when needed. We make sure the vectorstore is updated with updated datasets and allows the model to work without the content in the dataset that the llm was trained on.

    • @excusemenoexcusemeno1671
      @excusemenoexcusemeno1671 Před 17 dny

      @@anuragtekam4588 ohhh