GraphRAG Ollama: 100% Local Setup, Keeping your Data Private

Sdílet
Vložit
  • čas přidán 5. 09. 2024

Komentáře • 83

  • @SullyOrchestration
    @SullyOrchestration Před 2 měsíci +11

    Can you please show a way to visualize the knowledge graph with an interactive UI?

  • @maxs6128
    @maxs6128 Před 2 měsíci +10

    Hey! Cool video. I actually built a full local solution using Ollama, no need for LM Studio at all. Here's what I did: I created a proxy that translates between OpenAI API embeddings and Ollama's format, both ways.
    The cool thing is, it works flawlessly for both global and local queries. I'd be happy to share the script with you if you're interested!

  • @anubisai
    @anubisai Před 2 měsíci +6

    Great work as usual. Humble. Concise. Helpful. Perfect. 👌

  • @ignaciopincheira23
    @ignaciopincheira23 Před měsícem +1

    It is essential to conduct a thorough preprocessing of the documents before entering them into the RAG. This involves extracting the text, tables, and images, and processing the latter through a vision module. Additionally, it is crucial to maintain content coherence by ensuring that references to tables and images are correctly preserved in the text. Only after this processing should the documents be entered into a LLM.

  • @nikhielsingh748
    @nikhielsingh748 Před 2 měsíci +2

    best GenAI CZcamsr, I mean it .

  • @MattJonesYT
    @MattJonesYT Před měsícem

    Looking forward to more on this, it is the most interesting cutting edge tech in AI and almost no one else on youtube is talking about it

  • @macjonesnz
    @macjonesnz Před 2 měsíci +8

    Mervin, Hi from New Zealand, I see that took 20 minutes to index…. what are the specs of your machine?

    • @d.d.z.
      @d.d.z. Před 2 měsíci +2

      Same question. My computer runs Gemma 2 quite slow and I prefer to use Llama3 or Phi. The results will be the same?

    • @dudicrous
      @dudicrous Před 2 měsíci

      @@d.d.z. What's your pc spec?

    • @d.d.z.
      @d.d.z. Před 2 měsíci +1

      @@dudicrous Intel core i5 8th gen 8gb RAM I have a HP pavilion 13-an0012la.

    • @MervinPraison
      @MervinPraison  Před 2 měsíci

      It’s very slow for me too, I use Mac M2 32GB
      In the video I had to cut that part, because it took 26 mins just for indexing and considering it’s small chunk of data

  • @girijeshthodupunuri1300
    @girijeshthodupunuri1300 Před 2 měsíci

    This is not all feasible on my computer but I would love move graph rag videos aiming more at how we can get this technology production ready.

  • @Gwaboo
    @Gwaboo Před měsícem +1

    can you please show or explain how to get the visualization of the data ? looks verry good, and thanks for the tutorial

  • @GeertBaeke
    @GeertBaeke Před měsícem

    Good stuff. As expected, on a Mac M2, indexing and global queries are quite slow. Local queries are doable because it's usually just one LLM call after the similarity & graph search.

  • @MeinDeutschkurs
    @MeinDeutschkurs Před 2 měsíci +1

    A start, great!

  • @xinzhang3502
    @xinzhang3502 Před měsícem +1

    So compared to GPTs, his search generation effect will be better?

  • @Gurdershan
    @Gurdershan Před 2 měsíci +1

    I was eagerly waiting for this, big thanks

  • @NimaAmini2000
    @NimaAmini2000 Před 2 měsíci

    Another great video about GraphRAG, good job.

  • @yingzhu505
    @yingzhu505 Před měsícem

    Great tutorial! Thank you!

  • @alamgirqazi1
    @alamgirqazi1 Před 2 měsíci +2

    You are amazing 🎉

  • @BatukeshwarVats
    @BatukeshwarVats Před měsícem

    Nice and useful video, but still not getting one thing. You made this video around 3 weeks ago, but in april , ollama released some embedding models. Then how we are saying it is not having embedding compatibility?.

  • @song8775
    @song8775 Před měsícem

    Hi, how do you fix the issues with running local search using command line?

  • @nikhielsingh748
    @nikhielsingh748 Před 2 měsíci +1

    quick question, I already have a folder of embeddings and chunks, can I just pass the documents and embeddings to GraphRAG ?

  • @sharankumar31
    @sharankumar31 Před 2 měsíci

    Thank you for this tutorial. Very useful..

  • @MattJonesYT
    @MattJonesYT Před měsícem

    At 7:10 I believe the reason it's giving errors is the url in the settings file is missing the word embeddings at the end. It probably tested some different urls until it figured it out.

  • @JV61
    @JV61 Před 2 měsíci

    This is what I was looking for

  • @nz6291
    @nz6291 Před měsícem

    Thanks this really helped!

  • @lesptitsoiseaux
    @lesptitsoiseaux Před 2 měsíci

    What a perfect video to wake up to after yesterday's video :) I'm starting to think that we're abusing graphRAG here, all of us. You see, and I may be wrong I'm still a n00b here, we are not using semantic chunking and also, for those of us with thousands of files, say transcripts, feeding graphRAG a summary and tags might be good enough for a recommendation engine and if the user wants to dive in, then you use rag but you create a rag for each main collection of documents. So the graph rag may be able to list say what cooking classes you can take much faster and then querying each class that is its own rag for details should be also much faster and overall cheaper? What do you think?

    • @MervinPraison
      @MervinPraison  Před 2 měsíci +1

      Basic RAG is fine for basic tasks but this GraphRAG is for advanced and more meaningful response.

  • @mllearning-qc6dt
    @mllearning-qc6dt Před měsícem

    What is the average query time that you were experiencing with the global/local search?

  • @SonGoku-pc7jl
    @SonGoku-pc7jl Před 2 měsíci

    thanks. coming soon local vs. global

  • @codelucky
    @codelucky Před měsícem

    Can you create a video on how to use GraphRAG with the GROQ API? Looks like nobody has done it yet. Thank you.

  • @eduardoconcepcion4899
    @eduardoconcepcion4899 Před 2 měsíci

    Can't we use nomic-embed-text provided by ollama for the embeddings?

  • @Yannick-ei2tz
    @Yannick-ei2tz Před 2 měsíci

    Thx for the prez. It is about graph so is it possible to get a grip on the underlying graph db and vizualise it using a ds tool ?

  • @jcalonsoc
    @jcalonsoc Před 2 měsíci

    Great video! Can you export a CSV file to visualize the graph using an external tool like Gephi?

  • @shiweiluo
    @shiweiluo Před měsícem

    Why do you read the settings.yaml file by default when you create an index, but mine reads the .env file?

  • @debarghyadasgupta1931
    @debarghyadasgupta1931 Před 2 měsíci

    Would you recommend GraphRAG for structured data as well like Postgres or MySql? I am still stuck between LLM SQL agent vs Vectors. I did explored Vanna already and like it. Appreciate your thoughts on this.

  • @changtimwu
    @changtimwu Před 2 měsíci

    Is LM Studio necessary? I believe OpenWebUI should also suffice, potentially creating a completely open-source graphrag solution.

  • @Gwaboo
    @Gwaboo Před měsícem

    how can we use graph rag on data that's in another language? i have hundreds of documents i want to put into a rag but i cant get a good result out of the rag. the only things that come out are generic and often even things that i didn't ask about. do i have to use a German llm and embeddings model and translate all the prompts into German or do i have to translate all the data into English and live with it being in English?

  • @mohsenghafari7652
    @mohsenghafari7652 Před měsícem

    Thanks

  • @bradmiles358
    @bradmiles358 Před 2 měsíci

    Great stuff, id be really interested to see csv rather than txt import as it looks like it might possibly give higgher reasoning by leveraging the structure. Ie better temporal reasoning etc...
    source_column: "author" # the column containing the source/author of the data text_column: "message" # the column containing the text of the data timestamp_column: "date(yyyyMMddHHmmss)" # optional, the column containing the timestamp of the

  • @mamadousow7323
    @mamadousow7323 Před 2 měsíci

    Didn’t understand the final sentence, running things in llmstudio , what about many pdf documents

  • @rockypunk91
    @rockypunk91 Před měsícem

    I want to understand, how can we use it in a actual application. I reality users will upload their documents anytime they want, If I run indexer for different documents seperately it creates a seperate timestamp based folders in output, now how will the graph rag work when we have multiple artifacts? Our do I have to run indexer on entire documents even if one new document is added? and how do we trigger it programitically

  • @chending5419
    @chending5419 Před měsícem

    After a few tries, my conclusion on graphRAG was that it is buggy when running local, took too long to process - as such that it is not practical to run locally.

  • @lemon268
    @lemon268 Před měsícem

    is this completely free? or are there open api calls?

  • @dove8998
    @dove8998 Před 2 měsíci +2

    You need an OpenAI API Key to run it. It is unfortunatly not 100% local. Is there a way to run it without an OpenAI paid subscription?

    • @linmichael7935
      @linmichael7935 Před měsícem

      yes you can, just export the "string or some text" for the OpenAI API Key. And then you can run it locally.

  • @dawn_of_Artificial_Intellect

    can you use NIM to keep your Data Private?

  • @KS-tj6fc
    @KS-tj6fc Před 2 měsíci

    Do a video that finds the balance of speed (local takes forever) and cost by using DeepSeek-V2, which is only $0.14/million tokens input and $0.28/million tokens output.
    Once you get your results, I would say that at RAG = 2 the Deepseek would be at least equal to or slightly better than GPT 4o at =1.
    Then add an additional python step, prompt to take overall points and run limited amounts of tokens via GPT 4o or even Gemini Flash 1.5, which is quite good, for improvement to 1.6~1.7 level outputs at almost “free token costs”.

  • @mrchongnoi
    @mrchongnoi Před 2 měsíci

    Good video. Please do work out the error you encountered. Do you have a GPU on your laptop. 20+ minutes makes this unusable for a company with 100s or Thousands of documents.

  • @JrTech-rw6wj
    @JrTech-rw6wj Před 2 měsíci +1

    can i use gemini model with graph rag ??

    • @MervinPraison
      @MervinPraison  Před 2 měsíci +2

      Yes, but not straight out of the box as I can see. You might need to modify the code slightly

    • @takshitmathur2761
      @takshitmathur2761 Před 2 měsíci +2

      @@MervinPraison can you make a video on this please.

  • @ShadowyLane
    @ShadowyLane Před měsícem

    Something is off. It works with openapi, no problem... But local models and embedding models - no chance. Followed all instructions.
    EDIT: httpS made me problems for local model, amateur issue. Sorry. It works locally now. Thank you.

  • @PradeepKumar-zy6cd
    @PradeepKumar-zy6cd Před 2 měsíci

    The problem is LM studio getting error

  • @zz-dy7bz
    @zz-dy7bz Před 27 dny

    Are you serious? It's not working at all.

  • @iham1313
    @iham1313 Před 2 měsíci +2

    as long as ms does not add other formats besides txt and csv, this graphrag is useless to me and all the business use cases i know and have.

    • @eggmaster88
      @eggmaster88 Před 2 měsíci +1

      Why? Can't you transform the data?

    • @iham1313
      @iham1313 Před 2 měsíci +1

      @@eggmaster88 can you? including images, diagrams, tables, metadata like page numbers, … please tell me, if there is a good solution

    • @eggmaster88
      @eggmaster88 Před 2 měsíci

      @@iham1313 for tables you can use .csv from what i know but it seems that you don't really need graph database for your data, more like sql.

    • @littledaddi3
      @littledaddi3 Před 2 měsíci

      @@iham1313 look up LangChain GraphRag vids on the topic - they show how to do images diagrams tables and metadata.

  • @zhengwu-jw6fm
    @zhengwu-jw6fm Před 2 měsíci

    GRAPHRAG_API_KEY= "ollama"?big thanks