Llama-3 🦙 with LocalGPT: Chat with YOUR Documents in Private

Поделиться
HTML-код
  • Опубликовано: 27 ноя 2024

Комментарии • 39

  • @engineerprompt
    @engineerprompt  6 месяцев назад

    If you are interested in learning more about how to build robust RAG applications, check out this course: prompt-s-site.thinkific.com/courses/rag

  • @engineerprompt
    @engineerprompt  6 месяцев назад +1

    Want to learn RAG beyond basics? Make sure to sign up here: tally.so/r/3y9bb0

  • @vetonrushiti19
    @vetonrushiti19 5 месяцев назад +1

    does localgpt work in an ubuntu machine without nvidia gpu?

  • @soarthur
    @soarthur 3 месяца назад

    This is very interesting and great work. There is the Mozilla project called llamafile which makes running local LLM with one simple executable file. It also can use CPU instead of GPU intensive. LLamafile makes running LLMs on older hardware possible. It has great performance improvement. It will be great if LocalGPT can work with LLamafile. Thank you.

  • @TraveleroftheSoul7674
    @TraveleroftheSoul7674 6 месяцев назад

    there is a problem in the code. Even when I ingest new files it's still gives answer and make mess with the last file I deleted. How to handle this. I tried different prompts but it's not working for me?

  • @thegooddoctor6719
    @thegooddoctor6719 6 месяцев назад +3

    By Far the LocalGPT is the most robust RAG system out there - Thank you - But I'm running it on a i9 13900/4090 GPU system - Is there any plans on making the RAG system a bit faster - It can take up to 5 minutes to come back with a response...... Thanks again - Very Cool...

    • @engineerprompt
      @engineerprompt  6 месяцев назад +1

      Yes, I am experimenting with using ollama for the LLM and I think that will increase the speed. Working on major updates, stay tuned :)

    • @laalbujhakkar
      @laalbujhakkar 6 месяцев назад +2

      on m2 mbp 16gb with ollama+llama38b+anythingllm is returning in. seconds …

    • @thegooddoctor6719
      @thegooddoctor6719 6 месяцев назад +1

      @@laalbujhakkar Then again I'm having it search 300 MB of documents.........

  • @Ankara_pharao
    @Ankara_pharao 6 месяцев назад +2

    May i use llama3 with languages other then english?

    • @sauravmukherjeecom
      @sauravmukherjeecom 6 месяцев назад +3

      Yes you can. Out of the total training data around 5 or 10 percent (forgot now) is languages other than English. Which is close to the total training data for llama 2.

    • @engineerprompt
      @engineerprompt  6 месяцев назад +1

      Yes, you can as pointed out. You also want to make sure to use a multi-lingual embedding model.

  • @zahidahmad1894
    @zahidahmad1894 6 месяцев назад

    I want a specific conversational chatbot with very few amount of data. How can I do it?

  • @EDRM-my5rd
    @EDRM-my5rd 6 месяцев назад

    I tested the ingest and query model with PDF edition of FINANCIAL ACCOUNTING International Financial Reporting Standards ELEVENTH EDITION using default parameters and answers were 80% wrong, particularly with sample journal entries from the context:
    > Question:
    provide example of VAT journal entries
    > Answer
    * The sales revenue is recorded as a debit to the "Sales Revenue" account, which increases the company's assets.

  • @kingfunny4821
    @kingfunny4821 6 месяцев назад +1

    can use this offline
    and
    Can I save the conversation so that I can refer to it after a period of time or when creating a new conversation?

    • @sauravmukherjeecom
      @sauravmukherjeecom 6 месяцев назад +1

      Yes,
      For memory you will have to send the past conversation as context. Try looking into one of the rope trained models with longer context length.

    • @bobby-and2crows
      @bobby-and2crows 6 месяцев назад

      Yeah fella

    • @engineerprompt
      @engineerprompt  6 месяцев назад +2

      This is for offline use. localgpt has a flag save_qa that will enable you to save your conversations and you can load them.

  • @pablolbrown
    @pablolbrown 6 месяцев назад

    Any idea when support for Apple Silicon M3 is coming?

    • @engineerprompt
      @engineerprompt  6 месяцев назад

      It already supports Apple Silicon. Make sure you correctly install the llamacpp version. Instructions are in the Readme

  • @azizjaffrey123
    @azizjaffrey123 6 месяцев назад

    Please keep this code version for future use, if you update code and if people cannot find code from this video they skip , which i personally did on your old video on LocalGPT and started watching this but for my gpu old code was compatable but cannot clone, since that version doesnt exist

  • @NovPiseth
    @NovPiseth 6 месяцев назад

    Hello thanks for great video you help me alot about this. Could you help me to add Panda and PandaAI? it could help me to analys the data from the excel and/or csv file. Thanks

  • @adityamishra611
    @adityamishra611 6 месяцев назад

    I am getting this error: You are trying to offload the whole model to the disk

  • @o1ecypher
    @o1ecypher 6 месяцев назад +1

    a .exe or a gui for windows would me nice gradio like stable diffusion please

  • @Player-oz2nk
    @Player-oz2nk 6 месяцев назад

    Very interested in how to correctly ingest csv files and formats and limitations

    • @sauravmukherjeecom
      @sauravmukherjeecom 6 месяцев назад

      Csvs are tricky. You can either go by adding the data to a database and then querying on it. Or create text chunks out of it.

    • @Player-oz2nk
      @Player-oz2nk 6 месяцев назад

      @@sauravmukherjeecom assuming foe larger cvs importing directly to db would make more sense and smaller file we could chunk

  • @shaonsikder556
    @shaonsikder556 6 месяцев назад

    Which screen recorder do you use?

  • @zahidahmad1894
    @zahidahmad1894 6 месяцев назад

    4gb gpu 16 gb ram. Will llama3 work fine?

  • @colosys
    @colosys 6 месяцев назад

    Could you help me configure localGPT with pgvector embeddings? :$ I'm seriously struggling

  • @ai-folk-music
    @ai-folk-music 6 месяцев назад +1

    Why use this over something like AnythingLLM?

    • @engineerprompt
      @engineerprompt  6 месяцев назад

      They solve the same problem. My goal with localgpt is to be a framework for testing different components of RAG as lego blocks.

  • @FranchGuy
    @FranchGuy 6 месяцев назад

    Hi , is there way to contact you for privet project ?

    • @engineerprompt
      @engineerprompt  6 месяцев назад

      There is a link in the video description or email me at engineerprompt at gmail

  • @kunalr_ai
    @kunalr_ai 6 месяцев назад

    😂kuch samaj nahi aa raha .. kaha se start karna hai

    • @engineerprompt
      @engineerprompt  6 месяцев назад

      there is a playlist on localgpt on the channel. that will be a good starting point :)