RAG Implementation using Mistral 7B, Haystack, Weaviate, and FastAPI

Поделиться
HTML-код
  • Опубликовано: 16 сен 2024

Комментарии • 74

  • @jorgerios4091
    @jorgerios4091 11 месяцев назад +3

    Oh Man, you rock! it seems like you read your followers' minds. Thank you so much for this great video.

    • @AIAnytime
      @AIAnytime  11 месяцев назад +2

      Yeah Non Invasive models in brain has been implanted 😂.. thanks btw

  • @shivamroy1775
    @shivamroy1775 11 месяцев назад +3

    Absolutely great quality content. Please continue the great work, These videos are extremely useful.

    • @AIAnytime
      @AIAnytime  11 месяцев назад +1

      Thanks Shivam... sure more on the cards.

    • @MohamElY
      @MohamElY 4 месяца назад

      Haystack the same time ​@@AIAnytime

  • @talhaabdulqayyum193
    @talhaabdulqayyum193 11 месяцев назад +5

    PLease upload on the Fine tuning of the Llama 2 models (7B, 13B et), especially regarding how to create a data set for a chatbot to model fine tuning and model inference

  • @sneharoy3566
    @sneharoy3566 11 месяцев назад +1

    This is just next level

  • @jorgerios4091
    @jorgerios4091 11 месяцев назад +4

    Hello.
    I've been thinking about ways to improve response speed in this example, and I came across fine-tuning the model using autotrain. I was wondering if fine-tuning the model with GPU support from platforms like Kaggle or Colab could be a viable option for achieving faster responses.
    I think it would be really interesting if you could explore this topic in one of your videos and show us how to fine-tune a model using GPU support (this same example). It would be great to see the process in action and learn more about the benefits and considerations involved in fine-tuning models for faster response times.
    Looking forward to your insights on this topic! Keep up the great work on your channel.
    Jorge.

  • @Techonsapevole
    @Techonsapevole 11 месяцев назад

    Thanks, great tutorial

    • @AIAnytime
      @AIAnytime  11 месяцев назад

      Glad it was helpful!

  • @afj4ojfa340
    @afj4ojfa340 11 месяцев назад

    Thank you for doing a video on this!

    • @AIAnytime
      @AIAnytime  11 месяцев назад

      Glad it was helpful!

  • @rajkachhadiya6192
    @rajkachhadiya6192 11 месяцев назад

    really very informative video. Thank you dear for your efforts and informative video. I appreciate your effort and content dear. Keep it up. Once again thank you ❤❤❤

    • @AIAnytime
      @AIAnytime  11 месяцев назад

      It's my pleasure

  • @giridharreddy7011
    @giridharreddy7011 11 месяцев назад +4

    Sir, can you please show how to run it in Mac M1 system as well

    • @ajayanilkumar1645
      @ajayanilkumar1645 11 месяцев назад +1

      Yes please sir, it would be helpful for me also

  • @Sulayman.786
    @Sulayman.786 11 месяцев назад +2

    AI now?! Let's go!

  • @ancient_living
    @ancient_living 10 месяцев назад +1

    Hi,
    The embedding dimension specified is 768 and you are using a 384 model. It throws error, can you help?

  • @TrueWarship
    @TrueWarship 7 месяцев назад

    hi i really love you video weldone but am finding it hard to install haystack on my system it saying haystack.nodes is not found.
    please any help

  • @manishpatil09
    @manishpatil09 9 месяцев назад +1

    At 1:40 you said you are going to use "Vector Database" not "Vector Store". Can somebody please tell me the difference between them.

    • @nogueirad
      @nogueirad 6 месяцев назад

      Hi! Duda from Weaviate here. They are the same. Langchain calls the. Vector Store, but they are the same :)

  • @sampriti6026
    @sampriti6026 8 месяцев назад

    Which sentence transformer are we using exactly? And why not any top model from mteb ? What are the drawbacks? Would be helpful if you can clarify.
    Great video tho❤

  • @rpmknowledge3067
    @rpmknowledge3067 11 месяцев назад +1

    thankyou sir sir i want use sqldatabase inplace of pdf and etc
    so please make video for mysql

  • @AminovJP
    @AminovJP 11 месяцев назад

    Thank you very much, you are fantastic, you could be a teacher your natural way of talking is very captivating !!!!, what do you think about PGVector ?

    • @AIAnytime
      @AIAnytime  11 месяцев назад

      Thank you! 😃 yes You can try Supabase. It has a PG extension i guess.

  • @casualdiscord
    @casualdiscord 8 месяцев назад

    It seems like there's a mismatch that I can't figure out. The model has dimensions 384 and the datastore 768. So it won't update any dimensions:
    RuntimeError: Embedding dimensions of the model (384) don't match the embedding dimensions of the document store (768). Initiate WeaviateDocumentStore again with arg embedding_dim=384.
    What am I missing?

  • @RedCloudServices
    @RedCloudServices 11 месяцев назад +1

    thanks again this is my exact use case here in DC! Just testing different opensource stacks there is a great langchain interview w Weviate discussing the technique of using a summary vector in each prompt to improve LLM responses. How would you improve the response time?

    • @AIAnytime
      @AIAnytime  11 месяцев назад

      Yeah that's a very good idea. Those can be used in memory as well. Give it a try.
      Let me know what you build. Thanks

  • @lukasvanuden6049
    @lukasvanuden6049 6 месяцев назад

    Can you do it without Docker?
    Since Docker needs a commercial licence for bigger companies..

  • @khanmahmuna
    @khanmahmuna 7 месяцев назад

    thanks for the informative video,can you upload a same tutriol video for different types of document summarization implementing RAG and open source LLM not OpenAI asap...it would be greatful need help in that and not found any related video for summarization .

  • @networks6111
    @networks6111 6 месяцев назад

    Hey bro! very nice video, what are the system specs for running this project? Is 8gb RAM enough, to produce results if we use Quantised Bloke mistral model

  • @tarik1895
    @tarik1895 Месяц назад

    Can please explain where you download the two models you have the mistral and the llama

  • @rachadnajjar
    @rachadnajjar 8 месяцев назад

    Hi , thank you for the tutorial , when i'm install the requirements.txt , I got this error message : ERROR: Could not find a version that satisfies the requirement torch (from versions: none)
    ERROR: No matching distribution found for torch . Can you please advise and help? Thank you

  • @davidfa7363
    @davidfa7363 9 месяцев назад +1

    Hi. I am building a chatbot with my pdf document. I created an endpoint in Fastapi. If i send 1 request the answer is really fast, but if i try to send 100 requests in the same time the response time is so much bigger. How can i run those requests paralell using gpu?

  • @VaibhavPatil-rx7pc
    @VaibhavPatil-rx7pc 8 месяцев назад

    excellent !

  • @TJ-hs1qm
    @TJ-hs1qm 11 месяцев назад

    awesome thank you!

    • @AIAnytime
      @AIAnytime  11 месяцев назад

      You're welcome!

  • @rohanghogare5766
    @rohanghogare5766 3 месяца назад

    This git repository has not been updated. Can you upload models, template files/folders

  • @Sundarampandey
    @Sundarampandey 9 месяцев назад

    bhai thoda mike ki quality sai krr lo yaar
    aawaz chubh rai h kasam se
    Baaki great knowledge
    keep up the good work

  • @SonGoku-pc7jl
    @SonGoku-pc7jl 10 месяцев назад +1

    i try finishing this pending video, and i have a question. today i see your video of jina v2. in this video you says about model of incrustation in huggingface that name is multilingüe. Is possible to fusion? you can make a video for best implementations cases that jina 2 or ideas for other cases of language spanish or catalan jeje :P thanks for your content! is increible!

    • @AIAnytime
      @AIAnytime  10 месяцев назад

      Great suggestion. Will create one soon...

  • @tareeshmuluguru395
    @tareeshmuluguru395 11 месяцев назад

    i am facing an issue with embedding dimension. this is the error: Embedding dimensions of the model (384) don't match the embedding dimensions of the document store (768). Initiate WeaviateDocumentStore again with arg embedding_dim=384

    • @SanjayBalikar
      @SanjayBalikar 10 месяцев назад

      I had same issue try delete the docker image and again rerun everything starting from creating docker image.

  • @MohamElY
    @MohamElY 4 месяца назад

    ❤❤Viagra

  • @gogodo9934
    @gogodo9934 11 месяцев назад

    Bro when’s the “deployment on aws” video coming out ?. It’s been more than a month since the 2 nd part released

  • @wtfJonKnowNothing
    @wtfJonKnowNothing 11 месяцев назад

    I just love how you pronounce massine 😂

  • @vivekmishra69
    @vivekmishra69 10 месяцев назад

    Hey did you try fine-tuning mistral on custom domain data? Can you recommended GPU for fine tuning 7B models such as Mistral, Llama etc..?

    • @AIAnytime
      @AIAnytime  10 месяцев назад +1

      Take a rental GPU on Runpod or A100 or v100 Google Colab with high RAM.

  • @usmanyousaaf
    @usmanyousaaf 11 месяцев назад

    Hi sir
    Am making a pdf questions answers chatbot but the issues ia my dataset is to dirty and aslo very small
    So what's u r suggestion without changing the dataset how can i achieve the average accuracy
    I know its a silly question but ...

  • @oryxchannel
    @oryxchannel 11 месяцев назад

    _fire_

  • @atulya.anurag
    @atulya.anurag 11 месяцев назад +2

    help !!
    Tried everything i know to resolve this problem (I have cloned your git repo)
    Error: ImportError: cannot import name 'Document' from 'haystack' (E:\ATULYA\Haystack\myenv\Lib\site-packages\haystack\__init__.py)

    • @pratheeshp.m.4078
      @pratheeshp.m.4078 11 месяцев назад

      im getting same error on mac python 3.11

    • @pratheeshp.m.4078
      @pratheeshp.m.4078 11 месяцев назад +1

      is there any particular version of python or haystack we have to use?

    • @tareeshmuluguru395
      @tareeshmuluguru395 11 месяцев назад

      did you get the solution for this error??

    • @andrejusb1
      @andrejusb1 11 месяцев назад +1

      I had same problem, fixed it by updating requirements.txt. Remove "farm-haystack" and keep only "farm-haystack[weaviate]". You need to pip uninstall farm-haystack. It seems there is some conflict, when both libs are installed and some classes are not visible

  • @ashishkgp
    @ashishkgp 6 месяцев назад +1

    THIS GITHUB REPO IS NOT UPDATED AND IS GIVING MULTIPLE ISSUES!

  • @SaiRohith08
    @SaiRohith08 10 месяцев назад

    Hello! can i add multiple pdfs and can it read images in pdfs?

  • @souvickdas5564
    @souvickdas5564 11 месяцев назад

    Please provide the link for downloading the quantized Mistral model

    • @AIAnytime
      @AIAnytime  11 месяцев назад +3

      Here you go: huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF

  • @TJ-hs1qm
    @TJ-hs1qm 11 месяцев назад +1

    Dependency management in Python is a pain. Which exact package version to install? Nobody knows 😂

    • @AIAnytime
      @AIAnytime  11 месяцев назад +2

      Use poetry for managing the versions.

    • @Udayanverma
      @Udayanverma 8 месяцев назад

      same! ImportError: cannot import name 'send_event' from 'haystack.telemetry' (/usr/local/lib/python3.8/site-packages/haystack/telemetry/__init__.py)

  • @TJ-hs1qm
    @TJ-hs1qm 11 месяцев назад

    check your mic or audio settings. totally over-driving the sound 🤗

    • @AIAnytime
      @AIAnytime  11 месяцев назад

      Thanks for the tip!

  • @jw200
    @jw200 7 месяцев назад

    can you explain why this all is good thing?
    people talk about it as awesome stuff. but why?

  • @user-iu4id3eh1x
    @user-iu4id3eh1x 11 месяцев назад

    Weaviate is better or Pinecone?

    • @AIAnytime
      @AIAnytime  11 месяцев назад

      Pinecone premium is better. Weaviate is also good as open sourced.