7-End To End Advanced RAG Project using Open Source LLM Models And Groq Inferencing engine

Поделиться
HTML-код
  • Опубликовано: 4 июн 2024
  • In this video we will develop an end to end advanced Rag project using open source llm models such as Llama2, Mistral using groq inferencing engine
    code github: github.com/krishnaik06/Update...
    ---------------------------------------------------------------------------------------------
    Support me by joining membership so that I can upload these kind of videos
    / @krishnaik06
    -----------------------------------------------------------------------------------
    Fresh Langchain Playlist: • Fresh And Updated Lang...
    ►LLM Fine Tuning Playlist: • Steps By Step Tutorial...
    ►AWS Bedrock Playlist: • Generative AI In AWS-A...
    ►Llamindex Playlist: • Announcing LlamaIndex ...
    ►Google Gemini Playlist: • Google Is On Another L...
    ►Langchain Playlist: • Amazing Langchain Seri...
    ►Data Science Projects:
    • Now you Can Crack Any ...
    ►Learn In One Tutorials
    Statistics in 6 hours: • Complete Statistics Fo...
    End To End RAG LLM APP Using LlamaIndex And OpenAI- Indexing And Querying Multiple Pdf's
    Machine Learning In 6 Hours: • Complete Machine Learn...
    Deep Learning 5 hours : • Deep Learning Indepth ...
    ►Learn In a Week Playlist
    Statistics: • Live Day 1- Introducti...
    Machine Learning : • Announcing 7 Days Live...
    Deep Learning: • 5 Days Live Deep Learn...
    NLP : • Announcing NLP Live co...
    ---------------------------------------------------------------------------------------------------
    My Recording Gear
    Laptop: amzn.to/4886inY
    Office Desk : amzn.to/48nAWcO
    Camera: amzn.to/3vcEIHS
    Writing Pad:amzn.to/3OuXq41
    Monitor: amzn.to/3vcEIHS
    Audio Accessories: amzn.to/48nbgxD
    Audio Mic: amzn.to/48nbgxD

Комментарии • 49

  • @krishnaik06
    @krishnaik06  Месяц назад +4

    Support me by joining membership so that I can upload these kind of videos
    ruclips.net/channel/UCNU_lfiiWBdtULKOw6X0Digjoin

    • @dr8544
      @dr8544 Месяц назад

      make a patreon

    • @r1ckmav
      @r1ckmav Месяц назад +1

      Just became member of this amazing channel.

  • @DoomsdayDatabase
    @DoomsdayDatabase Месяц назад +2

    Can't wait till Groq introduces option for running almost any HuggingFace models! Never came across such a simple, fast and cost effective endpoint setup!

  • @avixek
    @avixek Месяц назад +3

    I am using these series to learn langchain and deploy it in our fortune 10 company.. One of the key challenges in knowledge retrieval for all of the engineers in our team. I ran this scenario and we may have a savings of $500K/per yr. Please keep doing what you are doing.

    • @EkNidhi
      @EkNidhi Месяц назад

      Error initializing FAISS index: Error raised by inference endpoint: HTTPConnectionPool(host='localhost', port=11434): Max retries exceeded with url: /api/embeddings (Caused by NewConnectionError(': Failed to establish a new connection: [WinError 10061] No connection could be made because the target machine actively refused it'))
      can you solve this error?

  • @shalabhchaturvedi6290
    @shalabhchaturvedi6290 Месяц назад +1

    Krish, you are doing an excellent job! Great job, bro!

  • @r1ckmav
    @r1ckmav Месяц назад

    Hi Krish, Thanks a lot for such an amazing content. You are doing a great help to humanity by publishing such an amazing content available free for everyone.

  • @nishantchoudhary3245
    @nishantchoudhary3245 Месяц назад +2

    Best langchain series by best teacher. Krish sir ❤

  • @canyouvish
    @canyouvish Месяц назад

    Very helpful set of videos. And yes, when I type words that have 'ng' in them - the 'g' almost always come before 'n'.

  • @maazidris4435
    @maazidris4435 Месяц назад

    Great work sir, Like from Pakistan

  • @rabbanimogal1786
    @rabbanimogal1786 Месяц назад +1

    Great work!!

  • @DavidHelmen-sb9wf
    @DavidHelmen-sb9wf Месяц назад +2

    Hey Krish I hope you are doing well, can you make a video about full stack data scientist AKA Unicorn data scientist role?

  • @andrespineiro7609
    @andrespineiro7609 Месяц назад

    You are the best, dude!

  • @ShayansCodeCommunity
    @ShayansCodeCommunity Месяц назад +1

    Nice 🙂

  • @YorkYongYeo
    @YorkYongYeo 29 дней назад

    Will it be possible to process the vector embeddings before hand and allow the app to run faster in terms of reply?

  • @adityavipradas3252
    @adityavipradas3252 Месяц назад

    Is Groq LPU hardware needed to see the actual inference performance boost?

  • @dharmendersingh7565
    @dharmendersingh7565 Месяц назад +1

    Great work, but make a video on implementing RAG on complex pdf (like extracting financial statements tables)

  • @mithunshet5922
    @mithunshet5922 Месяц назад

    Hi.. I am interested in predictive analytics where I mainly use ML models.. How will LLM be useful in this field?

  • @NairNitinvenugopal
    @NairNitinvenugopal 7 дней назад

    instead of chains, can we use agent here?

  • @mohsenghafari7652
    @mohsenghafari7652 Месяц назад

    Hi dear friend .
    Thank you for your efforts .
    How to use this tutorial in PDFs at other language (for example Persian )
    What will the subject ?
    I made many efforts and tested different models, but the results in asking questions about pdfs are not good and accurate!
    Thank you for the explanation

  • @r1ckmav
    @r1ckmav Месяц назад

    Hi Krish, I just tried replacing the WebBaseLoader to PyPDFLoader somehow the same code piece does not seem to be working. Any suggestions to how to replace the WebBaseLoader with PyPDFLoader?

  • @varshavaran2558
    @varshavaran2558 Месяц назад +1

    Sir, the latest advancements in AI and data science trigger anxiousness since I am in early stages of Data science of roadmap. Will I ever catch-up with the latest technology or the basics remain relevant still

    • @ankaleshganorkar
      @ankaleshganorkar Месяц назад

      yes, tools will always change, domain knowledge will remain, enjoy the anxiousness.

  • @tootemakan
    @tootemakan Месяц назад

    Please build a conversational chain system using open source llms

  • @chinnibngrm272
    @chinnibngrm272 Месяц назад

    Sir want to build chatbot for Fintuned llama2 model in my local system
    Is it possible to build with streamlit like ur projects
    Please give us solution for this 😊😊

  • @_itsvinay
    @_itsvinay Месяц назад

    Hi @krishnaik06 pls make a video on how to connect it with cloud database like pinecone or astra db. After version update it's not working with langchain.

  • @amritsubramanian8384
    @amritsubramanian8384 Месяц назад

    awesomeee ;)

  • @EkNidhi
    @EkNidhi Месяц назад +1

    Error initializing FAISS index: Error raised by inference endpoint: HTTPConnectionPool(host='localhost', port=11434): Max retries exceeded with url: /api/embeddings (Caused by NewConnectionError(': Failed to establish a new connection: [WinError 10061] No connection could be made because the target machine actively refused it'))
    i m getting this error why?

    • @allinoneofficial5300
      @allinoneofficial5300 19 дней назад

      i got the same error check your .env file paths of keys it is not given properly

    • @samagrashrivastava8042
      @samagrashrivastava8042 17 дней назад

      @@allinoneofficial5300 I got the same error! can you explain the solution.

    • @allinoneofficial5300
      @allinoneofficial5300 17 дней назад

      @@samagrashrivastava8042problem with api key initialisation. It is not given properly.initialise it properly

  • @captionjunior6959
    @captionjunior6959 20 дней назад

    how to develop a plagrism checker?

  • @TarabUTK
    @TarabUTK 17 дней назад

    out of all the ollama models:
    Llama 3 8B 4.7GB ollama run llama3
    Llama 3 70B 40GB ollama run llama3:70b
    Phi-3 3.8B 2.3GB ollama run phi3
    Mistral 7B 4.1GB ollama run mistral
    Neural Chat 7B 4.1GB ollama run neural-chat
    Starling 7B 4.1GB ollama run starling-lm
    Code Llama 7B 3.8GB ollama run codellama
    Llama 2 Uncensored 7B 3.8GB ollama run llama2-uncensored
    LLaVA 7B 4.5GB ollama run llava
    Gemma 2B 1.4GB ollama run gemma:2b
    Gemma 7B 4.8GB ollama run gemma:7b
    Solar 10.7B 6.1GB ollama run solar
    on a 16 GB RAM 'ok' desktop, Phi-3, Gemma-2b seems to run in reasonable time .... but I could not get Groq to work with either. Please advise what changes are needed for this part of the series. As of now, wait time is too long.
    Thanks

  • @mohamedkeddache4202
    @mohamedkeddache4202 5 дней назад

    what is the purpose of using open source LLMs with paid groq api.
    just use directly openAi api key
    the only purpose of free LLMs are the ability to run them locally fro free

  • @shankar9063
    @shankar9063 Месяц назад

    llm=ChatGroq(groq_api_key=groq_api_key,
    model_name="mixtral-8x7b-32768")
    I initialized the mixtral but i 'am getting the following error.
    ValueError: Error raised by inference API HTTP code: 404, {"error":"model 'llama2' not found, try pulling it first"}

  • @sarveshudapurkar6235
    @sarveshudapurkar6235 Месяц назад

    SIr I'm getting some error
    ValueError: Error raised by inference API HTTP code: 404, {"error":"model 'llama2' not found, try pulling it first"}

  • @sivanthik8288
    @sivanthik8288 Месяц назад

    Thanks!

  • @ramdasvk0716
    @ramdasvk0716 Месяц назад +1

    Open source is 🫶🫶🫶