What is Agentic RAG?

Поделиться
HTML-код
  • Опубликовано: 7 янв 2025

Комментарии • 53

  • @norbertschmidt
    @norbertschmidt 2 месяца назад +19

    You make the best LLM + RAG explainers in the world. Thanks so much.

  • @lesmoe524
    @lesmoe524 2 месяца назад +10

    Fantastic description, it seems using agents to help you route to potentially different databases is a game changer, thank you.

  • @sqweepsrussell9412
    @sqweepsrussell9412 2 месяца назад +3

    Thanks for this well explained video. This is the most simplified explanation of agentic rag that sank into my grey matter

  • @ChielReemer
    @ChielReemer 28 дней назад +4

    The most amazing thing about all this is that he can write so easily in reverse.

  • @vincenttima
    @vincenttima 22 дня назад +2

    Thanks David, well done presentation.

  • @nbamastermind
    @nbamastermind Месяц назад +2

    Excellent explanations! Simple without additional fluff. Thank you.

  • @TiếnNguyễnMinh-t8h
    @TiếnNguyễnMinh-t8h 8 дней назад

    Thanks so much, youre talented. I really interested in the way you share your useful knowledge, information. I love it, the best wishes to you. Im looking forward to your video everyday, once again, thanks a lot.

  • @whitesaladchips
    @whitesaladchips 24 дня назад +1

    can someone explain that this is basically instead of using similarity search, here we are instead using llm to decide which database is more appropriate to query from? am i correct here

  • @afzaalawan
    @afzaalawan 10 дней назад

    Excelllent concept -- many thanks.

  • @jaffarbh
    @jaffarbh 2 месяца назад +1

    Thanks. This is pretty useful and much better than the naïve approach of overloading the model with lot's of irrelevant data from the vector DB.

  • @PriyeshYadav
    @PriyeshYadav 2 месяца назад +3

    So the Agent is also a pretrained LLM with those 2 vector db ???

  • @averlyrolf7970
    @averlyrolf7970 21 день назад +1

    Where are these agents installed/located? How do they get into the pipeline connected to the LLM?

  • @digambardagade288
    @digambardagade288 11 дней назад

    This will help you in case when you have multiple documents or subsections and you are storing them into different vector databases.
    In that case, based on user's query, agent will understand which vector database model needs to access.

  • @ErickM.Joseph
    @ErickM.Joseph Месяц назад

    Makes sense I’ll be implanting this into our agent workflow

  • @ssdm444
    @ssdm444 Месяц назад

    Simple, clear explanation! Thanks!!

  • @sofluzik
    @sofluzik 13 дней назад

    What is the logic or rule to help llm decide where to go ?? Is it only key value pairs or looking for keywords , context based understanding....

  • @SasoCasablanca63
    @SasoCasablanca63 6 дней назад

    I want to know about the display you are using . I like the way this white board works and gives the illusion of being there . What tool is that ?

  • @Plife-507
    @Plife-507 17 дней назад

    This this would be good for training an agentic swarm for a trading system based on on depth guides curated from pdf?

  • @Trazynn
    @Trazynn 28 дней назад

    4:23 it doesn't have to decline to answer. It probably is able to answer such questions by virtue of it already being a sufficiently big enough LLM. It could say "Though this isn't part of our internal database (creating a clear signal that it goes beyond its domain) the world series was won by..."

    • @shubham.1172
      @shubham.1172 27 дней назад

      Yes but it's best to decline for most customer facing use cases, right? For example I wouldn't want my Amazon bot to answer questions about Superbowl, it's best if it just declines to answer. Also reduces legal exposure?

  • @sterlingabbot695
    @sterlingabbot695 2 месяца назад +8

    Please do RAG + Big Data next

    • @IBMTechnology
      @IBMTechnology  2 месяца назад +1

      An interesting suggestion, any specific element or question you have about the topic?

  • @AbdullahTBukhari
    @AbdullahTBukhari Месяц назад

    Easy to Understand explanation. 👍🏼

  • @Ilovepotatoes-t4t
    @Ilovepotatoes-t4t 2 месяца назад

    thank u!! One question: Do you know how to evaluate an agentic rag? Do I have to take anything into account compared to a normal rag evaluation? (e.g. with RAGAs..) Best regards

  • @imVishalKS
    @imVishalKS Месяц назад

    Really good explanation.

  • @CDR4445
    @CDR4445 2 месяца назад

    We need to know how to implement agent and what is this? is this another llm to determine the context and route to right db

  • @BizAutomation4U
    @BizAutomation4U Месяц назад

    I asked Perplexity how best to decide if seperate vector DBs is preferred over a large vector DB with metadata to contextualize the pool of vector data. I guess it's tantemount to asking if we should normalize vector tables similar to OLTP or use a data warhouse approach. It recommended the metadata approach to help the LLM decide which portion of the vector data to contextualize, which makes sense to me. Maybe it's the same thing described in this video in a different way.

  • @BernardoCortinaCeballos
    @BernardoCortinaCeballos Месяц назад

    Que bien está explicado este video, gracias.

  • @ugurkaraaslan9285
    @ugurkaraaslan9285 Месяц назад

    There should be huge response delay due to additional step?

  • @StalinDeLaTorre
    @StalinDeLaTorre 2 месяца назад +1

    Where can I learn this approach?

  • @MikewasG
    @MikewasG 2 месяца назад +2

    What is the difference between this and semantic routing?

    • @IBMTechnology
      @IBMTechnology  2 месяца назад +1

      Semantic routing uses more straightforward methods like cosine similarity and other predefined rules to make the decision on which route to take. The example I drew used an LLM agent, which can understand and interpret more complex/nuanced queries, understand context, but is much heavier (compute and latency) because it is using an LLM.
      -David

  • @RohitGulati309
    @RohitGulati309 Месяц назад

    Im confused ... Andrew Ng says just RAG is Agentic AI. But RAG does not need LLM during the query phase. So what's correct here?

  • @nitingaurz
    @nitingaurz Месяц назад

    Since the agent make use of LLM in decision making, it would be more clear to show the connection in the drawing.

  • @Agent221314
    @Agent221314 Месяц назад

    What framework support this? Ag2?

  • @Ijmeisner
    @Ijmeisner 2 месяца назад +2

    In the reflection I think that’s the teleprompter… or maybe I am hallucinating 🤣

  • @simonthompson1099
    @simonthompson1099 Месяц назад

    How about a citation that shows it works?

  • @vichupayyan
    @vichupayyan 3 дня назад

    every day new concept keep popping up, too hard to get the concept and follow before the concept gets into another form :)

  • @peabody3000
    @peabody3000 4 дня назад

    retro encabulator vibes 😊

  • @hi5wifi-s567
    @hi5wifi-s567 2 месяца назад +1

    “More responsible, more accurate, more adaptable, “
    plus more secure as well?

  • @sergiogonzart
    @sergiogonzart 24 дня назад

    I'd love a recommendation for an LLM that actually admits it does't have an answer instead of just making something up.

  • @sjmediaonline
    @sjmediaonline 2 месяца назад +5

    IBM’s main task now is to create the catchup teaching videos. No innovation or breakthroughs are coming out from them. I see so many IBM old timers are watching with amazement how smaller open source innovators are moving lightning fast in GenAI and then they start recording teaching videos… I do not see any new things / new demos coming out from IBM.

    • @JC-kb8vb
      @JC-kb8vb 22 дня назад +1

      Yeah. On the other hand, your company is booming… 😆

  • @AK-be7jh
    @AK-be7jh 2 месяца назад +2

    So the agent will acts like a controller here .

  • @marcomaiocchi5808
    @marcomaiocchi5808 2 месяца назад +9

    This pipeline doesnt make a lot of sense.

    • @JustinKahrs
      @JustinKahrs Месяц назад +1

      skill issue

    • @scycer
      @scycer Месяц назад +19

      Got a particular question?
      A standard rag pipeline just takes your question convert it into a vector and searches the data source to find content that is similar to it.
      By adding an agent in front of it, the question can be interpreted to figure out which data sources should be used to fetch that content.
      It's like the difference between asking a single book a question or asking a question to a librarian who can find the right book for you before looking into the book for the answer to that question.
      Its hard with all the unique fancy naming they give these architectures, it abstracts the simplicity of what is really happening underneath.

    • @Whynot83848
      @Whynot83848 Месяц назад +1

      @@scycervery cool example. Thanks

    • @egopgogo-job8614
      @egopgogo-job8614 15 дней назад

      @@scycerreally great explanation. Thanks!

    • @ClifHDunn
      @ClifHDunn 4 дня назад

      @@scycertotally going to swipe this librarian analogy and may even give you credit on occasion 😊

  • @box4soumendu4ever
    @box4soumendu4ever Месяц назад

    👏👏🥰🥰🕊🕊🙏🙏👍👍🪻

  • @jojojojojojojo2
    @jojojojojojojo2 5 дней назад

    Why do you say "We'll" instead of "We".. are you just starting to understand it theoretically just yet? This stuff is more than 2 years old already... And on top of that - this stuff is not only old but outdated as well...

  • @funkfreeze
    @funkfreeze 2 месяца назад

    Overkill for most consumer facing applications and, as the answer to generalist queries, not specific enough a system for internal tooling. Lots of noise introduced here.