Introduction to Query Pipelines (Building Advanced RAG, Part 1)

Поделиться
HTML-код
  • Опубликовано: 22 май 2024
  • This episode gives you an introduction to using Query Pipelines in LlamaIndex to setup basic to advanced orchestration over your data. This ranges from simple prompt chains to full DAGs that contain a lot of advanced retrieval components.
    This video mini-series around advanced RAG orchestration is designed to help you get a handle of how to build complex workflows over your data for different use cases.
    EDIT: at 26:19 we erroneously mention `FunctionComponent`, which doesn't exist. Try `FnComponen` instead!
    Colab notebook: colab.research.google.com/dri...
    Timeline:
    00:00-07:00 - Intro
    07:00-15:50 - Prompt Chaining
    15:50-24:13 - RAG Workflows
    24:13-27:50 - Custom Components
    27:50 - Async/Parallel

Комментарии • 24

  • @AI__Spectrum
    @AI__Spectrum 2 месяца назад

    OMG THIS SERIE IS JUST AMAZING!!!!! It's truly incredible! We are waiting for more Multi-Agent videos :)

  • @geoffreygordonashbrook1683
    @geoffreygordonashbrook1683 3 месяца назад +7

    Many thanks for the tools and videos! Request: could you sometime cover in extra detail how to locally run locally downloaded (and locally made / fine tuned) models as part of RAG systems? Being able to swap in and out local models (and fine tune those) is crucial in many use-cases. But available guides nearly all use a non-modular workflow where either a cloud-api mode or a fresh-hugging-face download are required, neither of which are suitable for many production or even research cases. Many thanks!

  • @olivermarohn5741
    @olivermarohn5741 Месяц назад

    Great Tutorial!

  • @michaelfloyd2525
    @michaelfloyd2525 3 месяца назад +1

    Jerry you are my GOAT

  • @bwc1007
    @bwc1007 3 месяца назад

    Thank you for the video... here I am the next morning ~2M tokens in to a test use case (thanks for the phoenix tracing tip too..)

  • @debasmitroy4513
    @debasmitroy4513 Месяц назад

    Great !

  • @unclecode
    @unclecode 3 месяца назад

    Many thank fro this video series, how to get the output of each module individually after running the chain for inspection. Thx

  • @HaiNanZhifa
    @HaiNanZhifa 3 месяца назад +1

    Hi Jerry, when I have a corpus of 100 million legal documents, what are the effective indexing strategies available? How can I apply the techniques you showed here and in your other videos?

  • @nikhilshrestha7475
    @nikhilshrestha7475 3 месяца назад

    Great video. Just a few questions:
    1. Does the speed of this system remain so much ?
    2. I tried the same by including 40MB and 20 MB book with small kbs and mbs PDFs. The speed was somewhat similar.

  • @navanshukhare
    @navanshukhare 3 месяца назад

    Hi, Can you please make an advanced video on different types of engines, especially SQLAutoVectorQueryEngine? A lot of features and we love the llama index frameowrk

  • @SuperLiberty2008
    @SuperLiberty2008 3 месяца назад

    Hey! I'm looking for effective strategy how to handle structured data. Could you advise tutorial or paper to study? Regards!

  • @zugbob
    @zugbob 3 месяца назад

    Is this all possible in the TS version?

  •  2 месяца назад

    So many thanks for your hard work guys!! I'm following most of the guides but there is something that's killing me. Just imagine loading a document with "chapters", "subchapters" and, obvioulsy a parragraph inside... RAG will work flawlessly extracting chunks of these parragraphs but if you ask for a resume of all the subchapters of certaing chapter then the RAG don't work so well if the context is huge (due missed chunks).
    ¿There is any kind of metadata/chunking way of for this kind of questions?
    For more context, I'm using Sherpa for reading a PDF. The PDF has a list of chapters about certain things. The "query" is, "what's included in this chapter" and the expected response is a list of all subchapters... Obtained response is not complete due lack of context.

  • @ajg3951
    @ajg3951 3 месяца назад +1

    I'm curious about something here. As we're generating indexes from a file, it's possible to create diverse types of indexes, such as VectorStoreIndex, Summary Index, Tree Index, Keyword Table Index, and Knowledge Graph Index, etc ...
    What are the advantages of utilizing these various indexes? How can we determine which indexing method is most suitable for our specific use case?

    • @lloganm1234
      @lloganm1234 3 месяца назад

      In general, the most production ready ones are vector and summary
      Use vector index for semantic search. Combine with hybrid + reranking + query rewrite for improved performance.
      SummaryIndex is best for generating summaries on the fly as you need them

  • @thetagang6854
    @thetagang6854 Месяц назад

    I know the data used for the code is PG essays, but there's quote a few of them.... is there somewhere to download this??

  • @sritharan20
    @sritharan20 3 месяца назад

    Hey Jerry, first off llama_index is awesome, pls revisit or change the docs side tree panel, its very annoying,same doc is referenced in multiple places and it keeps jumping around, a good example will be langchain docs, but i love using llama_index. thx

  • @austinmw89
    @austinmw89 3 месяца назад

    What happened to FunctionComponent? I don't see it in the latest LlamaIndex version 🧐

    • @LlamaIndex
      @LlamaIndex  3 месяца назад

      it is FnComponent 😬 typo in video

  • @iloos7457
    @iloos7457 3 месяца назад

    differrence between this and langgraph?

    • @LlamaIndex
      @LlamaIndex  3 месяца назад +5

      Query pipelines is actually mostly for orchestrating DAGs very easily (chains, retrieval workflows), and not meant for loops!
      To do agentic loops we'll have a followup video showing how you can build an agent around a query pipeline to do step-by-step reasoning such as ReAct/query planning

    • @austinmw89
      @austinmw89 3 месяца назад

      Please link follow up video when possible!

  • @tusharganguli
    @tusharganguli 2 месяца назад

    Why would we use a hallucinated passage as an input to the retriever to return the actual passage? ruclips.net/video/CeDS1yvw9E4/видео.htmlfeature=shared&t=1111