Will the New GEMINI PDF Feature Replace RAG?

Поделиться
HTML-код
  • Опубликовано: 13 сен 2024

Комментарии • 47

  • @engineerprompt
    @engineerprompt  Месяц назад +2

    Check out the RAG Beyond Basics Course: prompt-s-site.thinkific.com/courses/rag

  • @artur50
    @artur50 Месяц назад +8

    It’d be excellent if you could test gpt4o and Flash against your RAG and show the results like you did in this video. That would be a nice demonstration of different capabilities and results of course with the use of local LLM

  • @marcomeyer7545
    @marcomeyer7545 Месяц назад +4

    Hi, can you do a video on this:
    In a typical AI workflow, you might pass the same input tokens over and over to a model. Using the Gemini API context caching feature, you can pass some content to the model once, cache the input tokens, and then refer to the cached tokens for subsequent requests. At certain volumes, using cached tokens is lower cost than passing in the same corpus of tokens repeatedly.

  • @MeinDeutschkurs
    @MeinDeutschkurs Месяц назад +3

    What if Gemma 2 is also able to do this. How could we test this?

  • @durand101
    @durand101 Месяц назад +1

    Impressive model. Thank you for the video.
    I think the main benefit from classic RAG so far for me has been citations and clear sourcing (where the llm can return which page it is using for information). How well does Gemini Flash return this kind of info?

    • @engineerprompt
      @engineerprompt  Месяц назад

      I haven't tested it on multiple files yet but I suspect that should be possible. I will put together a new tutorial on it when I get a chance.

  • @perelmanych
    @perelmanych Месяц назад

    In scientific papers tables are usually in text format. Latex just uses fancy formatting of text to make tables, so table content extraction is not test of visual capabilities of a model.

  • @RedCloudServices
    @RedCloudServices 26 дней назад

    Thanks for your videos and course. You said at the beginning Gemini 1.5 was only good for small docs what would you recommend for a large corpus of multi-modal PDF requirements? Would an agentic approach work to breakup the PDFs into buckets and a single agent to combine responses?

  • @vitalis
    @vitalis 25 дней назад +1

    What about using Gemini Flash to parse the PDFs into markdown and optimally structure it for LLMs and then embedding for RAG?

    • @wesleymogaka
      @wesleymogaka 5 дней назад +1

      Pursuing this idea

    • @vitalis
      @vitalis 5 дней назад

      @@wesleymogaka report back once you do it. Maybe send the RUclipsr a link so he can also review it and give you some exposure

  • @KumR
    @KumR Месяц назад +2

    Hi. Can u show us how to get to the UI ?

  • @gregsLyrics
    @gregsLyrics Месяц назад

    One Q that I missed: when making API calls to our pdf, does our private data become publicly available in any way? Another amazing vid. Really appreciate all the work you put into making great content.

    • @engineerprompt
      @engineerprompt  Месяц назад

      For free api, Google does say, they can use it for training. For paid api, that doesn't seem to be case. Now just like the other api providers, really it's on your own comfort level and how much you trust their words :)

  • @ashimov1970
    @ashimov1970 5 дней назад

    your Colab link doesn't work. It doesn't open

  • @IdPreferNot1
    @IdPreferNot1 Месяц назад

    love the meta paper choice to scan

  • @ryshabh11
    @ryshabh11 Месяц назад +1

    Thanks

  • @intellect5124
    @intellect5124 Месяц назад +1

    small number of pdf means how many? whats ur assumption?

    • @engineerprompt
      @engineerprompt  Месяц назад

      As long as they fit in the context, which is 1M, although I would suggest using about 50-70% of that. Using more can result in lost in the middle

  • @KGIV
    @KGIV Месяц назад

    I don't like using libraries to parse my PDF files. I found it to be more complex and less robust than writing the parsing services myself. I will defintely give flash a try though.

  • @muhammadsaqib453
    @muhammadsaqib453 Месяц назад

    Please run any ad compaign for your channel as your channel has the potential to get 500k subscribes in a hour.

  • @freddiechipres
    @freddiechipres 24 дня назад

    Why testing Gemini flash? Does Gemini Pro not work better?

    • @engineerprompt
      @engineerprompt  24 дня назад

      Pro is better but has more limitations for free usage.

  • @ebandaezembe7508
    @ebandaezembe7508 Месяц назад

    thank you so much for this video

  • @micbab-vg2mu
    @micbab-vg2mu Месяц назад +1

    great i will test it -:)

  • @lavericklavericklave
    @lavericklavericklave Месяц назад +1

    This review is basically pointless. Youre running it on one pdf. The whole pdf can easily be dumped into the context (oai default is 20 x 1000 token chunk). You should be doing it on much larger datasets

  • @interspacer4277
    @interspacer4277 Месяц назад +1

    RAG in general has been slowly dying as context increases are combined with cost decrease. On top of that, folk are getting better at compression and database use (LLMs understand SQL, etc), and agentic flows.
    The speed loss and cost to maintain a vector database, just isnt always worth it when I can simply task a flow itself for semantic search and feed it to whatever needs it.

    • @Hisma01
      @Hisma01 Месяц назад +3

      RAG is not dying. It merely depends on the use-case. It was even mentioned several times in this video where this is not a replacement for RAG where there is a large corpus of information (millions of docs). It certainly is evolving however, and quite rapidly. I would love to get to the point where I can avoid having to parse pdfs and documents completely, and just feed docs to a vision model & have that the chunks stored directly in a db. But getting rid of RAG completely? Nah. Not yet. I would say RAG would only go away if there's some way where model training reaches a point you can just throw docs at it and rather than feeding them into a vector db, you can feed docs directly into the llm itself.

  • @CryptoMaN_Rahul
    @CryptoMaN_Rahul Месяц назад

    i wanted to build a previous year paper analysis system for my colllege ( engineering ) , there are total 7 departments , all subjects come upto 7*6*8. Can you just guide fine tuning or Rag ??

    • @engineerprompt
      @engineerprompt  Месяц назад +1

      For this, my recommendation will be to use RAG for it.

    • @CryptoMaN_Rahul
      @CryptoMaN_Rahul Месяц назад

      Cool thanks ​@@engineerprompt

  • @rodrigolumi
    @rodrigolumi Месяц назад

    Great video.

  • @mohammad-xy9ow
    @mohammad-xy9ow Месяц назад

    Is there demand of rag in the market ?

    • @engineerprompt
      @engineerprompt  Месяц назад

      RAG is the only real application of GenAI at the moment that businesses are actually widely using.

  • @ebandaezembe7508
    @ebandaezembe7508 Месяц назад

    gemini 1.5 pro also has this new feature i think

    • @engineerprompt
      @engineerprompt  Месяц назад

      Yes, it does. Its relatively more expensive though if you put it in production.

  • @nguyenanhnguyen7658
    @nguyenanhnguyen7658 Месяц назад

    Why would u want to pay for cloud GPT !?!? Do it yourself.

  • @NeuroScientician
    @NeuroScientician Месяц назад +1

    As usual I will wait for third parties to verify which google's claims are real and which are just another scam.