Summarization with LangChain using LLM - Stuff - Map_reduce - Refine

Поделиться
HTML-код
  • Опубликовано: 24 янв 2025

Комментарии • 12

  • @zerofive3699
    @zerofive3699 9 месяцев назад +2

    Really useful info mam , keep up the good work

  • @vijaygandhi7313
    @vijaygandhi7313 8 месяцев назад +2

    In the abstractive summarization use-case, usually a lot of focus is given to the LLMs being used and its performance. Limitations of LLM including context length and ways to overcome this issue are often overlooked. Its important to make sure that our application is scalable when dealing with large document sizes. Thank you for this great and insightful video.

    • @AboniaSojasingarayar
      @AboniaSojasingarayar  8 месяцев назад +1

      Thank you Vijay Gandhi, for your insightful comment! You've raised an excellent point about the importance of considering the limitations of LLMs in the context of abstractive summarization, especially regarding their context length and scalability issues when dealing with large documents.
      Indeed, one of the significant challenges in using LLMs for abstractive summarization is their inherent limitation in processing long texts due to the maximum token limit imposed by these models. This constraint can be particularly problematic when summarizing lengthy documents or articles, where the full context might not fit within the model's capacity.

  • @alvaroaraujo7945
    @alvaroaraujo7945 4 месяца назад +1

    Hey, Abonia..Thanks for the amazing content. I just had one issue though: on executing the 'map_reduce_outputs' function, I had the ConnectionRefusedError: [Errno 61].
    Hope someone know what it is

    • @AboniaSojasingarayar
      @AboniaSojasingarayar  4 месяца назад

      @@alvaroaraujo7945 Hello , thanks for your kind words.
      It may be related to your ollama serve.Are you sure Ollama is running ?

  • @evellynnicolemachadorosa2666
    @evellynnicolemachadorosa2666 8 месяцев назад +1

    hello! Thanks for the video. I am from Brazil. What would you recommend for large documents, averaging 150 pages? I tried map-reduce, but the inference time was 40 minutes. Are there any tips for these very long documents?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar  8 месяцев назад +1

      Thanks for you kind words and glad this helped.
      Implement a strategy that combines semantic chunking with K-means clustering to address the model’s contextual limitations. By employing efficient clustering techniques, we can extract key passages effectively, thereby reducing the overhead associated with processing large volumes of text. This approach not only significantly lowers costs by minimizing the number of tokens processed but also mitigates the recency and primacy effects inherent in LLMs, ensuring a balanced consideration of all text segments.

    • @VirtualMachine-d8x
      @VirtualMachine-d8x 4 месяца назад +1

      ​@@AboniaSojasingarayar Video was great and very useful.. can you make the small video on this clustering method using embedding ?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar  3 месяца назад +1

      @@VirtualMachine-d8x Sure will do, happy to hear from you again. Thanks for the feedback.

  • @Coff03
    @Coff03 8 месяцев назад +1

    Did you use OpenAI API key here?

    • @AboniaSojasingarayar
      @AboniaSojasingarayar  8 месяцев назад

      Here we use open-source Mixtral from ollama.But, yes we can use OpenAI models as well.