Is AutoGen just HYPE? Why I would not use AUTOGEN in a REAL use case, Yet

Поделиться
HTML-код
  • Опубликовано: 14 май 2024
  • In this video, we delve into AutoGen and learn about its functioning while distinguishing between the hype and reality. Although AutoGen is an excellent resource for creating multi-agent workflows, it is still unsuitable for production. Watch this video if you want to know what AutoGen can do beyond the hype.
    Read the blog post that complements this RUclips content: / autogen-isnt-practical...
    Need to develop some AI? Let's chat: www.brainqub3.com/book-online
    Register your interest in the AI Engineering Take-off course: • Building Chatbots with...
    Hands-on project (build a basic RAG app): www.educative.io/projects/bui...
    Stay updated on AI, Data Science, and Large Language Models by following me on Medium: / johnadeojo
    Multi-hop questions paper: arxiv.org/pdf/2108.00573.pdf
    Chapters
    Intro: 00:00
    Multi-hop questions: 01:43
    Multi-agent architecture: 05:02
    Testing the AutoGen workflow: 11:20
    Other AutoGen limitations: 33:10
  • НаукаНаука

Комментарии • 43

  • @denisblack9897
    @denisblack9897 3 месяца назад +8

    Very refreshing after a year of “shocking shock is changing everything” type of videos

  • @ObservingBeauty
    @ObservingBeauty 3 месяца назад +4

    I was waiting for someone to do exactly this video! Well explained. Clear. Thanks!!

  • @JiHa-Kim
    @JiHa-Kim 3 месяца назад +3

    Thanks for keeping it real.

  • @gileneusz
    @gileneusz 2 месяца назад +4

    I would say most of AI projects are far from being production ready. We are judging them by our established standards for regular non-AI stuff. After some time all these ai apps will get better, and we as an early adopters will have advantage of the experience of using them so we will utilize them faster than others. So don’t worry! We are on the right track.

    • @Data-Centric
      @Data-Centric  2 месяца назад +1

      I agree. I think the other part to consider is that LLMs are getting more powerful, we might get to a stage where an LLM could be powerful enough to make these agent frameworks viable.

  • @nicklove8857
    @nicklove8857 2 месяца назад +1

    Subbed. This is great. Thank you

  • @Juansonofgun
    @Juansonofgun 3 месяца назад +2

    I love your content. Keep up the excellent work.

    • @Data-Centric
      @Data-Centric  3 месяца назад +1

      Glad you enjoy it!

    • @synaestesia-bg3ew
      @synaestesia-bg3ew 3 месяца назад +2

      ​@@Data-Centric3 second only listening to you, i knows you are very smart.
      This is why i will finish the 38 minutes left.

  • @lmcfificoco
    @lmcfificoco Месяц назад

    Excellent content. Thks you

  • @mikew2883
    @mikew2883 2 месяца назад +2

    Excellent analysis! 👍 Have you had a chance to analyze CrewAI or TaskWeaver in comparison by chance?

    • @Data-Centric
      @Data-Centric  2 месяца назад

      Thank you! Haven't had a chance to yet, they're on my list. I suspect there might be similar issues...

  • @starblaiz1986
    @starblaiz1986 3 месяца назад +4

    Yeh, I've had very much the same experience with AutoGen and AutoGen Studio. It's very inconsistent to the point where I feel like I'm herding cats even for fairly simple workflows. Overall, at least right now, I feel its much better just to define your workflow using the raw python code and calling the LLM directly. AutoGen just feels like an unnecessary middleman that just gets in the way.

    • @Data-Centric
      @Data-Centric  3 месяца назад +2

      Yes, I agree! And herding cats is a great analogy. I had my fingers crossed the entire time hoping AutoGen would deliver the response I wanted.

  • @user-ep4ow9xd7y
    @user-ep4ow9xd7y 2 месяца назад

    Thanks for sharing.
    I still don’t see why would we use Autogen when semantic kernel is there and has support for assistants.
    What can autogen do in addition which makes it a better framework?

    • @Data-Centric
      @Data-Centric  2 месяца назад

      I haven't used semantic kernel before so it's impossible to say.

  • @issair-man2449
    @issair-man2449 2 месяца назад +2

    could i ask,
    what if we add multiple agents using multiple local LLM that are trained on different setup, by doing so i imagine that if each agent supervise the other and encourage it for better context and reasoning, perhaps the outcome may be more beneficial?
    what is your opinion

    • @Data-Centric
      @Data-Centric  2 месяца назад +1

      I find that mixing LLMs doesn't work too well. In a GroupChat, AutoGen shares the entire conversation context with every agent. When you're working with different LLMs (for example Mistral or Llama), the prompt formats play an important role in getting coherent responses. So mixing LLMs means mixing up prompt formats (given the shared context), which will almost certainly lead to unfavorable outcomes. On your point on reasoning, I'm not sure that we have solved the reasoning problem with any open-source LLM yet either.

    • @issair-man2449
      @issair-man2449 2 месяца назад

      @@Data-Centric thank you for sharing your insight, your videos are very interesting, eager to see whats next idea you will be sharing with us.

  • @jbo8540
    @jbo8540 2 месяца назад

    You had better results than most youtubers from a multi-agent framework. Most i've seen got purely hallucinated results and didn't realize it. I am curious what you are considering as an alternative to multi agent frameworks, if anything.

    • @Data-Centric
      @Data-Centric  2 месяца назад +1

      I think that when models become better at planning and reasoning, we might not even need multiple agents. As for now, I anticipate the problems I mentioned will be there with any of the current frameworks.

  • @jasonsting
    @jasonsting 3 месяца назад +1

    Thoughts on Langchain/Langraph+Agents or Open Interpreter

  • @Daviddd10
    @Daviddd10 2 месяца назад +1

    Do you have this on a github? I would look to look through your code and see how you organized this project with chainlit

    • @Data-Centric
      @Data-Centric  2 месяца назад

      I'll be releasing the code as a part of an online course.

  • @JC.72
    @JC.72 3 месяца назад +2

    How does autogen compare to CrewAI? Or any other potential tools?

    • @Data-Centric
      @Data-Centric  3 месяца назад +2

      I haven't tested CrewAI yet; I might do something on Crew shortly.

  • @notnotandrew
    @notnotandrew 3 месяца назад +1

    Is this necessarily an issue with AutoGen per se? It seems like it may be more of an issue with the underlying LLMs that drive it.

    • @Data-Centric
      @Data-Centric  3 месяца назад

      I think so; ultimately, they are unable to follow the reasoning traces. The entire conversation context is shared across each LLM in the agent workflow. So, as you can imagine, the context gets long after a while. Maybe there is some "forgetting" of the context in for more complex workflows. Anyhow, I agree it might be an LLM issue.

  • @lm-gu1ki
    @lm-gu1ki 3 месяца назад +2

    So gpt-4 turbo isnt' smart enough to be a good Group Chat Manager. How about the other roles? Perhaps you can keep the Chat Manager , but also add (hard-coded) logic to go back to the planner if the answer wasn't achieved (limiting total number of calls to avoid spending too much money)? It may still fail, but you'll see what the other weaknesses of this setup are.
    It would also be interesting to see some more realistic multi-hop questions, e.g., "'What events in the US corporate history resemble November 2023 OpenAI drama?" etc.

    • @Data-Centric
      @Data-Centric  3 месяца назад +2

      I think I agree with your logic, you need to hardcode the agent execution order. But essentially you're just building your own multi-agent framework from scratch, so just ditch AutoGen entirely. The other issues with costs, memory management, etc., would remain, I suspect.

  • @dipteshbosedb
    @dipteshbosedb 2 месяца назад +3

    Excellent video. I've have begun to perceive that these frameworks are being engineered to boost token consumption by model-as-a-service providers. It's crucial to recall that token utilization constitutes a significant portion of revenue for these companies. Unfortunately, we often find ourselves spending dollars just to assess the functionality and viability of these frameworks.

    • @Data-Centric
      @Data-Centric  2 месяца назад +1

      Thank you. I'll be honest, I'm not sure OpenAI or any proprietary LLM service has much to do with these frameworks. However, I wouldn't be surprised if there was some collaboration. I don't think it makes too much sense to purposely create frameworks that are token intensive and therefore impractical in production.

    • @jmsuhy
      @jmsuhy Месяц назад

      ⁠​⁠Microsoft is behind AutoGen …. It is in their best interest to ensure it first addresses OpenAI ‘stuff’ first. Creating autogen was probably on a strategic initiative document somewhere at MS.

    • @beauzero
      @beauzero 27 дней назад

      You can run a local. The framework supports it.

  • @xaxfixho
    @xaxfixho 3 месяца назад +2

    🙋‍♂️🙋🙋‍♀️Can read sheet, can you zoom in please 😮
    Otherwise why show the code at all 🤷‍♀️

  • @GreenmeResearch
    @GreenmeResearch 2 месяца назад

    Just run them locally on a Macbook Pro Max.....

  • @yoyartube
    @yoyartube 9 дней назад

    Have you had more or less success with other FWs?

    • @Data-Centric
      @Data-Centric  8 дней назад +1

      Doing one on CrewAI

    • @yoyartube
      @yoyartube 8 дней назад

      @@Data-Centric I subscribed and will look for it. Thanks!