Deploying AI
Deploying AI
  • Видео 9
  • Просмотров 59 580
Stream an Agent to a Copilot UI with Groq’s Fast Inference (Llama 3, Groq, LangGraph)
What does an agent-powered user experience feel like using today's fastest inference (Groq) and best open source model (Llama 3 70b)?
Inference speeds will get faster. Open source models will get better. But in this demo, it's obvious that agents are ready today.
This prototype covers:
- Experience an agentic workflow powered by Groq and Llama-3-70b
- Live stream data from every node in an agent workflow
- Build a collaborative, Copilot-like UI
I used the following tools:
AI Models - Llama-3-70b
Inference Provider - Groq
Agent Coordination - LangGraph, LangChain
Frontend - React, Radix UI, Tailwind
Interested in talking about a project? Reach out!
LinkedIn: linkedin.com/in/christianerice
Email: christ...
Просмотров: 1 593

Видео

Building Agents: Visualize a Multi-Agent Workflow that Outperforms a Single SOTA Prompt
Просмотров 4,5 тыс.Месяц назад
In this demo, I combine several agentic patterns - reflection, planning, and multi-agent workflows - to replace a complex prompt. I was able to match results from GPT-4 by combining multiple steps utilizing only GPT-3.5 and Claude Haiku. This video was inspired by Andrew Ng's recent work on agentic workflows, in which he demonstrates the potential to exceed state-of-the-art performance in LLMs ...
Building Agents: Self-Improving Prompt Engineering Agent that Runs Evals and Iterates on a Prompt
Просмотров 6 тыс.2 месяца назад
In this video, we'll build an agent that mimics a prompt engineer: it will iterate on a prompt, run evaluations against the prompt, and reflect and iterate on those eval findings as if it were a real Prompt Engineer. I'll also cover how to more generally set up, run, and score prompt evaluations, which is essential for this agent to work. You might want to watch the previous video on building a...
Build an Agent with Long-Term, Personalized Memory
Просмотров 24 тыс.2 месяца назад
This video explores how to store conversational memory similar to ChatGPT's new long-term memory feature. We'll use LangGraph to build a simple memory-managing agent to extract pertinent information from a conversation and store it as long-term memory via parallel tool calling. Interested in talking about a project? Reach out! Email: christian@botany-ai.com LinkedIn: linkedin.com/in/christianer...
Use the OpenAI API to call Mistral, Llama, and other LLMs (works with local AND serverless models)
Просмотров 3,5 тыс.3 месяца назад
An incredibly easy way to call local or hosted models using the exact same OpenAI API (or LangChain). Now that Ollama is compatible with the OpenAI API, we can use the same API to call either local models via Ollama or hosted models via most of the serverless LLM providers. Interested in talking about a project? Reach out! Email: christian@botany-ai.com LinkedIn: linkedin.com/in/christianerice ...
LangGraph Deep Dive: Agents with Parallel Function Calling
Просмотров 3,8 тыс.3 месяца назад
I reduced token usage by 87% by configuring my LangGraph agent to use parallel function calling. In this video, I walk through the basics of parallel function calling in LLMs, then dive into code on how to build a custom LangGraph agent from scratch that utilizes parallel function calling with OpenAI and LangChain. And in the second half of the video, I also run through a step-by-step walkthrou...
Run Code Llama 70B in the cloud
Просмотров 7 тыс.3 месяца назад
My laptop can't handle the new 70B parameter version of Code Llama. In this video, I'll walk through how you can still easily run Code Llama 70B (and other open models) via a hosted LLM using serverless APIs. This allows for token-based usage at a cost that's a fraction of OpenAI. We'll cover setup on Together (www.together.ai/) and Anyscale, (www.anyscale.com/) but there are many more provider...
Hands on with LangGraph Agent Workflows: Build a LangChain Coding Agent with Custom Tools
Просмотров 6 тыс.4 месяца назад
In this video, I walk through how to build a LangChain-writing agent using LangGraph. I'll build up from the basics of manually managing a conversation with OpenAI Tools and then walk through how to handle the same workflow with a custom agent built with LangGraph. Interested in talking about a project? Reach out! Email: christian@botany-ai.com LinkedIn: linkedin.com/in/christianerice Part 1: R...
Build an AI code generator w/ RAG to write working LangChain
Просмотров 2,4 тыс.4 месяца назад
Most AI models don’t have working knowledge of LangChain, let alone LangChain Expression Language. To resolve AI’s knowledge gap, I built a retrieval-augmented generation (RAG) development tool that passes relevant, instructive code examples to GPT-4’s context to successfully generate working LangChain and LCEL code View the GitHub repo here: github.com/christianrice/rag-coding-assistant And re...

Комментарии

  • @RamiAwar
    @RamiAwar 5 дней назад

    Hey Christian! Really enjoy your videos, are you on Twitter by any chance? Would love to share some stuff with you

  • @kayalvizhi8174
    @kayalvizhi8174 7 дней назад

    Did you forget to add the graphQl frontend code to Github ?

  • @trantrungduc2558
    @trantrungduc2558 10 дней назад

    your channel is really really good. Can you share what tool you use to create the web UI?

  • @st.3m906
    @st.3m906 13 дней назад

    You're an excellent teach Christian, thank you for this video!

  • @allenlawson9872
    @allenlawson9872 14 дней назад

    Hey I'm building a health tech app rn and your videos are so helpful. Thank you !

  • @gary3548
    @gary3548 15 дней назад

    “I won’t go through how I actually built the application” what’s the point ? I switched off here.

  • @flamed7s
    @flamed7s 17 дней назад

    I like your videos! Keep on doing more videos 😊

  • @dinugakasun5218
    @dinugakasun5218 18 дней назад

    Superb content, Thanks for sharing ✨

  • @HarpaAI
    @HarpaAI 21 день назад

    🎯 Key Takeaways for quick navigation: 00:00 *🧩 Swapping Between Local and Serverless Models* - Easily switch between local LLMs and serverless models using the OpenAI API. - Challenges and considerations when using serverless options like AMA for open AI compatibility. - Demonstrating how to toggle and set up models for different requests effortlessly. 03:37 *🚀 Utilizing Various Models in Tool Calling* - Exploring the models compatible with tool calling in serverless options like Together. - Highlighting the importance of model support for tool calling functionalities. - Demonstrating how to effectively run requests through different models and analyze their responses. 08:25 *💻 Optimizing the Model Selection Process* - Comparing the efficiency of different models for tool calling, including MRA, M, and Mistral. - Discussing the flexibility of interchanging models based on performance and cost factors. - Sharing insights on the future landscape of model swapping and the competitive advantages of serverless models. Made with HARPA AI

  • @HarpaAI
    @HarpaAI 21 день назад

    🎯 Key Takeaways for quick navigation: 00:00 *🤖 Agent Prototype Overview* - Developing a sophisticated agentic workflow involving multiple steps. - Streaming data back in bits instead of waiting for completion. - Exploring agent workflows in co-pilot experiences with Llama 3 and Groq for fast inference. 02:04 *🚀 Building AI Tools for Product Teams* - Creating an AI tool for rapid brainstorming and product vision establishment. - Extracting core ideas such as business objectives, customer problems, and enabling technologies. - Using a step-by-step approach with vision writing agents to refine the product vision. 05:48 *⚡️ Fast Inference with Llama 3 and Groq* - Leveraging Groq's super-fast inference engine for agent workflows. - The speed advantage of Groq's 300 tokens per second compared to other models. - The future trend towards faster and cheaper inference models like Llama 3 challenging traditional models. Made with HARPA AI

  • @HarpaAI
    @HarpaAI 21 день назад

    🎯 Key Takeaways for quick navigation: 00:01 *🛠️ Building a self-improving prompt engineering agent* - Explained the concept of developing an agent capable of running evaluations, revising prompts, and learning from its own evaluations iteratively. - Shared the process of setting up evaluations, scoring runs, and building the agent to handle prompt iteration. - Emphasized the need for thorough evaluations to ensure confidence in shipping the prompt. 08:12 *🔄 Utilizing AI for prompt generation and evaluation loops* - Described a loop mechanism involving an agent and tools for prompt revision and evaluation. - Discussed the importance of using specific models for prompt development, evaluation, and validation. - Introduced the concept of passing learnings from evaluations back to the agent for continuous improvement. 13:54 *🤖 Automated prompt engineering process overview* - Provided an overview of generating synthetic data for evaluations and validating prompt performance. - Demonstrated the loop where the agent iteratively revises prompts, runs evaluations, and updates based on learnings. - Highlighted the significance of leveraging different AI models for specific stages within the prompt engineering process. 20:12 *🔄 Iterative prompt generation and testing process* - Described the iterative loop process involving running evaluations, revising prompts, and testing new versions. - Explained the role of the prompt writer in generating updated prompts based on evaluation results. - Highlighted the importance of using smart models for prompt writing to improve accuracy over time. 22:04 *🧰 Tools and datasets for prompt generation and evaluation* - Shared details on the code structure, including agents, data graphs, and tools for prompt generation. - Demonstrated the process of generating synthetic data for evaluations and fine-tuning models. - Discussed the flexibility to choose between different prompt writing tools and methodologies for evaluation. 30:02 *🧠 Final optimization steps and considerations for prompt engineering* - Discussed potential risks of over-optimization and the need to counteract them in prompt engineering. - Emphasized the significance of reserving a fresh, untouched dataset for final confirmation testing. - Encouraged experimentation while iterating on the process and seeking improvements or existing tools to enhance prompt engineering techniques. Made with HARPA AI

  • @HarpaAI
    @HarpaAI 22 дня назад

    🎯 Key Takeaways for quick navigation: 00:00 *🧠 Building an agentic workflow from a complex prompt* - Demonstrated the process of dividing a single prompt into multiple agentic steps. - Divided the prompt into memory extraction, reflective review, action assignment, and category assignment steps. - Discussed the importance of breaking down prompts for improved accuracy and cost efficiency. 02:44 *🔄 Andrew Yang's Four Agentic Workflow Methods* - Shared insights from Andrew Yang's work on improving agentic workflows using reflection, tool use, planning, and multi-agent collaboration. - Explained how combining these methods can enhance the quality of results in workflows. - Highlighted the benefits of incorporating reflection, tool use, planning, and multi-agent collaboration in workflows. 05:17 *🚀 Enhancing Accuracy with Multi-Agent Workflows* - Demonstrated the implementation of a multi-agent workflow in processing prompts. - Showcased the division of the prompt into memory extraction, action assignment, and category assignment steps with reflective feedback loops. - Discussed the trade-offs between accuracy, cost efficiency, and processing speed in multi-agent workflows. Made with HARPA AI

  • @HarpaAI
    @HarpaAI 23 дня назад

    🎯 Key Takeaways for quick navigation: 00:00 *🧠 Building a demo of a memory feature similar to ChatGPT's long-term memory:* - Building a cooking application to learn about personal habits and preferences. - Extracting information from conversations about allergies, food likes, and family attributes. - Demonstrating information extraction and memory feature in a conversation simulation. 03:28 *🤖 Implementing a memory management system inspired by GPT research:* - Exploring the combination of long-term and short-term memory in memory management. - Condensing messages into long-term memory after a certain threshold for optimization. - Utilizing archival and recall storage to create a comprehensive conversational memory chain. 05:18 *🔄 Setting up a memory system with separate agents for conversation management:* - Developing a conversational agent and a memory agent running in parallel. - Using Lang chain and Lang graph for memory management and reasoning. - Establishing a memory tool caller for managing memory creation, update, and deletion. 09:56 *🖥️ Implementing memory agents and memory management logic:* - Detailing the memory Sentinel's function to analyze and record relevant information. - Creating a memory manager tool for efficient memory tracking and database operations. - Demonstrating the process of memory creation, updates, and deletions using a tool caller. 14:22 *🧭 Setting up a memory graph for efficient memory and knowledge processing:* - Defining nodes and edges for memory agents, knowledge structures, and tool calling. - Illustrating the flow of information between different stages of memory processing. - Detailing the setup and execution of memory processing steps in the memory graph. 19:48 *🔍 Exploring memory optimization and customization possibilities:* - Discussing potential optimizations for storing and retrieving specific memory types. - Considering the selective retention of information based on agent requirements. - Addressing methods to tailor memory processing for different agents and use cases. 21:11 *🧠 Optimizing memory processing for efficiency and effectiveness:* - Analyzing when to condense message logs or save memories in a user session. - Exploring different strategies to optimize short-term and long-term memory retention. - Considering factors such as speed, cost, and accuracy in memory agent optimization. Made with HARPA AI

  • @Balphis
    @Balphis 26 дней назад

    interesting UX decisions. I've also considered the idea of streaming progress from these agentic workflows for the end-user to observe the system doing its job but also just making for a more snappy experience.

  • @user-dk8dm8db8t
    @user-dk8dm8db8t 26 дней назад

    Hi can you publish your langsmith traces for this? I am trying to implement this for models without tool calling. It will be incredibly helpful

  • @mynameisedu
    @mynameisedu 26 дней назад

    A code walkthrough would be awesome!

  • @varunmehra5
    @varunmehra5 26 дней назад

    Would really appreciate if you shared some code blocks for Langgraph and langchain

  • @user-dk8dm8db8t
    @user-dk8dm8db8t 27 дней назад

    Thanks for the super cool demo. Can you provide your langsmith traces for us to inspect? it would be really easy to track the flow engineering. Thanks!

  • @CaseyLisak
    @CaseyLisak 27 дней назад

    Is there a follow up video for the code walk through?

  • @ubaba06
    @ubaba06 27 дней назад

    what ui are you using?

    • @mr.daniish
      @mr.daniish 26 дней назад

      looks like a vite, react app

    • @ubaba06
      @ubaba06 26 дней назад

      @@mr.daniish thanks 🙏🏻

    • @deployingai
      @deployingai 26 дней назад

      Yes, that's right. It's a Vite React app, and I used Radix and Tailwind for components and styling.

  • @st.3m906
    @st.3m906 27 дней назад

    Super cool! What made you want to build it in typescript?

    • @mr.daniish
      @mr.daniish 26 дней назад

      the brower maybe?

    • @deployingai
      @deployingai 26 дней назад

      Once you use TypeScript it's hard to go back! Typing is just great to have for maintainability and error detection even for a team of one, and really makes sense as you scale a team or product. My backend is still currently Python since there seems to be better Python support for AI tools, but I use TypeScript for everything in the frontend.

  • @zyu5354
    @zyu5354 Месяц назад

    This langchain codes work well for one Q&A, but how to modify these to allow interactive conversations, and still able to handle the tools? Thanks!

  • @tomasbird
    @tomasbird Месяц назад

    Great content, one of the unfortunately underrated channels on youtube about LLMs real use cases. no bs, straight to action. love it. keep up the good work.

    • @deployingai
      @deployingai 26 дней назад

      Thanks for watching and for your feedback!

  • @joao.morossini
    @joao.morossini Месяц назад

    Excelent content! Thanks for sharing, man :D

  • @KristijanKL
    @KristijanKL Месяц назад

    my problem is that long term memory drastically increases prompt size. so you either need multiple long term memories depending on type of the prompt or local AI that server as AI router deciding what prompt needs from long term memory

  • @VitthalGusinge
    @VitthalGusinge Месяц назад

    what code examples did you pushed in weaviate ?

  • @Mikoto_401
    @Mikoto_401 Месяц назад

    Thank you for your videos, your videos help me much more than the official LangChain videos. Could you please also a video where you use LangChain Agents e.g. the Tool Calling Agent?

  • @al3030
    @al3030 Месяц назад

    Thanks for sharing! Have you experimented with asking the models to generate prompts for you for each step? It could accelerate the workflow building :)

  • @junilkim8901
    @junilkim8901 Месяц назад

    Hi! Really good walkthrough. Have one question. How did you create and deploy the ui front end part of this? Was wondering if you used the lang serve. As a non-developer wanting to create quick proof of concept for potential users, was wondering if there's a lean way to deploy on local the user interface. Realized your github only have ipynbs. Thank you so much!

  • @nipunj15
    @nipunj15 Месяц назад

    Amazing improvement on your last video on Memory. Any way I can get access to the Frontend you're using in your videos?

    • @NoWayFolding
      @NoWayFolding Месяц назад

      This! Are you able to share it? Would even pay a small fee for it.

  • @daviddamifogodaramola3596
    @daviddamifogodaramola3596 Месяц назад

    Thanks for this video, it's really helpful!!

  • @unclecode
    @unclecode Месяц назад

    Great work, don't you mind to share the code for front end? ;)

  • @jazearbrooks7424
    @jazearbrooks7424 Месяц назад

    Are you open to consulting? I just emailed you

  • @viky2002
    @viky2002 Месяц назад

    I did something similar, - generate triplets from the information - check / review triplets (if bad refine, if good go to next step) - save to neo4j as knowledge graph

    • @deployingai
      @deployingai Месяц назад

      Awesome! How did that approach work for you? If this were going to production, I'd definitely compare a few different workflow approaches.

  • @75M
    @75M Месяц назад

    Thanks for sharing this project!

  • @jeevs77
    @jeevs77 Месяц назад

    Thanks for this straightforward explanation of agentic memory formation. I’m very curious about how you chose which layers should use Anthropic Claude Haiku vs OpenAI GPT-3.5-turbo.

    • @deployingai
      @deployingai Месяц назад

      Great question! Since this was just exploratory, I didn't give it too much thought. In my first iteration, I used GPT-3.5 for each step, and I didn't find it to be sufficiently critical of itself for reflection. I chose GPT-3.5 since it was one of the best inexpensive models that supported reliable JSON output and tool calling, and building and trying demos like this is effectively free to do. But now that Claude supports tool calling, I pulled in Haiku for reflection to give that a try. Its output is a bit more critical, but the prompts could use some work to improve it. If this were going to production, I'd evaluate the model choices a lot more carefully than I did for this demo.

    • @jeevs77
      @jeevs77 Месяц назад

      @@deployingai Thanks for sharing your logic. I have been very impressed by the Claude 3 models’ new tool use capabilities too.

  • @ingoeichhorst1255
    @ingoeichhorst1255 Месяц назад

    What are you using to build the Frontend? looks neat

    • @thetagang6854
      @thetagang6854 Месяц назад

      Darude sandstorm

    • @deployingai
      @deployingai Месяц назад

      I used Radix UI and Tailwind, great for throwing something together quickly!

    • @VitthalGusinge
      @VitthalGusinge Месяц назад

      @@deployingai can you share the code for this UI

  • @mwdcodeninja
    @mwdcodeninja Месяц назад

    Grok would be fast as hell. I would be interested to see the performance looping grok in.

    • @hiranga
      @hiranga Месяц назад

      likwise. But currently find groq does not output consistently for tool_calls / JSON. Any experience with improving this?

    • @mwdcodeninja
      @mwdcodeninja Месяц назад

      @hiranga The only thing I can think of is you don't depend on grok for doing the main function calling and controlling of the application logic flow. So I would use gpt4 as the app controller / router and delegate work tasks to faster models. Each model is going to have its own Unique challenges.

    • @deployingai
      @deployingai Месяц назад

      Good idea! You're right, if speed is important then grok could offer a big gain. And you could probably rework the workflow to reduce its reliance on structured outputs if that proves to be a problem.

  • @jimbob3823
    @jimbob3823 Месяц назад

    Thinking if there would be a way for it to build a knowledge graph... Somehow?

    • @deployingai
      @deployingai Месяц назад

      Yeah this could be a great use case for a knowledge graph. It would easily make sense for family members and foods to be entities with relationships likes like/dislike/allergy, and the whole 'attributes' catchall could be expanded to encompass much richer data.

  • @benjaminbascary8219
    @benjaminbascary8219 Месяц назад

    Can you share the vite front end? or how you setted up the front and the backend?

  • @mindseye123
    @mindseye123 Месяц назад

    I liked the overview you did before jumping into the code - so helpful! Most tutorials just jump a s s backwards into the code.

  • @TK_Prod
    @TK_Prod Месяц назад

    🎯 Key Takeaways for quick navigation: 00:01 *🤖 Building a Self-Improving Prompt Engineer Agent* - Creating an agent capable of self-improving prompt engineering, running evaluations, revising prompts based on feedback, and learning independently. - The agent autonomously conducts prompt engineering without human intervention, relying on self-learning and iterative improvement. 02:08 *💡 Evaluation Setup and Testing* - Setting up evaluation scenarios involves creating diverse input messages, expected outputs, and testing conditions. - Synthetic data generation and validation are essential for building an accurate evaluation dataset. - Evaluations compare model outputs to expected results, determining their accuracy and guiding prompt revisions. 08:12 *🔄 Iterative Prompt Engineering Loop* - The agent continuously revises prompts based on evaluation feedback, improving its performance over successive iterations. - Prompt revisions are guided by insights from previous evaluation runs, aiming to enhance the model's effectiveness. - Leveraging feedback mechanisms and evolutionary algorithms, the agent refines prompt engineering autonomously. 17:40 *🚀 Autonomous Prompt Engineering Process* - The agent's goal is to generate effective prompts capable of handling diverse scenarios with confidence. - By cycling through evaluation iterations, the agent gains confidence in prompt engineering, leveraging extensive evaluation data. - The process culminates in the creation of prompts suitable for various conversational contexts, ensuring robust performance. 19:01 *🧩 Prompt Writing and Evaluation Process* - Describes the process of writing a new prompt and testing it against an evaluation dataset. - Components of the prompt include the user message, memories, and desired responses. - The evaluation involves comparing expected and actual outputs, iterating until a satisfactory accuracy score is achieved. 20:24 *🔄 Iterative Prompt Optimization* - Explains the iterative nature of the prompt optimization process. - Details how failures trigger revisiting the prompt writer with updated information. - Emphasizes the importance of a smart model underlying the prompt writer for better results over time. 23:49 *📊 Dataset Generation Process* - Outlines the process of generating the evaluation dataset. - Describes steps including prompt generation, memory creation, and output mutation. - Mentions the possibility of fine-tuning a model based on the dataset for comparison. 27:14 *🛠️ Graph Setup and Tools* - Discusses the setup of the Lang graph and the tools utilized within it. - Key components include the prompt writer, tester, and evaluator. - Explains the decision-making process within the graph and the tool calls for prompt writing and evaluation. 30:02 *💻 Code Overview and Considerations* - Provides an overview of the code structure and functionality. - Highlights potential areas for improvement and considerations, such as avoiding over-optimization and preserving a fresh evaluation dataset. - Encourages experimentation and feedback for further enhancements. Made with HARPA AI

  • @matthewwalker7063
    @matthewwalker7063 Месяц назад

    Is this using DSPy

    • @deployingai
      @deployingai Месяц назад

      I haven't actually built anything with DSPy yet, but it looks awesome. I used LangGraph for this demo because I was primarily curious about using reflection within an agent loop and wanted to share some info about evals, so I settled on prompt engineering as a good example to cover both. But I'll be exploring DSPy soon.

  • @GiovanneAfonso
    @GiovanneAfonso Месяц назад

    Great video!!!!

  • @GiovanneAfonso
    @GiovanneAfonso Месяц назад

    beautiful UI for evals

  • @genailearnerfver
    @genailearnerfver Месяц назад

    @deployingai amazing video, Is it possible to share the Miro link, Thank you !

  • @genailearnerfver
    @genailearnerfver Месяц назад

    Great video christiane, is it possible to share Miro link. Thanks

  • @bertobertoberto3
    @bertobertoberto3 Месяц назад

    That was excellent

  • @curiouslycory
    @curiouslycory 2 месяца назад

    I really appreciate all of your videos. For some reason the volume on all of the content that I've seen you post is like 50% more quiet than other videos on youtube. The "why" of how these things are put together is great in your videos and I'm always glad to have a better understanding of the underlying context.

    • @deployingai
      @deployingai Месяц назад

      Thanks for the feedback, I appreciate it! I'll have to look into the audio issue, I obviously haven't spent any time on production quality but hopefully that's a quick fix.

  • @thawab85
    @thawab85 2 месяца назад

    would love to hear your opinion about using dspy, you mentioned llm compiler in your parallel function video. dspy team released a paper called storm where langchain team implemented here: ruclips.net/video/1uUORSZwTz4/видео.html