John Tan Chong Min
John Tan Chong Min
  • Видео 225
  • Просмотров 219 334
AriGraph: Learning Knowledge Graph World Models with Episodic Memory for LLM Agents
Excited to have Petr and Nikita present their work on AriGraph! It is a way to use episodic and semantic memory jointly to aid in retrieval of key information for an agent to perform decision making on. It leverages the power of Knowledge Graphs to sufficiently constrain the context and make it relevant for the agent. It also features a dynamically updatable way of modifying the knowledge graph based on most recent observations, thereby enabling the agent to have the most recent information.
I think knowledge graphs could form a semantically meaningful abstraction space for memory, on which association can be done. It would likely be better than naive RAG over data chunks. This, combined w...
Просмотров: 589

Видео

NeoPlanner - Continually Learning Planning Agent for Large Environments guided by LLMs
Просмотров 36421 час назад
Excited to have Swarna present his work on state space graph-based Planning, and how he used an Explore-Exploit approach to build and use this state space graph for future action planning. Speaker Profile - Swarna Kamal Paul completed his PhD on general problem solving agents from Jadavpur University on 2023. He also have 15 years of work experience in IT industry including consultancy, researc...
Intelligence = Sampling + Filtering
Просмотров 34014 дней назад
Recent work has shown that increased sampling can solve more of the Abstraction and Reasoning Challenge, solve dynamically changing mazes, solve programming problems, solve math olympiad problems etc. Can just simply increasing sampling give massive performance gains, or is it more than that? Many thanks to Derek and others for the wonderful discussion! Related Video (A Roadmap for AI): ruclips...
Michael Hodel: Reverse Engineering the Abstraction and Reasoning Corpus
Просмотров 87214 дней назад
Had a great discussion with Micahel Hodel and a few others (Simon Strandgaard, Yassine and many more) about reverse engineering the ARC dataset, and possible approaches to solve ARC. Speaker Profile: Michael studied Computer Science in Zurich, Switzerland. He is currently working as a freelance programmer, but his main focus remains ARC. Currently, his team MindsAI with Jack Cole and Mohamed Os...
TaskGen Conversational Class v2: JARVIS, Psychology Counsellor, Sherlock Holmes Shop Assistant
Просмотров 12028 дней назад
Here, I share the latest feature of TaskGen - ConversableAgent Now you can link your Agent to a Conversation Interface, much like AutoGen (but the Agent uses its own task-completion procedure that is not conversation based) I showcase a few projects: - Psychology Counsellor - Sherlock Holmes Shop Assistant - JARVIS (Simulated Iron Man Suit firing missiles and checking enemy status etc.) - Math ...
CodeAct: Code As Action Space of LLM Agents - Pros and Cons
Просмотров 461Месяц назад
Using Code as Action for LLM Agents improves accuracy by up to 20% compared to using JSON or Text for modular function calling approaches! Code allows for intermediate state caching, and draws on training data for native pythonic structures which can solve the problem more easily (e.g. for loop, min, max). Code also allows for multiple actions to be done together, which can solve potential plan...
TaskGen Conversation with Dynamic Memory - Math Quizbot, Escape Room Solver, Psychology Counsellor
Просмотров 1502 месяца назад
This is a showcase of what the Conversation Class of TaskGen contains, and how the memory is dynamically updated with each message. We use it to create a Math Quizbot, an Escape Room Solver, and a Psychology Counsellor. Notebook can be found here: github.com/simbianai/taskgen/blob/main/Tutorial 7 - Conversation Class.ipynb 0:00 Introduction 1:18 TaskGen Installation Conversation Class 11:59 Mat...
Integrate ANY Python Function, CodeGen, CrewAI tool, LangChain tool with TaskGen! - v2.3.0
Просмотров 2562 месяца назад
In this session, I show how to use LangChain tools, CrewAI tools, your own tools for a very flexible Agent in TaskGen! Spent 2 months creating a very versatile prompt that works with memory / plug and play functions. So, this is a showcase for TaskGen v2.3.0! Why TaskGen? Most existing Agentic frameworks use free-text conversation as the base for Agentic interactions. TaskGen uses StrictJSON, a...
Empirical - Open Source LLM Evaluation UI
Просмотров 2812 месяца назад
Had a great conversation with Empirical's CEO, Arjun Attam today. He has built a great open source tool to enable anyone to evaluate across any LLM, dataset and workflow procedure, as all you have to do is to put the LLM prompt / python script to a .json file, as well as whatever input/output dataset you would be using to evaluate. Essentially, Empricial's business model is to provide value for...
TaskGen Ask Me Anything #1
Просмотров 3872 месяца назад
TaskGen is an agentic framework which attempts to solve complex arbitrary problems by breaking them down into subtasks, and mapping each subtask to an equipped function to execute. In order to reduce verbosity (and hence token usage) that is typical of conversational-based agents like AutoGen or BabyAGI, it uses a JSON format for outputs. Specifically, TaskGen uses StrictJSON, a Large Language ...
StrictJSON (LLM Output Parser) Ask Me Anything #1
Просмотров 4073 месяца назад
StrictJSON is a python package I created last year in order to parse LLM outputs in a structured format (JSON), with optional type checking! How to use? Simply type in a system prompt (what the LLM should do), user prompt (the input to the LLM), and the output_format (the JSON format which you want the output to be in). It is configurable to use your own LLMs with the llm parameter, be it other...
Tutorial #14: Write latex papers with LLMs such as Llama 3!
Просмотров 2123 месяца назад
Here, I show you how to use GenAI to write latex code for easy inserting of figures, tables, equations into your paper. Note: While entirely possible, I do not advocate using GenAI to write the actual content of your paper - A paper is sacred and should reflect original and intellectual thought, do not use GenAI for writing your main ideas. If you want the reference links to have a hyperlink to...
SORA Deep Dive: Predict patches from text, images or video
Просмотров 3113 месяца назад
SORA by OpenAI is one of the best models I've seen for generating videos from text-based descriptions. Not only that, it can take a single image frame and make it a video, change a video according to some text instruction, or even interpolate between two videos. It is simply magical to see how large scale web-based video-text and image-text training can do to create this model. One of the key r...
OpenAI CLIP Embeddings: Walkthrough + Insights
Просмотров 5703 месяца назад
If there is one thing that has been impactful ever since its launch, it has to be CLIP Embeddings. CLIP stands for Contrastive Language-Image Pre-training. From Stable Diffusion to DALL-E to Robotics Tasks involving Vision and Text, CLIP bridges the gap between image and text using an embedding space common to both of them. Granted, CLIP is not able to do everything well - it struggles with the...
TaskGen - LLM Agentic Framework that Does More, Talks Less: Shared Variables, Memory, Global Context
Просмотров 6124 месяца назад
Want LLM Agents that focus on doing the task, and not be too verbose with "Thank You" and "I'm sorry"? TaskGen is the answer for you. Using StrictJSON at the core to ensure that the functions output in structured JSON with no output key missing, it makes the agents' output more concise. Problem: Too many possible action paths to solve a task Solution: Agent is only allowed to use the constraine...
CRADLE (Part 2): An AI that can play Red Dead Dedemption 2. Reflection, Memory, Task-based Planning
Просмотров 2764 месяца назад
CRADLE (Part 2): An AI that can play Red Dead Dedemption 2. Reflection, Memory, Task-based Planning
CRADLE (Part 1) - AI that plays Red Dead Redemption 2. Towards General Computer Control and AGI
Просмотров 6704 месяца назад
CRADLE (Part 1) - AI that plays Red Dead Redemption 2. Towards General Computer Control and AGI
TaskGen - A Task-based Agentic Framework using StrictJSON at the core
Просмотров 1,2 тыс.4 месяца назад
TaskGen - A Task-based Agentic Framework using StrictJSON at the core
SymbolicAI / ExtensityAI Paper Overview (Part 2) - Evaluation Benchmark Discussion!
Просмотров 1255 месяцев назад
SymbolicAI / ExtensityAI Paper Overview (Part 2) - Evaluation Benchmark Discussion!
SymbolicAI / ExtensityAI Paper Overview (Part 1) - Key Philosophy Behind the Design - Symbols
Просмотров 2935 месяцев назад
SymbolicAI / ExtensityAI Paper Overview (Part 1) - Key Philosophy Behind the Design - Symbols
Embeddings Walkthrough (Part 2): Context-Dependent Embeddings, Shifting Embedding Space
Просмотров 5205 месяцев назад
Embeddings Walkthrough (Part 2): Context-Dependent Embeddings, Shifting Embedding Space
Embeddings Walkthrough (Part 1) - Bag of Words to word2vec to Transformer contextual embeddings
Просмотров 7005 месяцев назад
Embeddings Walkthrough (Part 1) - Bag of Words to word2vec to Transformer contextual embeddings
V* - Better than GPT-4V? Iterative Context Refining for Visual Question Answer!
Просмотров 3326 месяцев назад
V* - Better than GPT-4V? Iterative Context Refining for Visual Question Answer!
AutoGen: A Multi-Agent Framework - Overview and Improvements
Просмотров 2,7 тыс.6 месяцев назад
AutoGen: A Multi-Agent Framework - Overview and Improvements
AppAgent: Using GPT-4V to Navigate a Smartphone!
Просмотров 5076 месяцев назад
AppAgent: Using GPT-4V to Navigate a Smartphone!
Tutorial #13: StrictJSON, my first Python Package! - Get LLMs to output into a working JSON!
Просмотров 6426 месяцев назад
Tutorial #13: StrictJSON, my first Python Package! - Get LLMs to output into a working JSON!
"Are you smarter than an LLM?" game speedrun
Просмотров 2527 месяцев назад
"Are you smarter than an LLM?" game speedrun
Is Gemini better than GPT4? Self-created benchmark - Fact Retrieval/Checking, Coding, Tool Use
Просмотров 3947 месяцев назад
Is Gemini better than GPT4? Self-created benchmark - Fact Retrieval/Checking, Coding, Tool Use
Learning, Fast and Slow: 10 Years Plan - Memory Soup, Hier. Planning, Emotions, Knowledge Sharing
Просмотров 6797 месяцев назад
Learning, Fast and Slow: 10 Years Plan - Memory Soup, Hier. Planning, Emotions, Knowledge Sharing
Tutorial #12: Use ChatGPT and off-the-shelf RAG on Terminal/Command Prompt/Shell - SymbolicAI
Просмотров 7798 месяцев назад
Tutorial #12: Use ChatGPT and off-the-shelf RAG on Terminal/Command Prompt/Shell - SymbolicAI

Комментарии

  • @tanchienhao
    @tanchienhao 5 часов назад

    awesome work! thanks for sharing

  • @yashtiwari3565
    @yashtiwari3565 День назад

    Every paper you talk about is a gem. I am a fellow AI enthusiast trying to create Jarvis from Marvel! and your videos always help me to get closer to my goal. Please never stop making youtube content!!

    • @johntanchongmin
      @johntanchongmin 22 часа назад

      Thanks for the affirmation. Also kudos to swarna for the great sharing :)

  • @simonstrandgaard5503
    @simonstrandgaard5503 День назад

    Impressive work. Many good insights. Thank you.

  • @johntanchongmin
    @johntanchongmin 2 дня назад

    Paper at 1:23:52 in which we talked about human priors in Atari Games, and how human priors are important for humans to solve games. Investigating Human Priors for Playing Video Games: arxiv.org/pdf/1802.10217

  • @dariusduesentrieb
    @dariusduesentrieb 6 дней назад

    Nice presentation! I found this in the wake of the new DeepMind AlphaProof announcement. It seems they are doing something very similar with AlphaProof like here or with GPT-f, but their way of generating data seems fancier: Reinforcement learning on problem statements that are translated from natural language into a formal language, and if I'm interpreting their writing correctly, they even do some online learning. Anyway. For the sampling from the policy network, I could even imaging the usage of MCTS inside MCTS :D Instead of greedily selecting next tokes, or using beam search, the most promising token sequences could be found using a monte carlo tree search, a bit like in "AlphaZero-Like Tree-Search can Guide Large Language Model Decoding and Training". I kinda like ideas about combining LLMs and MCTS.

  • @oncedidactic
    @oncedidactic 8 дней назад

    Nice thanks :) I like the discussion at 1:40:00

  • @iaman6047
    @iaman6047 14 дней назад

    you are talking no sense

    • @johntanchongmin
      @johntanchongmin 14 дней назад

      Thanks for your feedback. If you could elaborate which part you need clarification, I'll be happy to help

    • @pi5549
      @pi5549 13 дней назад

      @@johntanchongmin Recommend you simply delete such comments as these

  • @johntanchongmin
    @johntanchongmin 15 дней назад

    Slides: github.com/tanchongmin/TensorFlow-Implementations/blob/main/Paper_Reviews/Intelligence_Is_Sampling_And_Filtering.pdf

  • @ai-cowboy
    @ai-cowboy 15 дней назад

    Great video John

  • @architsch
    @architsch 16 дней назад

    Your videos are such a rich source of inspirations. Thank you for presenting these, it really helps me learn a great deal about the interplay between mathematics, language, and other areas of knowledge.

  • @mulderbm
    @mulderbm 18 дней назад

    Curious how far you are 😅 i relistened to this and now second probably we are trying to enhance for the same solution

    • @johntanchongmin
      @johntanchongmin 17 дней назад

      Hi there, if you are talking about the context-dependent embeddings, I have an implementation of it on github.com/tanchongmin/strictjson/tree/main/Experiments I found the most performant way is to re-adjust the text in the form that it incorporates the context. That way, the embedding model can be used without change. Check out the video here: ruclips.net/video/YZTThlPg0rU/видео.html

  • @oncedidactic
    @oncedidactic 19 дней назад

    great to have this interview!

  • @johntanchongmin
    @johntanchongmin 20 дней назад

    0:00 Speaker Introduction 0:45 Introduction to ARC-DSL 8:26 Data Generation 12:13 How close is DSL to human priors 14:42 How to decide which DSL to keep and which to add in 19:53 Introduction to RE-ARC 23:45 Overview of RE-ARC 25:32 Task Generalisation in RE-ARC 26:00 Example Verification in RE-ARC 26:47 Example Difficulty in RE-ARC 31:48 Limitations of RE-ARC 33:37 Examples of RE-ARC 35:22 Using RE-ARC to gauge model learning 37:48 Vision for meta-learning beyond RE-ARC 39:08 Can arbitrary DSL be generated with RE-ARC? 43:17 Discussion

  • @johntanchongmin
    @johntanchongmin 20 дней назад

    Repo: github.com/michaelhodel/re-arc Paper: arxiv.org/abs/2404.07353

  • @imsomeonesomebody
    @imsomeonesomebody 25 дней назад

    Regarding to the official paper: 1. In inference part, we're stacking R,s,a but a is empty, therefore how can we stack them? 2. If we're using a placeholder, i.e. R1, S1, NOP (that will stand for "no action" before a1 because it's not taken yet), we're getting a1 as a prediction in the inference 3. But in training we're not shifting actions, i.e. there's no mention that we're passing into the loss something like [NOP, a1, a2, ..] with [a_pred1, a_pred2,, a_pred3 ...] Therefore, the model is trained to predict current action 4. Since (R, s, a) after stacking stands for one token with dimension size which has tripled, inferring implies that we're getting (R1, s1, NOP) => a1 but the model clearly is trained (R1, s1, a1) => a1 because no shift is mentioned in paper. Moreover, if model is a function (R,s,a) => a, there's no point of R, s, it will just make it 1 and pass a => a. In teacher forcing, shifting is necessary 5. Nothing in the model points that the target is to maximize the return; it just generalizes relations between return, state, action tuples and some actions, unclear, the very same actions or the next actions. So I'm deeply interested in this topic but this concrete paper seems questionable despite being fundamental

    • @johntanchongmin
      @johntanchongmin 21 день назад

      If you look at my other videos on Decision Transformers (ruclips.net/video/qEoiLgdQC9k/видео.html), you will see that I am not a fan of reward-driven approaches. I would rather use goal-driven approaches, and I proposed a Goal-Directed Decision Transformer. Also, with regards to the actual implementation details, feel free to differ from the paper's implementation. The key thing is given a trajectory of states and actions, we can predict the next state/action using some model. That model is a Transformer for now but it can be some other model too. It is very cool how we can use self-supervised learning from your own trajectory to do the learning.

  • @doobie_doobie_doo
    @doobie_doobie_doo 25 дней назад

    What is the difference between this and the symbolic equation that KANs give?

    • @johntanchongmin
      @johntanchongmin 25 дней назад

      KANs are just backpropagation with different kind of basic units (MLP uses addition of input nodes with a non-linear activation, KAN uses different kinds of continuous functions applied to input) Symbolic Regression first does a genetic algorithm to select the equation parameters (like sin, cos, e^x and so on), then uses backprop to tune the weightages. In some sense, Symbolic Regression is less general than KANs, but I would say it is more stable. The intuition is that if you know more about what your final equation is, the less time you need to train. I'll let Kei Sen reply using his expert opinion as well :)

    • @doobie_doobie_doo
      @doobie_doobie_doo 12 дней назад

      Ohh okay. I think I understood. Thanks

  • @bastabey2652
    @bastabey2652 26 дней назад

    "maybe the text itself is way more expressive than the knowledge graph" ouch 1:16:10 in the age of LLM, it seems that any representation that deviates from the source text is a serious gamble

  • @johntanchongmin
    @johntanchongmin 29 дней назад

    Check out the repo here (Tutorial 7): github.com/simbianai/taskgen

  • @johntanchongmin
    @johntanchongmin 29 дней назад

    This Conversation Class shown here is for TaskGen v2.6.0. TaskGen/strictjson has been greatly revamped in v3.0.0 to work better for smaller LLMs

  • @rajathslr
    @rajathslr Месяц назад

    Forgive me if my question is not correct, Are we using LLMs to build a Knowledge graph here?

    • @johntanchongmin
      @johntanchongmin 27 дней назад

      Indeed, I showcased how to use strict_json (LLM) to build a Knowledge Graph. We also can use KGs to provide context for LLMs, so it works both ways.

    • @rajathslr
      @rajathslr 27 дней назад

      @@johntanchongmin 'We also can use KGs to provide context for LLMs' Can you please send me any video/blog link where I can read more about this

    • @johntanchongmin
      @johntanchongmin 26 дней назад

      @@rajathslr You can refer to the paper I covered in this video: arxiv.org/html/2306.08302v3 Can look at an example KG-RAG workflow here: www.nebula-graph.io/posts/KG_RAG_Bridging_Gap_Knowledge_and_Creativity_of_LLM_based_Agents

    • @rajathslr
      @rajathslr 26 дней назад

      @@johntanchongmin Thanks for your time, you really rock!!

  • @snehotoshbanerjee1938
    @snehotoshbanerjee1938 Месяц назад

    Hi John, any reason not going with yaml vis-a-vis JSON because yaml format consume less token consumption compared to json.

    • @johntanchongmin
      @johntanchongmin 27 дней назад

      I have considered yaml before. I don't think the llms are ready for it yet. I may change my mind in the future and do a yaml version of strictjson haha.

  • @snehotoshbanerjee1938
    @snehotoshbanerjee1938 Месяц назад

    Cool!

  • @johntanchongmin
    @johntanchongmin Месяц назад

    Update. This is live in TaskGen 4.2.1. Refer to Tutorial 6 for example of how to do Code Actions. My take: The code generated can be quite flimsy and harder to error-correct as compared to modular function calling. I think this method of using code is not very robust XD.

  • @snehotoshbanerjee1938
    @snehotoshbanerjee1938 Месяц назад

    John, do you have perspective on other multiagent frameworks like TaskWeaver, CrewAI (based on LC) and LangGraph?

    • @johntanchongmin
      @johntanchongmin Месяц назад

      I have yet to use TaskWeaver. For CrewAI, I think it is one of the better agentic frameworks out that, but it can also be too verbose since it is conversational-based. LangGraph tries to do what TensorFlow did instead of native Python. It feels unnatural to use, I recommend not using it.

  • @snehotoshbanerjee1938
    @snehotoshbanerjee1938 Месяц назад

    Great content!!

  • @lhc6340
    @lhc6340 Месяц назад

    great content and thorough walk through. cheers!

  • @shimotown
    @shimotown Месяц назад

    Agents can eat breakfast three times in a row 😂🎉🎉 great video and accent 😅

  • @syndicator
    @syndicator Месяц назад

    that is freaking awesome

  • @geoffcounihan7093
    @geoffcounihan7093 Месяц назад

    thanks for sharing!

  • @snehotoshbanerjee1938
    @snehotoshbanerjee1938 Месяц назад

    As always, Fantastic video!!. Your content is just awesome. As you say "Food for thought", your video gives me a lot of content to explore :)

    • @johntanchongmin
      @johntanchongmin Месяц назад

      Hope you enjoy the exploration process, come join the discord group for more intriguing conversations!

  • @JonnyBanana21
    @JonnyBanana21 Месяц назад

    Hi John can ChatDev help develop a Webapp/Website?

    • @johntanchongmin
      @johntanchongmin Месяц назад

      Hey there, you can try it out. Website and webapp, if they are part of Python programming, could be done too. I think the baseline is whether the LLM used is robust enough to generate code for your particular application. For cases that are too complex involving multiple files, it is best to design each file separately as the context length of ChatDev is the total sum of all the content of each file you need it to generate and more.

  • @johntanchongmin
    @johntanchongmin Месяц назад

    TaskGen repo: github.com/simbianai/taskgen

  • @johntanchongmin
    @johntanchongmin Месяц назад

    My slides: github.com/tanchongmin/TensorFlow-Implementations/blob/main/Paper_Reviews/Code%20as%20Actions.pdf

  • @johntanchongmin
    @johntanchongmin Месяц назад

    Github: github.com/xingyaoww/code-act Paper: arxiv.org/abs/2402.01030

  • @DynamicGenesis
    @DynamicGenesis Месяц назад

    Thank you! Excellent teaching style 😊

  • @gristlelollygag
    @gristlelollygag 2 месяца назад

    This seems very useful! Can't wait to get deeper into this, I will be needing it soon for an upcoming project

  • @johntanchongmin
    @johntanchongmin 2 месяца назад

    Fun fact, at 1:11:50 onwards I repeated "I am such a failure with life" three times because of a buggy code XD (had to try 3 times to get it right) Haha live coding is fun.

  • @johntanchongmin
    @johntanchongmin 2 месяца назад

    Key Highlights: 29:31 Escape Room Solver 1:09:30 Psychology Counsellor The Math Quizbot it is better done with rule-based methods, you can see 41:26 for how it can be done

  • @johntanchongmin
    @johntanchongmin 2 месяца назад

    Notebook can be found here: github.com/simbianai/taskgen/blob/main/Tutorial%207%20-%20Conversation%20Class.ipynb

  • @gemini_537
    @gemini_537 2 месяца назад

    Here is a similar research for the modulo operator. project-archive.inf.ed.ac.uk/msc/20172390/msc_proj.pdf

  • @gemini_537
    @gemini_537 2 месяца назад

    Gemini 1.5 Pro: This video is about whether a neural network can model if-else statements. The speaker, John Tan Chong Min, investigates this by coding a neural network to model a simple if-else statement. The code defines a function that takes in an input x. If x is greater than 30, then the function returns 0, otherwise it returns 1. The speaker then trains the neural network on this function. The speaker finds that the neural network has difficulty modeling the if-else statement. The network can only learn a simple linear function, rather than the piecewise linear function required for the if-else statement. The speaker concludes that traditional MLPs may not be well-suited for modeling if-else statements. Here are the key points: * The speaker defines a function that takes in an input x. If x is greater than 30, then the function returns 0, otherwise it returns 1. * The speaker trains a neural network on this function. * The neural network has difficulty modeling the if-else statement. * Traditional MLPs may not be well-suited for modeling if-else statements.

  • @sinaardehali7295
    @sinaardehali7295 2 месяца назад

    Can you compare Autogen, task and langchain agents ? Which one is more mature ? Do you know any other one ?

    • @johntanchongmin
      @johntanchongmin 2 месяца назад

      I think most mature is LangChain, simply because it is around longer. AutoGen is better than LangChain in getting Agents to talk to one another in a group setting. Most promising one so far I've seen is Crew.ai. TaskGen is relatively new, but it will eventually be the better option as I plan to incorporate a lot more things about memory and continual learning.

  • @faisalIqbal_AI
    @faisalIqbal_AI 2 месяца назад

    Thanks

  • @user-wr4yl7tx3w
    @user-wr4yl7tx3w 2 месяца назад

    Is it function calling for Llama3?

    • @johntanchongmin
      @johntanchongmin 2 месяца назад

      It uses StrictJSON, which parses the LLM output into a specified json format. Check it out here: github.com/tanchongmin/strictjson

  • @johntanchongmin
    @johntanchongmin 2 месяца назад

    Companion Notebook to this session: github.com/simbianai/taskgen/blob/main/TaskGen%20AMA_18May2024.ipynb

  • @johntanchongmin
    @johntanchongmin 2 месяца назад

    After this session, I went to massively improve TaskGen so that it works better for non-structured tasks and also not return so many errors due to list constraining or enum constrainer in Agent subtask generation logic. The Agent is also much more receptive to environment outputs and can modify it's plan more dynamically as it is not constrained to a list of steps anymore. The notebook for this session has been updated accordingly. Try out the new TaskGen v2.0.1 today!

  • @johntanchongmin
    @johntanchongmin 2 месяца назад

    Note: The information in this Notebook is largely correct, but for Functions, do take note variable names need to be enclosed by <> now, to let LLM understand better that it is an input. I will be releasing more updated tutorial videos in my StrictJSON/TaskGen channel once things become more concrete.

  • @FahmiFauzi-kf1wl
    @FahmiFauzi-kf1wl 2 месяца назад

    Great stuff! I've been looking for ways to generate output in JSON, and this helps me a lot. Thank you

    • @johntanchongmin
      @johntanchongmin 2 месяца назад

      Do refer to the most updated Tutorial for more information before you use it for your use case: ruclips.net/video/1N-znDTlhNc/видео.html

  • @Karl-Asger
    @Karl-Asger 2 месяца назад

    Looking forward to more on task gen. Great work

    • @johntanchongmin
      @johntanchongmin 2 месяца назад

      Tomorrow is TaskGen Ask Me Anything session, join my discord group for link!

    • @Karl-Asger
      @Karl-Asger 2 месяца назад

      @@johntanchongmin I joined the discord and saw that! Really happy for that, I spent my entire Sunday catching up on your projects through your great videos. Unfortunately I'm in CET so it will be 4am for me, but I'll catch up on it after! Thanks for sharing your work so enthusiastically 😎

  • @Karl-Asger
    @Karl-Asger 2 месяца назад

    Amazing work John, I hope I can be part of future meetings, I love the points that you are most focused on in these developments.