Analytics Camp
Analytics Camp
  • Видео 33
  • Просмотров 44 954
🤯 Mind-blowing AI Agents Systems That Will Change The Industry: OpenAI Swarm, Google RIG, Adala
#aiagents #llm #rig #swarm
Here are new and super-interesting AI agent systems and LLM research that Fortune 500 companies plan to introduce that will impact industry and research, including Google RIG technique in DataGemma, Open AI’s Swarm system, and Adala's automatic labelling system.
Related Terms and Concepts:
Retrieval Augmented Generation (RAG), Retrieval Interleaved Generation (RIG), LLM hallucination, AI agent coordination, automatic LLM labelling, Swarm agent execution and testing, Data Commons, Internet of Agents (IOA) system
Latest AI Agents research and agentic frameworks:
ruclips.net/video/lA3Tju4VUho/видео.html
A list of the AI Agentic systems and research and video tutorials:...
Просмотров: 61

Видео

Mastering LLM Fine-Tuning with QLoRA: Quantization on a Single GPU + Code
Просмотров 601День назад
#llm #finetuning #qlora Fine-tune like a pro: A complete beginner's tutorial on fine-tuning Large Language Models (LLMs) with QLoRA or Quantized Low-Rank Adaptation technique, on a single GPU! If you are GPU-poor and low on memory, this video is for you. :) Bonus steps: Save your fine-tuned model locally and test or use it for your own tasks. Key Terms and Concepts: Parameter-efficient fine-tun...
✅ All You Need to Fine-tune LLMs With LoRA | PEFT beginner’s tutorial & code
Просмотров 300Месяц назад
#finetuning #llm #lora #peft The easiest way to fine-tune a large language model with LoRA or low-rank adaptation method: Parameter Efficient Fine Tuning ( PEFT) beginner’s hands-on tutorial with code link. The code & process of fine-tuning with LoRA: github.com/Maryam-Nasseri/Fine-tuning-with-QLoRA Full fine-tuning large language models: ruclips.net/video/oG0jsMVTg9w/видео.html Key Terms & Con...
✅ Easiest Way to Fine-tune LLMs Locally + Code - NO GPU Needed!
Просмотров 433Месяц назад
#ai #llm #finetuning #nlp This is the easiest way to fine-tune a large language model 100% free and locally, without using any API or third-party platform in just 20 minutes! You don't need GPU, we will run this model training on CPU only! The code and process for fine-tuning are in my GitHub repository; the link is below :) Transformer Language Models Simplified in JUST 3 MINUTES! ruclips.net/...
Internet Of Agents (IOA) Can Dominate AI Agents Systems! 🤯 Here’s Why
Просмотров 4242 месяца назад
#aiagents #llm #ai #ioa The Internet of Agents or IOA system for large language models is inspired by how the Internet works, more specifically decentralised collaborative projects such as Wikipedia and Linux, to build an Internet-like system for autonomous agents. This design has the potential to dominate the AI Agents systems in the near future so watch the video to see why. Based on the pape...
Retrieval Augmented Generation (RAG) vs In-Context-Learning (ICL) vs Fine-Tuning LLMs
Просмотров 3793 месяца назад
#ai #rag #llm #prompt This video is a simplified explanation of Retrieval Augmented Generation (RAG) vs In-Context-Learning (ICL) vs Fine-Tuning LLMs for beginners; three terms and concepts related to the ways to use Large Language Models and increase the accuracy of their responses. I have already explained these in my previous videos separately so I thought I would give you all in one place. ...
🔴 Mixture of Agents (MoA) Method Explained + Run Code Locally FREE
Просмотров 2,7 тыс.3 месяца назад
#aiagents #moa #llm #mistral This video is an easy explanation of the Mixture of Agents (MoA) method and algorithm and a tutorial on how to run a MoA multi-LLM AI agents locally and 100% FREE. The method is discussed in this paper: Wang et al. (2024). Mixture-of-Agents Enhances Large Language Model Capabilities. Full process to run the MoA system in this GitHub repo: github.com/Maryam-Nasseri/M...
✅ Easiest Way to Build AI Agents With RAG & CrewAI Locally
Просмотров 5 тыс.4 месяца назад
#aiagents #crewai #rag #llama3 This tutorial is the easiest way to build AI agents with crewai and LLAMA3 that use a RAG or Retrieval-Augmented Generation framework for setting up a repeatable workflow for your specific tasks with your own documents, which reduces hallucinations and irrelevant responses and improves the results of the LLMs. Set up your virtual environment for crewai with this e...
Is ICL The Next Big Thing In AI? 🤯 Replace Fine-tuning LLMs
Просмотров 3154 месяца назад
#ai #gpt4o #llm When GPT-4o or omni was released, there was so much excitement about the super cool voice and visual recognition and generation that we overlooked an important aspect of it that I will be discussing today based on this paper by several Stanford University researchers including Andrew NG: Jiang et al. (2024). Many-Shot In-Context Learning in Multimodal Foundation Models Related v...
💯 FREE Local LLM - AI Agents With CrewAI And Ollama Easy Tutorial 👆
Просмотров 3,9 тыс.5 месяцев назад
#aiagents #crewai #tutorials Follow along with this super easy code tutorial to set up your local agentic workflow which is 100% free. I will use VSCode and show you how to install CrewAI and Ollama to work in your virtual environment. Key Terms and Concepts: agentic, AI agents, LLM, crewai, self-refine, refinement method, mistral, Mistral AI, API, ollama, crew, task, multiagent Check out this ...
🔴 This Agentic AI Workflow Will Take Over 🤯 Algorithm + Papers Explained
Просмотров 13 тыс.6 месяцев назад
#ai #llm #aiagents #agentic What Language Model To Choose For Your Project? 🤔 LLM Evaluation: ruclips.net/video/PXX2OO7s8wY/видео.html : evaluation of Hugging Face models Please subscribe to support this channel :) Explanation of the papers and algorithms of LLM agents in the Agentic AI systems (see timestamps below) using the following concepts and papers: Iterative feedback and refinement for...
How Sora AI works: OpenAI Text-To-Video Model and LLM
Просмотров 6657 месяцев назад
#ai #llm #sora The Model Behind Sora AI Simplified! OpenAI Text-To-Video LLM Have you wondered what the main reason is that makes Sora so good at generating videos? What ingredient is added to the base diffusion model that has taken video generation to a whole new level, setting the scene for unlimited creativity? I will discuss this text-to-image and video model in this video, along with the L...
Beginner's Tutorial: Locally Run Models With Ollama 🤫 5 Steps to Improve LLM Results
Просмотров 1787 месяцев назад
#ai #llm #llama #orca I will test the 5-step method for improving the results of large language models on a famous reasoning test between llama 2 and orca 2. I have spent quite some time getting to the bottom line so you don’t have to! Watch till the end to see the winner and how these steps help improve the models’ responses. This test is one example of the GSM8K test used to evaluate LLMs in ...
What Language Model To Choose For Your Project? 🤔 LLM Evaluation
Просмотров 5688 месяцев назад
#llm #huggingface #gpt4 #ai With more than 490,000 language models uploaded in the Hugging Face model repositories, how do you find the best language model for your personal or business projects? I have spent two weeks searching for the best models so you don’t have to. In this video, you will get to know all details about Hugging Face LLM Leaderboard and how it evaluates all the models objecti...
Is Mamba Destroying Transformers For Good? 😱 Language Models in AI
Просмотров 6 тыс.8 месяцев назад
#mamba #transformers #llm #mambaai The Transformer language model has one main problem that can make it go extinct even before it blasts off in 2024! The model architectures and performance differences of the Transformer and Mamba language models. I will compare the functionalities of the main AI and machine learning models, and show the necessary improvements in the Mamba AI model compared to ...
Mamba Language Model Simplified In JUST 5 MINUTES!
Просмотров 7 тыс.9 месяцев назад
Mamba Language Model Simplified In JUST 5 MINUTES!
The Concept of Backpropagation Simplified in JUST 2 MINUTES! -- Neural Networks
Просмотров 9229 месяцев назад
The Concept of Backpropagation Simplified in JUST 2 MINUTES! Neural Networks
Transformer Language Models Simplified in JUST 3 MINUTES!
Просмотров 51610 месяцев назад
Transformer Language Models Simplified in JUST 3 MINUTES!
5 AI Tools and Features to Expect in 2024 That You NEED to Know!
Просмотров 9610 месяцев назад
5 AI Tools and Features to Expect in 2024 That You NEED to Know!
This is how EXACTLY Language Models work in AI - NO background needed!
Просмотров 36610 месяцев назад
This is how EXACTLY Language Models work in AI - NO background needed!
Is Claude better than ChatGPT? SHOCKING Claude going viral
Просмотров 62Год назад
Is Claude better than ChatGPT? SHOCKING Claude going viral
ChatGPT Prompt Engineering 101: 5 Best Prompt Hacks Everyone SHOULD know!
Просмотров 90Год назад
ChatGPT Prompt Engineering 101: 5 Best Prompt Hacks Everyone SHOULD know!

Комментарии

  • @optiondrone5468
    @optiondrone5468 День назад

    That Adala framework looks like a game changer. Most AI devs spend a lot of time labeling data for their training.

    • @analyticsCamp
      @analyticsCamp День назад

      Yep, but not just for dev projects, but also for research in academia. I can tell from my own experience how long it takes for researchers to label data. Let me know if you try this system :)

  • @ElObredor
    @ElObredor 5 дней назад

    how can i access to the notebook? no entiendo nada ;C

    • @analyticsCamp
      @analyticsCamp 5 дней назад

      Hi, If you mean the code and process, then the process is explained in the video; you can access separate files and functions here: github.com/Maryam-Nasseri/Fine-tuning-LLMs-Locally

  • @Researcher100
    @Researcher100 12 дней назад

    "Basically", I really liked this tutorial! Does this setting work with a larger model, say 7B?

    • @analyticsCamp
      @analyticsCamp 12 дней назад

      Thanks for watching! Should work with a 7b-model too if you have more VRAM, set the cuda to True in the training arguments, and keep the batch size lower. Hope it helps :)

  • @yoyartube
    @yoyartube Месяц назад

    With no cude true do you think I could fine tune deepseek llm on my mac m2 with 16 gigs of ram?

    • @analyticsCamp
      @analyticsCamp Месяц назад

      Running LLMs is more about VRAM than RAM; I'd say you need 6+ VRAM (larger the model, larger the training dataset => more processing needed). Maybe start with DeepSeek base which is only 7B (see if any Q4 is available). Running on powerful CPU is possible ; I don't use MAC so I cannot comment on it :)

  • @Nathan-pu9um
    @Nathan-pu9um Месяц назад

    Using tools like n8n low code you can do this alot easier

    • @analyticsCamp
      @analyticsCamp Месяц назад

      I agree, but for deployment and wider use there is pricing for n8n, which could be beyond some users' budget (unlike CrewAI which can work with local LLMs free!). But thanks for watching :)

    • @Nathan-pu9um
      @Nathan-pu9um Месяц назад

      @@analyticsCamp I agree but, you can use n8n to create workflows connected to Pinecone or a vector database so you can make your own agentic custom workflow internally

  • @chadricheson1038
    @chadricheson1038 Месяц назад

    This channel is underrated.

  • @paradigmnnf
    @paradigmnnf Месяц назад

    OK, so where is the paper?

    • @analyticsCamp
      @analyticsCamp Месяц назад

      Hi, the full references of all the papers are cited in the description box :)

  • @optiondrone5468
    @optiondrone5468 Месяц назад

    Love it, thanks for demystifying many of the fine tuning terms and their use! 👋 keep up the good work👍

    • @analyticsCamp
      @analyticsCamp Месяц назад

      Thanks, will do and more to come!

  • @sai_ai-z1c
    @sai_ai-z1c Месяц назад

    SmythOS seems like a great way to increase productivity! I've been trying to find ways to make my process more efficient. What is its difference from other AI technologies that you have used? #SmythOS #AI #Productivity #AI

    • @analyticsCamp
      @analyticsCamp Месяц назад

      Hi, unfortunately I did not understand your question/comment :(

  • @EarthrightCanvas
    @EarthrightCanvas 2 месяца назад

    Cant follow.

    • @analyticsCamp
      @analyticsCamp 2 месяца назад

      Hi, the code and process is on my GitHub (link in the description box) so you can follow at your own pace :)

  • @bladealex1844
    @bladealex1844 2 месяца назад

    This video is an excellent deep dive into Mixture of Agents (MoA)! 🚀 As someone who's been working on implementing MoA concepts, I found the explanation and tutorial incredibly valuable. For those interested in a practical application of MoA principles, I've developed MALLO (MultiAgent LLM Orchestrator): github.com/bladealex9848/MALLO You can try it live here: mallollm.streamlit.app/ MALLO builds on the MoA concept, integrating local models, OpenAI and Together AI APIs, and specialized assistants. It's fascinating to see how the MoA architecture with its layers of agents, as explained in the video, can be adapted for specific use cases. The benchmarks comparing MoA against GPT-4/GPT-4o are particularly interesting. In MALLO, I've implemented a similar multi-layered approach, focusing on specialized domains like legal and constitutional law. The tutorial on running MoA locally is a game-changer for accessibility. In MALLO, I've also integrated local models using Ollama, which aligns well with the free and local approach demonstrated here. I'm curious about how others are adapting these MoA concepts in their projects. Has anyone else experimented with combining different model types or specialized agents in their implementations? Thanks for this comprehensive guide! It's exciting to see the AI community pushing the boundaries of what's possible with open-source and locally-run models. 🌟 #MixtureOfAgents #MALLO #AIInnovation #OpenSourceAI #LocalLLM

  • @stephenzzz
    @stephenzzz 2 месяца назад

    Thanks for all your videos! on a side note, my wife wants to create a membership site with a chat & RAG of sorts to answer questions from her bespoke sales content. Which system out there do you think would work best, that is low code.

    • @analyticsCamp
      @analyticsCamp 2 месяца назад

      Thanks for watching :) If this is for a simple QA chatbot, then CrewAI could do, but if you need a more robust system and you're willing to spend on it, then one of the paid frameworks such as Oracle may be better (I haven't used it personally, so do your research), and Good luck with your project :)

  • @Researcher100
    @Researcher100 2 месяца назад

    Thanks for bringing this system to our attention. I think this is the first YT video that talks about this new agentic work. ❤

    • @analyticsCamp
      @analyticsCamp 2 месяца назад

      Thanks for watching :) I also think this is an innovative approach!

  • @chadricheson1038
    @chadricheson1038 2 месяца назад

    Very interesting topic

  • @optiondrone5468
    @optiondrone5468 2 месяца назад

    This paper was a good find, thanks for your explanation, looks like future of AI internet is here!

  • @jeffg4686
    @jeffg4686 2 месяца назад

    what if the models you choose are different for different agents in the various layers. Such as layer 1 has Agent 1 (llama3.1), Agent 2(mixtral7b), Agent3(gemma) layer 2 has Agent 1(chatgpt 4), agent2(mixtral7b), agent3 (llama3) Also, can the layers have different numbers of agents - I assume so, but not sure.

    • @analyticsCamp
      @analyticsCamp 2 месяца назад

      Hi, yes you can effectively do all that. If you see the video at 06:05 you see there are four different models for the layers with Qwen2 acting as the aggregator. If you take a look at the MOA diagram, you'll see each agent/LLM is depicted with a different colour (A1, A2, A3) in each layer, so in their current set-up in each layer the number of 'reference models' defined by the user will separately produce the intended result, and they get aggregated at the final output (depicted as A4). Yes, I think you can tweak the code to have a different number of agents per layer too; please check their GitHub repository, the bot.py file ( I haven't done that personally). Thanks for your comment :)

  • @arielle-cheriepaterson7851
    @arielle-cheriepaterson7851 2 месяца назад

    Are you available for consulting?

    • @analyticsCamp
      @analyticsCamp 2 месяца назад

      Hi, could you please send me an email with more details? (my email address is in my channel's About section. Thanks :)

  • @OpenAITutor
    @OpenAITutor 3 месяца назад

    I love this! I did create a version using Groq and open-webui!

    • @analyticsCamp
      @analyticsCamp 3 месяца назад

      Thanks for your comment. I visited your channel and subed! Great videos :)

  • @thatsfantastic313
    @thatsfantastic313 3 месяца назад

    beautifully explained!

  • @soccerdadsg
    @soccerdadsg 3 месяца назад

    Another quality video from the channel!

  • @soccerdadsg
    @soccerdadsg 3 месяца назад

    Appreciate your effort to make this video.

    • @analyticsCamp
      @analyticsCamp 3 месяца назад

      My pleasure, thanks for watching :)

  • @soccerdadsg
    @soccerdadsg 3 месяца назад

    This is a very good video. It is a good summary of current development of agentic workflows with scientific paper support.

    • @analyticsCamp
      @analyticsCamp 3 месяца назад

      Thanks for your supportive words. Stay tuned, I have more of this coming :)

  • @optiondrone5468
    @optiondrone5468 3 месяца назад

    Medical images better than human operators! If we keep going at this rate soon many general practitioners in UK will have no jobs.

    • @analyticsCamp
      @analyticsCamp 3 месяца назад

      Now imagine if we combine this with the agentic power! But I still think it's too early to make a definitive judgement, as many of these papers report on their best results/round! Thanks for watching though :)

  • @DrRizzwan
    @DrRizzwan 3 месяца назад

    Good explanation 👏

  • @analyticsCamp
    @analyticsCamp 3 месяца назад

    Hey everyone, I have already explained RAG, ICL, and fine-tuning in the previous videos separately, so I thought I would give you all in one place!

  • @peralser
    @peralser 3 месяца назад

    Excelente Video!! Thanks!

  • @dreamphoenix
    @dreamphoenix 3 месяца назад

    Thank you.

  • @jonjeffers5153
    @jonjeffers5153 3 месяца назад

    Thanks for the video! I'm having an issue with the API key. I'm not a python programmer, FYI. The bot.py runs, but when I type something I get: OpenAIError: The api_key client option must be set either by passing api_key to the client or by setting the OPENAI_API_KEY environment variable

    • @analyticsCamp
      @analyticsCamp 3 месяца назад

      Hi, I think you have not set your environment variables correctly (Please follow the video steps). You should either have a valid OpenAI API key (can get it from their website) or just get a free key from the Together AI website for this project. Then, from your code editor terminal, export your environment API key by typing this exactly: echo "export OPENAI_API_KEY='yourkey'" >> ~/.bash_profile -but replace 'yourkey' with the key ID you got (it doesn't have to be only from OpenAI, but any partner of them like Together AI, etc). Then update the shell with the new variable by typing: source ~/.bash_profile -to confirm if it is set correctly, type: echo $TOGETHER_API_KEY -This should show your key, if it is correct you are set. I hope this helps :) PS: if you work on Windows, I think you should use \ instead of / in my code, and instead of bash_profile in the first two lines, use autoexec.bat. I don't work with Windows but this information that I found online may help: To set environment variables in Windows, you can follow these steps: Press Win + R to open the Run dialog. Type sysdm.cpl and press Enter to open the System Properties window. Go to the "Advanced" tab and click on the "Environment Variables" button. In the Environment Variables window, you can set system variables (for all users) or user variables (specific to the current user). Click on "New" to add a new variable, or select an existing variable to edit or delete it.

  • @travelingbutterfly4981
    @travelingbutterfly4981 3 месяца назад

    hi. I dont think the data it produced is correct did u try some method to validate it?

    • @analyticsCamp
      @analyticsCamp 3 месяца назад

      Thanks for your comment. You are right! I checked the top 3 manually in the CSV file and it looks different. With Mistral I get more accurate results. However, LLAMA3 produced a good synthesis of the career path. The video is basically meant as a tutorial (how to do), but the choice of LLM makes a difference. Thanks for watching :)

    • @travelingbutterfly4981
      @travelingbutterfly4981 3 месяца назад

      @@analyticsCamp Thanks for the reply. Actually I am trying to get insights from the dataset using crew ai. Can you suggest some ways to do it?

    • @analyticsCamp
      @analyticsCamp 3 месяца назад

      Is your dataset a CSV file? This video tutorial is a standard way of calling a CSV file within the CrewAI framework, if you don't get accurate results, change the model, e.g., to Mistral or Qwen2, or dbrx (from Databricks) on a sample dataset where you already know the results; any of the model's which produce accurate results, use that one on your target dataset. If you are doing a more serious data anlytics work, keep in mind that most of these LLMs are primarily language models (designed to predict the next word, not necessarily the 'correct' data), so in this case, using the traditional methods in Pandas, for example for data wrangling, or machine learning models from Scikit-learn will give you the most accurate results. If you insist on agentic method, then try asking one of those LLM agents to access Pandas or Scikit-learn and do the work for you. I haven't tried this honestly, so I don't know how it would turn out. But please keep me updated if this works for you. Hope this information helps :0

  • @gc1979o
    @gc1979o 3 месяца назад

    Awesome presentation!

  • @BooleanDisorder
    @BooleanDisorder 4 месяца назад

    Next will probably be: Mixture of Mixtures!

    • @analyticsCamp
      @analyticsCamp 4 месяца назад

      LOL :) Who knows? Maybe you're right!

    • @BooleanDisorder
      @BooleanDisorder 3 месяца назад

      ​@analyticsCamp seriously though, thanks for the excellent video.

  • @sr.modanez
    @sr.modanez 4 месяца назад

    top top top + + + + + +👏👏👏👏👏👏👏👏👏

    • @analyticsCamp
      @analyticsCamp 4 месяца назад

      Glad you liked it and thanks for watching :)

  • @JavierTorres-st7gt
    @JavierTorres-st7gt 4 месяца назад

    How to protect a company's information with technology ?

    • @analyticsCamp
      @analyticsCamp 4 месяца назад

      I'm not sure if I understand your question :( Apologies, but it'll be good if you give more context.

  • @BARD-no4wq
    @BARD-no4wq 4 месяца назад

    great video, you channel is underrated

  • @optiondrone5468
    @optiondrone5468 4 месяца назад

    Wow CSV file reading agent, this is so cool, does this mean that agent can also be programmed to generate SQL and access data from a database and do additional analysis?

    • @analyticsCamp
      @analyticsCamp 4 месяца назад

      As far as I know, the only sql parser in crewai tools is PGSearchTool, which is specifically made for PostgreSQL database tables; yep, it can search and generate sql queries, I think they call it Retrieve and Generate RAG. I haven't tested it yet, but if enough viewers ask for it, I may make sth out of it :)

    • @optiondrone5468
      @optiondrone5468 4 месяца назад

      @@analyticsCamp thanks for the tool name hope enough people here ask for #SQLagent tutorial!

  • @analyticsCamp
    @analyticsCamp 4 месяца назад

    Thanks for all your helpful comments :) Here's a related video explaining AI agentic workflow: ruclips.net/video/lA3Tju4VUho/видео.html

  • @analyticsCamp
    @analyticsCamp 4 месяца назад

    Some of you asked for AI agents in action; here's a video with code to use 100% free local AI agents: ruclips.net/video/XkS4ifkLwwQ/видео.html

  • @analyticsCamp
    @analyticsCamp 4 месяца назад

    Hey, if you are new to LLMs and need to improve the responses, here's a related video that shows 5 ways to improve LLM results: ruclips.net/video/8IC8bWvORFU/видео.html

  • @optiondrone5468
    @optiondrone5468 4 месяца назад

    All very exciting things but how long do you think before everyone can have access to all these AI based new applications?

    • @analyticsCamp
      @analyticsCamp 4 месяца назад

      Thanks for watching :) You can use ICL with any LLM, especially the ones you can download directly from Hugging Face or via Ollama. Some other interfaces allow users to attach files to process, so you can write your prompts and instructions in those files plus any images you need to attach. I'm not sure about audio and video ICL at this moment, though.

  • @Researcher100
    @Researcher100 4 месяца назад

    The explanation was clear, thanks. Does this paper show how to use this method in practice? I think most llm users don't know ins and out of fine tuning so icl can be very helpful for ordinary users.

    • @analyticsCamp
      @analyticsCamp 4 месяца назад

      Thanks for the comment :) Yes, the paper comes with all those explanation. Yep, I also believe this can open a way for more ordinary AI users AND many researchers in other fields.

  • @jarad4621
    @jarad4621 5 месяцев назад

    Sorry another quesion, am i able to use LM studio with crewai as well, wanted to test some other models and its gpu accel allows models to run better then ollama for me, is it still going to have probems due to the issues you fix with the models file or is that issue not a problem for other local servers? Or is ollama the best way because you can actually edit those things to make it work well? Thanks

    • @analyticsCamp
      @analyticsCamp 5 месяцев назад

      I do not use LM Studio so I cannot comment on that. But Ollama via terminal is pretty sturdy, CrewAI it should work with all Ollama models, but I have not tested all. If you run into issues you can report it here so others can know and/or help you :)

  • @first-thoughtgiver-of-will2456
    @first-thoughtgiver-of-will2456 5 месяцев назад

    can mamba have its input rope scaled? It seems it doesnt require positional encoding but this might make it extremely efficient for second order optimization techniques.

    • @analyticsCamp
      @analyticsCamp 5 месяцев назад

      In Mamba sequence length can be scaled up to a million (e.g., a million-length sequences). It also computes the gradient (did not find any info on second-order opt in their method): they train for 10k to 20k gradient steps.

  • @jarad4621
    @jarad4621 5 месяцев назад

    Also never seen the mistral idea so this model would make a really good agent then better then the others? Really helpful to know, glad I found this. Also can you test agencu dwarm ans let us know what the best agent framewoek is currently? Apparently crew is not great for production?

    • @analyticsCamp
      @analyticsCamp 5 месяцев назад

      Thanks for watching :) I have tested multiple models from Ollama and mistral seems to have better performance overall, across most tasks. Agent Swarm can be useful for VERY specialised tasks in which general LLMs get it totally wrong. Other than that, it will add to the time/cost of build.But I'm not sure if I understood your question right!

  • @jarad4621
    @jarad4621 5 месяцев назад

    Awesome I've been looking for some of this info for ages, Best video on agents after watching dozens of vids, nobody explains the problems with other models or fixing model file and how to make sure the local models work, many of these YT Experts are just using local and other nodels snd wondering why it's not working well. Can i use phi 3 mini local as well and it needs same model setup? Also will llama 70b on openrouter api actually work as a good agent or does something need to be tweaked first nobody can answer these things, please help? Thanks!

    • @analyticsCamp
      @analyticsCamp 5 месяцев назад

      Sure, you can essentially use any models listed in Ollama as long as you make a model file, you can manage the temperature etc. I have used LLAMA 70b before but surprisingly, it did not show better response than its 7b and 13b on most tasks! I recommend LLAMA3 (I may be able to make a video on it if I get some free time, LOL ).

    • @jarad4621
      @jarad4621 5 месяцев назад

      @@analyticsCamp Awesome thanks ill test the smaller ones first then

  • @optiondrone5468
    @optiondrone5468 5 месяцев назад

    Thanks for sharing your thoughts and practical AI agent workflow. I also believe that this agentic workflow will fuel many LLM based development in 2024

    • @analyticsCamp
      @analyticsCamp 5 месяцев назад

      Thanks for watching :) If you have a specific LLM-based development/project in mind please let me know. With such easy agentic access, I am also surprised how many AI users are still hooked on zero-shot with paid interfaces!

    • @optiondrone5468
      @optiondrone5468 5 месяцев назад

      @@analyticsCamp ha ha it also never made sense to me why ppl don't look into open source LLM 🤔 its free, not limiting your token size, free to experiment with different models and most importantly your data (prompt) is yours and don't become automatically OpenAi's property. Keep up the good work, looking forward to your next video.

  • @milanpospisil8024
    @milanpospisil8024 5 месяцев назад

    Yeah, Pat and Mat, thats from Czech studios :-)