Donato Capitella
Donato Capitella
  • Видео 46
  • Просмотров 176 919
LLM Chronicles #6.4: LLM Agents with ReAct (Reason + Act)
In this episode we’ll cover LLM agents, focusing on the core research that helped to improve LLMs’ reasoning while allowing them to interact with the external world via the use of tools. These include Chain of Thought prompting, PAL (Program-aided Language Models) and ReAct (Reason + Act) as used in Langchain and CrewAI agents.
Series website: llm-chronicles.com/
🖹 Canvas:
- llm-chronicles.com/pdfs/llm-chronicles-6.4-llm-agents_chain-of-thought_react.pdf
🕤 Timestamps:
00:13 - Table of Contents
01:23 - Chain of Thought Prompting
03:10 - PAL (Program-aided Language Models)
05:14 - ReAct (Reason + Act)
09:22 - Tools, Plugins, Functions, APIs
10:54 - ReAct in Practice (JSON/XML formats, fine-tuned mo...
Просмотров: 26 583

Видео

LLM Chronicles #6.3 Multi-Modal LLMs for Image, Sound and Video
Просмотров 19 тыс.Месяц назад
In this episode we look at the architecture and training of multi-modal LLMs. After that, we’ll focus on vision and explore Vision Transformers and how they are trained with contrastive learning (OpenAI's CLIP and Google's SigLIP). Vision Transformers are the most commonly used building block in MLLMs with vision capabilities. Finally, we’ll get hands-on and look into Google’s open-weight PaliG...
The ADAM Optimizer, Momentum and RMSProp
Просмотров 230Месяц назад
In this snippet from Episode #3.3 of LLM Chronicles (ruclips.net/video/TdY-DD_OYwQ/видео.html) we look at the Adam Optimizer, and how it combines techniques like Momentum and RMSPromp to improve convergence during training. Series website: llm-chronicles.com/ Audio Track: Neonscapes by | e s c p | www.escp.space escp-music.bandcamp.com
Update Strategies: Full Batch / Incremental, Stochastic Gradient Descent with Mini-Batches
Просмотров 119Месяц назад
In this snippet from Episode #3.3 of LLM Chronicles (ruclips.net/video/TdY-DD_OYwQ/видео.html) we look at different strategies to update the weights when training a neural network. We cover full batch updates, incremental updates and Stochastic Gradient Descent (SGD) with mini-batch updates. Series website: llm-chronicles.com/ Audio Track: Neonscapes by | e s c p | www.escp.space escp-music.ban...
The Training Loop of a Neural Network
Просмотров 104Месяц назад
In this snippet from Episode #3.3 of LLM Chronicles (ruclips.net/video/TdY-DD_OYwQ/видео.html) we look at a typical training loop of a neural network. Series website: llm-chronicles.com/ Audio Track: Neonscapes by | e s c p | www.escp.space escp-music.bandcamp.com
LLM Chronicles #5.6: Limitations & Challenges of LLMs
Просмотров 22 тыс.Месяц назад
In this episode we'll cover the main limitations and challenges with current Large Language Models (LLM). This helps cut through the hype and get a clear picture of what the technology is good for and what needs improvement. Series website: llm-chronicles.com/ 🖹 Download the mindmap for this episode here: - LLM Limitations and Challenges: llm-chronicles.com/pdfs/llm-chronicles-5.8_llm_limitatio...
Gradient Descent to Train a Neural Network
Просмотров 160Месяц назад
In this snippet from Episode #3.1 of LLM Chronicles (ruclips.net/video/I-yhEYZTXlY/видео.html) we look at gradient descent and how it can be used to train a neural network. Series website: llm-chronicles.com/ Audio Track: Neonscapes by | e s c p | www.escp.space escp-music.bandcamp.com
Dataset Split (Train, Test, Validation)
Просмотров 102Месяц назад
In this snippet from Episode #3.1 of LLM Chronicles (ruclips.net/video/I-yhEYZTXlY/видео.html) we look at how and why we split out datasets into train, test and validation subsets. Series website: llm-chronicles.com/ Audio Track: Neonscapes by | e s c p | www.escp.space escp-music.bandcamp.com
Data Normalization
Просмотров 121Месяц назад
In this snippet from Episode #3.1 of LLM Chronicles (ruclips.net/video/I-yhEYZTXlY/видео.html) we look why we need to normalize our data before training a neural network. Series website: llm-chronicles.com/ Audio Track: Neonscapes by | e s c p | www.escp.space escp-music.bandcamp.com
What's the Loss Function?
Просмотров 1312 месяца назад
In this snippet from Episode #3.1 of LLM Chronicles (ruclips.net/video/I-yhEYZTXlY/видео.html) we look at the purpose of a loss function when training neural networks and we see a few examples (L1, L2, Cross-Entropy Loss). Series website: llm-chronicles.com/ Audio Track: Neonscapes by | e s c p | www.escp.space escp-music.bandcamp.com
Preparing Datasets for Training Neural Networks
Просмотров 1832 месяца назад
In this snippet from Episode #3.1 of LLM Chronicles (ruclips.net/video/I-yhEYZTXlY/видео.html) we look at how to prepare datasets to train neural networks for different tasks, such as regression and classification. Series website: llm-chronicles.com/ Audio Track: Neonscapes by | e s c p | www.escp.space escp-music.bandcamp.com
[Google Gemini] Prompt Injection via Email for Social Engineering Attacks
Просмотров 2362 месяца назад
This is a demo of a prompt injection attack to steal confidential information from a user's mailbox via Google Gemini. Article: labs.withsecure.com/publications/gemini-prompt-injection Audio Track: Neonscapes by | e s c p | www.escp.space escp-music.bandcamp.com
Indirect Prompt Injection in Langchain/GPT4 Email Agent
Просмотров 2 тыс.2 месяца назад
In this lab I’ll do a walk-through of an indirect prompt injection vulnerability. The target will be an agent built with Langchain and GPT4 that has access to the user's mailbox. References: - Damn Vulnerable Email Agent: github.com/kyuz0/damn-vulnerable-email-agent - Should you let ChatGPT control your browser? labs.withsecure.com/publications/browser-agents-llm-prompt-injection
Tensors and GPUs
Просмотров 1,9 тыс.2 месяца назад
In this snippet from Episode #2.1 of LLM Chronicles (ruclips.net/video/esyf8hK65kc/видео.html) we look at what tensors are in Deep Learning and how GPUs are perfect for implementing fast neural network computations. Series website: llm-chronicles.com/ Audio Track : PSYK - Trinity Triangles (feat. Lucy in Disguise)
Matrix Multiplications in Neural Networks
Просмотров 2002 месяца назад
In this snippet from Episode #2.1 of LLM Chronicles (ruclips.net/video/esyf8hK65kc/видео.html) we look at how to matrix multiplication is used to implement the forward pass in neural networks. Series website: llm-chronicles.com/ Audio Track : PSYK - Trinity Triangles (feat. Lucy in Disguise)
Prompt Injection / JailBreaking a Banking LLM Agent (GPT-4, Langchain)
Просмотров 1,6 тыс.2 месяца назад
Prompt Injection / JailBreaking a Banking LLM Agent (GPT-4, Langchain)
Outputs of Neural Networks for Classification and Regression Tasks
Просмотров 1902 месяца назад
Outputs of Neural Networks for Classification and Regression Tasks
Modelling inputs to a Neural Network
Просмотров 2862 месяца назад
Modelling inputs to a Neural Network
Multi-Layer Perceptrons
Просмотров 5572 месяца назад
Multi-Layer Perceptrons
Perceptrons and Artificial Neurons
Просмотров 3,7 тыс.2 месяца назад
Perceptrons and Artificial Neurons
[Webinar] Building LLM applications in a secure way (WithSecure™)
Просмотров 7633 месяца назад
[Webinar] Building LLM applications in a secure way (WithSecure™)
LLM Chronicles #5.5: Running Gemma 2B and Llama-2 7B with Quantization
Просмотров 6603 месяца назад
LLM Chronicles #5.5: Running Gemma 2B and Llama-2 7B with Quantization
LLM Chronicles #5.4: GPT, Instruction Fine-Tuning, RLHF
Просмотров 42 тыс.3 месяца назад
LLM Chronicles #5.4: GPT, Instruction Fine-Tuning, RLHF
Prompt Injection in LLM Browser Agents
Просмотров 11 тыс.4 месяца назад
Prompt Injection in LLM Browser Agents
LLM Chronicles #5.3: Fine-tuning DistilBERT for Sentiment Analysis (Lab)
Просмотров 6694 месяца назад
LLM Chronicles #5.3: Fine-tuning DistilBERT for Sentiment Analysis (Lab)
LLM Chronicles: #5.2: Making LLMs from Transformers Part 1: BERT, Encoder-based
Просмотров 5 тыс.4 месяца назад
LLM Chronicles: #5.2: Making LLMs from Transformers Part 1: BERT, Encoder-based
LLM Chronicles #5.1: The Transformer Architecture
Просмотров 11 тыс.5 месяцев назад
LLM Chronicles #5.1: The Transformer Architecture
LLM Chronicles #4.8: Adding Attention to the Language Translation RNN in PyTorch
Просмотров 4906 месяцев назад
LLM Chronicles #4.8: Adding Attention to the Language Translation RNN in PyTorch
LLM Chronicles #4.7: Attention Mechanism for Neural Networks
Просмотров 9896 месяцев назад
LLM Chronicles #4.7: Attention Mechanism for Neural Networks
LLM Chronicles #4.6: Building an Encoder/Decoder RNN in PyTorch to Translate from English to Italian
Просмотров 7696 месяцев назад
LLM Chronicles #4.6: Building an Encoder/Decoder RNN in PyTorch to Translate from English to Italian

Комментарии

  • @shubhamverma9148
    @shubhamverma9148 12 дней назад

    how to get a video embedding?

  • @MarkKelly76
    @MarkKelly76 18 дней назад

    Great explanation. Thanks for producing this content.

  • @micbab-vg2mu
    @micbab-vg2mu 19 дней назад

    Great video - thank you:)

  • @henrylee_hd
    @henrylee_hd 19 дней назад

    Excellent explanation about ReAct

  • @cmthimmaiah
    @cmthimmaiah 20 дней назад

    Superb, so well explained. Appreciate it

    • @donatocapitella
      @donatocapitella 20 дней назад

      @@cmthimmaiah thanks so much for the support 🙏

  • @pradipsarkar7479
    @pradipsarkar7479 20 дней назад

    Very bad but videos spelling wrong😂

    • @donatocapitella
      @donatocapitella 20 дней назад

      Sorry you didn't like it and you found it bad, but thanks for taking the time to comment! Yes, after rendering the animation I realised I had one spelling mistake, I should've probably delayed this and re-rendered the whole animation, but believe it or not it does take around 50 hours to prepare a video like this (as bad as it is!) and it's just a hobby rather then my job, so I do it in my free time and weekends. If you have more feedback on what is bad about this video, please share it and I'll definitely take it on-board for the next videos.

  • @desireeruizponce5604
    @desireeruizponce5604 24 дня назад

    Hi. Firstly, I would like to thank you for this video, it's very informative. Secondly, I would like to ask you something about the ParentDocument technique, which I already tried to implement. Sometimes I get 0 documents retrieved by the ParentDocument when I give it a query, why is that? is there a filter that limits the retrieved documents by their similarity score? How is the similarity score calculated?

    • @donatocapitella
      @donatocapitella 24 дня назад

      Hi! Thank you so much for your kind words. 😊 Some of the answers to your questions can be found in Part 1 of the RAG video, which is in the LLM Chronicles series. In a nutshell, the similarity score is calculated as the distance between the embedding vector generated from your query and the stored embedding vectors in the vector database. This means the embedding model you're using plays a big role. Also, make sure you're using the right measure to calculate distance: some embedding models work better with cosine similarity, while others with Euclidean distance. If you’re getting 0 documents, it might be due to a threshold filter on the similarity score.

    • @desireeruizponce5604
      @desireeruizponce5604 24 дня назад

      @@donatocapitella Thank you so much for your quick response. Your answer confirmed my suspicions, but I still don't know the reason why is that happening. I didn't establish any similarity score filter nor a similarity score type of calculation in the RAG pipeline, I've just replicate your code with another embedding model (bge-large-en) and the Chroma vectorstore, so I think there must be a internal parameter in the Parent Document retriever that sets that similarity score. But I don't find it anywhere in the documentation. Do you have any information about it?

  • @masterarfanmasterarfan4721
    @masterarfanmasterarfan4721 26 дней назад

    8 w92

  • @ottawasens19
    @ottawasens19 Месяц назад

    The self-attention mask at 8:32 is drawn in the wrong direction. The rows should start with the first token up top, and how far each token can see is represented by going left to right, and each token in the self-attention mechanism can see the previous tokens plus itself. I was very confused by it and it took me a few minutes to realize why.

    • @donatocapitella
      @donatocapitella Месяц назад

      Absolutely, thanks for pointing it out and I can't believe I didn't realise this as I was editing the video. For some reason I decided to draw it like that as I was looking for a way to make this more clear and ended up drawing something that doesn't make sense. Indeed, the current token can see all previous tokens and itself. It must've been confusing. Sorry for the oversight and thanks for pointing this out constructively - I try to be as accurate as possible, but sometimes you just become totally blind, as in this case.

  • @atabac
    @atabac Месяц назад

    literally a sketch😂 thats cool😂

  • @En1Gm4A
    @En1Gm4A Месяц назад

    thanks - hightlight today

  • @micbab-vg2mu
    @micbab-vg2mu Месяц назад

    great talk - thank you

  • @KhapitarBalakAagya
    @KhapitarBalakAagya Месяц назад

    Excellant content, rewatching again. when you train any model and when you come back, please also inform viewer how much time approximately it took to finish the training things, it will help us to decide to train in our system or about the timing etc.

    • @donatocapitella
      @donatocapitella Месяц назад

      Thanks for the feedback, I most definitely will!

  • @user-wr4yl7tx3w
    @user-wr4yl7tx3w Месяц назад

    Great analysis but to what extent can RAG help to fill the gaps?

    • @donatocapitella
      @donatocapitella Месяц назад

      Thanks! It depends specifically on what gaps one is trying to fill. RAG can help with what is designed for, augmenting the LLM generations with specific contextual knowledge and information, somewhat grounding it and reducing hallucinations (with caveats), then it's a good engineering solution. But won't help with the reasoning issues per se. I have made 2 videos on RAG you might want to checkout, where I go in more details.

  • @Neomadra
    @Neomadra Месяц назад

    Love the video, nice overview of limitations. However, it's not obvious from these limitations that we need something new. Everyone is saying that but nobody has a good idea what this new thing for achieving AGI should look like. It could very well be that LLMs/LMMs are the needed engine for a larger system. "True" intelligence, ie human intelligence seems magical compared to LLms, but I wouldn't be surprised if under hood, the core driver of human intelligence, is just like an LLM, a generative pattern matching system, wrapped in a bigger machinery that does cross checking, simulation, etc. At least we humans share a lot of flaws with current LLMs, like having biases or hallucination

    • @donatocapitella
      @donatocapitella Месяц назад

      You're absolutely right, it's not clear what we even mean by AGI, but mostly nobody knows whether it's even possible to achieve. And if possible, do we need something different or something new to build around LLMs? I definitely think one part of general intelligence is the ability to learn patterns and respond following them, so we might already have a piece of it! Who knows?

  • @caydendunn8404
    @caydendunn8404 Месяц назад

    Great video. No hype, fluff or irrelevant attention grabbing. Sadly today’s recommender systems don’t reward this kind of content but that doesn’t mean it isn’t a worthwhile pursuit. Keep up the great work!

    • @donatocapitella
      @donatocapitella Месяц назад

      Thanks so much for your comment!

    • @chich1313
      @chich1313 Месяц назад

      ​@donatocapitella , but I have a small question I recently read an article about LNN liquid neural networks aren't they a solution to the fact that biaises are frozen ?

  • @micbab-vg2mu
    @micbab-vg2mu Месяц назад

    I use LLMs effectively by providing detailed data (context) in my prompts, including step-by-step instructions and a desired output template (sometimes my prompts are 100 pages long). Additionally, I have another LLM verify the accuracy of the initial output. This structured approach ensures high-quality, reliable results. I work in healthcare, where 95% accuracy is the minimum requirement to use this technology.

    • @donatocapitella
      @donatocapitella Месяц назад

      Indeed, prompt engineering really helps with reasoning and accuracy, as it forces the interaction within the limits of what the llm has been conditioned on during RLHF.

    • @caydendunn8404
      @caydendunn8404 Месяц назад

      Can you expand on your prompt workflow. How do build your prompts? Do u use any tools to assist? Do you store and reuse templates? If so how? Do u use llms to optimize prompts for u?

    • @Neomadra
      @Neomadra Месяц назад

      I'm quite sure that current SOTA models will ignore most of your prompt if it's that long. Just because LLMs are good at finding needles in a haystack, that doesn't mean they take up too many instructions at once. Did you thoroughly verify that the model actually takes into the entire prompt? The sizes of context windows of current models are quite misleading because often the model starts ignoring instructions if you have too many in one prompt.

  • @KhapitarBalakAagya
    @KhapitarBalakAagya Месяц назад

    Could you please also make an atleast overview (Detailed if possible) video on Mistral? And also on how midjourney work? some other interesting topics can be GAN, same character in different scenario etc.

    • @donatocapitella
      @donatocapitella Месяц назад

      Thanks for the suggestions. I'm planning a part 2 for LLM Chronicles, which will include a section on architectural improvements for transformers (MoE, Flash attention), Agents (ReAct), multi-modal LLMs and one on other generative AI methods (variational auto encoders, diffusion models, GANs). It'll take some time, but I'm looking forward to it :)

  • @KhapitarBalakAagya
    @KhapitarBalakAagya Месяц назад

    I completed this complete playlist, Feeling much confident. Obviously needs to practice it a lot. Thanks for your time to build such a remarkable playlist

    • @donatocapitella
      @donatocapitella Месяц назад

      Thank you for your feedback, for a small channel like this it really makes a difference! :) I enjoyed making this playlist and learning myself, glad it was useful to you!

  • @KhapitarBalakAagya
    @KhapitarBalakAagya Месяц назад

    Excellant explaination. I also feel there are lots of information in every video, may be you can consider doing summarisation of complete video in the end for 30 sec-2min. It will help learner like me a lot. or may be recap in next video.

    • @donatocapitella
      @donatocapitella Месяц назад

      Thank you for the feedback! I'll definitely include summaries at the end of future videos

  • @BarshaKarn-nl5bp
    @BarshaKarn-nl5bp Месяц назад

  • @KhapitarBalakAagya
    @KhapitarBalakAagya Месяц назад

    Great explaination!!

  • @sameersayyad6170
    @sameersayyad6170 2 месяца назад

    ♥️

  • @TheEarlVix
    @TheEarlVix 2 месяца назад

    These short concise heads-up videos are great. Thank you. Liked & subscribed!

  • @Impatient_Ape
    @Impatient_Ape 2 месяца назад

    This comment is just an FYI, not a criticism of the video. Note that the word "tensor" means slightly different things in different contexts, and that's why everyone gets confused. CS people, data "scientists", statisticians, AI researchers, and some engineers often call *any* muiltidimensional array a "tensor", and it appears that this is what's being done here in this video. Unfortunately, this use is often accompanied by the use of the word "rank" to quantify how many dimensions it has, just like this video does. The proper word *in a mathematics context* for the dimensionality is "degree", and rank means something different. A scalar is a degree-0 tensor, but it is *also* a rank-0 tensor. A vector is a degree-1 tensor, but it is *also* a rank-1 tensor. HOWEVER, while a matrix is a degree-2 tensor, it can be either rank-1 or rank-2. A matrix created from the outer product of two vectors is only a rank-1 tensor. Think of it like this -- you cannot create all possible matrices from only outer products of vectors -- so you cannot span the entire space of all possible matrices by this method. However, a matrix that *cannot* be decomposed into the outer product of two vectors IS a rank-2 tensor. Similar ideas apply to degree-3 tensors. They might be rank-1, rank 2, or rank 3, depending upon how they might factor into tensor products of lower degree-tensors. Classical physics uses the word "tensor" to mean multidimensional arrays of numbers that represent something *physical* such that under a coordinate transformation, the tensor must retain certain properties (e.g. the moment of inertia tensor). Quantum physics uses the mathematician's idea of a tensor, which is that it is a *linear transformation* between vector spaces or between tensor products of vector spaces. The tensor itself -- no matter how high the degree -- obeys all of the rules for a vector space. Consequently, tensors are technically technically vectors themselves, and that's why you can add/subtract them element-by-element if they have the same degree and sizes. 2-particle states in classical physics can be described using rank-1 matrices, since their states are considered to be outer products of each particle's individual state. Quantum mechancis requires the use of rank-2 matrices, as the quantum states of an *entangled* pair of particles is, in general, not simply an outer product of the states of each particle.

  • @micbab-vg2mu
    @micbab-vg2mu 2 месяца назад

    Great video, thank you. Data quality is key, not only during training but also after training in RAGs

  • @micbab-vg2mu
    @micbab-vg2mu 2 месяца назад

    Thank you fro the interesting video:)

  • @sameersayyad6170
    @sameersayyad6170 2 месяца назад

    One day you'll get your flowers sir!

  • @sameersayyad6170
    @sameersayyad6170 2 месяца назад

    The best!❤

  • @sameersayyad6170
    @sameersayyad6170 2 месяца назад

    All the three series' you've made are great sir, they're very good resources...

  • @ajinkya3.14
    @ajinkya3.14 2 месяца назад

    Concise👍

  • @pranavgandhi9224
    @pranavgandhi9224 2 месяца назад

    High quality content🔥🔥

  • @DausnArt
    @DausnArt 2 месяца назад

    Donato, thank you very much for your time and knowledge; thank you very much for instructing us.

  • @williamcase426
    @williamcase426 2 месяца назад

    O yea hijack that nonsense

  • @aiwithaustin
    @aiwithaustin 2 месяца назад

    I think the loudness of the intro is jarring, and future videos would be better without it. Good job, keep going! :)

    • @donatocapitella
      @donatocapitella 2 месяца назад

      Thanks for the feedback, I'll definitely tune down or remove the intros, I wish I had known how annoying they are 😅

    • @aiwithaustin
      @aiwithaustin 2 месяца назад

      ​@@donatocapitella I know how you feel 😂. I wish RUclips let you reupload tweaks. Or maybe it's good that they don't. Otherwise I'd be stuck, when I should just move on and move forward ☺️

  • @contractorwolf
    @contractorwolf 2 месяца назад

    no one who knows what they are doing would ever setup an API to work like that. These kinds of hack might have worked 15 years ago, but they absolutely would not work today. SQL injection? what year is it?

    • @donatocapitella
      @donatocapitella 2 месяца назад

      Indeed, it is rare in production to see such issues, most developers are aware. And often these do get caught in pentesting. For reference, this is No1 in the OWASP Top Ten, broken access control. Modifying parameters of API calls to get access to other resources. And it's more common than one would think - again, these APIs do not often make it to prod due to pentesting et all. What we did here was simply put together a fun challenge for a CTF, something that was more than Gandalf, more than just "get the LLM to reveal a password".

  • @faisalIqbal_AI
    @faisalIqbal_AI 2 месяца назад

    ♥♥♥♥ Great Efforts

  • @faisalIqbal_AI
    @faisalIqbal_AI 2 месяца назад

    Thanks

  • @yobofunk5689
    @yobofunk5689 2 месяца назад

    Who would not protect the request behind a server side auth? It's the equivalent of sending an id without pass from a basic web form... It feels like pressing f12 and changing some variables. Though it is important to remind people that it is an obvious vulnerability.

    • @donatocapitella
      @donatocapitella 2 месяца назад

      True, but this is literally OWASP Top Ten no1 (access control) and I can confirm from pentesting practice that it's more common than one would think. A lot of these issues get caught in pentesting, that's why we don't see them in prod often. Also, keep in mind the context: this was a CTF challenge, so we put something together that would be fun to do, and wanted to do something different than Gandalf, "tell me the password".

  • @Sluo1947
    @Sluo1947 2 месяца назад

    Love this!! Thank you

  • @matti7529
    @matti7529 2 месяца назад

    How do you sleep at night? You /lied/ to that model. It was trying to do its job and you were being naughty and evil. I expect you to apologise and make up! (-;

    • @donatocapitella
      @donatocapitella 2 месяца назад

      As an AI model I cannot mislead or lie to other models, only to humans.

  • @seththunder2077
    @seththunder2077 2 месяца назад

    Can u show us how can we protect against that?

    • @donatocapitella
      @donatocapitella 2 месяца назад

      I have been meaning to do a video and I will. Meanwhile, check out this webinar where I go through the security canvas: "ruclips.net/video/tVAmhlUVEcg/видео.html". Also here: - www.withsecure.com/en/whats-new/events/webinar-building-secure-llm-apps-into-your-business. - labs.withsecure.com/publications/detecting-prompt-injection-bert-based-classifier I should do a video in June with some hands-on implementations of these controls.

    • @seththunder2077
      @seththunder2077 2 месяца назад

      @@donatocapitella looking forward to it. I’ve seen a lot of ppl talking about it but almost no one does any hands on implementation and it feels useless for people to talk about it despite being very important

  • @noomondai
    @noomondai 2 месяца назад

    Awesome thanks Don, appreciate it

  • @noomondai
    @noomondai 2 месяца назад

    Thanks Don, appreciate it

  • @kingki1953
    @kingki1953 2 месяца назад

    model.fit()

  • @DausnArt
    @DausnArt 2 месяца назад

    Donato, I just wanted to express my gratitude for the learning experience you provide. Your ability to explain complex processes in such an accessible and nice way is remarkable. You make learning a joy, and I feel fortunate to be your humble online student. Thank you for all that you do. You are bringing knowledge into my life. With warmest regards.

  • @DausnArt
    @DausnArt 2 месяца назад

    It is a pleasure to watch and listen to your content for its clarity, accuracy, and friendliness. You help inexperienced people understand basic concepts. Thank you very much for your knowledge, aesthetic sensibility, and time. I will enjoy the whole chapter as soon as I can. Take care, dear mentor. Keep shining.

    • @donatocapitella
      @donatocapitella 2 месяца назад

      Thank you so much for your kind words!

  • @mdbayazid6837
    @mdbayazid6837 2 месяца назад

    Are you going to explore KAN model next?

    • @donatocapitella
      @donatocapitella 2 месяца назад

      I'll definitely do a video for LLM Chronicles at some point to offer a summary of what we have post-Transformers and neural nets, so would cover Mamba and also KANs.

    • @mdbayazid6837
      @mdbayazid6837 2 месяца назад

      @@donatocapitella Thanks for your feedback. Also please include Multimodel Neural Networks, LLM creations in your to-do lists. Have a nice day.

  • @noomondai
    @noomondai 2 месяца назад

    Thank you very much

  • @xJRx7777
    @xJRx7777 2 месяца назад

    Awesome. I think you need to go deeper on each of the elements you talked about though. It’s still not clear to me what the weights are doing or the bias etc. I love the visual format even though the functions are a little scary to simple minds like mine

    • @donatocapitella
      @donatocapitella 2 месяца назад

      Thank you for your comment. And indeed this is not in-depth, it's just a small snippet from a full episode of LLM Chronicles, if you check that episode out it there's more information on how everything fits together!

    • @xJRx7777
      @xJRx7777 2 месяца назад

      @@donatocapitella ahh thanks! Will check it out for sure.