Chris Alexiuk
Chris Alexiuk
  • Видео 50
  • Просмотров 236 242
LangChain Series: Prompt Tools 101 - Simple Prompt Templates
This video introduces a critical piece of the LangChain puzzle - Prompt Templates!
If you're new to Langchain, this is a great way to dip your toes in and get started!
🔗 LangChain Prompt Template Docs: python.langchain.com/docs/modules/model_io/prompts/prompt_templates/
🔗 Colab: colab.research.google.com/drive/1GlGwE2ScSbPNbHxwLCVoYMpyW1zGXm9x?usp=sharing
About me:
Follow me on LinkedIn: www.linkedin.com/in/csalexiuk/
Twitter: c_s_ale
Check out what I'm working on: getox.ai/
Просмотров: 11 808

Видео

Chainlit 🔗🔥 - Build an Arxiv QA Chat Application in Minutes!
Просмотров 11 тыс.Год назад
Huge props to Michael Wright to highlighting this tool to me! Learn how to build slick apps and demos with your LLMs using Chainlit, a Python framework similar to Streamlit. In this video, I walk through creating a simple Arxiv QA app with OpenAI's GPT-4 in just a few lines of code. Chainlit lets you: Build fast - Create apps quickly with minimal code! This video will show you how to start Chai...
LangChain Series: LangChain Intro Project - Dungeons and Dragons Knowledge Base!
Просмотров 2,9 тыс.Год назад
This video builds an AI assistant powered by GPT-4 with Dungeons & Dragons 5th edition knowledge. This overview demonstrates building AI applications with GPT-4 and Langchain. Future videos will explore each component, from data embedding and QA chains to building agents and memories. Using a powerful model like GPT-4, we create an assistant with specialized knowledge for engaging exchanges. 🔗 ...
Tree of Thoughts: Deliberate Problem Solving with Large Language Models - Let Your LLMs Play Games!
Просмотров 5 тыс.Год назад
In this video, I explore Tree of Thoughts, a technique for helping large language models perform better at complex reasoning tasks! 🔗 Paper: arxiv.org/pdf/2305.10601.pdf 🔗 Repository: github.com/ysymyth/tree-of-thought-llm 🔗 Colab Notebook in Video: colab.research.google.com/drive/1LNnBsseeIXecfiwIQzIi4NJEG-uKhNXn?usp=sharing About me: Follow me on LinkedIn: www.linkedin.com/in/csalexiuk/ Check...
Lit-LLaMA: Freeing the LLaMA! - Another Permissively Licensed LLaMA Reproduction
Просмотров 1,9 тыс.Год назад
Lit-Llama is an open-source resource for training LLaMA-style language models. Lit-Llama is optimized for speed, precision, and commercial use with an Apache 2.0 license. Powered by Lit-LLaMA and using the pre-trained weights provided by OpenLM's OpenLLaMA (training on the RedPajama dataset) - you can instruct-tune a LLaMA style model in ~9hrs. in a Colab Pro instance. 🔗Lit-LLaMA Repository: gi...
MPT 7B - A marvel of MLOps, ML Engineering, and Innovation from MosaicML
Просмотров 2,5 тыс.Год назад
In this video, we explore Mosaic's new open-source language model MPT-7B. Mosaic is pushing the boundaries of open-scale AI and building tools to empower researchers and practitioners! 🔗 Mosaic's Blog Post: www.mosaicml.com/blog/mpt-7b 🔗 Instruct Demo: huggingface.co/spaces/mosaicml/mpt-7b-instruct 🔗 ALiBi Paper: arxiv.org/pdf/2108.12409.pdf About me: Follow me on LinkedIn: www.linkedin.com/in/...
Simple App to Question Your Docs: Leveraging Streamlit, Hugging Face Spaces, LangChain, and Claude!
Просмотров 5 тыс.Год назад
THIS IS A REUPLOAD: The original title/description/thumbnail of the video were not representative of the content, so I recreated the video to be more clear. This is a non-comprehensive tutorial - but you can look forward to more in-depth tutorials for LangChain in the coming weeks! We create an app to upload Canadian bills and ask the AI questions. Using Streamlit and Langchain, you can quickly...
Transformers Agent - Hugging Face enter the "AutoGPT" game!
Просмотров 686Год назад
Explore the new Transformers Agent from Hugging Face! This tool lets you build natural language interfaces to call on AI tools. In this overview, learn how to quickly generate images, summarize text, play audio and more with just a few lines of code. Build custom tools to give the Agent new superpowers! An incredible new tool that makes AI more accessible than ever. 🔗 Blog Post: huggingface.co/...
Exploring StarCoder: Open Source LLM for Code Completion
Просмотров 8 тыс.Год назад
StarCoder, the hottest new Open Source code-completion LLM, is based on GPT-2 architecture and trained on The Stack - which contains an insane amount of permissive code. Star Coder shows how open source AI is advancing fast. The model may not match GPT-4 but it highlights how the community is gaining capabilities that are on pace to match industry titans such as Google and OpenAI! Overall, Star...
May the 4th Be With You: YOLO-NAS Powered Jar Jar Binks Detector
Просмотров 393Год назад
Revolutionize your object detection game with YOLO-NAS! This open-sourced architecture uses Neural Architecture Search to enhance detection of small objects, improve localization accuracy, and achieve higher performance-per-compute ratio. Ideal for real-time edge-device applications. #AI #objectdetection #supergradients #yolonas GitHub repo: bit.ly/yolo-nas-launch Starter Notebook: bit.ly/yolo-...
Low-rank Adaption of Large Language Models Part 2: Simple Fine-tuning with LoRA
Просмотров 22 тыс.Год назад
In this video, I go over a simple implementation of LoRA for fine-tuning BLOOM 3b on the SQuADv2 dataset for extractive question answering! LoRA learns low-rank matrix decompositions to slash the costs of training huge language models. It adapts only low-rank factors instead of entire weight matrices, achieving major memory and performance wins. 🔗 LoRA Paper: arxiv.org/pdf/2106.09685.pdf 🔗 Intr...
Low-rank Adaption of Large Language Models: Explaining the Key Concepts Behind LoRA
Просмотров 101 тыс.Год назад
In this video, I go over how LoRA works and why it's crucial for affordable Transformer fine-tuning. LoRA learns low-rank matrix decompositions to slash the costs of training huge language models. It adapts only low-rank factors instead of entire weight matrices, achieving major memory and performance wins. 🔗 LoRA Paper: arxiv.org/pdf/2106.09685.pdf 🔗 Intrinsic Dimensionality Paper: arxiv.org/a...
HuggingChat - Is this open source LLMs "ChatGPT" moment?
Просмотров 864Год назад
Meet HuggingChat, an open-source tool built by @HuggingFace and powered by LAION-AI's OpenAssistant. Forget cost, hardware, and tech skills. Hugging Chat works in your browser using machine learning. We're talking AI that's simple, accessible, and for the people.👊 I genuinely believe this is one of the most exciting releases in the last...few weeks? Things are moving fast, that's for sure. Chat...
Exploring Mini GPT-4: Multimodal LLM with Open Source Tools
Просмотров 1,8 тыс.Год назад
In this video, we dive into MiniGPT-4, a powerful application that combines open source tools to describe images in text. We explore its model architecture, training process, and the fascinating concept of soft prompts. Discover how this application pushes the boundaries of large language models and their multimodal capabilities. 🔗 MiniGPT-4 Paper: arxiv.org/pdf/2304.10592.pdf 🔗 MiniGPT-4 Proje...
Cohere's Wikipedia Embeddings: A Short Primer on Embedding Models and Semantic Search
Просмотров 998Год назад
Learn about Wikipedia embeddings from Cohere! This video explains how Cohere embedded millions of Wikipedia articles and released them for open use. Embeddings represent text as numbers, allowing us to determine how semantically similar two pieces of text are. Using Cohere's embeddings, you can build applications like neural search, query expansion, and more. Check out the code example in Colab...
Exploring Stability AI's New Open Source Language Model (StableLM)
Просмотров 760Год назад
Exploring Stability AI's New Open Source Language Model (StableLM)
GPT4All Chat - A fun but limited AI chatbot 🤖 (1-click install)
Просмотров 1,9 тыс.Год назад
GPT4All Chat - A fun but limited AI chatbot (1-click install)
Animate Your Own Drawn Characters in Minutes! | Using Meta's Open Source Animated Drawings Repo!
Просмотров 1 тыс.Год назад
Animate Your Own Drawn Characters in Minutes! | Using Meta's Open Source Animated Drawings Repo!
AI Shell, a GPT powered alternative to Github Copilot X!
Просмотров 1 тыс.Год назад
AI Shell, a GPT powered alternative to Github Copilot X!
Exploring Databricks's Open Source Dolly 2.0 Language Model (Fine-Tuned on 15K Human Instructions!)
Просмотров 3 тыс.Год назад
Exploring Databricks's Open Source Dolly 2.0 Language Model (Fine-Tuned on 15K Human Instructions!)
How to Use Grounded Segment Anything for Image Segmentation and Inpainting! - WSL2 Tutorial
Просмотров 3,2 тыс.Год назад
How to Use Grounded Segment Anything for Image Segmentation and Inpainting! - WSL2 Tutorial
git good with Chris! - git revert AKA CTRL+Z +++
Просмотров 67Год назад
git good with Chris! - git revert AKA CTRL Z
Generative Agents: Interactive Simulacra of Human Behavior AKA "GPT-3.5 Meets The Sims" - Explained!
Просмотров 7 тыс.Год назад
Generative Agents: Interactive Simulacra of Human Behavior AKA "GPT-3.5 Meets The Sims" - Explained!
How to install and run Auto-GPT! Also, is it AGI?! No, it's not, and that's okay!
Просмотров 1,9 тыс.Год назад
How to install and run Auto-GPT! Also, is it AGI?! No, it's not, and that's okay!
Give Yourself an AI Sidekick - Program Alongside Tabby, Your Self-Hosted GitHub CoPilot!
Просмотров 3,3 тыс.Год назад
Give Yourself an AI Sidekick - Program Alongside Tabby, Your Self-Hosted GitHub CoPilot!
Running Alpaca-LoRA aka "Local ChatGPT" on Windows through Docker Desktop and WSL2!
Просмотров 2,8 тыс.Год назад
Running Alpaca-LoRA aka "Local ChatGPT" on Windows through Docker Desktop and WSL2!
git good with Chris! - github.com actions: how to enforce linting!
Просмотров 129Год назад
git good with Chris! - github.com actions: how to enforce linting!
Self-Refine: making GPT-4 prompt engineer itself for you
Просмотров 1,5 тыс.Год назад
Self-Refine: making GPT-4 prompt engineer itself for you
Train and Deploy Amazing Models in Less Than 6 Lines of Code with Ludwig (no, not that Ludwig) AI!
Просмотров 2,7 тыс.Год назад
Train and Deploy Amazing Models in Less Than 6 Lines of Code with Ludwig (no, not that Ludwig) AI!
git good with Chris! - git stash all your cares away! (just don't accidentally drop them)
Просмотров 275Год назад
git good with Chris! - git stash all your cares away! (just don't accidentally drop them)

Комментарии

  • @les-fauxmonnayeurs9887
    @les-fauxmonnayeurs9887 11 дней назад

    is there any SEO expert that isn't overly excited when speaking? someone dark maybe? a bit punk

  • @Invalid_Username404
    @Invalid_Username404 29 дней назад

    Short and to the point, and most importantly honest. Subscribed

  • @arnes.1328
    @arnes.1328 Месяц назад

    very cool. really liked the video !

  • @honglu679
    @honglu679 Месяц назад

    Thx for great video! so what is the better way to teach a model new knowledge, if FT is somehow only good for structure? thx much!

  • @joeybasile1572
    @joeybasile1572 Месяц назад

    Nice video.

  • @skyforever1000
    @skyforever1000 Месяц назад

    good one, thank you

  • @tommyshelby6277
    @tommyshelby6277 Месяц назад

    doing a f'cking god's work

  • @user-bb2ut7nu3l
    @user-bb2ut7nu3l 2 месяца назад

    Thank you for the explanation, It helps me a lot.

  • @ArunkumarMTamil
    @ArunkumarMTamil 2 месяца назад

    how is Lora fine-tuning track changes from creating two decomposition matrix? How the ΔW is determined?

  • @blackswann9555
    @blackswann9555 2 месяца назад

    Short videos are great

  • @sampadmohanty8573
    @sampadmohanty8573 2 месяца назад

    @15:28, there is nothing great about adding the extra LoRA parameters to the weights that makes it easier to swap the behaviour of the model at inference time because the difference between adding the matrices and loading the entirely new weight matrix from different finetuned models to the model architecture is negligible.

    • @chrisalexiuk
      @chrisalexiuk 2 месяца назад

      I actually think that winds up being largely incorrect, looking at platforms like Gradient - and initiatives like multi-LoRA (LoraX, etc), seem to be a testament to that.

  • @myselfremade
    @myselfremade 2 месяца назад

    My left ear enjoys this video

  • @Lzyue0092youtube
    @Lzyue0092youtube 3 месяца назад

    作为初学者,我想请问一下如何运行本篇文章的代码,这个复杂吗,谢谢你❤

  • @satyakebakshi7896
    @satyakebakshi7896 3 месяца назад

    Amazing explanation, Thank you!

  • @INTERN4AI
    @INTERN4AI 3 месяца назад

    @AlexG4MES 8 months ago This was beautifully explained. As someone who relies exclusively on self learning from online materials, the mathematical barriers of differig notations and overly complex wording is the most time consuming challenge. Thank you for such a distilled explanation, with only the notation and wording that makes sense for an intuitive and initial dive understanding. Subscribed!

  • @Mathschampion510
    @Mathschampion510 3 месяца назад

    I want to ask these template line we have to write

  • @Jeredriq
    @Jeredriq 3 месяца назад

    Outdated, not working

  • @rene9901
    @rene9901 3 месяца назад

    Hello ... new to the field of llm training .. thanks for putting up these two videos .. However , I am a bit confused by the comment ' ...you can do it during inference time' ? As per my understanding the weight updates are done during fine tuning ... and they are later used during inference ... if the task is changed we just revert back to the pertained weight by getting rid off the weight update for current task .. and fine tune the model based on the new task to get the new weight update ... the new update are then again used later during inference .. so the weight updates are during fine tuning only ... which I think why the authors mentioned that batch processing is not obtained by LoRA (base) ... though possible and difficult ... may be there is some future version where its implemented ? I am not sure .... but please correct me if I am conceptually wrong ...

  • @thisismambonumber5
    @thisismambonumber5 4 месяца назад

    why are u ourple

  • @yagneshbhadiyadra7938
    @yagneshbhadiyadra7938 4 месяца назад

    What's the difference between less parameters, and low intrinsic weights? because weights are parameters of Neural Net isn't it?

  • @Veptis
    @Veptis 4 месяца назад

    The one issue I have had is that this causes memory footprint to grow. But it sounds like you should be able to merge it into the base model at the end to keep the same footprint. Maybe that is something for me to try. I wonder if this low rank decomposition can be used for model distillation. Instead of just quantizing weights.

  • @szymonpogodzinach2495
    @szymonpogodzinach2495 5 месяцев назад

    Man that video is fireee! Thank you for your work!

  • @nmirza2013
    @nmirza2013 5 месяцев назад

    Thanks a lot for this amazing explanation. I am fine tuning Mixtral 8x7B and using QLoRA have been able to perform test runs on Colab Pro using A100 machine.

  • @mdraugelis
    @mdraugelis 5 месяцев назад

    Thanks for the great intro to LoRA. I liked your graphics and your take-aways, also you energetic presentation :)

  • @Machiuka
    @Machiuka 5 месяцев назад

    How to use it locally through gpt4all? Thank you!

  • @swfsql
    @swfsql 5 месяцев назад

    Thanks a lot for this video! This is the first time I see a good explanation on this LoRA thing! 14:45 One minor note, is that it would indicate that the model has a low intrinsic info only if you could get rid of the original weights and just stick to the lora. That is, during the lora finetune training, if you could get away with while decaying the original (non-lora) weights down to zero. So I think that what has a low intrinsic info is "what you have to change from the base model" for your finetuning needs - but not the base model itself.

  • @kashishmukheja7024
    @kashishmukheja7024 5 месяцев назад

    Probably a case where overfitting can be beneficial. 😁

  • @mattchinnock7064
    @mattchinnock7064 5 месяцев назад

    Any plans to continue the series? Sounds like exactly what I'm looking for.

  • @heirofsam
    @heirofsam 6 месяцев назад

    Do I need to "download" LLM?

  • @yasinyaqoobi
    @yasinyaqoobi 6 месяцев назад

    Attempting to run the notbook but I keep getting ValueError: Attempting to unscale FP16 gradients. Tried different colab envs but no luck.

  • @yasinyaqoobi
    @yasinyaqoobi 6 месяцев назад

    Great video. Wish you showed the comparison against the base model. Just to clarify, we are not able to use the LORA model generated from model A with a different base model?

  • @luisjoseve
    @luisjoseve 6 месяцев назад

    Christ, amazing job! please keep going!

  • @luisjoseve
    @luisjoseve 6 месяцев назад

    Impresive, you rock, waiting for more videos

  • @Robo-fg3pq
    @Robo-fg3pq 6 месяцев назад

    Getting "ValueError: Attempting to unscale FP16 gradients." when running the cell with trainer.train(). Any idea?

    • @shashankjainm5009
      @shashankjainm5009 2 месяца назад

      Even i'm getting the same error for "bloom-1b7". Did your problem resolved ?

    • @Jithendra0001
      @Jithendra0001 10 дней назад

      @@shashankjainm5009 I am getting the same error. did you fix that??.

  • @legionmars9491
    @legionmars9491 7 месяцев назад

    sudo docker run --gpus=all --shm-size 64g -p 7860:7860 -v ${HOME}/.cache/root/.cache --rm alpaca-lora-demo <- no errors, no output, server localhost doesnt work ;(

  • @monolizm
    @monolizm 7 месяцев назад

    this is a masterpiece

  • @alexandria6097
    @alexandria6097 7 месяцев назад

    Do you know how much GPU RAM the meta-llama/Llama-2-70b-chat model would take to fine-tune?

  • @user-qy9sx7bn1l
    @user-qy9sx7bn1l 7 месяцев назад

    I particularly appreciate the depth of research and preparation that clearly goes into this video. It's evident that you're passionate about the topics you cover, and your enthusiasm is contagious. Your dedication to providing accurate information while maintaining an accessible and entertaining format is commendable.

  • @moonly3781
    @moonly3781 7 месяцев назад

    Thank you so much! I'm working on a project to develop a chatbot for student advisory services and i am contemplating between two approaches: Fine-Tuning and Retrieval-Augmented Generation (RAG). Here are the key tasks the chatbot needs to address: Answering general queries about university courses and requirements. Providing personalized advice on study plans and schedules. Assisting with administrative processes (enrollment, documentation, etc.). Offering support for common academic and personal challenges faced by students. Given these responsibilities, which approach would be more suitable? Fine-Tuning could offer more precise, tailored responses, but RAG might better handle diverse, real-time queries with its information retrieval capabilities. Any insights or experiences with similar projects would be greatly appreciated!

  • @A1AutomotiveDrivetrains-nq4pl
    @A1AutomotiveDrivetrains-nq4pl 8 месяцев назад

    I like the videos and I'm trying to learn the whole AI thing which is no doubt a lot to grasp when you're coming into it new. the only thing that I will change is, if possible just let people know in the beginning that you will be explaining further details of certain things throughout the classes and then let it be. Because what I'm finding is that as you're explaining it, when I start to follow you and I'm synced into what you're saying you kind of break that concentration by repeating the fact that you're going to explain the details again later. if that makes any sense hopefully you understand what I'm saying. Because I am still so new at this if you have any other information that is more up to date I would greatly appreciate anything that you can help me along with. All in all I like the videos.

  • @simonnarang3369
    @simonnarang3369 8 месяцев назад

    Why does deltaW need to be represented by both WA x WB? Why couldn't it be represented using just smaller matrix?

    • @chrisalexiuk
      @chrisalexiuk 8 месяцев назад

      In order to preserve the original shape of the weights and to avoid needing to change the model architecture!

  • @user-gq2bq3zf1f
    @user-gq2bq3zf1f 8 месяцев назад

    When I run inpaintanything in StableDiffusionUI, especially when I run inpainting, I keep getting error Unexpected end of JSON input.I ran it through Google Labs, what should I do?

  • @glebmaksimov4885
    @glebmaksimov4885 8 месяцев назад

    Отличный туториал)

  • @98f5
    @98f5 8 месяцев назад

    any chance you can make an example of fine tuning code llama like this

    • @chrisalexiuk
      @chrisalexiuk 8 месяцев назад

      I might, yes!

    • @98f5
      @98f5 8 месяцев назад

      @chrisalexiuk itd be greaty appreciated. There is almost no implementation docs or examples around for using lora 😀

  • @kingdown7502
    @kingdown7502 8 месяцев назад

    Could you explain why this saves memory? Don't you need the pre-trained weights in backprop to calculate the difference matrixes and during the forward pass?

    • @chrisalexiuk
      @chrisalexiuk 8 месяцев назад

      We only need to pass through the frozen weights, which means we don't need them in the optimizer. That is where the significant memory load reduction comes from.

  • @randomthoughts7838
    @randomthoughts7838 8 месяцев назад

    I am able to inference the fine tuned model from lit llama but I want to do the conversation with the fine tuned model. How can I do it with the code in repository.

  • @fatvvsfatvvs5642
    @fatvvsfatvvs5642 8 месяцев назад

    Thank you for the good explanation of LoRA! Super!

  • @maxlgemeinderat9202
    @maxlgemeinderat9202 8 месяцев назад

    Great video! What wpuld be different if i do download the model not on colab but locally? Which lines do change in the code?

    • @chrisalexiuk
      @chrisalexiuk 8 месяцев назад

      You should be able to largely recreate this process locally - but you would need to `pip install` a few more dependencies. You can find which by looking at what the colab environment has installed - or using a tool like pipreqs!

  • @benjaminbear138
    @benjaminbear138 8 месяцев назад

    Dude. This video is amazing

    • @chrisalexiuk
      @chrisalexiuk 8 месяцев назад

      Thank you! That means a lot!

  • @davidromero1373
    @davidromero1373 8 месяцев назад

    Hi a question, can we use lora to just reduce the size of a model and run inference, or we have to train it always?

    • @chrisalexiuk
      @chrisalexiuk 8 месяцев назад

      LoRA will not reduce the size of the model during inference. It actually adds a very small amount extra - this is because the memory savings come from reduced number of optimizer states.