Ian Wootten
Ian Wootten
  • Видео 80
  • Просмотров 535 790
Instructor: The Best Way to get Typed Data from Ollama
Instructor makes getting structured data back from large language models simple. Here I explore why instructor is so useful when you want to use an LLM as part of a larger system.
00:00 What's the problem?
02:17 Formatting responses as JSON
04:15 Using Instructor
07:46 Web scraping demo
12:59 Conclusion
useinstructor.com
docs.pydantic.dev/latest/
Support My Work:
Check out my website: www.ianwootten.co.uk
Follow me on twitter: iwootten
Subscribe to my newsletter: newsletter.ianwootten.co.uk
Buy me a cuppa: ko-fi.com/iwootten
Learn how devs make money from Side Projects: niftydigits.gumroad.com/l/side-projects
Просмотров: 1 644

Видео

Using Ollama to Run Local LLMs on the Steam Deck
Просмотров 1,9 тыс.2 месяца назад
The Steam Deck is a fully fledged pc, which means it's possible to run Ollama on it too. Here I try to install it, run a few models and compare it to the speeds on the Raspberry Pi 5. 00:00 Intro 00:31 Installation 04:38 Model Runs 10:04 Conclusion ollama.com store.steampowered.com/steamdeck Support My Work: Check out my website: www.ianwootten.co.uk Follow me on twitter: iwootten S...
Choosing Your LLM Provider is a Whole Lot Easier with This
Просмотров 1,4 тыс.4 месяца назад
If you want to use a hosted language model you have so many options, how do you know which is the best? This is a great site that will help you. artificialanalysis.ai Support My Work: Check out my website: www.ianwootten.co.uk Follow me on twitter: iwootten Subscribe to my newsletter: newsletter.ianwootten.co.uk Buy me a cuppa: ko-fi.com/iwootten Learn how devs make money from Side ...
Will Your New Python Package Manager be Written in Rust?
Просмотров 2,9 тыс.4 месяца назад
"uv" is a new Python package installer written by astral, but this one is 100x faster and written in Rust. I compare it with pip 00:00 Intro 01:20 Speed Test with pip 05:25 Differences to pip 08:47 Conclusion #python #rustlang git: github.com/astral-sh/uv Blog post: astral.sh/blog/uv Support My Work: Check out my website: www.ianwootten.co.uk Follow me on twitter: iwootten Subscribe...
How Fast Will Your New Mac Run LLMs?
Просмотров 5 тыс.5 месяцев назад
How fast can the new Apple Silicon Mac you so desperately want run LLM's and is it worth the price? llama.cpp benchmarks: github.com/ggerganov/llama.cpp/discussions/4167 Ollama: ollama.ai 00;00 Intro 00:47 Benchmarks 05:06 Unbox 05:47 Results Support My Work: Check out my website: www.ianwootten.co.uk Follow me on twitter: iwootten Subscribe to my newsletter: newsletter.ianwootten.c...
Using Ollama to Run Local LLMs on the Raspberry Pi 5
Просмотров 53 тыс.5 месяцев назад
My favourite local LLM tool Ollama is simple to set up and works on a raspberry pi 5. I check it out and compare it to some benchmarks from more powerful machines. 00:00 Introduction 00:41 Installation 02:12 Model Runs 09:01 Conclusion Ollama: ollama.ai Blog: www.ianwootten.co.uk/2024/01/19/using-ollama-to-run-llms-on-the-raspberry-pi-5/ Support My Work: Check out my website: www.ianwootten.co....
Google Gemini AI: Cutting Through the Hype
Просмотров 9827 месяцев назад
We got a surprise announcement of Google's new Gemini model this week. I breakdown everything that's been announced about and why you should take notice. 00:00 Introduction 00:25 Gemini Examples 03:59 Alphacode 2 05:26 Benchmarks 07:23 Availability 08:03 Conclusion Gemini: deepmind.google/technologies/gemini/ Support My Work: Check out my website: www.ianwootten.co.uk Follow me on twitter: twit...
Draw My UI: Turn Wireframes into Tailwind/JS with One Click
Просмотров 3,4 тыс.7 месяцев назад
Text to code is so last month! We can now just draw what we want to build, just like proper designers do. #webdevelopment Website: drawmyui.com Github Link: github.com/tldraw/draw-a-ui 00:00 Intro 00:50 Newsletter Signup 03:15 Navbar/Login 06:21 Pricing Plans 09:54 Conclusion Support My Work: Check out my website: www.ianwootten.co.uk Follow me on twitter: iwootten Subscribe to my n...
Easily Create Voiceovers Using OpenAI's New Text to Speech and Vision Models
Просмотров 4,6 тыс.8 месяцев назад
OpenAI's latest updates include TTS and GPT4V allowing us to get image descriptions and create audio. In this video I check out some of the most novel ways devs have been using it and show how to generate voiceovers for images or video. #openai 00:00 Intro 01:11 Using TTS to create audio 04:51 Using GPT4V to describe images 08:58 Using TTS & GPT4V for Video voiceovers 14:20 Conclusion Github Li...
Amazon Bedrock: Generative AI on AWS without the Headaches
Просмотров 1,1 тыс.9 месяцев назад
Amazon Bedrock - AWS's answer to generative AI for a whole collection of foundational models has just been released. I show you how you can use it in both the AWS Console and the command line. 00:00 Intro 00:43 Enabling Models 01:52 Basic Text Completion 05:30 Streaming Text Completions 09:32 Other Use Cases 10:33 Conclusion Link: aws.amazon.com/bedrock/ Support My Work: Check out my website: w...
Ollama on Linux: Easily Install Any LLM on Your Server
Просмотров 24 тыс.9 месяцев назад
Ollama has just been released for linux, which means it's now dead simple to run large language models on any linux server you choose. I show you how to install and configure it on digitalocean. 00:00 Installation on DigitalOcean 03:30 Running Llama2 on a Server 05:43 Calling a Model Remotely 12:26 Conclusion #llm #machinelearning Link: ollama.ai/download/linux Support My Work: Get $200 credit ...
Code Llama: First Look at this New Coding Model with Ollama
Просмотров 4,8 тыс.10 месяцев назад
I install and try out Metas Code Llama. I in this video I walkthrough the process of setting it up on my Mac using Ollama. We'll also attempt some Python programming challenges. Is it really any good at Python? #llm #python Using Ollama with LLama 2: ruclips.net/video/tIRx-Sm3xDQ/видео.html 00:00 Intro 00:25 Install with Ollama 02:33 Better than Llama 2? 05:33 Coding Challenges 11:34 Conclusion...
FaceFusion: Creating Deepfakes is Now Shockingly Simple on a Mac
Просмотров 31 тыс.10 месяцев назад
Facefusion is a really easy way to create basic deepfakes on your local machine. It's incredibly simple to install and makes use of coreml in order to use Apple Silicon. In this video I set it up and try it out. Link: facefusion.io 00:00 Intro/Install 03.25 Creating Deepfakes 09:50 Conclusion Support My Work: Check out my website: www.ianwootten.co.uk Follow me on twitter: iwootten ...
Ollama: The Easiest Way to Run Uncensored Llama 2 on a Mac
Просмотров 31 тыс.11 месяцев назад
Ollama is the simplest way of getting Llama 2 installed locally on your apple silicon mac. I install it and try out llama 2 for the first time with minimal hassle. Even includes uncensored versions of models! Link: ollama.ai 00:00 Intro/Install 01:48 Trying Llama 2 Out for Programming 04:22 Logic Problem 05:10 Ollama backend 05:48 Custom Models 07:40 API 09:14 Uncensored Llama 2 11:02 Conclusio...
AI Showdown: Can Claude 2 from Anthropic Take On ChatGPT?
Просмотров 1,9 тыс.11 месяцев назад
I try out Anthropics Claude 2 beta and compare to ChatGPT. Upload multiple files, 100k context tokens and currently free to use if you're in the US or UK. #gpt #chatbot claude.ai 00:00 Intro 01:48 Upload 02:27 Creativity 03:08 Data Cut Off 04:26 Reasoning 05:16 Code Analysis 07:03 Data Analysis 10:15 Outtro Support My Work: Check out my website: www.ianwootten.co.uk Follow me on twitter: twitte...
How Developers Can Use ChatGPT's New Code Interpreter
Просмотров 1,9 тыс.Год назад
How Developers Can Use ChatGPT's New Code Interpreter
Aider: This AI Coder Can Create AND Update Git Codebases
Просмотров 6 тыс.Год назад
Aider: This AI Coder Can Create AND Update Git Codebases
GPT Engineer: Can AI Really Code a Complete Codebase?
Просмотров 16 тыс.Год назад
GPT Engineer: Can AI Really Code a Complete Codebase?
How to Replace Bose QC35 Ear Pads
Просмотров 70Год назад
How to Replace Bose QC35 Ear Pads
OpenAI Function Calling: Structured Data from GPT is Now Much Simpler
Просмотров 1,8 тыс.Год назад
OpenAI Function Calling: Structured Data from GPT is Now Much Simpler
Copilot Chat: First Impressions Reviewing My Code
Просмотров 4,7 тыс.Год назад
Copilot Chat: First Impressions Reviewing My Code
Copilot for CLI - GitHub AI is Now in Your Terminal
Просмотров 1,9 тыс.Год назад
Copilot for CLI - GitHub AI is Now in Your Terminal
A Deep Dive into How Python Virtual Environments Work
Просмотров 320Год назад
A Deep Dive into How Python Virtual Environments Work
ShellGPT: Never Leave the Terminal Again
Просмотров 5 тыс.Год назад
ShellGPT: Never Leave the Terminal Again
GPT4 is here! What's New and How You Can Use It
Просмотров 250Год назад
GPT4 is here! What's New and How You Can Use It
How to Code Custom ChatGPT API Interfaces (CLI & Voice)
Просмотров 856Год назад
How to Code Custom ChatGPT API Interfaces (CLI & Voice)
What's Microsoft's new Bing AI Chat like for Coding?
Просмотров 2,5 тыс.Год назад
What's Microsoft's new Bing AI Chat like for Coding?
A Proposal to Fix the Confusing World of Python Package Installation
Просмотров 92Год назад
A Proposal to Fix the Confusing World of Python Package Installation
Pythons Package Installer is Confusing - What can we Learn from Node.js?
Просмотров 385Год назад
Pythons Package Installer is Confusing - What can we Learn from Node.js?
FAST External Storage for Macs - Acasis USB4 NVMe SSD Enclosure
Просмотров 7 тыс.Год назад
FAST External Storage for Macs - Acasis USB4 NVMe SSD Enclosure

Комментарии

  • @crazytom
    @crazytom 9 дней назад

    Thanks for leaving all the errors in and correcting them. Excellent.

  • @ioaircraft
    @ioaircraft 10 дней назад

    tryin, 10 turing pi's with 40 nvidia jetson orin's on them, with primary linux computer, so 41 systems total, clustered using archer and kubermites. it will def rocket then... 4,000-6,000 tops, 1.2 terabytes of ram.

  • @SuperRia33
    @SuperRia33 17 дней назад

    How do you connect to server via Python Client or Fast APIs for integration with projects/notebook?

    • @IanWootten
      @IanWootten 16 дней назад

      If you simply want to make a request to an API from Python, there are plenty of options. You can use a package from Python itself like urlllib, or a popular library like requests.

  • @andikunar7183
    @andikunar7183 17 дней назад

    TG largely depends on memory-bandwidth (the SoC has to pump all of the parameters and the KV-caches from RAM into the SoC's caches for each token generated). PP (and ML) is dependent on compute (GPU-horsepower) because token-processing can be batched. The M4 has 20% faster memory-bandwidth in addition to the faster GPUs. Let's see when Apple will do MacBooks with these chips, maybe I will upgrade my M2. For me, the M3 is not interesting enough for an M2 upgrade.

  • @sb_2378
    @sb_2378 18 дней назад

    Please stop click baiting

  • @trapez_yt
    @trapez_yt 24 дня назад

    i cant run it on service ollama start, it says the following: $sudo: service ollama start ollama: unrecognized service

  • @user-ed4yp6eq5k
    @user-ed4yp6eq5k 24 дня назад

    Sir I'm off topic but how can I enable usb support on the steam deck Gnome Boxes

  • @exploratoria
    @exploratoria 27 дней назад

    Hi Ian, great clip - how do we get it to read the prompt answers aloud with reasonable low latency?

    • @IanWootten
      @IanWootten 27 дней назад

      You'd need to pipe the chat output into a text to speech model (TTS). MacOS has the built in "say" command, so you could send it straight into that, if you want to keep it all local but it won't be anywhere near as good as an external service.

  • @MeinDeutschkurs
    @MeinDeutschkurs 28 дней назад

    Great! But try Gemma model. Sounds strange, but it is really good on picking content.

  • @whatsbetter8457
    @whatsbetter8457 28 дней назад

    Hey Ian, there is a more Ollama native library for the same use case called ollama-instructor. It was inspired by instructor from Jason Liu.

    • @IanWootten
      @IanWootten 28 дней назад

      This is a very recent library and doesn't seem to offer most of the features of instructor. I'm not clear on why you'd use it over instructor itself.

  • @davidtindell950
    @davidtindell950 28 дней назад

    thank you.

    • @IanWootten
      @IanWootten 28 дней назад

      You're very welcome.

  • @galdakaMusic
    @galdakaMusic Месяц назад

    What about renew this video with the new Rpi Hat AI? Thanks

    • @IanWootten
      @IanWootten Месяц назад

      Could do, but I don't think Ollama would be able to leverage it, plus it's not out yet.

  • @perschinski
    @perschinski Месяц назад

    Great stuff, thanks a lot!

  • @AdarshSingh-rm6er
    @AdarshSingh-rm6er Месяц назад

    hello Ian, Its a very great video. I have some query, i will very thankful if you can help me. I am stuck since 3 days. Apparently, I am trying to host the ollama on my server. i am very new to linux and dont understand the whats wrong i am doing. I am using nginx to host the ollama on my proxies and configure the nginx file and yet getting access denied error. I can show you the code if you want, please respond.

  • @TheSeriousDog
    @TheSeriousDog Месяц назад

    I had no idea brew was also available for linux. Pretty cool way of going about it

  • @TheStallion1319
    @TheStallion1319 Месяц назад

    I want to start experimenting with llms and I have a budget for laptop or pc or a compromise of both , I was going for a great Mac or an ok one and a pc , what’s your advise ?

    • @IanWootten
      @IanWootten Месяц назад

      A lot of it will come down to personal preference. I'm familiar with Macs, really like that they are silent and have great battery. Most of my choice is based on that, the fact they're very good for llms too works in my favour. I'm sure there's some pretty good PCs out there too, and now Ollama works there too.

    • @TheStallion1319
      @TheStallion1319 Месяц назад

      @@IanWootten yes I like Mac OS much more than windows but my concern was the speed and size of the model , I am concerned with that 16gb of unified memory wouldn’t be enough

  • @DanteS-119
    @DanteS-119 Месяц назад

    That's sick!!!

  • @salah-eddineboucetta601
    @salah-eddineboucetta601 Месяц назад

    Very helpful thank you so much

  • @northerncaliking1772
    @northerncaliking1772 Месяц назад

    Says error when pulling list

  • @petehboy2
    @petehboy2 Месяц назад

    How much was the new compared to refurbished?

    • @IanWootten
      @IanWootten Месяц назад

      This was the 8c M1 512GB version and cost £1700 from Apple Refubished. I ended up plumping for a 10c M1 1TB from Amazon Warehouse a few months later that cost £1400 due to a very light dink in it's underside. You can see me unboxing that here: ruclips.net/video/KOn4BRoSB7A/видео.html

  • @ankitvaghasiya3789
    @ankitvaghasiya3789 Месяц назад

    thank you🦙

  • @MacGamePlayTrick
    @MacGamePlayTrick Месяц назад

    Traceback (most recent call last): File "/Users/*******/Desktop/facefusion/run.py", line 3, in <module> from facefusion import core File "/Users/*******/Desktop/facefusion/facefusion/core.py", line 9, in <module> import numpy ModuleNotFoundError: No module named 'numpy'

  • @devlok4841
    @devlok4841 Месяц назад

    Hi Ian, thank you for the video! When I issue the run command, I'm getting a "Frame processor .DS_Store could not be loaded" error. Do you have any suggestions to fix this?

  • @kayodep.i5012
    @kayodep.i5012 Месяц назад

    Not simple at all, you just go ahead with it assuming we already knew the basics

    • @wavi_2_d_world715
      @wavi_2_d_world715 26 дней назад

      He must be so stupid I guess it’s a scam… how do you clone what you don’t have on your system

  • @lrdass
    @lrdass Месяц назад

    I think your solution using brew is better than using distrobox. For some weird reason (I couldn´t find out why) more than a dozen times I had the whole distrobox "broke" for any weird reason. I could not understand why it happens, but all of the sudden distrobox would not initiate the image. Something like podman had lost track of where the volume of the image was. So for podman the container was still running. And it was indeed running an external container (which was the volume). This have happened to me at least 7 times. Dunno if it is when you switch to gaming mode some bad-process management happen and the process of podman gets lost -- or some file gets losts in some process of the updates. I've given up trying to figure out why this kept on happening. And once it does, it is almost impossible to retrieve the volume of your old container, so it is the same as starting over. So .. imo it is way easier and better setup to use brew. But if you installed distrobox and did not find any error, just go along! prob was on my deck only.

    • @IanWootten
      @IanWootten Месяц назад

      Sorry to hear that, I hit a an error when trying to enter the distro, but it didn't persist. Heads up that I recently returned to use the desktop on the deck and it appeared that most of my brew setup had been messed up due to os updates - distrobox however still seemed to be fine.

    • @lrdass
      @lrdass Месяц назад

      @@IanWootten gotcha! yea! I feel that something happening on brew after the updates but I dunno what. I'm trying to create a stable setup with nix this time. I hope I can get this right this time!

  • @sweetbb125
    @sweetbb125 2 месяца назад

    I've trie drunning OLLAMA on my Raspberry Pi 5, as well as an Intel Celeron based computer, and also an old Intel i7 based computer, and it worked everywhere. It is really behind impressive, thank you for this video to show me how to do it!

  • @mehmetbakideniz
    @mehmetbakideniz 2 месяца назад

    Does it automaticall detect and use apple m2 gpu? is there anything I need to configure to use it with gpu?

    • @IanWootten
      @IanWootten 2 месяца назад

      Nope, should automatically be making use of apple silicon.

  • @CFedits62
    @CFedits62 2 месяца назад

    Cool

  • @CFedits62
    @CFedits62 2 месяца назад

    i know this video is old but can you or someone make a video on the white dev board with the pin connections and how to use it?

    • @IanWootten
      @IanWootten 2 месяца назад

      It's called a "breadboard". Lots of great vids already on youtube to help with using it.

    • @CFedits62
      @CFedits62 2 месяца назад

      @@IanWootten thanks just learnt how to use leds and buttons you helped a ton (:

  • @itolond
    @itolond 2 месяца назад

    is there 2and 3 key versions of this board?

    • @IanWootten
      @IanWootten 2 месяца назад

      Just the 4 x 4 at the moment. There are a few 3 x 3 versions you can 3d print yourself though.

  • @ystrem7446
    @ystrem7446 2 месяца назад

    Does it run on CPU or GPU ? Thx

    • @IanWootten
      @IanWootten 2 месяца назад

      Hi there, I mentioned toward the end, but yeah it's running on the CPU.

    • @yuuleeyianmaiser2900
      @yuuleeyianmaiser2900 Месяц назад

      @@IanWootten I've also experimented with it and unfortunately wasn't able to get it running on the GPU. If you're successful, I'd be very interested in the results.

  • @internetcarson
    @internetcarson 2 месяца назад

    I just read your article on running the brew package manager on the Deck!

  • @ftlbaby
    @ftlbaby 2 месяца назад

    Thank you for this! The two main things that I dislike about LLMs is the middle school level answers and the nanny rails. Hopefully, running an uncensored LLM will at least make the low intelligence level less grating.

  • @74Gee
    @74Gee 2 месяца назад

    Run Pod is very affordable too. From 17c per hour for a Nvidea 3080

    • @IanWootten
      @IanWootten 2 месяца назад

      Yeah, I wanted to do a comparison of all the new services appearing.

  • @DaveParr
    @DaveParr 2 месяца назад

    2:54 love the honesty of hitting the bug and keeping it in the video ❤

    • @IanWootten
      @IanWootten 2 месяца назад

      Thanks Dave. Thought it was important to show that if you do go ahead and install stuff, it's going to potentially get wiped out with SteamOS updates.

  • @NicolasSilvaVasault
    @NicolasSilvaVasault 2 месяца назад

    that's super impressive even if it takes quite a while to respond, is a RASPBERRY PI

  • @donmitchinson3611
    @donmitchinson3611 2 месяца назад

    Thanks for video and testing. I was wondering if you have tried setting num_threads =3. I can't find video of where I saw this but I think they set it before calling ollama. Like environment variable. It's supposed to run faster. I'm just building a rpi5 test station now

  • @jondoe0x0x
    @jondoe0x0x 2 месяца назад

    Thannks Ian! I really wanted to try doing some simple stuff with my steam deck on the go and this is just the way

    • @IanWootten
      @IanWootten 2 месяца назад

      Glad you found it useful.

  • @complextheory9529
    @complextheory9529 2 месяца назад

    What terminal alternatives do you recommend?

  • @pandukawb
    @pandukawb 2 месяца назад

    If they built up an ARM based Framework Laptop, I would buy that in a heartbeat.

  • @AlexanderGriaznov
    @AlexanderGriaznov 2 месяца назад

    Am I the only one who noticed tiny llama response to “why sky is blue?” was shitty? What the heck rust causing blue color of the sky?

    • @IanWootten
      @IanWootten 2 месяца назад

      Others have mentioned it in the comments too. It is a much smaller model, but there are many others to choose from (albeit possibly slower).

  • @siciliandragon8654
    @siciliandragon8654 2 месяца назад

    That's reassuring...you're supposed to be the expert teaching the public, and you don't even cover all the standard classes in the module because you admit right off the bat that you 'only have experience' with a some?? I'll be sure to watch this until the end...or not.

    • @IanWootten
      @IanWootten 2 месяца назад

      Why do you think I am meant to be an expert? If only experts were allowed to make videos, there would be a whole lot less useful information out there.

  • @emir5146
    @emir5146 2 месяца назад

    Poetry is extremely slow when resolving the dependencies. And I m stucking this matter. poetry add hopsworks that it took for a long time. And finally it' not working. It's worked to forever

  • @metacob
    @metacob 2 месяца назад

    I just got a RPi 5 and ran the new Llama 3 (ollama run llama3). I was not expecting it to be this fast for something that is on the level of GPT-3.5 (or above). On a Raspberry Pi. Wow.

    • @brando2818
      @brando2818 Месяц назад

      I just recieved my pi, and I'm about to do the same thing.. Are you doing anything else on it?

  • @SethHarcrowmusic
    @SethHarcrowmusic 2 месяца назад

    shockingly simple? you just highlighted how technical skills are needed and it's not for beginners lol that tutorial was awful

    • @IanWootten
      @IanWootten 2 месяца назад

      Uploading a video and image seems pretty tame compared to sourcing a large collection of images for a traditional deepfake. Granted the setup process is the most involved thing here but should be familiar for most developers.

  • @josephkaisner4581
    @josephkaisner4581 2 месяца назад

    Very helpful thanks!

  • @jzam5426
    @jzam5426 2 месяца назад

    Do you know how to get it to run in LangChain while taking advantage of the M1/2 chips?

  • @nilutpolsrobolab
    @nilutpolsrobolab 2 месяца назад

    Such a calm tutorial but so informative💙

  • @technocorpus1
    @technocorpus1 3 месяца назад

    Awesome! I want to try this now! Can someone tell me if it necessary to install the model on an exterior SSD?

    • @IanWootten
      @IanWootten 2 месяца назад

      Not necessary, but may be faster. All the experiments here I was just using a microsd.

    • @technocorpus1
      @technocorpus1 2 месяца назад

      ​@@IanWootten That's just amazing to me. I have a Pi3, but am planning on upgrading to a pi5. After I saw your video, I downloaded ollama onto my windows pc. It only has 4 GB RAM, but I will still able to run several models!

  • @glittlehoss
    @glittlehoss 3 месяца назад

    I didnt think you could use a zero 2w in the original case

    • @IanWootten
      @IanWootten 3 месяца назад

      You can. This was right around when V2W of the case came out, so might be better off going for that given the nicer features (like rechargeable battery) it has.