Apple Join the AI Race with MLX

Поделиться
HTML-код
  • Опубликовано: 30 ноя 2024

Комментарии • 52

  • @darkwoodmovies
    @darkwoodmovies 11 месяцев назад +7

    I'm SO happy that Apple did this. The only thing we need from Apple now is to make their Apple Silicon capable of running AAA games, and then there will be no reason for Windows to exist.

    • @Yu-Fei-Hung
      @Yu-Fei-Hung 6 месяцев назад

      It is capable through Parallels and even Switch games with emulation

  • @lrkx_
    @lrkx_ 11 месяцев назад +7

    Nice coverage. I’m really keen to do a fine-tune on a Mac natively, taking full advantage of the hardware finally. :)

  • @AZisk
    @AZisk 11 месяцев назад +1

    Yay! Nice edit !

    • @engineerprompt
      @engineerprompt  11 месяцев назад

      Thanks, finally the right version :)

  • @DocuFlow
    @DocuFlow 10 месяцев назад +1

    Excellent content as always. Keep it up!

  • @ifysalabas
    @ifysalabas 7 дней назад

    Ok is it like pytorch tensorflow etc? Can we use trained llms on it. How we do ai model deployment and optimization pipeline on it?

  • @engineerprompt
    @engineerprompt  11 месяцев назад +4

    This is the EDITTED VERSION 😃

    • @The_8Bit
      @The_8Bit 11 месяцев назад

      i wanna see the blooper reel!

    • @engineerprompt
      @engineerprompt  11 месяцев назад +1

      @@The_8Bit haha, you missed the unedited one :D May be will release those for all the videos :D

  • @ishudshutup
    @ishudshutup 10 месяцев назад

    Oh very nice video! I hope you do more MLX coverage, for many people this will be the only viable option as high ram >16GB vram isn't accessible otherwise. Would be interesting to see a performance comparison between MLX and using llama.cpp based models like on LM Studio.

    • @engineerprompt
      @engineerprompt  10 месяцев назад

      Thanks, will look into it further if there is interest.

  • @gabrield4892
    @gabrield4892 11 месяцев назад +1

    I wish I was smart enough to understand everything you did in this video. Can someone maybe answer this instead? Does this mean I’ll eventually be able to run 7B model from huggingface with better performance on my M2 MacBook Air, once the models are updated/converted to this format?

  • @prestonmccauley43
    @prestonmccauley43 11 месяцев назад +3

    I use llama cpp on make to do a lot and it’s pretty fast. I’m wondering this actually will make it faster,

  • @billybob9247
    @billybob9247 11 месяцев назад

    You need an Activity Monitor that shows Neural Engine activity. Is the NE being used here?

    • @Sam16842
      @Sam16842 11 месяцев назад +1

      NE is only used when running ML models, not for building them. This framework is to build models.

    • @darkwoodmovies
      @darkwoodmovies 11 месяцев назад

      @@Sam16842 Does the training process only use the GPU??

    • @tacorevenge87
      @tacorevenge87 10 месяцев назад

      @@darkwoodmoviesyes

  • @qq-sh8sn
    @qq-sh8sn 11 месяцев назад

    When running I get the following error:
    [INFO] Loading model from disk.
    Traceback (most recent call last):
    File "/Users/MacBookAir/mlx-examples/llama/llama.py", line 356, in
    model = load_model(args.model)
    File "/Users/MacBookAir/mlx-examples/llama/llama.py", line 306, in load_model
    weights = mx.load(str(model_path / "weights.npz"))
    RuntimeError: [load] Failed to open file Llama-2-7b-chat-mlx/Llama-2-7b-chat.npz/weights.npz
    Why is it not able to locate the weights? Everything else has been completed according to the instructions.

    • @shen_7777
      @shen_7777 11 месяцев назад +1

      The github file has been modified and needs to wait for huggingface fix it

    • @donconkey1
      @donconkey1 11 месяцев назад

      Encountered similiar error. The runtime script auto appends weights.npz. If specified with path in llama.py run scripts mens two instances and command fials. This worked for me. Runs horribly slow on my mac m1 pro 16gb machine.
      python llama.py Llama-2-7b-chat-mlx/ Llama-2-7b-chat-mlx/tokenizer.model --prompt "hello"
      The terminal command as executed with environment mlx active in directory ~/Llama-2-7b-chat-mlx/llama

  • @asermauricio
    @asermauricio 11 месяцев назад +1

    Can you make a video explaining to use mlx and whisper in Mac device

  • @andresshamis4348
    @andresshamis4348 10 месяцев назад

    Is MLX the same as ferret?

    • @engineerprompt
      @engineerprompt  9 месяцев назад

      No, MLX is a framework. Ferret is a model :)

    • @andresshamis4348
      @andresshamis4348 9 месяцев назад

      @@engineerprompt i would really enjoy a basic tutorial on ferret and mlx

  • @kingfunny4821
    @kingfunny4821 11 месяцев назад

    can you explain how can use this model with local documentation

    • @engineerprompt
      @engineerprompt  11 месяцев назад +2

      Yes, looking into it

    • @kingfunny4821
      @kingfunny4821 11 месяцев назад

      @@engineerprompt
      Is there a video that explains chatting with documents other than this model?

    • @kingfunny4821
      @kingfunny4821 11 месяцев назад +1

      Is there a video that explains chatting with documents other than this model?
      offline

  • @tacorevenge87
    @tacorevenge87 10 месяцев назад

    Would be good to run on ANE

  • @benjaminalonso4630
    @benjaminalonso4630 11 месяцев назад +1

    muchas gracias

  • @HoyleGustave
    @HoyleGustave Месяц назад

    641 Sage Track

  • @abhiramgaruda9566
    @abhiramgaruda9566 9 месяцев назад

    Hi, It was very informative video.. Thanks a ton..
    I have a small question. When i run
    "[INFO] Loading model from Llama-2-7b-chat-mlx/weights.npz.
    Press enter to start generation"
    It doesn't show like yours and it also not using the full capacity of gpu nor cpu.. taking lot of time to give the response..
    System specs.. 16gb ram & M2 Pro chip.

    • @engineerprompt
      @engineerprompt  8 месяцев назад

      I think there were updates to the mlx package. I am thinking about making more content on the updated package. Coming soon :)

  • @tapos999
    @tapos999 11 месяцев назад

    whats your config may i know? m2 max 16gb/24gb? I was trying to run on m1 16gb air but i think i was a bit ambitiuous, its horribly slow :(

  • @WallaceJosephine-k6d
    @WallaceJosephine-k6d 2 месяца назад

    424 O'Conner Wells

  • @xt3708
    @xt3708 11 месяцев назад +14

    Please for the loves sake, make the advertisements come on in logical part not midsentence. Your videos are generally quiet so I pump up the volume and then an ad comes on Breaking my speakers and my ears. Pick a spot where an advertisement should start and then take a pause and say when we come back …. please

    • @zacharykosove9048
      @zacharykosove9048 11 месяцев назад +2

      Just use an adblocker

    • @xt3708
      @xt3708 11 месяцев назад +1

      Using the app

    • @Max-zr7hr
      @Max-zr7hr 11 месяцев назад +2

      You tube just adds this in the creator doesn’t say where the ads are

  • @KimberlyMyers-s3f
    @KimberlyMyers-s3f 3 месяца назад

    Martin Sarah Young Angela Rodriguez Gary

  • @AdrianScottty
    @AdrianScottty Месяц назад

    Rodriguez Jason Taylor Richard Lee Dorothy

  • @MariaThompson-d7y
    @MariaThompson-d7y 2 месяца назад

    Anderson Nancy Clark Elizabeth Young Charles

  • @crazytrain86
    @crazytrain86 11 месяцев назад

    Ummm isn't this exactly what llama.cpp was originally for? Literally for Apple chips native inference.

  • @RockefellerSidney
    @RockefellerSidney 2 месяца назад

    51768 Alana Meadow