Using the BUNNY Multimodal Model (Image description) with LLama.cpp

Поделиться
HTML-код
  • Опубликовано: 11 сен 2024
  • Hey my dudes! In this video, we dive into the super exciting world of multimodal models with LLama.cpp. Essentially that means we can upload images and get highly detailed information back. All from our command-line. Super cool!
    For this tutorial we use a 4b parameter model called "Bunny", which blows away all other Llava models I've used in the past. No joke.
    We're gonna show you how to get this awesome model running on the command line to generate image descriptions. Plus, we'll use Chat-GPT to create a cool, quick and dirty, Gradio app so we can use it right in our browser.
    For this tutorial to work you must have a working version of llama.cpp
    Model Repository: github.com/BAA...
    Code used in the tutorial: www.cognibuild...
    Patreon site: www.patreon.com/cognibuild

Комментарии • 2

  • @АльбертИванов-ц4х
    @АльбертИванов-ц4х Месяц назад

    llama_model_load: error loading model: error loading model hyperparameters: key not found in model: phi3.attention.sliding_window

    • @cognibuild
      @cognibuild  Месяц назад

      havent seen that error. ill look at it in a bit. Here's the discord channel so we can chat there: discord.gg/JfWJXXUu