6 Years of AI Progress: ModernBERT Finally Replaces BERT

Поделиться
HTML-код
  • Опубликовано: 6 янв 2025

Комментарии • 7

  • @botsknowbest
    @botsknowbest  8 дней назад +4

    Thank you for watching! What do you think about ModernBERT?

  • @focusonai
    @focusonai 7 дней назад +2

    Amazing video!!

  • @clay1958
    @clay1958 7 часов назад +1

    How do BERT and ModernBERT differ from embedding models such as OpenAI's text-embedding-3-small and text-embedding-3-large?

    • @botsknowbest
      @botsknowbest  4 часа назад +1

      @clay1958 BERT and ModernBERT are general-purpose NLP models. They were trained using masked language modeling (MLM), which involves predicting masked tokens in a sequence. Using MLM, the models can learn to understand the context and can capture lots of knowledge due to large pre-training corpora, but you usually want to fine-tune them for your downstream tasks to get the best results. For example, if you want to use ModernBERT for semantic search, you might want a version fine-tuned on MS-MARCO, a popular retrieval dataset with contrastive examples. (This is what I did in my video.)
      OpenAI's embedding models are already fine-tuned on some internal datasets. They're designed to work well for tasks like semantic search, clustering, and recommendations right out of the box, so you don't need additional training to get started.
      ModernBERT Embed (huggingface.co/nomic-ai/modernbert-embed-base) is probably a good alternative to OpenAI's embed models at the moment. It's trained on the Nomic Embed datasets and performs on par with OpenAI's text-embedding-3 models on benchmarks like MTEB. Plus, you will also get all the benefits of using open-source models.

    • @clay1958
      @clay1958 4 часа назад +1

      @@botsknowbest Thank you for this reply! And fantastic video by the way

    • @botsknowbest
      @botsknowbest  4 часа назад

      Thank you!!