Arthur
Arthur
  • Видео 35
  • Просмотров 25 488

Видео

[Webinar] LLMs and Misinformation: A Double-Edged Sword in the Digital Age
Просмотров 94Месяц назад
Misinformation-both human-made and, more recently, AI-generated-poses a significant societal challenge in the digital age. It is a serious threat to information ecosystems that causes public fear and anxiety, erodes trust in institutions, and exacerbates discrimination between social groups. In this session, Senior ML Engineer Cherie Xu explored the role that large language models have played a...
Ground Truth #6: OpenAI’s O1 Model Unpacked
Просмотров 250Месяц назад
In this episode of Ground Truth, co-hosts Adam Wenchel and John Dickerson dive into the groundbreaking release of OpenAI's O1 model. They explore its advancements in multi-step reasoning, cryptography, and its impact on complex tasks like math and programming. Join them as they discuss the model's capabilities, its comparison to previous releases, and its implications for the future of AI.
Ground Truth Podcast #5: Amazon's Alexa Revamp, Ilya Sutskever's $1B Raise, & The Antitrust Circus
Просмотров 67Месяц назад
Welcome to the fifth episode of Ground Truth, Arthur’s podcast where we delve into the latest happenings in the world of artificial intelligence. Join our hosts Adam Wenchel, CEO and Co-Founder of Arthur, and John Dickerson, Chief Scientist at Arthur, as they discuss the most recent AI trends, events, and their far-reaching implications. In this episode of Ground Truth, we dive into a whirlwind...
Ground Truth Podcast #4: Nvidia Earnings, AI Code Completion, and California’s AI Regulations
Просмотров 68Месяц назад
Welcome to the fourth episode of Ground Truth, Arthur’s podcast where we delve into the latest happenings in the world of artificial intelligence. Join our hosts Adam Wenchel, CEO and Co-Founder of Arthur, and John Dickerson, Chief Scientist at Arthur, as they discuss the most recent AI trends, events, and their far-reaching implications. In this episode, Adam and John dive into Nvidia's latest...
Ground Truth Podcast #3: Generative AI, Model Monitoring, and AI in Healthcare
Просмотров 782 месяца назад
Welcome to the third episode of Ground Truth, Arthur’s podcast where we dive into the latest happenings in the world of artificial intelligence. Join our hosts Adam Wenchel, CEO and Co-Founder of Arthur, and John Dickerson, Chief Scientist at Arthur, as they discuss current AI events, trends, and their implications. In this episode, Adam and John explore critical topics like the rise of generat...
Ground Truth Podcast #2: AI Privacy, Autonomous Vehicles, and Wall Street Insights
Просмотров 352 месяца назад
Welcome to the second episode of Ground Truth, Arthur’s podcast where we dive into the latest happenings in the world of artificial intelligence. Join our hosts Adam Wenchel, CEO and Co-Founder of Arthur, and John Dickerson, Chief Scientist at Arthur, as they discuss current AI events, trends, and their implications. In this episode, Adam and John tackle pressing topics such as data privacy, th...
[Webinar] Safeguarding AI Models: Exploring Prompt Injection Variants
Просмотров 3262 месяца назад
Prompt injections refer to a large category of attacks on large language models (LLMs) and multimodal models that are meant to elicit unintended behavior. But what are these types of “unintended behaviors”? And what techniques are used to accomplish this? In this session, ML Research Scientist Teresa Datta introduced a taxonomy to organize different types of prompt injection attacks-both direct...
Ground Truth Podcast #1: Mamba-based models, Apple’s 7B open-source model...
Просмотров 1102 месяца назад
Welcome to the first episode of Ground Truth, Arthur’s podcast where we dive into the latest happenings in the world of artificial intelligence. Join our hosts Adam Wenchel, CEO and Co-Founder of Arthur, and John Dickerson, Chief Scientist at Arthur and tenured faculty at the University of Maryland, as they discuss current AI events, trends, and their implications. In this episode, Adam and Joh...
[Webinar] A New Era of AI: Exploring Multimodal Embedding Observability
Просмотров 4323 месяца назад
2024 is the year of multimodal, and multimodal embeddings are some of the most exciting developments in this space. In this webinar, John Dickerson, Arthur's co-founder and Chief Scientist and Zach Nussbaum, Principal Machine Learning Engineer at Nomic discussed key concepts at the intersection of embeddings and ML observability, best practices for implementation, a behind-the-scenes look at bu...
[Webinar] What's The Deal With LLM App Development in 2024?
Просмотров 3364 месяца назад
LLMs are being used for all kinds of business applications. In this webinar, ML Research Engineer Max Cembalest and VP of Engineering Zach Fry shared an overview of a variety of tools & techniques to help you start experimenting with different uses of LLMs. They covered: - Touchpoints & Experimentation: Trends in getting set up quickly to start testing the newest models - Evaluation: How do you...
Ground Truth Episode 6 | Trailblazers in Tech: Women Pioneering AI Innovation
Просмотров 2286 месяцев назад
At our most recent Ground Truth event in celebration of Women’s History Month, we hosted an incredible lineup of female founders, builders, and academics: • Savannah Thais, Columbia University Data Science Institute • Diana Hsieh, Reality Defender • Rebekah Tweed, All Tech Is Human • Teresa Datta, Arthur • Tanya Dua, LinkedIn News They talked about their professional journeys, shared insights o...
Ground Truth Episode 5: AI Crashes the Science Party
Просмотров 5308 месяцев назад
At our most recent Ground Truth event, we hosted this incredible lineup of founders, builders, and academics who discussed the intersection of generative AI and STEM: • Andriy Mulyar (Co-Founder / CTO, Nomic) • John Dickerson (Co-Founder / Chief Scientist, Arthur) • Kyunghyun Cho (Professor, NYU and Senior Director, Genentech) • Micah Goldblum (Post-Doctoral Research Fellow, NYU / FAIR) • Raz B...
Model Monitoring for Recommender Systems | Arthur Feature Demo
Просмотров 1929 месяцев назад
In this demonstration, we showcase the capabilities around recommender system model monitoring in the Arthur Scope platform. Our capabilities are designed to provide real-time insights into your models' performance, ensuring the consistent delivery of high-quality suggestions and recommendations. Interested in learning more? 👉 Explore Arthur Scope: www.arthur.ai/product/scope 👉 Get in touch: ww...
[Webinar] LLMs & Generative AI in Editorial Content Creation & Media
Просмотров 3069 месяцев назад
Large language models (LLMs) and other forms of generative AI are being used in the media and in editorial content creation. In this webinar, Arthur's Chief Scientist John Dickerson covered: - The ways in which generative AI and LLMs are being used in editorial and creative content generation - Ethical considerations and challenges with generative AI and LLMs - The current frontier of LLMs incl...
[Webinar] Arthur LLM Product Demo & Research on Generative AI Challenges
Просмотров 53910 месяцев назад
[Webinar] Arthur LLM Product Demo & Research on Generative AI Challenges
[Webinar] Navigating the LLM Risk Landscape in Financial AI
Просмотров 1,2 тыс.11 месяцев назад
[Webinar] Navigating the LLM Risk Landscape in Financial AI
[Webinar] LLMs for Evaluating LLMs
Просмотров 10 тыс.Год назад
[Webinar] LLMs for Evaluating LLMs
[Webinar] Decoding the Machine Mind: The Quest for Explainability in ML
Просмотров 194Год назад
[Webinar] Decoding the Machine Mind: The Quest for Explainability in ML
Ground Truth Episode 4: The Future of LLMs with Arthur, MosaicML, LangChain, and Weaviate
Просмотров 3,8 тыс.Год назад
Ground Truth Episode 4: The Future of LLMs with Arthur, MosaicML, LangChain, and Weaviate
Pride in Tech: Embracing Responsible & Inclusive Innovation | Arthur x Out in Tech
Просмотров 72Год назад
Pride in Tech: Embracing Responsible & Inclusive Innovation | Arthur x Out in Tech
[Webinar] Considerations for Creating Fairer ML Systems
Просмотров 130Год назад
[Webinar] Considerations for Creating Fairer ML Systems
[Webinar] Lots of Little Mistakes: LLMs in Production
Просмотров 772Год назад
[Webinar] Lots of Little Mistakes: LLMs in Production
Ground Truth Episode 3: Jacopo Tagliabue on Recommender Systems
Просмотров 1,2 тыс.Год назад
Ground Truth Episode 3: Jacopo Tagliabue on Recommender Systems
[Webinar] Advanced Performance Monitoring with Arthur
Просмотров 183Год назад
[Webinar] Advanced Performance Monitoring with Arthur
Reckoning with the Disagreement Problem: Explanation Consensus as a Training Objective
Просмотров 71Год назад
Reckoning with the Disagreement Problem: Explanation Consensus as a Training Objective
Ground Truth Episode 2: Diego Oppenheimer on the Future of MLOps
Просмотров 302Год назад
Ground Truth Episode 2: Diego Oppenheimer on the Future of MLOps
Ground Truth Episode 1: Dr. Rachel Cummings on Differential Privacy
Просмотров 1,6 тыс.Год назад
Ground Truth Episode 1: Dr. Rachel Cummings on Differential Privacy
AI Summit New York 2022: Customer-First AI Panel with Arthur CEO Adam Wenchel
Просмотров 222Год назад
AI Summit New York 2022: Customer-First AI Panel with Arthur CEO Adam Wenchel
NeurIPS 2022: Teresa Datta on Tensions Between the Proxies of Human Values in AI
Просмотров 113Год назад
NeurIPS 2022: Teresa Datta on Tensions Between the Proxies of Human Values in AI

Комментарии

  • @TheUniversalAxiom
    @TheUniversalAxiom Месяц назад

    It was nicknamed Strawberry because it's an organic model, not a static GPT. What's buried in the straw, is their new jam. 🍎

    • @JohnPDickerson
      @JohnPDickerson Месяц назад

      Buried in the straw, or strraw, or strrraw, or ...

    • @TheUniversalAxiom
      @TheUniversalAxiom Месяц назад

      @@JohnPDickerson It's funny to watch people test this preview model, because it's not actually a test of o1's reasoning, it's a test of the users' reasoning. Those with imagination who can process information in creative ways, will see how different this model is compared to all precedent. And those who learned to just memorize information see it as an auto-complete or harder to use. They ask it to count letters in a word LMAO

  • @chrisogonas
    @chrisogonas Месяц назад

    I missed the live presentation, but this is just as good. Thanks for this incredible material. I am passionate about the disinformation and misinformation in the age of deluge of information, coupled with powerful tools at the disposal of everyday folks. It matters how these tools are deployed and how the unsuspecting vulnerable majority are protected from the potential harm. Thanks Cherie and Team!

  • @benny4013
    @benny4013 Месяц назад

    It was nothing about the Finance.

  • @Andrewthecommentator
    @Andrewthecommentator Месяц назад

    Subscribed

  • @ohmkaark
    @ohmkaark 4 месяца назад

    I was looking for a good summary around LLM evaluation metrics.. I see a lot of them captured here well

  • @brandiarakaki4071
    @brandiarakaki4071 8 месяцев назад

    💖 'Promo SM'

  • @ethanshub501
    @ethanshub501 9 месяцев назад

    Who is Diego M. Oppenheimer?

  • @rezgar482
    @rezgar482 9 месяцев назад

    Interesting

  • @bitcode_
    @bitcode_ 11 месяцев назад

    Awesome talk! So much cool info

  • @vincentkaranja7062
    @vincentkaranja7062 Год назад

    Fantastic presentation, Max and Rowan! The depth of your analysis and the clarity with which you presented the complexities of evaluating LLMs is truly commendable. It's evident that a lot of thought and effort went into this research. I'm particularly intrigued by your approach to using LLMs as evaluators. It opens up a plethora of possibilities but also brings forth some ethical considerations. How do you account for systemic biases in evaluation metrics when using LLMs as evaluators? Given that traditional metrics might not capture the fairness aspect adequately, have you considered incorporating fairness metrics or mitigation methods in your evaluation process?

  • @vincentkaranja7062
    @vincentkaranja7062 Год назад

    Excellent overview, Terry. The part about identifying age discrimination within machine learning models caught my attention. Could you share more about how Arthur.AI's platform sets the acceptable range for performance metrics in this context? Is it customizable based on industry or legal standards?

  • @benwilson1952
    @benwilson1952 Год назад

    Any way to boost the audio on this? Barely audible in some parts

  • @chrisogonas
    @chrisogonas Год назад

    That's a resourceful conversation, folks. Thanks for hosting.

  • @autumn993
    @autumn993 Год назад

    ✋ þrðmð§m