Delta Live Tables Demo: Modern software engineering for ETL processing

Поделиться
HTML-код
  • Опубликовано: 5 сен 2024

Комментарии • 18

  • @marcinsiara
    @marcinsiara Год назад +2

    Great video, easy to understand, good overview for the Databricks beginners. Thanks!

  • @amansehgal9917
    @amansehgal9917 2 года назад +10

    This is great. Can you share notebook for querying the transaction log and presenting it in reDash?

  • @simonhu5814
    @simonhu5814 7 месяцев назад

    Well explained. Thank you

  • @jolettin6408
    @jolettin6408 Год назад +1

    Do you have a link to show how the queries work for monitoring the data quality. Thanks

  • @tanushreenagar3116
    @tanushreenagar3116 Год назад

    GREAT VIDEO

  • @ozland7172
    @ozland7172 2 месяца назад

    Thank you for the explanation! Can please someone confirm that the files can only be read from FileStore and not blob storage by this approach (SQL)? Tried reading it from blob storage and cannot access using cloud_files() function. I guess it requires pyspark to access blob storage and create delta live tables.

  • @willf7493
    @willf7493 Год назад +1

    Nice demo, but why does the comment for the "cleaned_station_status" table say "partitioned by station_id" when the code actually uses the last_updated_date column? You should update the comment in that notebook. :-)

  • @joegenshlea6827
    @joegenshlea6827 Год назад

    Thank you for this video. I'm a little confused about what the "data.stations' refers to? Is it an array in the source json?

  • @mohammedsafiahmed1639
    @mohammedsafiahmed1639 Год назад +2

    am I missing something or does the video really doesnt show how he got all those files in the data lake in the first place?

    • @samgreene7961
      @samgreene7961 9 месяцев назад

      He mentions the python scripts/notebooks that get the data. Probably using an api and saving results to DBFS. I’m sure you can find how to do that in other videos.

    • @azazmir9340
      @azazmir9340 4 месяца назад

      Hes using autoloader to load the data probably from an s3 bucket, Azure cloud storage or volume

  • @nontapatsumalnop4740
    @nontapatsumalnop4740 Год назад +1

    Anyone knows how to create a live dashboard like this in databricks?

  • @gekodragon100
    @gekodragon100 Год назад

    Good video thank you. Quick question, is the DLT lineage also auotmatically avaialable and visisble in Unity Catalog?

  • @sid0000009
    @sid0000009 Год назад

    Can an API hosted on an App service in anyway fetch Delta live tables data ? thanks

  • @ibozhu
    @ibozhu 2 года назад +2

    It’s been in gated preview for too long, when will it be made GA?