End-to-End Data Pipeline Project with Change Data Capture Using Postgres, Airflow, and Snowflake!

Поделиться
HTML-код
  • Опубликовано: 22 ноя 2024

Комментарии • 11

  • @KAshIf0o7
    @KAshIf0o7 9 месяцев назад +1

    Thanks George and Manmeet for this. Very well explained 🙌

    • @thedataguygeorge
      @thedataguygeorge  9 месяцев назад +1

      Thank you, will pass the message along to Manmeet!

  • @rasmusandreasson1548
    @rasmusandreasson1548 9 месяцев назад +3

    Niccce! BIG thumbs up for some EndToEnd-projects!

  • @ManishJindalmanisism
    @ManishJindalmanisism 9 месяцев назад +1

    Hey.. Thanks for putting this. Lot of action. is this committed on github? This will be a learning project for people who are starters in airflow.
    Additionally I have one question : Why have we hardcoded customer in get_dag_conf, should that not be dynamic basis your list of 3 tables?
    Also in the tutorial, I think you didnt show customer.sql which you are using in extract task, and modifying to add where clause. Can you show how it looks like?

    • @thedataguygeorge
      @thedataguygeorge  9 месяцев назад +2

      It sure is! github.com/manmeetkaur/etl-airflow/tree/main
      Can see the customer.sql file there, and in practice you wouldn't want to hardcode, I just did that for ease of use for this example

    • @ManishJindalmanisism
      @ManishJindalmanisism 9 месяцев назад

      @@thedataguygeorge Thanks for replying. Very kind of you

  • @StephenRayner
    @StephenRayner 4 месяца назад

    Why snowflake for OLAP? Postgres, airflow, dbt are all open source (and cosmos). But snowflake isn't. I don't get why everyone uses it. Why not an open source alternative? What about Clickhouse?

    • @thedataguygeorge
      @thedataguygeorge  4 месяца назад

      Snowflake requires almost 0 set up or management expertise which is a big value add for a lot of people