Using Snowpipe | How to ingest data from AWS S3 | Snowflake Demo

Поделиться
HTML-код
  • Опубликовано: 18 ноя 2024

Комментарии • 39

  • @tao5198
    @tao5198 9 месяцев назад +1

    Thanks! It's so clear to explain steps with a sample and well documented.

  • @AaronWatkinsEKSELAN
    @AaronWatkinsEKSELAN 2 года назад +1

    Snowflake's documentation on this is quite good, but this video was critical for me to cover the gaps! Thank you!

  • @talhaiftikharchannel6153
    @talhaiftikharchannel6153 6 месяцев назад

    You have excellent Teaching attributes. Thanks for this.

  • @danielejiofor3493
    @danielejiofor3493 Год назад +1

    Best video out there on snowpipe

  • @kgolthi7787
    @kgolthi7787 2 года назад +1

    Very nicely explained. Amazing detail and clarity. Thank you!

  • @basilbrush7878
    @basilbrush7878 2 года назад +1

    Thanks for explaining this so clearly.

  • @pavannaikal
    @pavannaikal 2 года назад +1

    This is pretty good video. Can you please provide details on or share info if you have come across this scenario
    1. Once data is loaded into Snowflake stage, How to move the file from main S3 folder to an "archive" S3 folder. How to trigger that process after successful ingestion.
    2. If we have to do some custom validation (like file type, header etc..) of the files before ingestion into snowpipe. where can this validation be done.

    • @mastering_snowflake
      @mastering_snowflake  2 года назад +1

      1. You'll need to write a script yourself or use an ETL tool to do this. The COPY into statement can purge files once loaded, but it cannot move it to another folder.
      2. Use SQL scripts or your ETL tool to prepare and validate the files before you make it available for Snowflake to load.

    • @shashankm2859
      @shashankm2859 Год назад

      @@mastering_snowflake Regarding point 1, can that script be written in Java Script and called from Snowflake Procedure?

  • @marquinho1p
    @marquinho1p 2 года назад

    errr merr gersh! this was so helpful! Thank you sooo much! Mainly the bit for the sqs arn being in the snowpipe object and that one just needs to plug and play it - thank you! #10outof10 #wouldrecommend

  • @priyankamehra7146
    @priyankamehra7146 3 года назад +1

    Hi Adam.Have gone through your videos on snowflake.Good explanation.👍
    I have one query though.
    Can u help me understand if snowflake architecture would be good option to support oltp and olap systems and the reason being?

    • @dataarq945
      @dataarq945 3 года назад +2

      it is OLAP, SnowFlake is a Cloud Data Warehousing

    • @mastering_snowflake
      @mastering_snowflake  3 года назад

      Snowflake is primarily designed for large analytical queries, such as aggregations and calculations over very large data sets.
      They fare less well when users run, ‘point lookup’ queries, or in other words, queries which are highly selective, and therefore look up just a small number of records.
      If you have OLTP workloads I'd be exploring other technologies.

  • @אלכסנחשונוב-פ5ע
    @אלכסנחשונוב-פ5ע 2 года назад +1

    Hi Adam,
    I have a small question,
    I am trying to build a dwh pipeline, I got 2 json source files in on an s3 bucket, I created a pipe from s3 to my staging schema but I am not sure how do I continue from here, is there a way to create a pipline between two schemas\ tables or should I create a task + procedure to refresh every few minutes to load the data into prod?
    Thanks Alex

    • @mastering_snowflake
      @mastering_snowflake  Год назад +1

      Hi Alex, that should be all there is to it. If you have created the pipe with the auto ingest set to true it will periodically run as new files arrive in s3 and copy them into your target table.
      Hope this helps!

    • @אלכסנחשונוב-פ5ע
      @אלכסנחשונוב-פ5ע Год назад

      @@mastering_snowflake Thanks it works

  • @skateforlife3679
    @skateforlife3679 2 года назад +1

    Is it mandatory to use a storage integration ? Because my bucket is full public. When i copy the arn of my pipe into sqs arn, it tells me that is the wrong arn format ...

    • @mastering_snowflake
      @mastering_snowflake  2 года назад +1

      The storage integration isn’t mandatory, but it’s best practice within an enterprise scale environment.

    • @skateforlife3679
      @skateforlife3679 2 года назад

      @@mastering_snowflake All right thank you :)

  • @shashankm2859
    @shashankm2859 Год назад +1

    how can i configure a snowpipe to grab the same filename from an s3 bucket when the file is refreshed and re-uploaded?

    • @mastering_snowflake
      @mastering_snowflake  Год назад

      If the file contents is different it should recognise this and load it. You should be using a regex pattern to be picking up your files.
      If the contents are the same it won’t load it again by default within 64 days.

    • @shashankm2859
      @shashankm2859 Год назад

      @@mastering_snowflake Thank you, Adam for responding. Do you have some sample code for it?

    • @mastering_snowflake
      @mastering_snowflake  Год назад

      We have lots of code templates and examples available as part of the mastering snowflake program. If you’re interested in finding out more visit www.masteringsnowflake.com

  • @liberatingtruthseeking5513
    @liberatingtruthseeking5513 10 месяцев назад

    how to bring in incremental data ?

  • @pawanshrivastava1876
    @pawanshrivastava1876 2 года назад +1

    Can the text file from S3 be moved to snowflake?

    • @bouchard71
      @bouchard71 2 года назад +1

      Of course!

    • @mastering_snowflake
      @mastering_snowflake  Год назад

      The purpose of this is to load the data from your text file into a table in Snowflake. If your talking about the actually file and not the data then you’d leave the file in place and query it as an external table.
      Check out my video on this channel about how to do this

  • @mosa36
    @mosa36 Год назад

    How does it work if the data I want to ingest into Snowflake is on a RDS MySQL database?

    • @mastering_snowflake
      @mastering_snowflake  Год назад +1

      www.propeldata.com/blog/how-to-get-your-data-from-an-aws-rds-database-into-snowflake

    • @mosa36
      @mosa36 Год назад +1

      @@mastering_snowflake yes I found this article, it was really helpful. I was just wondering if there were a way to use directly snowpipe but it seems that I first need to bring the data to a stage ?

    • @mastering_snowflake
      @mastering_snowflake  Год назад

      Yes that’s right

  • @roblangdon404
    @roblangdon404 Год назад

    I have this in mind but I'm skeptical. I'm using Postman to pull data coming from an application API. I'd like to then push that data to individual Snowflake tables. Not sure if I need to Snowpipe or I can get away with just using the initial Postman to Snowflake (haven't even tested this yet). Using this template (in the video), I'd have to push the data from Postman to Snowflake and then grab it with Snowpipe. @mastering_snowflake what do you think?

  • @deepthikotipalli
    @deepthikotipalli Год назад

    sir,how can we know the name of the snowpipe while loading data

    • @mastering_snowflake
      @mastering_snowflake  Год назад

      Sure you name the pipe when you create it. Check this article out: docs.snowflake.com/en/sql-reference/sql/create-pipe

  • @sahillohiya7658
    @sahillohiya7658 7 месяцев назад

    helpful !!

  • @markcuello5
    @markcuello5 Год назад +1

    HELP

  • @dataarq945
    @dataarq945 3 года назад

    #SnowPipe #AWSS3