Azure Data Factory V2 - Copying On-Premise SQL Server Data to Azure Data Lake

Поделиться
HTML-код
  • Опубликовано: 5 сен 2024

Комментарии • 83

  • @tarvinder91
    @tarvinder91 4 года назад +1

    Very nice tutorial. Thanks for this. However two things are not clear..1. How does data factory link itself to app registration. 2. How exactly it filtered data for current date . And if there was a way to get the data for 2 days.? Thanks

    • @DineshPriyankara
      @DineshPriyankara  4 года назад +1

      Hi Singh, thanks for the comment.
      1. App Registration is used as a Service Principal, not only Data Factory, any service can use this Service Principal for accessing the data lake. I have added some details on it with my one of Databricks videos - Databricks - Accessing Data Lake using a Service Principal. You may get some additional info from it.
      2. Since we select the filter column as ModifiedDate column, records will be filtered as per current date, it is something similar to "WHERE ModifiedDate = GetDate()" in TSQL. If you want to change the filters as you want, then you need to go in to the pipeline and make changes. Watch my other Data Factory videos, I discuss setting things like that inside the pipeline.

  • @akshatsharmasre
    @akshatsharmasre 2 года назад

    This explanation is worth subscribing to your channel.
    Thanks a lot for this video.

  • @venkatx5
    @venkatx5 4 года назад +1

    The ModifiedDate filter value is not passed anywhere. Is it taking current System Date to filter?

  • @ambarmanu
    @ambarmanu 4 года назад +1

    I loved this one. Thank you Dinesh for taking efforts to explain this in so detailed manner.

  • @vijayashirurmuralidhar3007
    @vijayashirurmuralidhar3007 5 лет назад +2

    Excellent!! one of the best detailed tutorial for people with Non-cloud Background. Thank you!!

  • @darta1094
    @darta1094 6 лет назад +2

    Great content. Watched all three sessions. Clear-to the point-very useful. Thank you, Dinesh.

    • @DineshPriyankara
      @DineshPriyankara  6 лет назад

      Thanks Dimitri.

    • @gurubhaskarmulinti1564
      @gurubhaskarmulinti1564 6 лет назад

      Yes, its great content .. made it so simple.. I want to watch other 2 videos, can you please share other 2 as well

    • @DineshPriyankara
      @DineshPriyankara  6 лет назад

      Thanks Guru, working on few more videos, will publish them soon :)

  • @alpeshjain532
    @alpeshjain532 5 лет назад

    Thank you Dinesh. It is nicely illustrated. Please keep posting more videos on azure. I am beginner and it boost my confidence by learning from your videos.

  • @kingmaker2117
    @kingmaker2117 6 лет назад +2

    Excellent explanation Dinesh, tried a couple of videos prior to this, not much use even with Microsoft ignite, but your explanation on the step to step is awesome. keep doing the same work in other topics.

  • @vamsi.reddy1100
    @vamsi.reddy1100 Год назад

    Your voice and pronunciation is like kud venkat ...,

  • @aftabhussain5833
    @aftabhussain5833 5 лет назад +1

    This guy is awesome, I'm impressed the way he explained.

  • @GrowthMindsetGlobal
    @GrowthMindsetGlobal 2 года назад

    Very well explained

  • @biswajitsarkar5538
    @biswajitsarkar5538 5 лет назад +1

    You are awesome....Thanks a lot...carry on the good work.

  • @askuala
    @askuala 5 лет назад +2

    Great works! I really learned a lot from your videos. Keep posting, great job!

  • @Saket_Srivastava
    @Saket_Srivastava 3 года назад

    explained really well

  • @MatthewFoos
    @MatthewFoos 5 лет назад +1

    Very well done Dinesh! Thanks for the tutorial brother.
    PS - Loving that piano intro :)

  • @suhel.choudhury
    @suhel.choudhury 6 лет назад +1

    Dinesh thankyou so much this is a fantastic video. I am just starting a new project which will be doing exactly what you explained in this video. Very well presented and explained. I have subscribed to your channel, looking forward to viewing more of your videos.

  • @ananthasubramanian7355
    @ananthasubramanian7355 3 года назад

    Thank you for this video. Neat and clear. Very helpful.

  • @kiranreddyallam
    @kiranreddyallam 5 лет назад +1

    Nice explanation.. keep post good work.. if possible, plz make a video on ADF pipelines deployment

  • @nayagadiatoka
    @nayagadiatoka 5 лет назад

    Hello Sir Thanks for uploading such wonderful videos.Is is possible for you to create videos on below activities in details like you have explained for lookup or for copy data with scenarios.
    1. DataBricks
    i) Notebook
    ii) Jar
    iii) Ptython
    2. General
    ii) Append variable
    iii) Execute Pipeline
    iv) Execute SSIS Package
    v) Get Metadata
    vi) Set variable
    vii) Web-For sending mail
    viii) Wait
    3. HD Insight
    i) Hive
    ii) Map reduce
    iii) Pig
    iv) Spark
    v) Streaming
    5. Iteration and Conditional
    i) Filter
    iii) Until
    iv) If condition
    6. Machine Learning
    i) ML Batch Execution
    ii) ML Update resource

  • @demudunaidu3657
    @demudunaidu3657 5 лет назад +1

    Excellent explanation! very useful!

  • @birukbobo7398
    @birukbobo7398 4 года назад

    Thanks Dinesh for your awesome presentation. Can you please posting some additional demo videos that relate ADB migration from SQL Server to ADB by using SQL language. In addition how to run SQL
    SP logic quires run in ADB by using spark SQL in notebook.

  • @ksathish11
    @ksathish11 5 лет назад +2

    Thank you so much Bro, very well explained.

  • @RajivGuptaEverydayLearning
    @RajivGuptaEverydayLearning 3 года назад

    How you validate the data you just transferred using Azure Data Factory once you import same into say Azure sql database ?

  • @SQLTalk
    @SQLTalk 4 года назад

    Very well done video. Easy to understand. Thank you very much.

  • @SonuKumar-xy1mp
    @SonuKumar-xy1mp 3 года назад

    Hi, I couldn't find things, could you please help me to find "Data Lake Store", app registration "Settings" to create "access key" , "keys", set "Access" for datalake store, "Your Permissions"?

  • @pratikfutane8131
    @pratikfutane8131 5 лет назад +1

    Thanks for the great explanation!! Helpful

  • @kamlakarpawar6671
    @kamlakarpawar6671 2 года назад

    How to set up recurring migration through queries/script from SQL server(On-premise) to Azure SQL database
    I need a help to sync the data on Azure SQL from SQL server(On-premise).
    Available recourses:
    2 Database (SQL Server) on premise available on different server
    Azure SQL database on cloud
    Migration script/queries are ready to fetch data from on-premise sql server
    Requirements:
    Set up a scheduler which will run every 12 hours means two times in a day on Azure SQL.
    In scheduler, using migration scripts data will be fetch from the On-premise SQL server and insert into Azure SQL Database.

  • @abramswee
    @abramswee 4 года назад +1

    good tutorial video.

  • @sanwrites718
    @sanwrites718 5 лет назад

    How was today (24th Jan) filtered? No date was given nor was any 'todays date' provided . Only the column name was provided in filter. Just by providing a column name , how are the rows filtered? What if I only want 'yesterdays'. And nice explanation, thanks

  • @vishal259
    @vishal259 4 года назад

    GUI looks very different now, I could not write expression to parse my destination. Destination Options mentioned in this video do not exist.

  • @sivaprasad6393
    @sivaprasad6393 6 лет назад +1

    Awesome Dinesh..thank you.

  • @terryliu3635
    @terryliu3635 5 лет назад +1

    Great work! Thanks!

  • @tv90channel11
    @tv90channel11 5 лет назад +2

    Exllecnt , can you please give you a index for your videos to continity , so that it helps for to follow step by step

  • @09kashifshaikh
    @09kashifshaikh 4 года назад

    Hi, thank you for the video. I had a question to ask with respect to data transfer from on-premise SQL on to Azure DW. In this problem statement, we have an on-premise SQL server table that is changed by a product interface so no way to event it from the app code, so the question I wanted to put forth is that, is there any out of the box Azure service that can help us track changes of the on-premise table and inform a service to migrate the change on to the Azure DW?
    In a traditional model, one can use SQL dependency to get notified on a .Net Client of any changes on the source table which in turn transfers the change to a destination table.

  • @mhshbht12
    @mhshbht12 5 лет назад

    superb explanation.Thanks

  • @shallychoudhary4969
    @shallychoudhary4969 6 лет назад +1

    Nicely explained....very useful

  • @mr.prasadyadav
    @mr.prasadyadav 2 года назад

    Dinseh ,Will I get complete Videos of ADF teach by You ?

  • @ajaykulkarni9297
    @ajaykulkarni9297 6 лет назад

    how to manage the service principal key ? do we need to keep it with us all the time ? can it be stored in a .ini or XML file and referred dynamically in the ADF ?

  • @dinesh1kumaar
    @dinesh1kumaar 4 года назад

    Hi Dinesh,is it possible to move data from on prem sql to az storage account directly instead of az dw

  • @snehataori3388
    @snehataori3388 4 года назад

    Hello Dinesh,
    I am new to Azure Data Factory. I see all your videos related to ADF V2 and all are very knowledgeable .I have one requirement so can you help me on hoe can i achieve that.
    I want to read the .xlsx file and load into Azure DB/.csv file.
    What should i do please suggest.
    Thanks in advance .Waiting for your response .

  • @satya.b1b955
    @satya.b1b955 6 лет назад

    Hi Dinesh,
    Thanks for explaining so clearly.can you also help us understand how data bricks works in transforming data that would be very helpful.
    Thanks again!!

    • @DineshPriyankara
      @DineshPriyankara  6 лет назад

      Hi Satya,
      I am working on a video related to data bricks, will update you once it is published.

    • @satya.b1b955
      @satya.b1b955 6 лет назад

      Dinesh Priyankara
      Thanks Dinesh !!

  • @tarunjain5246
    @tarunjain5246 5 лет назад

    Will Integration pipeline help to access other Cloud Network like AWS/GCP?

  • @jaiprakashchauhan5725
    @jaiprakashchauhan5725 6 лет назад

    Well done. Keep it up!

  • @kammarabheemanna4544
    @kammarabheemanna4544 5 лет назад

    please add more vedious abt ADF with diffrent senarious

  • @clokeshreddy
    @clokeshreddy 6 лет назад

    Good info Dinesh. Is there a way the destination data will be stored directly in Database tables in Azure Data Lake?or we need to store them as files initially and write the U-SQL jobs to move from files to catalog in data lake?

    • @DineshPriyankara
      @DineshPriyankara  6 лет назад

      Hi Lokesh,
      As far as I know, you cannot directly access ADLA Database tables from outside. However, you can access external sources like Azure SQL DB, Azure SQL DW or Azure SQL VM using U-SQL (As External Tables). Considering that fact, if you source is one of those three, data can be taken directly to ADLA and insert them to ADLA tables. I have not done it but I am sure that it is possible. However, if your source is not one of those (like On-Premise SQL DB), then no other option but get them into ADLS as files and then send them to ADLA tables.

    • @clokeshreddy
      @clokeshreddy 6 лет назад

      Thankyou dinesh. i am trying to do incremental load of on-premise database to Data-Lake and i initially thought to move database using ADF to files in ADLS and then using U-SQL to move it to database tables in ADLS.
      But As per your above statement do you think moving from on-premise to Azure SQL DB and then to Data Lake Store tables is the best option?

    • @DineshPriyankara
      @DineshPriyankara  6 лет назад

      Hi Lokesh,
      If you data is not in a cloud-database, then I do not think that local->AzureSQLDB->ADLADB is the way. My preference is local->ADLS->ADLADB because you do not need to worry about creating a new AzureSQLDB and maintaining it. Anyway, the architecture should be decided based on many factors and requirements; like whether you need another AzureSQLDB, you need to have data in ADLADB or having data in file format is enoght, etc.

    • @clokeshreddy
      @clokeshreddy 6 лет назад +1

      Thanks for the info dinesh.will see which option best suits to my need and proceed accordingly.

  • @ramkumarthangavel8061
    @ramkumarthangavel8061 5 лет назад

    Excellent

  • @BrunoSantos-xz5hp
    @BrunoSantos-xz5hp 6 лет назад +1

    Congratulations for the excellent video, do you have any blog that I can follow?

    • @DineshPriyankara
      @DineshPriyankara  6 лет назад

      Thanks Bruno, this is my blog dinesql.blogspot.com/ but could not write anything for last 2-3 months :(

  • @manasmohanty5754
    @manasmohanty5754 4 года назад

    Sir.. can you pls show a demo project ..

  • @sparshbilawat5808
    @sparshbilawat5808 6 лет назад

    Can somebody help me solving the error am getting oin concatinating the two columns i.e first name and last name and choosing first letter of the column gender that is M or F for male and female respectively?
    {
    "name": "CopyPipeline2",
    "properties": {
    "activities": [
    {
    "name": "Copy_pzt",
    "type": "Copy",
    "policy": {
    "timeout": "7.00:00:00",
    "retry": 0,
    "retryIntervalInSeconds": 30,
    "secureOutput": false
    },
    "userProperties": [
    {
    "name": "Source",
    "value": "secondsamplecontainer/SampleForAzure1.txt"
    },
    {
    "name": "Destination",
    "value": "[dbo].[Contact]"
    }
    ],
    "typeProperties": {
    "source": {
    "type": "BlobSource",
    "recursive": true
    },
    "sink": {
    "type": "SqlSink",
    "writeBatchSize": 10000
    },
    "enableStaging": false,
    "dataIntegrationUnits": 0,
    "translator": {
    "type": "TabularTranslator",
    "columnMappings": {
    "PersonID": "PersonID",
    $concat("First Name","Last Name"): "Name",
    "Mobile": "Mobile",
    "Email": "Email",
    $substring("Gender",1,1): "Gender",
    "Address": "Address"
    }
    }
    },
    "inputs": [
    {
    "referenceName": "SourceDataset_pzt",
    "type": "DatasetReference"
    }
    ],
    "outputs": [
    {
    "referenceName": "DestinationDataset_pzt",
    "type": "DatasetReference"
    }
    ]
    }
    ]
    },
    "type": "Microsoft.DataFactory/factories/pipelines"
    }
    Plus am getting error wrong json type for name and gender column.

  • @muhammadfayyaz7134
    @muhammadfayyaz7134 5 лет назад

    Great video sir can u please explain how to install integration runtime on linux machine ? I am trying to move data from oracle in linux to azure data storage.

    • @PPFigueiredo
      @PPFigueiredo 4 года назад

      Hello. It only work in a windows machine. You can use a Oracle in Linux as a source database, but the integration runtime must be installed in a windows machine.

  • @joejoe570
    @joejoe570 5 лет назад

    good stuff

  • @phanichaitanya8829
    @phanichaitanya8829 3 года назад

    Those who disliked I think they are from non IT background....

  • @RavikiranS
    @RavikiranS 6 лет назад

    Good video. Can you please tell how the sign-on URL is used? You gave dinesqltest02.com, but can it be anything? If anything, what is its role?

    • @DineshPriyankara
      @DineshPriyankara  6 лет назад

      Hi Ravikiran,
      Sorry for not replying immediately. In this context, you do not need to worry about that URL because what we try to do is service-to-service authentication (DataLake to Data Factory). This is useful when you need to get the authentication done for other parties and token has to be sent for them. Since it is a mandatory field, it should be filled with something in this case.

    • @RavikiranS
      @RavikiranS 6 лет назад

      Thank You, You used self hosted integration run time to pull on-prem sql data. Can we use Azure SSIS IR connected to Point-to-site VPN to pull on-prem data? I have seen that you can do that with Site-to-Site and couldn't find references to do point-to-site

  • @SA-mh7uy
    @SA-mh7uy 4 года назад

    Please work on pronouncing

  • @texasnadars
    @texasnadars 6 лет назад

    Not sure how the filter on 'modified date' is configured to use the current date? It is not explained. What if I want to run this for a specific date?

    • @DineshPriyankara
      @DineshPriyankara  6 лет назад

      Hi Texasanders, sorry for not replying immediately, you do not need to do anything specific as we select "ModifiedDate" column as the filter for "Datetime filter", means it gets filtered for the current date, see this timeline: ruclips.net/video/KrB2JcTu21w/видео.html&lc=z23gg5vwbo3ygdbyo04t1aokg0fcwnouvj4cqjn1xx31bk0h00410

    • @rohitsethi5696
      @rohitsethi5696 6 лет назад

      hi dinesh i it not clear but i want the data of 21th of modified date. if you have selected the columns not used where clause pls guide me 9971364908

    • @DineshPriyankara
      @DineshPriyankara  6 лет назад

      Hi Rohit, if you do not need to get data selected based on current date and all you need is to pass a specific date for filtering, then 1) you need to maintain the date to be loaded somewhere (like Highwatermark table), 2) load the date using ADF Lookup and 3) use it for filtering with ADF Copy. I am making a video on that as well but will take some time for publishing it. Hope I have answered your question.

    • @vinpal2507
      @vinpal2507 6 лет назад

      Dinesh Priyankara Is this video(using ADF lookup) done yet? Thanks- Vineet