30.Access Data Lake Storage Gen2 or Blob Storage with an Azure service principal in Azure Databricks

Поделиться
HTML-код
  • Опубликовано: 14 окт 2024
  • In this Video, I discussed about accessing ADLS Gen2 or Blob Storage with an Azure Service Principal using OAuth.
    Code Used:
    spark.conf.set("fs.azure.account.auth.type.storage-account.dfs.core.windows.net", "OAuth")
    spark.conf.set("fs.azure.account.oauth.provider.type.storage-account.dfs.core.windows.net", "org.apache.hadoop.fs.azurebfs.oauth2.ClientCredsTokenProvider")
    spark.conf.set("fs.azure.account.oauth2.client.id.storage-account.dfs.core.windows.net", "application-id")
    spark.conf.set("fs.azure.account.oauth2.client.secret.storage-account.dfs.core.windows.net", service_credential)
    spark.conf.set("fs.azure.account.oauth2.client.endpoint.storage-account.dfs.core.windows.net", "login.microsof...")
    Link for Python Playlist:
    • Python Playlist
    Link for Azure Synapse Analytics Playlist:
    • 1. Introduction to Azu...
    Link for Azure Data bricks Play list:
    • 1. Introduction to Az...
    Link for Azure Functions Play list:
    • 1. Introduction to Azu...
    Link for Azure Basics Play list:
    • 1. What is Azure and C...
    Link for Azure Data factory Play list:
    • 1. Introduction to Azu...
    Link for Azure Data Factory Real time Scenarios
    • 1. Handle Error Rows i...
    Link for Azure Logic Apps playlist
    • 1. Introduction to Azu...
    #Azure #Databricks #AzureDatabricks

Комментарии • 28

  • @geethamca88
    @geethamca88 6 месяцев назад

    Really explained well
    Pls make some interview questions and answers series pls

  • @mallik_cmc3262
    @mallik_cmc3262 2 года назад +2

    content explanation is very nice .but one suggestion is that use proper naming conventions . that will give more understanding to new users.

  • @surajbasha9062
    @surajbasha9062 4 месяца назад +1

    Recent interview questions:
    1. If you are using unity catelog in your project then can we use service principals to connect adf to batabricks?
    Sir can you please explain in depth.

  • @surajbasha9062
    @surajbasha9062 4 месяца назад +1

    2. Can we use yarn as cluster manager or resource manager in spark in databricks?
    In real time?

  • @ravikumarkumashi7065
    @ravikumarkumashi7065 2 года назад

    Very crisp and clean thank you for this vedio

  • @gosmart_always
    @gosmart_always Год назад

    Thank you so much for your video. It was a much needed help.

  • @bhromonindia
    @bhromonindia Год назад +1

    love it, very well explained.

  • @joestopansky6375
    @joestopansky6375 2 года назад

    Thanks for the video; it is very informative. Using this method, do you need to execute the spark.conf.set() commands every time you restart the cluster? My guess is that you would since you are only affecting configs of this specific spark session.

    • @ravikumarkumashi7065
      @ravikumarkumashi7065 2 года назад

      Yes in real time these configurations will be part of your application code and once your cluster restarts,it kills your application due to driver unavailability and you need to start from beginning

  • @ramum4684
    @ramum4684 Год назад

    I want to know to what is the benefit of using this service principle I'd ,name and value under oauth function while we can access files from blobstorage using just secret scope itself directly. Is there any advantage of using this ?

  • @johnpaulprathipati153
    @johnpaulprathipati153 2 года назад

    Hi sir i have been following every video in this db playlist.
    could you tell me how many more videos can be there to complete this DB playlist??

  • @kanhashukla6265
    @kanhashukla6265 4 месяца назад

    Thanks. This helped a ton

  • @dineshdeshpande6197
    @dineshdeshpande6197 2 года назад

    Sir, how can we configure Azure data bricks Hive metastore to some external etl tool like informatica . The purpose is to fetch data from hive tables and use databricks engine for push down optimization to improve the performance of the data fetching.

  • @ishaangupta4941
    @ishaangupta4941 2 года назад

    Hi, One question . Dont you have to mount the file system again by using these Azure service principal's configuration ?
    I think you are able to read the Data coz your storage is already mounted by direct access keys ??

    • @ravikumarkumashi7065
      @ravikumarkumashi7065 2 года назад

      I think he is not using dbfs mount here and once your are authorized using service principal you can directly read from storage account but yes you can mount your adls to databricks file system once and it is set workspace level and then you can start reading from dbfs directly instead of adls

    • @alonsom.donayre1992
      @alonsom.donayre1992 Год назад

      @@ravikumarkumashi7065 Mounting is a deprecated pattern for storing and accessing data , it's not recommended anymore , using abfs driver is the best way right now! docs.databricks.com/external-data/azure-storage.html

  • @gosmart_always
    @gosmart_always 11 месяцев назад

    I followed your instructions. Still it is throwing error "Unsupported Azure scheme: abfss". May I know why and what the steps?

  • @bamidelejames375
    @bamidelejames375 2 года назад

    Hi, can you treat how to set up a shared external Hive metastore to be used across multiple Databricks workspaces, the purpose is to be able to reference dev workspace data in prod instance

  • @shivambansal3560
    @shivambansal3560 4 месяца назад

    We do not need to mount after setting the spark config ?

  • @sgr8280
    @sgr8280 Год назад

    connecting to data lake storage videos are confusing, why should we use service principal when we can access through Azure KeyVault directly?

  • @Amolinfo
    @Amolinfo 2 года назад

    Please make video on polybase and jdbc approach.

  • @ConCom665
    @ConCom665 7 месяцев назад

    What if the storage account has HNS disabled and I still want to use an SPN?

  • @BaBa_Ji-x8b
    @BaBa_Ji-x8b 2 года назад +2

    Thank a lot sir

  • @edwardr8826
    @edwardr8826 6 месяцев назад

    wafa ur the best

  • @sewastudies6835
    @sewastudies6835 2 года назад

    10 api access through adf in azure 6 success and 4 failed how to get only failed api?

  • @MBA_ANALYST
    @MBA_ANALYST Год назад

    ♥♥