Data World Solution
Data World Solution
  • Видео 154
  • Просмотров 23 542
Mastering ForEach Loops in Microsoft Fabric: Automate Repetitive Tasks in Data Pipelines!
In Microsoft Fabric, the ForEach activity allows looping over a collection (like a list of items) in a Data Factory pipeline. You can use it to perform actions on each item in the collection. For example, after using the Get Metadata activity to retrieve a list of files in a folder, ForEach can iterate through each file. Within each iteration, you can add activities like setting a variable to store each file's name. This loop structure executes the activities for each item in the list, enabling flexible and repeated processing within a pipeline.
---------------------
#MicrosoftFabric #DataFactory #ForEachActivity #DataPipelines #Automation #Azure #DataEngineering #CloudComputing #TechTutori...
Просмотров: 123

Видео

Mastering Pipeline Variables: Dynamic Data Handling in Your Workflows in ADF/Microsoft Fabric
Просмотров 5321 день назад
Pipeline variables are modifiable values within a pipeline run, unlike pipeline parameters that remain constant. You can set these variables using the Set variable activity. #DataPipeline #Variables #DataEngineering #WorkflowAutomation #TechTips #Azure #DataManagement #CloudComputing #DataIntegration #ETL #TechTutorials #DynamicData #microsoftfabric #dataworldsolution
How to uses Parameters in ADF/Microsoft Fabric for allow passing external values into pipelines!
Просмотров 6521 день назад
Parameters in Data Factory allow passing external values into pipelines, making resources reusable with different values. In a pipeline, a parameter can dynamically specify values During pipeline execution, can be easily added, edited, or deleted in the pipeline settings. #AzureDataFactory #DataPipelines #Parameterization #DynamicData #DataEngineering #DataIntegration #CloudData #microsoftfabri...
How to Dynamic Content and Functions in Microsoft Fabric | ADF | Flexible Data Pipelines
Просмотров 6228 дней назад
The demo shows how to use dynamic content in Azure Data Factory by configuring a pipeline to copy data. It employs system variables date functions (utcnow(), formatDateTime()), and the concat() function to customize file paths and names. Each pipeline run generates a unique file, demonstrating the flexibility of dynamic content with expressions and functions. DATA SOURCE - s3.amazonaws.com/trip...
How to use the Lookup Activity in Microsoft Fabric Data Factory: A Step-by-Step Guide
Просмотров 8728 дней назад
This lecture introduces the Lookup activity in Data Factory, which reads and returns content from a file, table, query, or stored procedure, outputting a single value or an array for use in subsequent activities. A pipeline named pl_lookup is created to demonstrate this. First, the activity is configured to return only the first row of data in JSON format. Then, all rows are returned as an arra...
Understanding the Get Metadata Activity in Microsoft Fabric Data Factory
Просмотров 8628 дней назад
The Get Metadata activity in Azure Data Factory retrieves metadata like column count, structure, and existence status for various data sources. When configured for a table, it returns details like the number of columns and schema. If set for a file or folder path, it provides additional attributes, including item name, type, size, and child items in folders. The output is displayed in JSON form...
Changing Column Data Types in a Spark DataFrame Using the Cast Method
Просмотров 30Месяц назад
This lecture demonstrates how to change column data types in a Spark DataFrame using the cast() method, converting columns between types like string, integer, and double, with references to Spark's data types API and documentation. #SparkDataFrame #DataTypes #CastMethod #BigData #DataEngineering #PySpark #DataTransformation #ProgrammingTips #DataScience #ETL #dataworldsolution
How to Use Copy Data Activity to Move and Convert Files in a Lakehouse
Просмотров 55Месяц назад
In this lecture, the instructor demonstrates how to use the Copy Data activity in a data pipeline to move files between locations within a lakehouse. The process involves selecting a source file, configuring the file format (such as delimited text or parquet), and specifying a destination path. The lecture covers file format settings, validation, and running the pipeline. Additionally, it expla...
Mirroring Data from Snowflake to Microsoft Fabric: A Step-by-Step Tutorial
Просмотров 120Месяц назад
Learn how to seamlessly mirror data from Snowflake to Microsoft Fabric using Fabric's low-cost, low-latency mirroring feature. This tutorial covers setting up connections, replicating tables, and querying mirrored data in Microsoft Fabric's OneLake environment. #MicrosoftFabric #Snowflake #DataMirroring #OneLake #CloudData #DataReplication #AzureSQL #CosmosDB #DataWarehouse #TechTutorial #dataw...
Adding New Columns to DataFrames in Databricks Using PySpark
Просмотров 26Месяц назад
In this tutorial, learn how to add new columns to DataFrames in Databricks using PySpark's withColumn method. You'll explore adding columns with literal values, functions like current_date, and performing arithmetic operations to create new columns such as population in millions. Perfect for enhancing your data manipulation skills in Databricks! #Databricks #PySpark #DataFrames #BigData #DataSc...
How to Select Required column and Rename Columns in Databricks Using Pyspark
Просмотров 33Месяц назад
In this tutorial, learn how to efficiently select and rename columns in Databricks using the withColumnRenamed function. Perfect for beginners looking to enhance their data transformation skills in Spark. #Databricks #withColumnRenamed #SparkSQL #DataTransformation #BigData #DataEngineering #ApacheSpark #DataScience #AzureDatabricks #ETL #DataFrame #DataProcessing #Python #DataTutorial #TechTut...
Reading CSV Files into DataFrames with Custom Schemas Using Spark in Azure Databricks
Просмотров 58Месяц назад
Learn how to read files into DataFrames and define custom schemas using Spark in Azure Databricks. This tutorial covers step-by-step instructions to efficiently manage and process data. #AzureDatabricks #ApacheSpark #DataFrames #SchemaDefinition #BigData #DataEngineering #DataScience #CloudComputing #SparkTutorial #Azure #dataworldsolution #databricks video code file - dws4dl4gen2.blob.core.win...
How to Mount Azure Data Lake Storage Gen2 to Databricks Using a Service Principal |Azure |Databricks
Просмотров 102Месяц назад
Learn how to securely mount Azure Data Lake Storage Gen2 to Databricks using a Service Principal, configure Spark settings, and explore key file system utilities like listing and unmounting storage. URL :- learn.microsoft.com/en-us/azure/databricks/dbfs/mounts #AzureDataLake #Databricks #ServicePrincipal #DataStorage #BigData #CloudStorage #DataEngineering #SparkConfig #AzureKeyVault #DataManag...
Accessing Azure Data Lake Storage Using Service Principal in Databricks: A Step-by-Step Guide
Просмотров 144Месяц назад
Learn how to securely access Azure Data Lake Storage from Databricks using an Azure Service Principal. This guide covers registering the Service Principal in Azure Active Directory, configuring Spark settings in Databricks, and assigning appropriate permissions for seamless data access. #AzureDataLake #Databricks #ServicePrincipal #AzureAD #DataAccess #RBAC #BigData #CloudStorage #DataSecurity ...
How to Load Multiple Files into Snowflake Using the Wildcard Method | ADF | Snowflake
Просмотров 181Месяц назад
In this tutorial, we'll guide you through the process of loading data from Azure Blob Storage to Snowflake from scratch. Learn how to leverage the wildcard method to efficiently load multiple files into Snowflake. Perfect for anyone looking to streamline their data integration between Azure and Snowflake. #SnowflakeData #AzureBlobStorage #DataIntegration #WildcardMethod #SnowflakeTutorial #Azur...
Securely Accessing Azure Data Lake Using Databricks Secrets Utility | Read data from blob storage
Просмотров 101Месяц назад
Securely Accessing Azure Data Lake Using Databricks Secrets Utility | Read data from blob storage
Creating and Linking Databricks Secret Scope with Azure Key Vault
Просмотров 2592 месяца назад
Creating and Linking Databricks Secret Scope with Azure Key Vault
Creating and Managing Secrets in Azure Key Vault: Step-by-Step Guide | Azure | Databricks
Просмотров 1212 месяца назад
Creating and Managing Secrets in Azure Key Vault: Step-by-Step Guide | Azure | Databricks
Creating and Managing Azure Data Lake Storage for Your Databricks Projects!
Просмотров 892 месяца назад
Creating and Managing Azure Data Lake Storage for Your Databricks Projects!
5| How to use Databricks Utilities | dbutils | Simplifying File Operations | Workflow Automation
Просмотров 472 месяца назад
5| How to use Databricks Utilities | dbutils | Simplifying File Operations | Workflow Automation
4 | Mastering Databricks Magic Commands: Switch Languages & Boost Productivity
Просмотров 472 месяца назад
4 | Mastering Databricks Magic Commands: Switch Languages & Boost Productivity
3| Mastering Databricks Notebooks: Development, Collaboration, and Automation Essentials
Просмотров 642 месяца назад
3| Mastering Databricks Notebooks: Development, Collaboration, and Automation Essentials
2|Setting Up a Databricks Compute Cluster: A Step-by-Step Guide
Просмотров 652 месяца назад
2|Setting Up a Databricks Compute Cluster: A Step-by-Step Guide
How to Retrieve the Last Query ID in Snowflake Using SQLID
Просмотров 872 месяца назад
How to Retrieve the Last Query ID in Snowflake Using SQLID
Using SQLROWCOUNT, SQLFOUND, and SQLNOTFOUND in Snowflake Scripting to Track DML Statement Effects
Просмотров 422 месяца назад
Using SQLROWCOUNT, SQLFOUND, and SQLNOTFOUND in Snowflake Scripting to Track DML Statement Effects
How to Handle Exceptions in Snowflake Scripting: A Step-by-Step Guide
Просмотров 842 месяца назад
How to Handle Exceptions in Snowflake Scripting: A Step-by-Step Guide
1| Azure Databricks Setup: How to Create a Databricks Workspace
Просмотров 1312 месяца назад
1| Azure Databricks Setup: How to Create a Databricks Workspace
Efficient Data Handling with Resultsets in Snowflake Scripting | Snowflake SQL
Просмотров 512 месяца назад
Efficient Data Handling with Resultsets in Snowflake Scripting | Snowflake SQL
How to Use Cursors in Snowflake: Step-by-Step Tutorial
Просмотров 1452 месяца назад
How to Use Cursors in Snowflake: Step-by-Step Tutorial
How to Create and Attach Event Triggers to Pipelines | Type of Trigger |Azure Data Factory Tutorial
Просмотров 1333 месяца назад
How to Create and Attach Event Triggers to Pipelines | Type of Trigger |Azure Data Factory Tutorial

Комментарии

  • @NitinKumar_up_se
    @NitinKumar_up_se День назад

    Fantastic Sir❤ Please update file link for good practices.

  • @nomansadiq3441
    @nomansadiq3441 6 дней назад

    Is it possible to do the integration without the Tenant ID when creating "STORAGE INTEGRATION" ? I am using student Azure account provided by the university and don't have an access to tenant ID, is there any alternative for that?

    • @dataworldsolution
      @dataworldsolution 5 дней назад

      Hi bro, I will look into this and get back to you.

  • @Ram.MovieTalks
    @Ram.MovieTalks 20 дней назад

    Pls keep in sequential order so that it will be easy for viewer to see the videos step by step

  • @KrishnaReddy-zr4no
    @KrishnaReddy-zr4no 23 дня назад

    Good

  • @RajeevKumar_SQH
    @RajeevKumar_SQH Месяц назад

    Super bro..

  • @skpin13
    @skpin13 Месяц назад

    Hi @dataworldsolution, can you provide the dataset you have used in this video?

  • @skpin13
    @skpin13 Месяц назад

    @dataworldsolution, could you please make a video on how create metadata catalog in azure databricks ?

  • @RajeevKumar_SQH
    @RajeevKumar_SQH 2 месяца назад

    Ok sir

  • @vshannkarind
    @vshannkarind 2 месяца назад

    Please explain use of Databricks

  • @mohammedvahid5099
    @mohammedvahid5099 3 месяца назад

    Amaxing ❤bro

  • @rajeevjosephk
    @rajeevjosephk 3 месяца назад

    how to conver 0,00 as 0.00

    • @dataworldsolution
      @dataworldsolution Месяц назад

      Hi bro, try with replace function to replace comma with decimal

  • @rajeevjosephk
    @rajeevjosephk 3 месяца назад

    how to conveert 0,00 as 0.00 ??

  • @mohammedvahid5099
    @mohammedvahid5099 3 месяца назад

    Ur awesome bro and explaining is very well thnk u❤

  • @RajeevKumar_SQH
    @RajeevKumar_SQH 4 месяца назад

    This is helpful.. Thank you 😊

  • @suryasurendrareddy
    @suryasurendrareddy 4 месяца назад

    How to implement scd type2 in multiple tables

  • @Prashanthi248
    @Prashanthi248 4 месяца назад

    Good projection

  • @NomanKhan80USA
    @NomanKhan80USA 4 месяца назад

    Hi , Please can you upload SCD Type 3 Implementation in Azure Data Factory, there is no vidoe on this . thanks

  • @mohammedvahid5099
    @mohammedvahid5099 4 месяца назад

    Thnk u bro

  • @NirmalJoseph-r4p
    @NirmalJoseph-r4p 4 месяца назад

    Thank you

  • @ryanhancox6921
    @ryanhancox6921 4 месяца назад

    Thank you this was very helpful!

  • @AnkitBilone
    @AnkitBilone 5 месяцев назад

    Super helpful!

  • @mohammedvahid5099
    @mohammedvahid5099 5 месяцев назад

    Pleas bro can u make videos on how to bigger files like large volume TB's of data into small chunks(files) through python code scripts on python languge make that video's of playlist bro🙏

  • @mohammedvahid5099
    @mohammedvahid5099 5 месяцев назад

    Nice video bro thnk u

  • @cherukurid0835
    @cherukurid0835 5 месяцев назад

    Hi Bro , Thanks for video , which one we have to select if client ask loading data using ADF or loading the data using snowflake copy command from azure storage ? and what is the reason ?

    • @dataworldsolution
      @dataworldsolution 5 месяцев назад

      If you want to automate data ingestion process through ETL than we need ADF option as well, it’s also depend on work process.

    • @cherukurid0835
      @cherukurid0835 5 месяцев назад

      @@dataworldsolution Bro , can you elaborate on the work process means

  • @c.4800
    @c.4800 5 месяцев назад

    Fantastic tutorial, thank you!

  • @mohammedvahid5099
    @mohammedvahid5099 5 месяцев назад

    Excellent bro thnk u ❤🎉

  • @mohammedvahid5099
    @mohammedvahid5099 5 месяцев назад

    Great job bro well explained ❤

  • @vikaspanchal1746
    @vikaspanchal1746 6 месяцев назад

    Nice brother 🎉

  • @Thehonestreview99
    @Thehonestreview99 6 месяцев назад

    Help me with resume building for snowflake developer

  • @Thehonestreview99
    @Thehonestreview99 6 месяцев назад

    You are doing a great job keep posting

  • @mohammedvahid5099
    @mohammedvahid5099 6 месяцев назад

    BRO PLEASE PROVIDE orders dataset csv FILE for practice....where u get that csv file pleas give the link bro

    • @dataworldsolution
      @dataworldsolution 6 месяцев назад

      drive.google.com/drive/folders/1E0rGggx27gZdcfLeR0g5zpP1JqaSbli-?usp=sharing

    • @dataworldsolution
      @dataworldsolution 6 месяцев назад

      # python scripts to generate dataset -------------------------------------------------------- import random import csv # Define the categories and subcategories categories = ["Electronics", "Books", "Clothing"] electronics_subcategories = ["Laptops", "Headphones", "Smartphones", "Accessories"] books_subcategories = ["Fiction", "Non-Fiction", "Science"] clothing_subcategories = ["T-Shirts", "Jeans", "Hoodies"] # Function to generate random data for a row def generate_row(): order_id = "ORD" + str(random.randint(1, 10000)).zfill(4) amount = random.randint(50, 500) profit = int(amount * random.uniform(0.1, 0.3)) quantity = random.randint(1, 5) category = random.choice(categories) if category == "Electronics": subcategory = random.choice(electronics_subcategories) elif category == "Books": subcategory = random.choice(books_subcategories) else: subcategory = random.choice(clothing_subcategories) return [order_id, amount, profit, quantity, category, subcategory] # Generate 1000 rows of data rows = [] for _ in range(25): rows.append(generate_row()) # Write the data to a CSV file with open("OrdersAzuredataset2.csv", "w", newline="") as csvfile: writer = csv.writer(csvfile) writer.writerow(["ORDER_ID", "AMOUNT", "PROFIT", "QUANTITY", "CATEGORY", "SUBCATEGORY"]) writer.writerows(rows) print("Dataset created successfully!")

  • @mohammedvahid5099
    @mohammedvahid5099 6 месяцев назад

    Nice explain bro thnk u

  • @shanezinno8077
    @shanezinno8077 6 месяцев назад

    🎶 'Promo sm'

  • @mohammedvahid5099
    @mohammedvahid5099 6 месяцев назад

    Ur video on tasks is excellent 👌 bro want to also know how to use SCD TYPE 1, 2 AND 3 THNK U

  • @mohammedvahid5099
    @mohammedvahid5099 6 месяцев назад

    Thnk u bro... BHAI kya ADF aur bolo ge kya ADF videos ka intezar kar rha hu bro. Pls make ADF completely ❤

    • @dataworldsolution
      @dataworldsolution 6 месяцев назад

      Busy in some family functions bro , i will prepared also video on ADF Completely , thanks bro

  • @ramabeeravalli
    @ramabeeravalli 7 месяцев назад

    i am not geetin your point i check my wh as well as i use C:\Users\ANJI>snowsql -a UK82993.AWS_AP_SOUTHEAST_1 -u RAMA

  • @ramabeeravalli
    @ramabeeravalli 7 месяцев назад

    I am getting this type of error could you please give me any suggestion !!!! 250001 (n/a): Could not connect to Snowflake backend after 2 attempt(s).Aborting If the error message is unclear, enable logging using -o log_level=DEBUG and see the log to find out the cause. Contact support for further help. Goodbye!

    • @dataworldsolution
      @dataworldsolution 7 месяцев назад

      Hi Which statement are you used, look like this is connectivity issue, plz check your data warehouse activated

  • @ashok6644
    @ashok6644 7 месяцев назад

    good one

  • @mohammedvahid5099
    @mohammedvahid5099 7 месяцев назад

    Superb bro ur like master pro ❤

  • @mohammedvahid5099
    @mohammedvahid5099 7 месяцев назад

    NICE APPROACH BRO THNK U

  • @mohammedvahid5099
    @mohammedvahid5099 7 месяцев назад

    WELL EXPLAINED BRO THNK U

  • @astrohomedecor3516
    @astrohomedecor3516 7 месяцев назад

    Nice sanu bhai

  • @mohammedvahid5099
    @mohammedvahid5099 7 месяцев назад

    Amazing pls work on more azure to and also how to SCD TYPE 1 AND SCD 2 TWO ALSO BRO AFTR FINISHING STREAMS AND TASKS..AND I WANT TO SEE HOW U WORK ON QUERY OPTIMIZATION FOR QUERY PERFORMANCE TUNING UR SCENARIO BRO...THNKS UR WORKING AMAZING BRO

  • @cherukurid0835
    @cherukurid0835 7 месяцев назад

    Hi Bro , Thanks for videos and which method is preferable to copy data from Azure storage to Snowflake tables : Using snowflake copy command or using ADF tool ?

    • @dataworldsolution
      @dataworldsolution 7 месяцев назад

      Hi I am using ADF for automation and transformation with databricks but for add hoc copy command is best. So it depend on type of work.

    • @cherukurid0835
      @cherukurid0835 7 месяцев назад

      Bro , do we have to provide access permission to storage for ADF to all the storage accounts we created ?

  • @mohammedvahid5099
    @mohammedvahid5099 7 месяцев назад

    please make to more videos on i just created azure aaccount and adf also ples post further videos on how to integrete with snowflake and migrating on adf to snowflake..am waitng bro thnk u so much for this video

  • @mohammedvahid5099
    @mohammedvahid5099 7 месяцев назад

    Nice session bro thnk u ples make on videos validation_mode optjon on copy into command and also how to use with azure blob on full video content bro ❤ using full load and incremental also wanto know more about merge on STREAMS on scd type 1 and SCD TYPE 2 ALSO🎉❤❤❤

  • @vru5696
    @vru5696 8 месяцев назад

    Suppose we have varchar field date col where the value is MM/DD/YYYY how can we convert it to the date datatype while loading into target in the select query? Any function?

    • @dataworldsolution
      @dataworldsolution 8 месяцев назад

      Bro , use cast function. Or to_date

    • @dataworldsolution
      @dataworldsolution 8 месяцев назад

      Thanks for asking

    • @vru5696
      @vru5696 8 месяцев назад

      @@dataworldsolution not able to give required format in cast. To_date is also not working - select to_date(datecol,'DD/MM/YYYY'). Can you help with code?

  • @mohammedvahid5099
    @mohammedvahid5099 8 месяцев назад

    Nice session bro where can I get the the source loan datasets 1 and 2 bro can tel me pls..

    • @dataworldsolution
      @dataworldsolution 8 месяцев назад

      I have python script to create dataset , I will provide you.

    • @dataworldsolution
      @dataworldsolution 8 месяцев назад

      use python script to generate dataset :- ------------------ # conda install Faker import csv from faker import Faker import random from datetime import datetime, timedelta fake = Faker() # Define loan statuses loan_statuses = ['PAIDOFF', 'COLLECTION', 'COLLECTION_PAIDOFF'] # Define education levels education_levels = ['High School or Below', 'College', 'Bechalor', 'Master or Above'] # Define genders genders = ['male', 'female'] # Generate 50000 loan payment records num_records = 4000 with open('loan_payment_dataset4.csv', mode='w', newline='') as file: writer = csv.writer(file) writer.writerow(["Loan_ID", "loan_status", "Principal", "terms", "effective_date", "due_date", "paid_off_time", "past_due_days", "age", "education", "Gender"]) for i in range(num_records): loan_id = fake.uuid4() loan_status = random.choice(loan_statuses) principal = str(random.randint(500, 5000)) terms = str(random.randint(7, 30)) + " days" effective_date = fake.date_time_between(start_date="-1y", end_date="now", tzinfo=None) due_date = effective_date + timedelta(days=int(terms.split()[0])) paid_off_time = '' past_due_days = '' if loan_status == 'PAIDOFF': paid_off_time = fake.date_time_between(start_date=effective_date, end_date="now", tzinfo=None) elif loan_status == 'COLLECTION_PAIDOFF': paid_off_time = fake.date_time_between(start_date=due_date, end_date="now", tzinfo=None) past_due_days = str((paid_off_time - due_date).days) elif loan_status == 'COLLECTION': past_due_days = str((datetime.now() - due_date).days) age = str(random.randint(18, 70)) education = random.choice(education_levels) gender = random.choice(genders) writer.writerow([loan_id, loan_status, principal, terms, effective_date, due_date, paid_off_time, past_due_days, age, education, gender]) print("Dataset generation completed.")

  • @jaga8956
    @jaga8956 8 месяцев назад

    Wow, learnt sub query concept with detailed explanation. Thanks BRO, it helped me alot. Hope to view further videos.

  • @mohammedvahid5099
    @mohammedvahid5099 8 месяцев назад

    Nice session bro❤