Hi Madam, Great video. I am using ADF for Bulk and Incremental loads. In an instance i need to do a REST API call for bulk loads. per call i am able to retrieve only 20 records from REST API. where the max can be at 50 records which i am not able to apply in ADF. so for this i started to look into adding a custom python module to hit the API. Is there way we can do this ? please let me know madam
Remember Microsoft's PROJECT REAL from 2005? In it they write a script component to work around the limitations of the Lookup Transform when it comes to implementing Inferred Members without having to keep making calls to the RDBMS for each value not in a lookup transform. They wanted the best of the full cache mode and partial cache mode. If you remember it, this would be a GREAT USE CASE for , upon encountering a new value that is not in the cache of either an SSIS lookup or ADF Lookup transform.....and routing the value to another branch.......call the RDBMS, insert the new value, bring back the surrogate key in the pipeline, add it to the cache as a sorted in-memory dictionary and continue to the next row in the pipeline. The next time this same value is encountered in the same pipeline run, it can then reference that dicitionary, pick up the surrogate key, and rejoin the main ETL branch without AGAIN having to make ANOTHER RDBMS call. Does this make sense? It would be an awesome demo.....if, indeed you could use the external call transform to either run an Azure Function or call a custom activity or REST API that did the same. Not sure we could achieve the same speed as the SSIS script component, but it would still be cool to see.
Thanks a lot madam. This helped a lot for my project. Only the last step the custom output part is not working. If there is some modification in those points do let me know. It's not a show stopper for me anyways
@@AllAboutBI Hello madam. It's working perfectly fine now. The write file path I had missed out. I thought that was irrelevant. Added those and it's working perfectly
Hello Ma'am.... I'll be really grateful if u cud extend this topic further and display Pandas and ML libraries etc in action on Batch Pool... Was just doing an interview task and found it difficult. In the end I went ahead with Databricks Python activity.
Hello madam. Is there any automated way to down scale and up scale the dedicated nodes in a pool? This would help reduce the azure cost. Please let me know your thoughts
Your videos are the best material on Azure, Thanks a lot...I have a question, can we restart the ADF pipeline from its point of failure? kind of checkpointing...
@@AllAboutBI Thanks for the swift reply mam...For example I have 500 files to copy from source to destination and after 400 files pipeline or Copy activity got failed, then is it possible that pipelines resumes for exactly for the remaining 100 files?
Hello Madam Good informational video Does setting up of pool and batch service adds cost to azure billing? Does batch pool resources are counted only when a python script from ADF is run or is it billed for all the time it is there even if no adf batch executions are invoked?
@@AllAboutBI Thank you very much for the quick response. I wish azure allowed using existing VM already set up for some other activities. This is where all these cloud providers try get revenue 🙂 So its just the VM gets charged or are there any charges such as pool etc? I am just asking as to get an idea, the info need not be accurate.
Good video. Very clear. However when executing these steps I am getting an error {"errorCategory":0,"code":"CommandProgramNotFound","message":"The specified command program is not found","details":[{"Name":"Message","Value":"The system cannot find the file specified."}]} Unable to run the program. Any pointers please. stderr.txt is empty.
Hello madam. Had a question. I have a customized job created in batch service but how can I make adf to make use of that job. Eventhough I've created the job for the pool adf uses, the adf creates its own new job and uses it. I want adf to make use of the existing job in pool. Please let me know if you've any idea
When we try to manipulate ads data from python script. Do we need to establish any authentication.. between Pool and adls.. if you have any video for that..
Hi I have tried this example of executing python script. I have used same script as yours, but the pipeline failed saying The system can not find file specified Any help on this. Thanks
How do i import the python libraries like shareplum and pandas, i have a python script but i need to integrate it to a pipeline , so i can grab the files from sharepoint using python
Hi I need details, how I can pass one file as a parameter into our python scripts. I have my source file in storage account..what my script doing converting a parquet(source file) to Xml file..can u pls tell me..
after setting up the batch account and all other details, when i run the pipeling the custom activity errors out saying "Cannot create Shared Access Signature unless Account Key credentials are used." Any clue might help. Thanks
@@AllAboutBI Hi Mam thanks for the replay, Pool Irukura Virtual machine and Azure Virtual machine different illa same ma mam, Ethu same na Pool Azure Virtual machine na use panalama ?
Great!! Best videos on ADF covering almost all the concepts and are very easy to understand
Why is this better than an automation account
Sorry, what's an automation acct
This was a great overview! I was thinking about this today and was happy to find your overview. It's a great starting point. Thank you!
Thanks Robert!
Very informative. I am currently attempting to execute in-process python to alter a files contents. This is very helpful.
Glad it's of use stephen
Thanks for a wonderful explanation. Indeed a great flow
Thank you
Hi Madam, Great video. I am using ADF for Bulk and Incremental loads. In an instance i need to do a REST API call for bulk loads. per call i am able to retrieve only 20 records from REST API. where the max can be at 50 records which i am not able to apply in ADF. so for this i started to look into adding a custom python module to hit the API. Is there way we can do this ? please let me know madam
Remember Microsoft's PROJECT REAL from 2005? In it they write a script component to work around the limitations of the Lookup Transform when it comes to implementing Inferred Members without having to keep making calls to the RDBMS for each value not in a lookup transform. They wanted the best of the full cache mode and partial cache mode.
If you remember it, this would be a GREAT USE CASE for , upon encountering a new value that is not in the cache of either an SSIS lookup or ADF Lookup transform.....and routing the value to another branch.......call the RDBMS, insert the new value, bring back the surrogate key in the pipeline, add it to the cache as a sorted in-memory dictionary and continue to the next row in the pipeline. The next time this same value is encountered in the same pipeline run, it can then reference that dicitionary, pick up the surrogate key, and rejoin the main ETL branch without AGAIN having to make ANOTHER RDBMS call.
Does this make sense? It would be an awesome demo.....if, indeed you could use the external call transform to either run an Azure Function or call a custom activity or REST API that did the same. Not sure we could achieve the same speed as the SSIS script component, but it would still be cool to see.
pls upload english subtitles for this video . very good concept
Thanks a lot madam. This helped a lot for my project. Only the last step the custom output part is not working. If there is some modification in those points do let me know. It's not a show stopper for me anyways
Glad to know. Did u use any library? Like pandas in ur script.
@@AllAboutBI Hello madam. It's working perfectly fine now. The write file path I had missed out. I thought that was irrelevant. Added those and it's working perfectly
Thanks for the video and can you run the same code in azure batch itself without the azure data factory
Great video !! Can we execute shell script as well ??
Hello Ma'am.... I'll be really grateful if u cud extend this topic further and display Pandas and ML libraries etc in action on Batch Pool... Was just doing an interview task and found it difficult. In the end I went ahead with Databricks Python activity.
Thanks a lot. Could you please put a video on how to execute a PowerShell script from ADF?
Hello madam. Is there any automated way to down scale and up scale the dedicated nodes in a pool? This would help reduce the azure cost. Please let me know your thoughts
Thanks a lot :) from chile
Your videos are the best material on Azure, Thanks a lot...I have a question, can we restart the ADF pipeline from its point of failure? kind of checkpointing...
Thanks for the feedback.
Yes it's possible using powershe'll. Pl check microsoft documentation on it
@@AllAboutBI Thanks for the swift reply mam...For example I have 500 files to copy from source to destination and after 400 files pipeline or Copy activity got failed, then is it possible that pipelines resumes for exactly for the remaining 100 files?
Yes possible
@@AllAboutBI Thanks again!!!
Hi mam, I tried to search the same on net but could not get it. If you have idea then could you pls share any link or steps to achieve this?
Hello Madam
Good informational video
Does setting up of pool and batch service adds cost to azure billing?
Does batch pool resources are counted only when a python script from ADF is run or is it billed for all the time it is there even if no adf batch executions are invoked?
From the moment you spin up a vm charges start to accumulate even if the nodes are idle.
@@AllAboutBI
Thank you very much for the quick response. I wish azure allowed using existing VM already set up for some other activities. This is where all these cloud providers try get revenue 🙂
So its just the VM gets charged or are there any charges such as pool etc? I am just asking as to get an idea, the info need not be accurate.
Good video. Very clear. However when executing these steps I am getting an error
{"errorCategory":0,"code":"CommandProgramNotFound","message":"The specified command program is not found","details":[{"Name":"Message","Value":"The system cannot find the file specified."}]}
Unable to run the program. Any pointers please. stderr.txt is empty.
Pls watch latest video 114 I have given steps
@@AllAboutBI it is not that easy to find video 114. Can you post a full name or the link?
maam in order run python script in ADF using custom activity, should we enable anything?
Very helpful content.... I tried to experiment and I'm getting the error "CommandProgramNotFound"; did I miss something in my python script?
Seems like the python file is not found. Pls check the path of the file you are trying to invoke
Hello madam. Had a question. I have a customized job created in batch service but how can I make adf to make use of that job. Eventhough I've created the job for the pool adf uses, the adf creates its own new job and uses it. I want adf to make use of the existing job in pool. Please let me know if you've any idea
Thanks a lot
When we try to manipulate ads data from python script. Do we need to establish any authentication.. between Pool and adls.. if you have any video for that..
Will this incur any cost?
Thanks a lot for the Vedio.Could you please share how to import numpy and pandas through adf.Can you please make a Vedio and upload it for us.
Hi
I have tried this example of executing python script. I have used same script as yours, but the pipeline failed saying
The system can not find file specified
Any help on this.
Thanks
Hi Nandana, were you able to fix this error? I'm also getting same error...can you please help?
Hi check this video. ruclips.net/video/_3_eiHX3RKE/видео.html
Thanks for video. Is it possible to handle CSV file with custom + python script ?
I want to know this as well. Kindly let me know solution if found
How do i import the python libraries like shareplum and pandas, i have a python script but i need to integrate it to a pipeline , so i can grab the files from sharepoint using python
Hi
I need details, how I can pass one file as a parameter into our python scripts. I have my source file in storage account..what my script doing converting a parquet(source file) to Xml file..can u pls tell me..
Thanks a lot. Iam trying to do some numpy operations. Can you pls make a video to download numpy and run some numpy codd
Sure I shall
what should i do if my python code imports something like below,
import pandas as pd
import os
from glob import glob
How to bring that pandas refrence
after setting up the batch account and all other details, when i run the pipeling the custom activity errors out saying "Cannot create Shared Access Signature unless Account Key credentials are used." Any clue might help. Thanks
What are you trying to do with the batch job.. if u can send details to funlearn0007@gmail.com i will see if I know something
Great
Thanks 🙏
@@AllAboutBI Madam can you please make a video for doing SQL to SQl copy activity using python
Please show output . That will help us
We have a Virtual machine, Pool level Virtual machine is different or the same, it is the same meaning how do I use it?
Sorry, I don't understand your question selva. Can you pls explain a bit
@@AllAboutBI Hi Mam thanks for the replay, Pool Irukura Virtual machine and Azure Virtual machine
different illa same ma mam,
Ethu same na Pool Azure Virtual machine na use panalama ?
Hi mam, can I execute power shell code using same method
Yes I believe
Great content. But I get some error. Can u help
Sure. Mail me the details to funlearn0007@gmail.com
can we try this on free tier account ?
Yes boss
@@AllAboutBI thank you
Hi mam