Earlier I watched some videos regarding this topic ,no one can explained in this way ,I am glad to see this video,now clearly understood spark optimization techniques
This is excellent and valuable knowledge sharing... Easily one can make out these trainings are coming out of personal deep hands-on experience and not the mere theory ..Great work
Awesome explanation of the optimisation techniques. If possible please create a video to cover the realtime challenges which you faced in your project and the solution you provided. That will be really helpful.
thanks a lot , i have case where someother modules write parquet file , i need to process in my module by reading it, so how should i apply bucketing on that day ...can it be possible without writing ???
@@BigDataThoughts thanks! Can you build an end to end project or some mini project where one can see how and where these properties arte getting implemented? Just watching these in silos only give half knowledge. Thanks.
how i join small table with big table but i want to fetch all the data in small table like the small table is 100k record and large table is 1 milion record df = smalldf.join(largedf, smalldf.id==largedf.id , how = 'left_outerjoin') it makes out of memory and i cant do broadcast the small df idont know why what is best case here pls help
Absolutely amazing, the way content is delivered!!
Earlier I watched some videos regarding this topic ,no one can explained in this way ,I am glad to see this video,now clearly understood spark optimization techniques
Thanks
I am searching this kind of topics from the web in long time.. Thank you very much.. And i appreciate your stuff..
@@durairaj4296 thanks
best video on internet for spark performace...
Thanks harsh
This is excellent and valuable knowledge sharing... Easily one can make out these trainings are coming out of personal deep hands-on experience and not the mere theory ..Great work
thanks
Very informative 👏👍
Keep such videos coming
what a wonderfull explanation to the point... thank you
Thanks
perfectly went into my brain, what a clean explanation.
can you please do videos on Pyspark from scratch.
Very good explanation keep create more videos in spark
Thanks
One of the best explanation on RUclips 😊
I only have one word for this video
"Awesome!!"
Thanks
It's very very easy to understand whatever you explained, thank you so much
Thanks namrata
Very Good vedio, awesome work. To the point and one can understand easily
Thanks sahil
Great expectations Mam ... eagerly waiting for your upcoming video
Thanks Rajni
Many thanks for making such informative video
All yours tutorials are too good!
Thanks for sharing. Very informative .
Thanks karan
This is excellent maam. Looking forward to watching more videos from you
Thanks Chethan
Another much needed video on Spark optimizations. point to point. Thank you very much for the video.
Thanks Kiran
Neat and clean explanation. Looking forward to the videos on Spark Optimization.
Thanks Gunasekaran
Very well explained. Loving your videos!❤
Thanks
REALLY helped me get better at my work
Thanks
Nice video it was really good 👍🏻 Thank you
Thanks
Thanks a lot mam for making these videos. These are extremely useful. One of the best videos I have come across.
Thanks Ram
Amazing Job Shreya... Keep it Up..
Thanks Ayush
Thanks for video ma'am., you made it very simple to understand.. waiting for more video's on this topic and spark
Thanks Arjun
ruclips.net/video/snPYj3TqM1g/видео.html you can chk this too. There are others videos on spark that i have posted
So nicely explained Shreya..
Thanks Abhishek
Awesome explanation of the optimisation techniques. If possible please create a video to cover the realtime challenges which you faced in your project and the solution you provided. That will be really helpful.
Awesome Shreya, if possible could you please upload realtime challenges which we faced in realtime environment
Thanks Khader
where do I get practical session on this optimization technique of spark..?
If count is not adviced, how can we count the number of rows in data frame?
thanks a lot , i have case where someother modules write parquet file , i need to process in my module by reading it, so how should i apply bucketing on that day ...can it be possible without writing ???
Good explanation 😊
Thanks
Great explaination !!
Thanks Puneet
Tnks Amigo
It's very helpful
thanks
very well explained. but you are telling everything i know to exclude. ?? We need count, group,agg
Ultimate 👏👏👏
Thanks
Great explanation
Thanks
Nice video mam
Coalesce doesn't do Shuffle and that's why it's less expensive than repartition. I believe.
It does but not as much as repartition. Repartition does entire data shuffle as it can reduce or increase no of partitions.
@@BigDataThoughts thanks! Can you build an end to end project or some mini project where one can see how and where these properties arte getting implemented? Just watching these in silos only give half knowledge. Thanks.
Excellent.
good optimization tips
Thanks Neethu
so nice it helped a lot
Thanks tanushree
What are configurations file in spark ?
Plz anyone can ans me please
❤
Keep the good work #Prinetechs.
I can clearly see all the good reviews about you man…I never believed my account can fixed after 7 months hahaha
how i join small table with big table but i want to fetch all the data in small table like
the small table is 100k record and large table is 1 milion record
df = smalldf.join(largedf, smalldf.id==largedf.id , how = 'left_outerjoin')
it makes out of memory and i cant do broadcast the small df idont know why what is best case here pls help
Keep the good work #Prinetechs.
I can clearly see all the good reviews about you man…I never believed my account can fixed after 7 months hahaha
Keep the good work #Prinetechs.
I can clearly see all the good reviews about you man…I never believed my account can fixed after 7 months hahaha