Delta Lake for Apache Spark - Why do we need Delta Lake for Spark?

Поделиться
HTML-код
  • Опубликовано: 18 окт 2024

Комментарии • 49

  • @ScholarNest
    @ScholarNest  3 года назад

    Want to learn more Big Data Technology courses. You can get lifetime access to our courses on the Udemy platform. Visit the below link for Discounts and Coupon Code.
    www.learningjournal.guru/courses/

  • @rameshthamizhselvan2458
    @rameshthamizhselvan2458 4 года назад +3

    No words! $imply excellent.I found the definition of crystal clear from your explanation. I had watched all your playlists. 2 year$ ago I got a super confident after watching your spark videos and I got a job just because of you....

  • @hrishabhg
    @hrishabhg 2 года назад

    Superb way of making things clear

  • @abu-yousuf
    @abu-yousuf 3 года назад

    great job brother. beautifully explained. thanks from Pakistan.

  • @meenakshipreethi5131
    @meenakshipreethi5131 2 года назад

    You are awesome dheivame! Fantastic video

  • @sandeepverma8350
    @sandeepverma8350 4 года назад

    Superb!! Excellent way of describing and summarizing.. Thanks

  • @saurabhdsawant
    @saurabhdsawant 3 года назад

    I was hooked since the first minute . Thank you for uploading this. Please push for more . Cheers !

  • @RahulDevanarayanan
    @RahulDevanarayanan 4 года назад +1

    Great job breaking down the content and explaining it clearly. I understood very well how everything fit together, on the first try too!

  • @taglud
    @taglud 2 года назад

    Hello Sir, i really love your video the explanation are really details. thank you :)

  • @davehartley3874
    @davehartley3874 2 года назад

    Great video, this really laid out the problem space perfectly 👍

  • @muthusamyperiyasamy5306
    @muthusamyperiyasamy5306 4 года назад +1

    Nice explanation.. the way you articulate is good. Thanks youtube for suggesting this video :)
    Now i'm going to have some basic knowledge about delta lake.

  • @BigBossInd7236
    @BigBossInd7236 2 года назад

    Very nicely explained

  • @范振-v2e
    @范振-v2e 4 года назад +1

    Excellent video for delta. It helps a lot and I love it.

  • @muhammadrahman3510
    @muhammadrahman3510 3 года назад

    This is just great to me! He explained the facts just the way I wanted. Thanks for this great presentation.

  • @samsamhuns928
    @samsamhuns928 4 года назад

    Excellent video and well done with actually showing the issues with demos!

  • @ladakshay
    @ladakshay 4 года назад

    Superb video, all points are valid, we do face such issues in our projects

  • @patrickzeng5668
    @patrickzeng5668 4 года назад +3

    classic quote of Johnny Depp's "if nobody sees it, it didn't happen" in the context of Spark write

  • @suhel.choudhury
    @suhel.choudhury 3 года назад

    Great video, very well explained.

  • @2007selvam
    @2007selvam 5 лет назад +1

    Excellent explanation for ACID Concept.

  • @dorababug4274
    @dorababug4274 4 года назад +1

    Could you please make a video on SCD implementation Using Databricks Spark

  • @adityasarin16
    @adityasarin16 4 года назад

    This is an excellent video .. MUST WATCH ... do you have your own tutorial site I would love to learn from those ... you are too good

  • @bishnupratikdebashram5911
    @bishnupratikdebashram5911 4 года назад

    Thank you so much ..excellent excellent explanation ..of the real need of Delta Lake in Datalake

  • @santhoshsandySanthosh
    @santhoshsandySanthosh 4 года назад

    Dear Prashanth, your videos are addicting.. I wonder how you explain things with that much deep diving into the concepts so easily.. Thanks for all your explanation.. 🙏 do you have any plans on similar explanation on apache Hudi or atleast comparision of features ?

  • @sachinthelearner1894
    @sachinthelearner1894 5 лет назад

    Hello sir, brilliant video with nicely articulated contents. One small suggestion, practical code flashes quite fast - good if you could zoom them in some slow motions or something. Looking forward for next part.

  • @ravikatiyar6530
    @ravikatiyar6530 5 лет назад +1

    Thank you so much ..excellent excellent explanation ..

  • @rohitcs1987
    @rohitcs1987 4 года назад

    Great detailed video, thanks for sharing .

  • @dorababug4274
    @dorababug4274 4 года назад

    Really nice and demystifying . Thank you so much. Subscribed.

  • @daughnutdaddy1594
    @daughnutdaddy1594 3 года назад

    Well explained mate

  • @charlesfe
    @charlesfe 5 лет назад

    Excellent explanation and clear examples.

  • @yellajosyulakameswar2548
    @yellajosyulakameswar2548 5 лет назад +1

    Great video.Please keep more such videos coming. I have a question. Is this an accurate summarization - If I use Append-only mode for my data frame writes, SPARK in a way complies with Atomicity using HDFS file committer version 1, but not when I use overwrite mode?

    • @ScholarNest
      @ScholarNest  5 лет назад

      As per documentation, even append is not atomic. But it just works due to the simplicity in the append and job level commit protocol.

  • @saurabhgulati2505
    @saurabhgulati2505 4 года назад

    Bawa u rocks 🙌

  • @sravyapagadala7154
    @sravyapagadala7154 4 года назад

    Very Helpful ! Thank you :)

  • @NikhileshwarYanamanagandla
    @NikhileshwarYanamanagandla Год назад

    where can we find more videos on delta lake with spark?

  • @debasish2332
    @debasish2332 5 лет назад

    Excellent Video !!

  • @Sunkarakrwutarth9389
    @Sunkarakrwutarth9389 4 года назад

    Thanks Prashanth for your good videos,
    i have observerd few things.
    dataframe writer is consistant in databricks spark cluster, where as it is not consistent in local spark cluster.
    as you said older data is not beeing deleted in the databricks spark cluster
    ,but it is in local spark cluster
    what might be the reason?
    keen to know the Anser.
    thanks,
    Mallik.

  • @bhushanbhange7065
    @bhushanbhange7065 4 года назад +1

    Are those Spark Architecture videos deleted :(

  • @vijaydas2962
    @vijaydas2962 5 лет назад

    Very useful content... Thanks for you effort

  • @ashishlimaye2408
    @ashishlimaye2408 4 года назад

    Really great Video!

  • @dn9416
    @dn9416 4 года назад

    Thank you

  • @alogicalhuman849
    @alogicalhuman849 4 года назад

    Thank you must watch content.

  • @sandyjust
    @sandyjust 5 лет назад

    Great videos

  • @bobdorous3945
    @bobdorous3945 4 года назад

    Could you share the slides deck and sample codes used for testing?

  • @ishanbhawantha
    @ishanbhawantha 5 лет назад

    Could you please do complete KSQL tutorial with a project

  • @SaimanoharBoidapu
    @SaimanoharBoidapu 5 лет назад

    Hi Sir, Thank you for such a nice video. Your videos are really helpfull. :)
    Would bucketing help in problem of reading all files in a partition for a filtered query. Kindly clarify.
    Eagerly waiting for the part-2.
    Thank you :)

    • @ScholarNest
      @ScholarNest  5 лет назад +7

      Bucketing in Spark has a different purpose. I will do a video on bucketing.

    • @KoushikPaulliveandletlive
      @KoushikPaulliveandletlive 4 года назад

      Partitions sole purpose is for filtering only .
      Similarly bucketing is for efficient joining,
      Ex two tables with 10^6 rows and 10^7 ( 10000000) rows while joining will check 10^(6+7)=10^13 combinations.
      Now if we bucket both these tables in 10^3 i.e 1000 buckets, each bucket from left side will only be compared with a single bucket from the right. and each bucket will hold close to 10^3 and a0^4 rows of data respectively for the two tables.
      Now the number of combination will be 10^3*10^4 =10^7 multiplying by number of buckets * 10^3 .
      So total combination =10^(3+4+3) =10^10 which is a 10^3x or 1000x less combination, so the join will be this much faster.

  • @linkrachitg
    @linkrachitg Год назад

    But that never happened so far! Spark write in overwrite mode failed many times but never lost data in cloudera cluster in each company

  • @KoushikPaulliveandletlive
    @KoushikPaulliveandletlive 4 года назад

    Can anyone suggest few pet projects on spark.
    Doing pet projects/hand on is a good chance to face all the different problems that happen in real projects also increase confidence.