Kafka Tutorial Offset Management

Поделиться
HTML-код
  • Опубликовано: 4 окт 2024
  • Spark Programming and Azure Databricks ILT Master Class by Prashant Kumar Pandey - Fill out the google form for Course inquiry.
    forms.gle/Nxk8...
    -------------------------------------------------------------------
    Data Engineering using is one of the highest-paid jobs of today.
    It is going to remain in the top IT skills forever.
    Are you in database development, data warehousing, ETL tools, data analysis, SQL, PL/QL development?
    I have a well-crafted success path for you.
    I will help you get prepared for the data engineer and solution architect role depending on your profile and experience.
    We created a course that takes you deep into core data engineering technology and masters it.
    If you are a working professional:
    1. Aspiring to become a data engineer.
    2. Change your career to data engineering.
    3. Grow your data engineering career.
    4. Get Databricks Spark Certification.
    5. Crack the Spark Data Engineering interviews.
    ScholarNest is offering a one-stop integrated Learning Path.
    The course is open for registration.
    The course delivers an example-driven approach and project-based learning.
    You will be practicing the skills using MCQ, Coding Exercises, and Capstone Projects.
    The course comes with the following integrated services.
    1. Technical support and Doubt Clarification
    2. Live Project Discussion
    3. Resume Building
    4. Interview Preparation
    5. Mock Interviews
    Course Duration: 6 Months
    Course Prerequisite: Programming and SQL Knowledge
    Target Audience: Working Professionals
    Batch start: Registration Started
    Fill out the below form for more details and course inquiries.
    forms.gle/Nxk8...
    --------------------------------------------------------------------------
    Learn more at www.scholarnes...
    Best place to learn Data engineering, Bigdata, Apache Spark, Databricks, Apache Kafka, Confluent Cloud, AWS Cloud Computing, Azure Cloud, Google Cloud - Self-paced, Instructor-led, Certification courses, and practice tests.
    ========================================================
    SPARK COURSES
    -----------------------------
    www.scholarnes...
    www.scholarnes...
    www.scholarnes...
    www.scholarnes...
    www.scholarnes...
    KAFKA COURSES
    --------------------------------
    www.scholarnes...
    www.scholarnes...
    www.scholarnes...
    AWS CLOUD
    ------------------------
    www.scholarnes...
    www.scholarnes...
    PYTHON
    ------------------
    www.scholarnes...
    ========================================
    We are also available on the Udemy Platform
    Check out the below link for our Courses on Udemy
    www.learningjo...
    =======================================
    You can also find us on Oreilly Learning
    www.oreilly.co...
    www.oreilly.co...
    www.oreilly.co...
    www.oreilly.co...
    www.oreilly.co...
    www.oreilly.co...
    www.oreilly.co...
    www.oreilly.co...
    =========================================
    Follow us on Social Media
    / scholarnest
    / scholarnesttechnologies
    / scholarnest
    / scholarnest
    github.com/Sch...
    github.com/lea...
    ========================================

Комментарии • 43

  • @ScholarNest
    @ScholarNest  3 года назад +1

    Want to learn more Big Data Technology courses. You can get lifetime access to our courses on the Udemy platform. Visit the below link for Discounts and Coupon Code.
    www.learningjournal.guru/courses/

  • @JemiloII
    @JemiloII 6 лет назад +4

    Exactly what I was looking for and very clear of what to do, even though I am using a different programming language. Thank you for the knowledge shared.

  • @shyamyadav-qk4zb
    @shyamyadav-qk4zb 2 месяца назад

    Very good video sir , my doubt is clear now much more helpful 🙏

  • @mukundsridhar4250
    @mukundsridhar4250 5 лет назад +1

    awesome... Thats the one word that i can think of to describe your work.

  • @sreekanthmunigati6450
    @sreekanthmunigati6450 7 лет назад

    excellent explanation ... I liked your playlist for Kafka , very clear and covered several points in each short video .. just play for an hour you will get to know lot about Kafka !!

  • @ganesans4957
    @ganesans4957 Год назад

    Great explanation. Started exploring your other videos.

  • @ravishankar4521
    @ravishankar4521 6 лет назад +1

    kudos to you sir for such a wonderful effort in making these videos.

  • @mohanbabubanavaram5211
    @mohanbabubanavaram5211 3 года назад

    Excellent explanation, Covering various scinarios

  • @jaineshmodi
    @jaineshmodi 6 лет назад +2

    sir your videos are really good.

  • @decaalv
    @decaalv 4 года назад

    Thank you Mr. Indian guy. I love you.

  • @drindianVlogs
    @drindianVlogs 6 лет назад

    Nice tutorial..... Learnt lot of things here... Thanks a lot sir for the videos

  • @krushnachandrasahoo5034
    @krushnachandrasahoo5034 6 лет назад

    exemplary teaching..really excellent...learnt a lot of things...

  • @rajareddy47444
    @rajareddy47444 6 лет назад

    Hi,Thank you sir for sharing your knowledge.The way u take the concept and explaining is awesome. After watching your spark videos i got the confidence that i can face interviews with more confidence .Thanks for showing way how to do in real time using GCP.
    Now i started learning kafka parallel with spark. These videos are in terminal.Can you please explain kafka also in real time with GCP. That will be a great advantage for those who are moving to this ecosystem. Thank you

  • @boycy69
    @boycy69 4 года назад

    Great explanation, thanks!

  • @saurav0777
    @saurav0777 4 года назад +2

    What happens if commitAsync has been failing from longer time let say more than 1 hours and if we restart the application after that it will process the duplicates records . How to handle such scenarios?

  • @kalpanabejawada2451
    @kalpanabejawada2451 6 лет назад

    Nice explanation! Found it useful

  • @lokeshwarreddy3
    @lokeshwarreddy3 3 года назад

    Excellent sir

  • @mayankwassup
    @mayankwassup 5 лет назад

    Hi, let's assume we have 3 topics with 1 partition each. There are 10 consumer groups, each group has a single consumer and each consumer is going to read all the three topics. So in that scenario, how offset will be maintained by each consumer?

  • @jkiran2020
    @jkiran2020 7 лет назад +1

    Very nice content. Is the video streaming slow only for me?

  • @jeetendrasingh1972
    @jeetendrasingh1972 3 года назад

    Great

  • @MohammadRasoolShaik
    @MohammadRasoolShaik 4 года назад

    Is Kafka allows to commit a higher offset before committing a lower offset?

  • @poorvivijay
    @poorvivijay 5 лет назад

    How does kafka handle, multiple consumers subscribing to the same topic? single publisher multiple consumers problem, as there is a restriction that only one consumer can read the partition.

  • @vivektwentysix9064
    @vivektwentysix9064 7 лет назад

    Very good

  • @pheiroz6307
    @pheiroz6307 7 лет назад +4

    Hi, In the example why would one want to commit sync in case of an exception. If there is an exception in mid processing i.e. one has processed only 10 messages out of 50 messages pulled from kafka won't a commit sync incorrectly set the committed offset at 50. However since there was an exception messages 11 to 50 never get processed because commit sync was invoked. Can you please clarify. Thanks!

    • @ScholarNest
      @ScholarNest  7 лет назад +4

      Yes, you are correct. Thanks for pointing this out. You watched carefully :-)
      I should use a particular offset in the commit sync instead of committing current offset. Using appropriate offset will make sure that only 10 messages that I successfully processed are Committed.

    • @pheiroz6307
      @pheiroz6307 7 лет назад +1

      Sure :).Your tutorials are extremely clear in communication. Can you please do a tutorial on the use of schema registry and it's implementation when consuming avro messages in Kafka. Thanks!

    • @ScholarNest
      @ScholarNest  7 лет назад +6

      I am working on Schema Registry Example and Avro Schema, Will upload it soon. Thanks.

    • @pheiroz6307
      @pheiroz6307 7 лет назад

      Awesome! eagerly looking forward to it. Thanks

  • @karthikragunaths
    @karthikragunaths 7 лет назад

    After 100 records are fetched by a consumer for processing, and 50 records are processed, for whatever reason a re-balance is triggered in the broker, the consumer still processes the remaining 50 records - is this correct ? ( I assuming that there's no way for the broker to stop the consumer because consumer is a stand-alone application and the communication is uni-directional from consumer to broker and not the other way )

    • @ScholarNest
      @ScholarNest  7 лет назад

      You are correct in saying that broker has no way of stopping the consumer, but the Broker blocks the poll call during the rebalance.

  • @rameshthamizhselvan2458
    @rameshthamizhselvan2458 5 лет назад

    Is there anyway to see the committed offsets???

  • @vishvaspatel34
    @vishvaspatel34 7 лет назад

    Thank you sir for valuable information.
    What if while reading from broker there will be an exception occurs on consumer side in async commit. so required to data to be read once again. But in async communication may be higher ordered can be committed on top of that, than ?

  • @TheNayanava
    @TheNayanava 7 лет назад

    sir. I have a doubt. In my code, I have set my enable auto commit to true and the interval as 1000 ms, yet I get the following error
    ERROR o.a.k.c.c.i.ConsumerCoordinator - Error UNKNOWN_MEMBER_ID occurred while committing offsets for group
    From the error I am assuming that its because the coordinator would have called for a partition rebalance, before the consumer goes to commit the offset, whereas the heartbeat interval is set to 3000. What could be the possible reason for this error then??

  • @karthikragunaths
    @karthikragunaths 7 лет назад

    Is the number of records fetched by poll() configurable ?

  • @Daniel_wwf
    @Daniel_wwf 7 лет назад

    At 1:08, shouldn't it be 99 for M100?

    • @ScholarNest
      @ScholarNest  7 лет назад

      It always points to the record that should be sent on next request. For 100 records delivered, offset 0-99 are done, the next one is 100. So 100 is correct.

    • @Daniel_wwf
      @Daniel_wwf 7 лет назад

      I meant the following: We assumed having 100 messages in the partition, so offsets should be 0-99. As M1 has offset 0, M100 should have offset 99.

  • @VishalYadav-di7ip
    @VishalYadav-di7ip 7 лет назад

    Can we delete specific offset ?

    • @ScholarNest
      @ScholarNest  7 лет назад +1

      Why would you want to do that?

    • @VishalYadav-di7ip
      @VishalYadav-di7ip 7 лет назад

      Sorry for the late...Sir, actually interviewer asksed me to do that.