How to Run a Spark Cluster with Multiple Workers Locally Using Docker

Поделиться
HTML-код
  • Опубликовано: 26 фев 2024
  • In this video, I'll show you how you can use Docker Desktop to run a Spark cluster that mimics a production environment on your local machine!
    / setting-up-a-spark-sta...

Комментарии • 24

  • @rasmusandreasson1548
    @rasmusandreasson1548 5 месяцев назад +1

    The king! Thank u for good content!

  • @not_saboor
    @not_saboor 5 месяцев назад +1

    Thanks for this !

  • @Levy957
    @Levy957 5 месяцев назад +1

    i love your videos

  • @ccc_ccc789
    @ccc_ccc789 5 месяцев назад

    Thanks

  • @early-riser18
    @early-riser18 2 месяца назад

    Thank you for the rundown - very helpful. Could you add a link to the code written please? Some code in the Dockerfile is hidden by the right-side screen fold and has to be guessed. Thanks :)

  • @mayowaoludoyi5425
    @mayowaoludoyi5425 3 месяца назад +1

    Thank you for this walkthrough video. How can I establish connection to a relational database like Oracle from "dockerised" spark like this. I understand there is a different set up that requires JDBC. Where does it fits in this your setup?

    • @thedataguygeorge
      @thedataguygeorge  3 месяца назад +2

      Hey, you would add it similarly to how I connect to snowflake in other scripts, where you use the python ODBC drivers to establish connections to relational db's like Oracle

  • @not_saboor
    @not_saboor 2 месяца назад +1

    Can you explain the part on Jinja Templating you mentioned in 3:40

    • @thedataguygeorge
      @thedataguygeorge  2 месяца назад

      Sure! What specifically about it are you interested in learning more about?

  • @imanitrecruiterineurope4142
    @imanitrecruiterineurope4142 2 месяца назад

    Hi!
    It seems that the applications aren't taking any resources and are stuck in a loop on my end. What could be the cause?

  • @Jalabulajunx
    @Jalabulajunx 2 месяца назад +1

    you don't have a directory like requirement, how will req/req.txt work?

    • @thedataguygeorge
      @thedataguygeorge  2 месяца назад

      With Spark, you'll typically initiate a spark session and provide a list of requirements you need for that particular session

  • @csmithDevCove
    @csmithDevCove 5 месяцев назад +1

    What about connecting sparp-nlp to this

    • @thedataguygeorge
      @thedataguygeorge  5 месяцев назад

      You would just want to add it to be installed within the docker image!

  • @rafaellourenco4599
    @rafaellourenco4599 3 месяца назад +2

    Bro, you skiped all the bugs stuff

    • @thedataguygeorge
      @thedataguygeorge  3 месяца назад +1

      Sorry was solving them off camera but will make sure to show more of the troubleshooting process next time!

    • @rafaellourenco4599
      @rafaellourenco4599 3 месяца назад +1

      @@thedataguygeorge can you share a repo with this project?

  • @Jalabulajunx
    @Jalabulajunx 2 месяца назад

    I am always getting entry point.sh not found, has anyone figured it out?