How to Run a Spark Cluster with Multiple Workers Locally Using Docker
HTML-код
- Опубликовано: 26 фев 2024
- In this video, I'll show you how you can use Docker Desktop to run a Spark cluster that mimics a production environment on your local machine!
/ setting-up-a-spark-sta...
The king! Thank u for good content!
Thanks so much Rasmus!
Thanks for this !
No problem Saboor!
i love your videos
Thanks Levy! Love your support!
Thanks
No problem!
Thank you for the rundown - very helpful. Could you add a link to the code written please? Some code in the Dockerfile is hidden by the right-side screen fold and has to be guessed. Thanks :)
Thank you for this walkthrough video. How can I establish connection to a relational database like Oracle from "dockerised" spark like this. I understand there is a different set up that requires JDBC. Where does it fits in this your setup?
Hey, you would add it similarly to how I connect to snowflake in other scripts, where you use the python ODBC drivers to establish connections to relational db's like Oracle
Can you explain the part on Jinja Templating you mentioned in 3:40
Sure! What specifically about it are you interested in learning more about?
Hi!
It seems that the applications aren't taking any resources and are stuck in a loop on my end. What could be the cause?
you don't have a directory like requirement, how will req/req.txt work?
With Spark, you'll typically initiate a spark session and provide a list of requirements you need for that particular session
What about connecting sparp-nlp to this
You would just want to add it to be installed within the docker image!
Bro, you skiped all the bugs stuff
Sorry was solving them off camera but will make sure to show more of the troubleshooting process next time!
@@thedataguygeorge can you share a repo with this project?
I am always getting entry point.sh not found, has anyone figured it out?