Taking you to the next exercise in seconds...
Want to create exercises like this yourself? Click here.
Note - Having trouble with the assessment engine? Follow the steps listed here
No hints are availble for this assesment
Answer is not availble for this assesment
Please login to comment
5 Comments
Hello
Can you please explain the above question indetail?
Upvote ShareWhen you start
spark-shell
orpyspark
, the shell environment initializes and starts a Spark Driver locally.The Driver process creates a
SparkContext
(orSparkSession
in the case of Spark 2.x and later).The Driver then interacts with the cluster manager (e.g., YARN, Mesos, or the built-in standalone cluster manager) to allocate resources for the executors.
The Driver translates the user commands (such as transformations and actions on RDDs or DataFrames) into a DAG (Directed Acyclic Graph) of tasks.
It then schedules these tasks to be executed on the cluster's worker nodes (executors).
Therefore, when you start
Upvote Sharespark-shell
orpyspark
, you are launching a Spark Driver process.Please elaborate?
Upvote ShareWhen we launch an spark-shell or pyspark, How does it start Spark Driver?
Hello,
Upvote ShareHow does it start spark-application and who triggers it?
spark-Application: It is installed on each and every data nodes. where as spark driver is on name node only which controls all the data nodes. So, It is the spark-drivers which triggers.
Lets join a group to discuss more.
Upvote Sharehttps://chat.whatsapp.com/G...