Registrations Closing Soon for DevOps Certification Training by CloudxLab | Batch Starts on 18th AprilEnroll Now
Normally, when a function passed to a Spark operation (such as map or reduce) is executed on a remote cluster node, it works on separate copies of all the variables used in the function. These variables are copied to each machine, and no updates to the variables on the remote machine are propagated back to the driver program. Supporting general, read-write shared variables across tasks would be inefficient. However, Spark does provide two limited types of shared variables for two common usage patterns: broadcast variables and accumulators.
No hints are availble for this assesment
Answer is not availble for this assesment