Apache Spark Basics with Python

83 / 86

We have an RDD containing a list of words. Each record of this RDD is a word. You can assume that the RDD is similar to what gets created when we run the following code:

rdd = sc.parallelize(["this", "this", "is", "good", "is"])

If we have to find the count of each unique word in an RDD having each word as a record. Which approach gives correct results?


Note - Having trouble with the assessment engine? Follow the steps listed here


No hints are availble for this assesment

Answer is not availble for this assesment

Loading comments...