diff --git a/core/src/main/scala/org/apache/spark/api/python/PythonRDD.scala b/core/src/main/scala/org/apache/spark/api/python/PythonRDD.scala index 78a5794bd557b..a36f4a1a221c8 100644 --- a/core/src/main/scala/org/apache/spark/api/python/PythonRDD.scala +++ b/core/src/main/scala/org/apache/spark/api/python/PythonRDD.scala @@ -760,8 +760,6 @@ private[spark] object PythonRDD extends Logging { /** * A helper to convert java.util.List[Double] into Array[Double] - * @param list - * @return */ def listToArrayDouble(list: JList[Double]): Array[Double] = { val r = new Array[Double](list.size) diff --git a/python/pyspark/rdd.py b/python/pyspark/rdd.py index f29af793737f8..a4cf1ead5bfa9 100644 --- a/python/pyspark/rdd.py +++ b/python/pyspark/rdd.py @@ -322,7 +322,7 @@ def randomSplit(self, weights, seed=None): :param weights: weights for splits, will be normalized if they don't sum to 1 :param seed: random seed - :return: split RDDs in an list + :return: split RDDs in a list >>> rdd = sc.parallelize(range(10), 1) >>> rdd1, rdd2, rdd3 = rdd.randomSplit([0.4, 0.6, 1.0], 11)