Skip to content

Commit

Permalink
review comment
Browse files Browse the repository at this point in the history
  • Loading branch information
liguoqiang committed Mar 5, 2014
1 parent d0a6005 commit b0d5c07
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions core/src/main/scala/org/apache/spark/SparkContext.scala
Original file line number Diff line number Diff line change
Expand Up @@ -848,7 +848,7 @@ class SparkContext(
allowLocal: Boolean,
resultHandler: (Int, U) => Unit) {
val outIndex = partitions.toSet.diff(rdd.partitions.map(_.index).toSet)
require(outIndex.isEmpty,"Partition index out of bounds: "+ outIndex.mkString(","))
require(outIndex.isEmpty,"Partition index out of bounds: " + outIndex.mkString(","))
val callSite = getCallSite
val cleanedFunc = clean(func)
logInfo("Starting job: " + callSite)
Expand Down Expand Up @@ -953,7 +953,7 @@ class SparkContext(
resultFunc: => R): SimpleFutureAction[R] =
{
val outIndex = partitions.toSet.diff(rdd.partitions.map(_.index).toSet)
require(outIndex.isEmpty,"Partition index out of bounds: "+ outIndex.mkString(","))
require(outIndex.isEmpty,"Partition index out of bounds: " + outIndex.mkString(","))
val cleanF = clean(processPartition)
val callSite = getCallSite
val waiter = dagScheduler.submitJob(
Expand Down

0 comments on commit b0d5c07

Please sign in to comment.