From 277043eaac8f168da5d04715334fc7916afa577d Mon Sep 17 00:00:00 2001 From: Joshi Date: Tue, 30 Jun 2015 10:49:26 -0700 Subject: [PATCH] Fix for SparkContext stop behavior --- core/src/main/scala/org/apache/spark/SparkEnv.scala | 7 +------ 1 file changed, 1 insertion(+), 6 deletions(-) diff --git a/core/src/main/scala/org/apache/spark/SparkEnv.scala b/core/src/main/scala/org/apache/spark/SparkEnv.scala index e49779b3b094d..1b133fbdfaf59 100644 --- a/core/src/main/scala/org/apache/spark/SparkEnv.scala +++ b/core/src/main/scala/org/apache/spark/SparkEnv.scala @@ -22,7 +22,6 @@ import java.net.Socket import akka.actor.ActorSystem -import scala.collection.JavaConversions._ import scala.collection.mutable import scala.util.Properties @@ -45,8 +44,6 @@ import org.apache.spark.storage._ import org.apache.spark.unsafe.memory.{ExecutorMemoryManager, MemoryAllocator} import org.apache.spark.util.{RpcUtils, Utils} -import scala.util.control.NonFatal - /** * :: DeveloperApi :: * Holds all the runtime environment objects for a running Spark instance (either master or worker), @@ -95,8 +92,7 @@ class SparkEnv ( if (!isStopped) { isStopped = true - - pythonWorkers.foreach { case (key, worker) => worker.stop()} + pythonWorkers.values.foreach(_.stop()) Option(httpFileServer).foreach(_.stop()) mapOutputTracker.stop() shuffleManager.stop() @@ -107,7 +103,6 @@ class SparkEnv ( outputCommitCoordinator.stop() rpcEnv.shutdown() - // Unfortunately Akka's awaitTermination doesn't actually wait for the Netty server to shut // down, but let's call it anyway in case it gets fixed in a later release // UPDATE: In Akka 2.1.x, this hangs if there are remote actors, so we can't call it.