Skip to content

Commit

Permalink
Merge branch 'branch-1.5' of github.com:apache/spark into csd-1.5
Browse files Browse the repository at this point in the history
  • Loading branch information
markhamstra committed Oct 26, 2015
2 parents 87ce08a + a355d0d commit ca12e62
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 1 deletion.
2 changes: 2 additions & 0 deletions core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala
Original file line number Diff line number Diff line change
Expand Up @@ -328,6 +328,8 @@ object SparkSubmit {
case (STANDALONE, CLUSTER) if args.isR =>
printErrorAndExit("Cluster deploy mode is currently not supported for R " +
"applications on standalone clusters.")
case (LOCAL, CLUSTER) =>
printErrorAndExit("Cluster deploy mode is not compatible with master \"local\"")
case (_, CLUSTER) if isShell(args.primaryResource) =>
printErrorAndExit("Cluster deploy mode is not applicable to Spark shells.")
case (_, CLUSTER) if isSqlShell(args.mainClass) =>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -48,8 +48,9 @@ private[spark] object TestClient {
val url = args(0)
val conf = new SparkConf
val rpcEnv = RpcEnv.create("spark", Utils.localHostName(), 0, conf, new SecurityManager(conf))
val executorClassname = TestExecutor.getClass.getCanonicalName.stripSuffix("$")
val desc = new ApplicationDescription("TestClient", Some(1), 512,
Command("spark.deploy.client.TestExecutor", Seq(), Map(), Seq(), Seq(), Seq()), "ignored")
Command(executorClassname, Seq(), Map(), Seq(), Seq(), Seq()), "ignored")
val listener = new TestListener
val client = new AppClient(rpcEnv, Array(url), desc, listener, new SparkConf)
client.start()
Expand Down

0 comments on commit ca12e62

Please sign in to comment.