diff --git a/dev/lint-java b/dev/lint-java
index 2f9b09a61575c..08992f42e6e4c 100755
--- a/dev/lint-java
+++ b/dev/lint-java
@@ -20,7 +20,7 @@
SCRIPT_DIR="$( cd "$( dirname "$0" )" && pwd )"
SPARK_ROOT_DIR="$(dirname $SCRIPT_DIR)"
-ERRORS=$($SCRIPT_DIR/../build/mvn -Pkubernetes -Pkinesis-asl -Pmesos -Pyarn -Phive -Phive-thriftserver checkstyle:check | grep ERROR)
+ERRORS=$($SCRIPT_DIR/../build/mvn -Phadoop-cloud -Pkubernetes -Pkinesis-asl -Pmesos -Pyarn -Phive -Phive-thriftserver checkstyle:check | grep ERROR)
if test ! -z "$ERRORS"; then
echo -e "Checkstyle checks failed at following occurrences:\n$ERRORS"
diff --git a/dev/make-distribution.sh b/dev/make-distribution.sh
index 5c767b3118252..ea10fb850ad2f 100755
--- a/dev/make-distribution.sh
+++ b/dev/make-distribution.sh
@@ -142,7 +142,7 @@ fi
echo "Spark version is $VERSION"
VERSION_SET=$("$MVN" versions:set -DnewVersion=$VERSION | tail -n 1)
if [ "$MAKE_TGZ" == "true" ]; then
- echo "Making spark-dist-$VERSION-$NAME.tgz"
+ echo "Making spark-dist_2.11-$NAME-$VERSION.tgz"
else
echo "Making distribution for Spark $VERSION in '$DISTDIR'..."
fi
diff --git a/dev/publish-local.sh b/dev/publish-local.sh
index ed5ddb6339e46..908bde4bb1068 100755
--- a/dev/publish-local.sh
+++ b/dev/publish-local.sh
@@ -1,31 +1,26 @@
#!/usr/bin/env bash
set -euo pipefail
-VERSION=$(git describe --tags)
-HADOOP_VERSION=$(./build/mvn help:evaluate -Phadoop-palantir -Dexpression=hadoop.version 2>/dev/null\
- | grep -v "INFO"\
- | tail -n 1)
-
+version=$(git describe --tags)
PALANTIR_FLAGS=(-Phadoop-cloud -Phadoop-palantir -Pkinesis-asl -Pkubernetes -Phive -Pyarn -Psparkr)
MVN_LOCAL="~/.m2/repository"
publish_artifacts() {
- ./build/mvn versions:set -DnewVersion=$VERSION
+ ./build/mvn versions:set -DnewVersion=$version
./build/mvn -DskipTests "${PALANTIR_FLAGS[@]}" install clean
}
make_dist() {
- dist_name="$1"
- build_flags="$2"
- shift 2
- dist_version="${VERSION}-${dist_name}"
- file_name="spark-dist-${dist_version}.tgz"
- ./dev/make-distribution.sh --name $dist_name --tgz "$@" $build_flags
- mkdir -p $MVN_LOCAL/org/apache/spark/spark-dist/${dist_version} && \
- cp $file_name $MVN_LOCAL/org/apache/spark/spark-dist/${dist_version}/
+ build_flags="$1"
+ shift 1
+ artifact_name="spark-dist_2.11-hadoop-palantir"
+ file_name="${artifact_name}-${version}.tgz"
+ ./dev/make-distribution.sh --name "hadoop-palantir" --tgz "$@" $build_flags
+ mkdir -p $MVN_LOCAL/org/apache/spark/${artifact_name}/${version} && \
+ cp $file_name $MVN_LOCAL/org/apache/spark/${artifact_name}/${version}/
}
publish_artifacts
-make_dist hadoop-$HADOOP_VERSION "${PALANTIR_FLAGS[*]}"
+make_dist "${PALANTIR_FLAGS[*]}"
diff --git a/dev/publish.sh b/dev/publish.sh
index fbab21ef84627..b575d9c73c082 100755
--- a/dev/publish.sh
+++ b/dev/publish.sh
@@ -17,15 +17,13 @@ publish_artifacts() {
}
make_dist() {
- dist_name="$1"
- build_flags="$2"
- shift 2
- dist_version="${version}-${dist_name}"
- file_name="spark-dist-${dist_version}.tgz"
- ./dev/make-distribution.sh --name $dist_name --tgz "$@" $build_flags
- curl -u $BINTRAY_USERNAME:$BINTRAY_PASSWORD -T $file_name "https://api.bintray.com/content/palantir/releases/spark/${version}/org/apache/spark/spark-dist/${dist_version}/${file_name}"
+ build_flags="$1"
+ shift 1
+ artifact_name="spark-dist_2.11-hadoop-palantir"
+ file_name="${artifact_name}-${version}.tgz"
+ ./dev/make-distribution.sh --name "hadoop-palantir" --tgz "$@" $build_flags
+ curl -u $BINTRAY_USERNAME:$BINTRAY_PASSWORD -T $file_name "https://api.bintray.com/content/palantir/releases/spark/${version}/org/apache/spark/${artifact_name}/${version}/${file_name}"
}
publish_artifacts
-make_dist hadoop-2.8.0-palantir.8 "${PALANTIR_FLAGS[*]}" --clean
-make_dist without-hadoop "-Phadoop-provided -Pkubernetes -Phive -Pyarn -Psparkr" --clean
+make_dist "${PALANTIR_FLAGS[*]}" --clean
diff --git a/dev/sparktestsupport/modules.py b/dev/sparktestsupport/modules.py
index 9bbc9b9a84cb6..34ace5e64b001 100644
--- a/dev/sparktestsupport/modules.py
+++ b/dev/sparktestsupport/modules.py
@@ -521,6 +521,14 @@ def __hash__(self):
sbt_test_goals=["kubernetes/test"]
)
+cloud = Module(
+ name="hadoop-cloud",
+ dependencies=[],
+ source_file_regexes=["hadoop-cloud/"],
+ build_profile_flags=["-Phadoop-cloud"],
+ sbt_test_goals=["hadoop-cloud/test"]
+)
+
# The root module is a dummy module which is used to run all of the tests.
# No other modules should directly depend on this module.
root = Module(
diff --git a/dists/without-hadoop/pom.xml b/dists/without-hadoop/pom.xml
deleted file mode 100644
index d41de26ebdc18..0000000000000
--- a/dists/without-hadoop/pom.xml
+++ /dev/null
@@ -1,168 +0,0 @@
-
-
-
-
- 4.0.0
-
- org.apache.spark
- spark-parent_2.11
- 2.3.0-SNAPSHOT
- ../../pom.xml
-
-
- spark-dist_2.11-without-hadoop
- Spark Project Dist Provided Hadoop
- http://spark.apache.org/
- pom
-
-
- spark-dist-hadoop-palantir
- none
- none
- provided
-
-
-
-
-
- org.apache.spark
- spark-core_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-mllib_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-streaming_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-graphx_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-sql_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-repl_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-yarn_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-mesos_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-kubernetes_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-hive_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-hive-thriftserver_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-ganglia-lgpl_${scala.binary.version}
- ${project.version}
-
-
- org.apache.spark
- spark-dist_${scala.binary.version}-without-hadoop
- ${project.version}
-
-
- org.apache.spark
- spark-dist
- ${project.version}-without-hadoop
-
-
-
-
-
-
-
- org.spark-project.spark
- unused
- 1.0.0
- provided
-
-
- org.apache.spark
- spark-core_${scala.binary.version}
-
-
- org.apache.spark
- spark-mllib_${scala.binary.version}
-
-
- org.apache.spark
- spark-streaming_${scala.binary.version}
-
-
- org.apache.spark
- spark-graphx_${scala.binary.version}
-
-
- org.apache.spark
- spark-sql_${scala.binary.version}
-
-
-
-
- com.google.guava
- guava
- ${hadoop.deps.scope}
-
-
-
- org.apache.spark
- spark-yarn_${scala.binary.version}
-
-
- org.apache.spark
- spark-kubernetes_${scala.binary.version}
-
-
- com.palantir.spark.influx
- spark-influx-sink
-
-
-
diff --git a/pom.xml b/pom.xml
index 67764aeafbc85..f057e8b034b5d 100644
--- a/pom.xml
+++ b/pom.xml
@@ -107,7 +107,6 @@
external/kafka-0-10
external/kafka-0-10-assembly
external/kafka-0-10-sql
- dists/without-hadoop
dists/hadoop-palantir