diff --git a/dev/lint-java b/dev/lint-java index 2f9b09a61575c..08992f42e6e4c 100755 --- a/dev/lint-java +++ b/dev/lint-java @@ -20,7 +20,7 @@ SCRIPT_DIR="$( cd "$( dirname "$0" )" && pwd )" SPARK_ROOT_DIR="$(dirname $SCRIPT_DIR)" -ERRORS=$($SCRIPT_DIR/../build/mvn -Pkubernetes -Pkinesis-asl -Pmesos -Pyarn -Phive -Phive-thriftserver checkstyle:check | grep ERROR) +ERRORS=$($SCRIPT_DIR/../build/mvn -Phadoop-cloud -Pkubernetes -Pkinesis-asl -Pmesos -Pyarn -Phive -Phive-thriftserver checkstyle:check | grep ERROR) if test ! -z "$ERRORS"; then echo -e "Checkstyle checks failed at following occurrences:\n$ERRORS" diff --git a/dev/make-distribution.sh b/dev/make-distribution.sh index 5c767b3118252..ea10fb850ad2f 100755 --- a/dev/make-distribution.sh +++ b/dev/make-distribution.sh @@ -142,7 +142,7 @@ fi echo "Spark version is $VERSION" VERSION_SET=$("$MVN" versions:set -DnewVersion=$VERSION | tail -n 1) if [ "$MAKE_TGZ" == "true" ]; then - echo "Making spark-dist-$VERSION-$NAME.tgz" + echo "Making spark-dist_2.11-$NAME-$VERSION.tgz" else echo "Making distribution for Spark $VERSION in '$DISTDIR'..." fi diff --git a/dev/publish-local.sh b/dev/publish-local.sh index ed5ddb6339e46..908bde4bb1068 100755 --- a/dev/publish-local.sh +++ b/dev/publish-local.sh @@ -1,31 +1,26 @@ #!/usr/bin/env bash set -euo pipefail -VERSION=$(git describe --tags) -HADOOP_VERSION=$(./build/mvn help:evaluate -Phadoop-palantir -Dexpression=hadoop.version 2>/dev/null\ - | grep -v "INFO"\ - | tail -n 1) - +version=$(git describe --tags) PALANTIR_FLAGS=(-Phadoop-cloud -Phadoop-palantir -Pkinesis-asl -Pkubernetes -Phive -Pyarn -Psparkr) MVN_LOCAL="~/.m2/repository" publish_artifacts() { - ./build/mvn versions:set -DnewVersion=$VERSION + ./build/mvn versions:set -DnewVersion=$version ./build/mvn -DskipTests "${PALANTIR_FLAGS[@]}" install clean } make_dist() { - dist_name="$1" - build_flags="$2" - shift 2 - dist_version="${VERSION}-${dist_name}" - file_name="spark-dist-${dist_version}.tgz" - ./dev/make-distribution.sh --name $dist_name --tgz "$@" $build_flags - mkdir -p $MVN_LOCAL/org/apache/spark/spark-dist/${dist_version} && \ - cp $file_name $MVN_LOCAL/org/apache/spark/spark-dist/${dist_version}/ + build_flags="$1" + shift 1 + artifact_name="spark-dist_2.11-hadoop-palantir" + file_name="${artifact_name}-${version}.tgz" + ./dev/make-distribution.sh --name "hadoop-palantir" --tgz "$@" $build_flags + mkdir -p $MVN_LOCAL/org/apache/spark/${artifact_name}/${version} && \ + cp $file_name $MVN_LOCAL/org/apache/spark/${artifact_name}/${version}/ } publish_artifacts -make_dist hadoop-$HADOOP_VERSION "${PALANTIR_FLAGS[*]}" +make_dist "${PALANTIR_FLAGS[*]}" diff --git a/dev/publish.sh b/dev/publish.sh index fbab21ef84627..b575d9c73c082 100755 --- a/dev/publish.sh +++ b/dev/publish.sh @@ -17,15 +17,13 @@ publish_artifacts() { } make_dist() { - dist_name="$1" - build_flags="$2" - shift 2 - dist_version="${version}-${dist_name}" - file_name="spark-dist-${dist_version}.tgz" - ./dev/make-distribution.sh --name $dist_name --tgz "$@" $build_flags - curl -u $BINTRAY_USERNAME:$BINTRAY_PASSWORD -T $file_name "https://api.bintray.com/content/palantir/releases/spark/${version}/org/apache/spark/spark-dist/${dist_version}/${file_name}" + build_flags="$1" + shift 1 + artifact_name="spark-dist_2.11-hadoop-palantir" + file_name="${artifact_name}-${version}.tgz" + ./dev/make-distribution.sh --name "hadoop-palantir" --tgz "$@" $build_flags + curl -u $BINTRAY_USERNAME:$BINTRAY_PASSWORD -T $file_name "https://api.bintray.com/content/palantir/releases/spark/${version}/org/apache/spark/${artifact_name}/${version}/${file_name}" } publish_artifacts -make_dist hadoop-2.8.0-palantir.8 "${PALANTIR_FLAGS[*]}" --clean -make_dist without-hadoop "-Phadoop-provided -Pkubernetes -Phive -Pyarn -Psparkr" --clean +make_dist "${PALANTIR_FLAGS[*]}" --clean diff --git a/dev/sparktestsupport/modules.py b/dev/sparktestsupport/modules.py index 9bbc9b9a84cb6..34ace5e64b001 100644 --- a/dev/sparktestsupport/modules.py +++ b/dev/sparktestsupport/modules.py @@ -521,6 +521,14 @@ def __hash__(self): sbt_test_goals=["kubernetes/test"] ) +cloud = Module( + name="hadoop-cloud", + dependencies=[], + source_file_regexes=["hadoop-cloud/"], + build_profile_flags=["-Phadoop-cloud"], + sbt_test_goals=["hadoop-cloud/test"] +) + # The root module is a dummy module which is used to run all of the tests. # No other modules should directly depend on this module. root = Module( diff --git a/dists/without-hadoop/pom.xml b/dists/without-hadoop/pom.xml deleted file mode 100644 index d41de26ebdc18..0000000000000 --- a/dists/without-hadoop/pom.xml +++ /dev/null @@ -1,168 +0,0 @@ - - - - - 4.0.0 - - org.apache.spark - spark-parent_2.11 - 2.3.0-SNAPSHOT - ../../pom.xml - - - spark-dist_2.11-without-hadoop - Spark Project Dist Provided Hadoop - http://spark.apache.org/ - pom - - - spark-dist-hadoop-palantir - none - none - provided - - - - - - org.apache.spark - spark-core_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-mllib_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-streaming_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-graphx_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-sql_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-repl_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-yarn_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-mesos_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-kubernetes_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-hive_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-hive-thriftserver_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-ganglia-lgpl_${scala.binary.version} - ${project.version} - - - org.apache.spark - spark-dist_${scala.binary.version}-without-hadoop - ${project.version} - - - org.apache.spark - spark-dist - ${project.version}-without-hadoop - - - - - - - - org.spark-project.spark - unused - 1.0.0 - provided - - - org.apache.spark - spark-core_${scala.binary.version} - - - org.apache.spark - spark-mllib_${scala.binary.version} - - - org.apache.spark - spark-streaming_${scala.binary.version} - - - org.apache.spark - spark-graphx_${scala.binary.version} - - - org.apache.spark - spark-sql_${scala.binary.version} - - - - - com.google.guava - guava - ${hadoop.deps.scope} - - - - org.apache.spark - spark-yarn_${scala.binary.version} - - - org.apache.spark - spark-kubernetes_${scala.binary.version} - - - com.palantir.spark.influx - spark-influx-sink - - - diff --git a/pom.xml b/pom.xml index 67764aeafbc85..f057e8b034b5d 100644 --- a/pom.xml +++ b/pom.xml @@ -107,7 +107,6 @@ external/kafka-0-10 external/kafka-0-10-assembly external/kafka-0-10-sql - dists/without-hadoop dists/hadoop-palantir