Skip to content

Commit

Permalink
Merge pull request #1625 from apache/master
Browse files Browse the repository at this point in the history
Create a new pull request by comparing changes across two branches
  • Loading branch information
GulajavaMinistudio authored Feb 28, 2024
2 parents 5d0d07c + 5d69388 commit 3efb087
Show file tree
Hide file tree
Showing 404 changed files with 5,858 additions and 2,622 deletions.
1 change: 0 additions & 1 deletion .github/labeler.yml
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,6 @@ INFRA:
- changed-files:
- any-glob-to-any-file: [
'.github/**/*',
'appveyor.yml',
'tools/**/*',
'dev/create-release/**/*',
'.asf.yaml',
Expand Down
5 changes: 3 additions & 2 deletions .github/workflows/build_and_test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -456,7 +456,7 @@ jobs:
fi
- name: Upload coverage to Codecov
if: fromJSON(inputs.envs).PYSPARK_CODECOV == 'true'
uses: codecov/codecov-action@v2
uses: codecov/codecov-action@v4
with:
files: ./python/coverage.xml
flags: unittests
Expand Down Expand Up @@ -574,8 +574,9 @@ jobs:
git -c user.name='Apache Spark Test Account' -c user.email='sparktestacc@gmail.com' merge --no-commit --progress --squash FETCH_HEAD
git -c user.name='Apache Spark Test Account' -c user.email='sparktestacc@gmail.com' commit -m "Merged commit" --allow-empty
- name: Install Buf
uses: bufbuild/buf-setup-action@v1
uses: bufbuild/buf-setup-action@v1.29.0
with:
version: 1.29.0
github_token: ${{ secrets.GITHUB_TOKEN }}
- name: Protocol Buffers Linter
uses: bufbuild/buf-lint-action@v1
Expand Down
1 change: 1 addition & 0 deletions .github/workflows/build_coverage.yml
Original file line number Diff line number Diff line change
Expand Up @@ -36,6 +36,7 @@ jobs:
hadoop: hadoop3
envs: >-
{
"PYTHON_TO_TEST": "python3.11",
"PYSPARK_CODECOV": "true"
}
jobs: >-
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/build_maven.yml
Original file line number Diff line number Diff line change
Expand Up @@ -33,5 +33,5 @@ jobs:
with:
envs: >-
{
"SKIP_SPARK_RELEASE_VERSIONS": "3.3.4,3.4.2,3.5.0"
"SKIP_SPARK_RELEASE_VERSIONS": "3.4.2"
}
4 changes: 4 additions & 0 deletions .github/workflows/build_maven_java21_macos14.yml
Original file line number Diff line number Diff line change
Expand Up @@ -33,3 +33,7 @@ jobs:
with:
java: 21
os: macos-14
envs: >-
{
"OBJC_DISABLE_INITIALIZE_FORK_SAFETY": "YES"
}
91 changes: 91 additions & 0 deletions .github/workflows/build_sparkr_window.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,91 @@
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
#
name: "Build / SparkR-only (master, 4.3.2, windows-2019)"

on:
schedule:
- cron: '0 17 * * *'

jobs:
build:
name: "Build module: sparkr"
runs-on: windows-2019
timeout-minutes: 300
if: github.repository == 'apache/spark'
steps:
- name: Download winutils Hadoop binary
uses: actions/checkout@v4
with:
repository: cdarlint/winutils
- name: Move Hadoop winutil into home directory
run: |
Move-Item -Path hadoop-3.3.5 -Destination ~\
- name: Checkout Spark repository
uses: actions/checkout@v4
- name: Cache Maven local repository
uses: actions/cache@v4
with:
path: ~/.m2/repository
key: build-sparkr-windows-maven-${{ hashFiles('**/pom.xml') }}
restore-keys: |
build-sparkr-windows-maven-
- name: Install Java 17
uses: actions/setup-java@v4
with:
distribution: zulu
java-version: 17
- name: Install R 4.3.2
uses: r-lib/actions/setup-r@v2
with:
r-version: 4.3.2
- name: Install R dependencies
run: |
Rscript -e "install.packages(c('knitr', 'rmarkdown', 'testthat', 'e1071', 'survival', 'arrow', 'xml2'), repos='https://cloud.r-project.org/')"
Rscript -e "pkg_list <- as.data.frame(installed.packages()[,c(1, 3:4)]); pkg_list[is.na(pkg_list$Priority), 1:2, drop = FALSE]"
shell: cmd
# SparkR build does not need Python. However, it shows warnings when the Python version is too low during
# the attempt to look up Python Data Sources for session initialization. The Windows 2019 runner
# includes Python 3.7, which Spark does not support. Therefore, we simply install the proper Python
# for simplicity, see SPARK-47116.
- name: Install Python 3.11
uses: actions/setup-python@v5
with:
python-version: '3.11'
architecture: x64
- name: Build Spark
run: |
rem 1. '-Djna.nosys=true' is required to avoid kernel32.dll load failure.
rem See SPARK-28759.
rem 2. Ideally we should check the tests related to Hive in SparkR as well (SPARK-31745).
rem 3. setup-java installs Maven 3.8.7 but does not allow changing its version, so overwrite
rem Maven version as a workaround.
mvn -DskipTests -Psparkr -Djna.nosys=true package -Dmaven.version=3.8.7
shell: cmd
- name: Run SparkR tests
run: |
set HADOOP_HOME=%USERPROFILE%\hadoop-3.3.5
set PATH=%HADOOP_HOME%\bin;%PATH%
.\bin\spark-submit2.cmd --driver-java-options "-Dlog4j.configurationFile=file:///%CD:\=/%/R/log4j2.properties" --conf spark.hadoop.fs.defaultFS="file:///" R\pkg\tests\run-all.R
shell: cmd
env:
NOT_CRAN: true
# See SPARK-27848. Currently installing some dependent packages causes
# "(converted from warning) unable to identify current timezone 'C':" for an unknown reason.
# This environment variable works around to test SparkR against a higher version.
R_REMOTES_NO_ERRORS_FROM_WARNINGS: true
20 changes: 10 additions & 10 deletions .github/workflows/maven_test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -73,13 +73,19 @@ jobs:
connector#kafka-0-10,connector#kafka-0-10-sql,connector#kafka-0-10-token-provider,connector#spark-ganglia-lgpl,connector#protobuf,connector#avro
- >-
sql#api,sql#catalyst,resource-managers#yarn,resource-managers#kubernetes#core
- >-
connect
# Here, we split Hive and SQL tests into some of slow ones and the rest of them.
included-tags: [ "" ]
excluded-tags: [ "" ]
comment: [ "" ]
include:
# Connect tests
- modules: connect
java: ${{ inputs.java }}
hadoop: ${{ inputs.hadoop }}
hive: hive2.3
# TODO(SPARK-47110): Reenble AmmoniteTest tests in Maven builds
excluded-tags: org.apache.spark.tags.AmmoniteTest
comment: ""
# Hive tests
- modules: sql#hive
java: ${{ inputs.java }}
Expand Down Expand Up @@ -178,13 +184,7 @@ jobs:
# Run the tests.
- name: Run tests
env: ${{ fromJSON(inputs.envs) }}
# The command script takes different options ubuntu vs macos-14, see also SPARK-47095.
shell: '[[ "${{ inputs.os }}" == *"ubuntu"* ]] && script -q -e -c "bash {0}" || script -q -e "bash {0}"'
run: |
# Fix for TTY related issues when launching the Ammonite REPL in tests.
export TERM=vt100
# `set -e` to make the exit status as expected due to use `script -q -e -c` to run the commands
set -e
export MAVEN_OPTS="-Xss64m -Xmx4g -Xms4g -XX:ReservedCodeCacheSize=128m -Dorg.slf4j.simpleLogger.defaultLogLevel=WARN"
export MAVEN_CLI_OPTS="--no-transfer-progress"
export JAVA_VERSION=${{ matrix.java }}
Expand All @@ -193,10 +193,10 @@ jobs:
./build/mvn $MAVEN_CLI_OPTS -DskipTests -Pyarn -Pkubernetes -Pvolcano -Phive -Phive-thriftserver -Phadoop-cloud -Pspark-ganglia-lgpl -Djava.version=${JAVA_VERSION/-ea} clean install
if [[ "$INCLUDED_TAGS" != "" ]]; then
./build/mvn $MAVEN_CLI_OPTS -pl "$TEST_MODULES" -Pyarn -Pkubernetes -Pvolcano -Phive -Phive-thriftserver -Phadoop-cloud -Pspark-ganglia-lgpl -Djava.version=${JAVA_VERSION/-ea} -Dtest.include.tags="$INCLUDED_TAGS" test -fae
elif [[ "$MODULES_TO_TEST" == "connect" ]]; then
./build/mvn $MAVEN_CLI_OPTS -Dtest.exclude.tags="$EXCLUDED_TAGS" -Djava.version=${JAVA_VERSION/-ea} -pl connector/connect/client/jvm,connector/connect/common,connector/connect/server test -fae
elif [[ "$EXCLUDED_TAGS" != "" ]]; then
./build/mvn $MAVEN_CLI_OPTS -pl "$TEST_MODULES" -Pyarn -Pkubernetes -Pvolcano -Phive -Phive-thriftserver -Phadoop-cloud -Pspark-ganglia-lgpl -Djava.version=${JAVA_VERSION/-ea} -Dtest.exclude.tags="$EXCLUDED_TAGS" test -fae
elif [[ "$MODULES_TO_TEST" == "connect" ]]; then
./build/mvn $MAVEN_CLI_OPTS -Djava.version=${JAVA_VERSION/-ea} -pl connector/connect/client/jvm,connector/connect/common,connector/connect/server test -fae
elif [[ "$MODULES_TO_TEST" == *"sql#hive-thriftserver"* ]]; then
# To avoid a compilation loop, for the `sql/hive-thriftserver` module, run `clean install` instead
./build/mvn $MAVEN_CLI_OPTS -pl "$TEST_MODULES" -Pyarn -Pkubernetes -Pvolcano -Phive -Phive-thriftserver -Phadoop-cloud -Pspark-ganglia-lgpl -Djava.version=${JAVA_VERSION/-ea} clean install -fae
Expand Down
1 change: 0 additions & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,6 @@ and Structured Streaming for stream processing.
<https://spark.apache.org/>

[![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_main.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_main.yml)
[![AppVeyor Build](https://img.shields.io/appveyor/ci/ApacheSoftwareFoundation/spark/master.svg?style=plastic&logo=appveyor)](https://ci.appveyor.com/project/ApacheSoftwareFoundation/spark)
[![PySpark Coverage](https://codecov.io/gh/apache/spark/branch/master/graph/badge.svg)](https://codecov.io/gh/apache/spark)
[![PyPI Downloads](https://static.pepy.tech/personalized-badge/pyspark?period=month&units=international_system&left_color=black&right_color=orange&left_text=PyPI%20downloads)](https://pypi.org/project/pyspark/)

Expand Down
75 changes: 0 additions & 75 deletions appveyor.yml

This file was deleted.

7 changes: 7 additions & 0 deletions assembly/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -266,6 +266,13 @@
<id>hive-provided</id>
<properties>
<hive.deps.scope>provided</hive.deps.scope>
<hive.jackson.scope>provided</hive.jackson.scope>
</properties>
</profile>
<profile>
<id>hive-jackson-provided</id>
<properties>
<hive.jackson.scope>provided</hive.jackson.scope>
</properties>
</profile>
<profile>
Expand Down
Loading

0 comments on commit 3efb087

Please sign in to comment.