Skip to content

Commit

Permalink
Make it neater
Browse files Browse the repository at this point in the history
  • Loading branch information
HyukjinKwon committed Jul 9, 2020
1 parent ba8029e commit a267705
Show file tree
Hide file tree
Showing 3 changed files with 10 additions and 8 deletions.
15 changes: 8 additions & 7 deletions .github/workflows/master.yml
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ on:
- master

jobs:
# TODO(SPARK-32248): Recover JDK 11 builds
# Build: build Spark and run the tests for specified modules.
build:
name: "Build modules: ${{ matrix.modules }} ${{ matrix.comment }} (JDK ${{ matrix.java }}, ${{ matrix.hadoop }}, ${{ matrix.hive }})"
Expand All @@ -22,7 +23,7 @@ jobs:
- hadoop3.2
hive:
- hive2.3
# TODO(SPARK-XXXXX): We don't test 'streaming-kinesis-asl' for now.
# TODO(SPARK-32246): We don't test 'streaming-kinesis-asl' for now.
# Kinesis tests depends on external Amazon kinesis service.
# Note that the modules below are from sparktestsupport/modules.py.
modules:
Expand Down Expand Up @@ -50,20 +51,20 @@ jobs:
# Here, we split Hive tests into some of heavy ones and the rest of them.
included-tags: [""]
excluded-tags: [""]
comment: ["- running all tests"]
comment: ["- all tests"]
include:
- modules: hive
java: 1.8
hadoop: hadoop3.2
hive: hive2.3
included-tags: org.apache.spark.tags.HeavyHiveTest
comment: "- running heavy tests"
comment: "- heavy tests"
- modules: hive
java: 1.8
hadoop: hadoop3.2
hive: hive2.3
excluded-tags: org.apache.spark.tags.HeavyHiveTest
comment: "- running non-heavy tests"
comment: "- light tests"
env:
TEST_ONLY_MODULES: ${{ matrix.modules }}
HADOOP_PROFILE: ${{ matrix.hadoop }}
Expand Down Expand Up @@ -106,7 +107,8 @@ jobs:
# PySpark
- name: Install PyPy3
# SQL component also has Python related tests, for example, IntegratedUDFTestUtils.
# Note that order here matters because default python3 is overridden by pypy3.
# Note that order of Python installations here matters because default python3 is
# overridden by pypy3.
uses: actions/setup-python@v2
if: contains(matrix.modules, 'pyspark') || matrix.modules == 'sql'
with:
Expand All @@ -127,13 +129,12 @@ jobs:
- name: Install Python packages
if: contains(matrix.modules, 'pyspark') || matrix.modules == 'sql'
# PyArrow is not supported in PyPy yet, see ARROW-2651.
# scipy installation with PyPy fails for an unknown reason.
# TODO(SPARK-32247): scipy installation with PyPy fails for an unknown reason.
run: |
python3 -m pip install numpy pyarrow pandas scipy
python3 -m pip list
python2 -m pip install numpy pyarrow pandas scipy
python2 -m pip list
# Installing NumPy is flaky in PyPy.
pypy3 -m pip install numpy pandas
pypy3 -m pip list
# SparkR
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -685,6 +685,7 @@ class MasterSuite extends SparkFunSuite
}
}

// TODO(SPARK-32250): Enable the test back. It is flaky in GitHub Actions.
ignore("SPARK-27510: Master should avoid dead loop while launching executor failed in Worker") {
val master = makeAliveMaster()
var worker: MockExecutorLaunchFailWorker = null
Expand Down
2 changes: 1 addition & 1 deletion dev/run-tests.py
Original file line number Diff line number Diff line change
Expand Up @@ -702,7 +702,7 @@ def main():

def _test():
if "TEST_ONLY_MODULES" in os.environ:
# Do not do anything except testing the targeted modules.
# TODO(SPARK-32252): Enable doctests back in Github Actions.
return

import doctest
Expand Down

0 comments on commit a267705

Please sign in to comment.