Skip to content

Commit a267705

Browse files
committed
Make it neater
1 parent ba8029e commit a267705

File tree

3 files changed

+10
-8
lines changed

3 files changed

+10
-8
lines changed

.github/workflows/master.yml

Lines changed: 8 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,7 @@ on:
99
- master
1010

1111
jobs:
12+
# TODO(SPARK-32248): Recover JDK 11 builds
1213
# Build: build Spark and run the tests for specified modules.
1314
build:
1415
name: "Build modules: ${{ matrix.modules }} ${{ matrix.comment }} (JDK ${{ matrix.java }}, ${{ matrix.hadoop }}, ${{ matrix.hive }})"
@@ -22,7 +23,7 @@ jobs:
2223
- hadoop3.2
2324
hive:
2425
- hive2.3
25-
# TODO(SPARK-XXXXX): We don't test 'streaming-kinesis-asl' for now.
26+
# TODO(SPARK-32246): We don't test 'streaming-kinesis-asl' for now.
2627
# Kinesis tests depends on external Amazon kinesis service.
2728
# Note that the modules below are from sparktestsupport/modules.py.
2829
modules:
@@ -50,20 +51,20 @@ jobs:
5051
# Here, we split Hive tests into some of heavy ones and the rest of them.
5152
included-tags: [""]
5253
excluded-tags: [""]
53-
comment: ["- running all tests"]
54+
comment: ["- all tests"]
5455
include:
5556
- modules: hive
5657
java: 1.8
5758
hadoop: hadoop3.2
5859
hive: hive2.3
5960
included-tags: org.apache.spark.tags.HeavyHiveTest
60-
comment: "- running heavy tests"
61+
comment: "- heavy tests"
6162
- modules: hive
6263
java: 1.8
6364
hadoop: hadoop3.2
6465
hive: hive2.3
6566
excluded-tags: org.apache.spark.tags.HeavyHiveTest
66-
comment: "- running non-heavy tests"
67+
comment: "- light tests"
6768
env:
6869
TEST_ONLY_MODULES: ${{ matrix.modules }}
6970
HADOOP_PROFILE: ${{ matrix.hadoop }}
@@ -106,7 +107,8 @@ jobs:
106107
# PySpark
107108
- name: Install PyPy3
108109
# SQL component also has Python related tests, for example, IntegratedUDFTestUtils.
109-
# Note that order here matters because default python3 is overridden by pypy3.
110+
# Note that order of Python installations here matters because default python3 is
111+
# overridden by pypy3.
110112
uses: actions/setup-python@v2
111113
if: contains(matrix.modules, 'pyspark') || matrix.modules == 'sql'
112114
with:
@@ -127,13 +129,12 @@ jobs:
127129
- name: Install Python packages
128130
if: contains(matrix.modules, 'pyspark') || matrix.modules == 'sql'
129131
# PyArrow is not supported in PyPy yet, see ARROW-2651.
130-
# scipy installation with PyPy fails for an unknown reason.
132+
# TODO(SPARK-32247): scipy installation with PyPy fails for an unknown reason.
131133
run: |
132134
python3 -m pip install numpy pyarrow pandas scipy
133135
python3 -m pip list
134136
python2 -m pip install numpy pyarrow pandas scipy
135137
python2 -m pip list
136-
# Installing NumPy is flaky in PyPy.
137138
pypy3 -m pip install numpy pandas
138139
pypy3 -m pip list
139140
# SparkR

core/src/test/scala/org/apache/spark/deploy/master/MasterSuite.scala

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -685,6 +685,7 @@ class MasterSuite extends SparkFunSuite
685685
}
686686
}
687687

688+
// TODO(SPARK-32250): Enable the test back. It is flaky in GitHub Actions.
688689
ignore("SPARK-27510: Master should avoid dead loop while launching executor failed in Worker") {
689690
val master = makeAliveMaster()
690691
var worker: MockExecutorLaunchFailWorker = null

dev/run-tests.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -702,7 +702,7 @@ def main():
702702

703703
def _test():
704704
if "TEST_ONLY_MODULES" in os.environ:
705-
# Do not do anything except testing the targeted modules.
705+
# TODO(SPARK-32252): Enable doctests back in Github Actions.
706706
return
707707

708708
import doctest

0 commit comments

Comments
 (0)