99 - master
1010
1111jobs :
12+ # TODO(SPARK-32248): Recover JDK 11 builds
1213 # Build: build Spark and run the tests for specified modules.
1314 build :
1415 name : " Build modules: ${{ matrix.modules }} ${{ matrix.comment }} (JDK ${{ matrix.java }}, ${{ matrix.hadoop }}, ${{ matrix.hive }})"
2223 - hadoop3.2
2324 hive :
2425 - hive2.3
25- # TODO(SPARK-XXXXX ): We don't test 'streaming-kinesis-asl' for now.
26+ # TODO(SPARK-32246 ): We don't test 'streaming-kinesis-asl' for now.
2627 # Kinesis tests depends on external Amazon kinesis service.
2728 # Note that the modules below are from sparktestsupport/modules.py.
2829 modules :
@@ -50,20 +51,20 @@ jobs:
5051 # Here, we split Hive tests into some of heavy ones and the rest of them.
5152 included-tags : [""]
5253 excluded-tags : [""]
53- comment : ["- running all tests"]
54+ comment : ["- all tests"]
5455 include :
5556 - modules : hive
5657 java : 1.8
5758 hadoop : hadoop3.2
5859 hive : hive2.3
5960 included-tags : org.apache.spark.tags.HeavyHiveTest
60- comment : " - running heavy tests"
61+ comment : " - heavy tests"
6162 - modules : hive
6263 java : 1.8
6364 hadoop : hadoop3.2
6465 hive : hive2.3
6566 excluded-tags : org.apache.spark.tags.HeavyHiveTest
66- comment : " - running non-heavy tests"
67+ comment : " - light tests"
6768 env :
6869 TEST_ONLY_MODULES : ${{ matrix.modules }}
6970 HADOOP_PROFILE : ${{ matrix.hadoop }}
@@ -106,7 +107,8 @@ jobs:
106107 # PySpark
107108 - name : Install PyPy3
108109 # SQL component also has Python related tests, for example, IntegratedUDFTestUtils.
109- # Note that order here matters because default python3 is overridden by pypy3.
110+ # Note that order of Python installations here matters because default python3 is
111+ # overridden by pypy3.
110112 uses : actions/setup-python@v2
111113 if : contains(matrix.modules, 'pyspark') || matrix.modules == 'sql'
112114 with :
@@ -127,13 +129,12 @@ jobs:
127129 - name : Install Python packages
128130 if : contains(matrix.modules, 'pyspark') || matrix.modules == 'sql'
129131 # PyArrow is not supported in PyPy yet, see ARROW-2651.
130- # scipy installation with PyPy fails for an unknown reason.
132+ # TODO(SPARK-32247): scipy installation with PyPy fails for an unknown reason.
131133 run : |
132134 python3 -m pip install numpy pyarrow pandas scipy
133135 python3 -m pip list
134136 python2 -m pip install numpy pyarrow pandas scipy
135137 python2 -m pip list
136- # Installing NumPy is flaky in PyPy.
137138 pypy3 -m pip install numpy pandas
138139 pypy3 -m pip list
139140 # SparkR
0 commit comments