Skip to content

Dev 132930 spark image research #51075

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Draft
wants to merge 983 commits into
base: master
Choose a base branch
from
Draft
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
983 commits
Select commit Hold shift + click to select a range
145c045
[SPARK-47910][CORE] close stream when DiskBlockObjectWriter closeReso…
JacobZheng0927 Jun 18, 2024
f57af38
[SPARK-47910][CORE][3.5][FOLLOWUP] Remove the use of MDC
JacobZheng0927 Jun 19, 2024
4741f7a
[SPARK-48498][SQL][3.5] Always do char padding in predicates
jackylee-ch Jun 24, 2024
0db5bde
[SPARK-48639][CONNECT][PYTHON] Add Origin to Relation.RelationCommon
HyukjinKwon Jun 24, 2024
405b23b
[SPARK-47172][CORE][3.5] Add support for AES-GCM for RPC encryption
sweisdb Jun 27, 2024
e6d5ddd
Revert "[SPARK-48639][CONNECT][PYTHON] Add Origin to Relation.Relatio…
HyukjinKwon Jun 27, 2024
6cee8e1
[SPARK-47927][SQL][FOLLOWUP] fix ScalaUDF output nullability
cloud-fan Jun 27, 2024
d35af54
[SPARK-48428][SQL] Fix IllegalStateException in NestedColumnAliasing
Jun 27, 2024
789ac5b
[SPARK-46957][CORE] Decommission migrated shuffle files should be abl…
Ngone51 Jun 27, 2024
7aa12b6
Revert "[SPARK-46957][CORE] Decommission migrated shuffle files shoul…
yaooqinn Jun 27, 2024
b28ddb1
[SPARK-46957][CORE][3.5][3.4] Decommission migrated shuffle files sho…
Ngone51 Jun 27, 2024
3d7a20a
[SPARK-48709][SQL][3.5] Fix varchar type resolution mismatch for Data…
wangyum Jun 27, 2024
ade9dbf
[SPARK-48494][BUILD][3.5] Update `airlift:aircompressor` to 0.27
bjornjorgensen Jun 28, 2024
686f59c
[SPARK-48586][SS][3.5] Remove lock acquisition in doMaintenance() by …
riyaverm-db Jun 28, 2024
df70cc1
[SPARK-48292][CORE][3.5] Revert [SPARK-39195][SQL] Spark OutputCommit…
AngersZhuuuu Jul 1, 2024
ef4e456
[SPARK-48759][SQL] Add migration doc for CREATE TABLE AS SELECT behav…
asl3 Jul 2, 2024
44eba46
[SPARK-48710][PYTHON][3.5] Limit NumPy version to supported range (>=…
codesorcery Jul 3, 2024
1cc0043
[SPARK-48806][SQL] Pass actual exception when url_decode fails
wForget Jul 4, 2024
67047cd
[SPARK-48719][SQL][3.5] Fix the calculation bug of RegrSlope & RegrIn…
wayneguow Jul 8, 2024
1e15e3f
[SPARK-48843] Prevent infinite loop with BindParameters
nemanja-boric-databricks Jul 10, 2024
8afce30
[SPARK-48863][SQL] Fix ClassCastException when parsing JSON with "spa…
sadikovi Jul 11, 2024
2bb46e0
[MINOR][DOCS] Add `docs/_generated/` to .gitignore
yaooqinn Apr 23, 2024
d517a63
[MINOR][SQL][TESTS] Remove a duplicate test case in `CSVExprUtilsSuite`
wayneguow Jul 11, 2024
b15a872
[SPARK-48871] Fix INVALID_NON_DETERMINISTIC_EXPRESSIONS validation in…
zhipengmao-db Jul 12, 2024
596f680
[SPARK-48845][SQL] GenericUDF catch exceptions from children
jackylee-ch Jul 12, 2024
56dec39
[SPARK-48666][SQL] Do not push down filter if it contains PythonUDFs
HyukjinKwon Jul 12, 2024
4d1bbfd
[SPARK-47172][DOCS][FOLLOWUP] Fix spark.network.crypto.ciphersince ve…
yaooqinn Jul 16, 2024
f1f5bb6
[SPARK-47307][SQL][3.5] Add a config to optionally chunk base64 strings
wForget Jul 16, 2024
033f6a3
[SPARK-47307][DOCS][FOLLOWUP] Add a migration guide for the behavior …
wForget Jul 17, 2024
ef1f377
[SPARK-48889][SS] testStream to unload state stores before finishing
siying Jul 17, 2024
443825a
[SPARK-48920][BUILD][3.5] Upgrade ORC to 1.9.4
williamhyun Jul 17, 2024
44f8766
[SPARK-48930][CORE] Redact `awsAccessKeyId` by including `accesskey` …
dongjoon-hyun Jul 17, 2024
b151012
Preparing Spark release v3.5.2-rc1
yaooqinn Jul 18, 2024
d556740
Preparing development version 3.5.3-SNAPSHOT
yaooqinn Jul 18, 2024
f07a547
[SPARK-48791][CORE][3.5] Fix perf regression caused by the accumulato…
Ngone51 Jul 18, 2024
e247915
[SPARK-48921][SQL][3.5] ScalaUDF encoders in subquery should be resol…
viirya Jul 18, 2024
66dce6d
[SPARK-48934][SS] Python datetime types converted incorrectly for set…
siying Jul 18, 2024
5a09a7d
Revert "[SPARK-47307][DOCS][FOLLOWUP] Add a migration guide for the b…
yaooqinn Jul 19, 2024
28d33e3
[SPARK-48498][SQL][FOLLOWUP] do padding for char-char comparison
cloud-fan Jul 19, 2024
ea53ea7
[SPARK-47307][SQL][FOLLOWUP][3.5] Promote spark.sql.legacy.chunkBase6…
wForget Jul 20, 2024
6d8f511
Preparing Spark release v3.5.2-rc2
yaooqinn Jul 22, 2024
9574cec
Preparing development version 3.5.3-SNAPSHOT
yaooqinn Jul 22, 2024
ada5dc2
[SPARK-48991][SQL] Move path initialization into try-catch block in F…
yaooqinn Jul 24, 2024
66b7cb9
[SPARK-48089][SS][CONNECT][FOLLOWUP][3.5] Disable Server Listener fai…
WweiL Jul 25, 2024
c4ef321
[SPARK-48308][CORE][3.5] Unify getting data schema without partition …
johanl-db Jul 25, 2024
ebda6a6
Preparing Spark release v3.5.2-rc3
yaooqinn Jul 25, 2024
61a90e5
Preparing development version 3.5.3-SNAPSHOT
yaooqinn Jul 25, 2024
4c38e90
[SPARK-49005][K8S][3.5] Use `17-jammy` tag instead of `17` to prevent…
dongjoon-hyun Jul 25, 2024
1edbddf
Preparing Spark release v3.5.2-rc4
yaooqinn Jul 26, 2024
75029e1
Preparing development version 3.5.3-SNAPSHOT
yaooqinn Jul 26, 2024
268c13e
[SPARK-49054][SQL][3.5] Column default value should support current_*…
gengliangwang Jul 30, 2024
c6df890
[SPARK-49000][SQL] Fix "select count(distinct 1) from t" where t is e…
uros-db Jul 31, 2024
a0f88f5
[SPARK-44638][SQL][TESTS][3.5] Add test for Char/Varchar in JDBC cust…
yaooqinn Jul 31, 2024
d23b70b
[SPARK-49066][SQL][TESTS] Refactor `OrcEncryptionSuite` and make `spa…
LuciferYang Jul 31, 2024
36f9a4b
Revert "[SPARK-49066][SQL][TESTS] Refactor `OrcEncryptionSuite` and m…
dongjoon-hyun Jul 31, 2024
94558f6
Revert "[SPARK-49000][SQL] Fix "select count(distinct 1) from t" wher…
yaooqinn Aug 1, 2024
a1e7fb1
[SPARK-49065][SQL] Rebasing in legacy formatters/parsers must support…
sumeet-db Aug 1, 2024
4f9dbc3
[SPARK-49066][SQL][TESTS][3.5] Refactor `OrcEncryptionSuite` and make…
LuciferYang Aug 1, 2024
0008bd1
[SPARK-49000][SQL][3.5] Fix "select count(distinct 1) from t" where t…
uros-db Aug 2, 2024
98eaaa5
[SPARK-49094][SQL] Fix ignoreCorruptFiles non-functioning for hive or…
yaooqinn Aug 3, 2024
b33a3ee
[SPARK-48791][CORE][FOLLOW-UP][3.5] Fix regression caused by immutabl…
Ngone51 Aug 5, 2024
f2e2601
[SPARK-49099][SQL] CatalogManager.setCurrentNamespace should respect …
amaliujia Aug 6, 2024
d13808c
[SPARK-49099][SQL][FOLLOWUP][3.5] recover tests in DDLSuite
cloud-fan Aug 6, 2024
bb7846d
Preparing Spark release v3.5.2-rc5
yaooqinn Aug 6, 2024
52ca284
Preparing development version 3.5.3-SNAPSHOT
yaooqinn Aug 6, 2024
0e2d757
[SPARK-49182][DOCS][PYTHON] Stop publish site/docs/{version}/api/pyth…
yaooqinn Aug 9, 2024
7bfb4f0
[SPARK-49176][K8S][DOCS] Fix `spark.ui.custom.executor.log.url` docs …
dongjoon-hyun Aug 9, 2024
204dd81
[SPARK-49183][SQL] V2SessionCatalog.createTable should respect PROP_I…
cloud-fan Aug 12, 2024
4a9dae9
[SPARK-48204][INFRA][FOLLOW] fix release scripts for the "finalize" step
cloud-fan Jun 3, 2024
deac780
[SPARK-46632][SQL] Fix subexpression elimination when equivalent tern…
zml1206 Aug 12, 2024
4d69021
[SPARK-49200][SQL] Fix null type non-codegen ordering exception
ulysses-you Aug 12, 2024
6a75640
[SPARK-49193][SQL] Improve the performance of RowSetUtils.toColumnBas…
wangyum Aug 11, 2024
7799811
[SPARK-49183][SQL][FOLLOWUP][3.5] Fix compilation
cloud-fan Aug 12, 2024
3237b8e
[SPARK-49197][CORE] Redact `Spark Command` output in `launcher` module
dongjoon-hyun Aug 12, 2024
d824219
[SPARK-49152][SQL] V2SessionCatalog should use V2Command
amaliujia Aug 13, 2024
6a250e7
[SPARK-49179][SQL] Fix v2 multi bucketed inner joins throw AssertionE…
ulysses-you Aug 13, 2024
6859ef9
Revert "[SPARK-49179][SQL] Fix v2 multi bucketed inner joins throw As…
dongjoon-hyun Aug 13, 2024
5463bfc
[SPARK-49179][SQL][3.5] Fix v2 multi bucketed inner joins throw Asser…
ulysses-you Aug 13, 2024
bd2cbd6
[SPARK-49038][SQL][3.5] SQLMetric should report the raw value in the …
cloud-fan Aug 14, 2024
8d05bf2
[SPARK-49152][SQL][FOLLOWUP][3.5] table location string should be Had…
cloud-fan Aug 15, 2024
3d96901
[SPARK-45590][BUILD][3.5] Upgrade okio to 1.17.6 from 1.15.0
roczei Aug 16, 2024
ac2219e
[SPARK-49039][UI] Reset checkbox when executor metrics are loaded in …
cxzl25 Aug 16, 2024
3148cfa
[SPARK-49211][SQL][3.5] V2 Catalog can also support built-in data sou…
amaliujia Aug 16, 2024
c06906d
[SPARK-49017][SQL][3.5] Insert statement fails when multiple paramete…
mihailom-db Aug 20, 2024
027a14b
[SPARK-49246][SQL] TableCatalog#loadTable should indicate if it's for…
cloud-fan Aug 21, 2024
481bc58
[SPARK-46444][SQL] V2SessionCatalog#createTable should not load the t…
cloud-fan Dec 26, 2023
eaadb39
[SPARK-49300][CORE][3.5] Fix Hadoop delegation token leak when tokenR…
zhangshuyan0 Aug 22, 2024
b2a25be
[SPARK-49333][SQL] Shutdown timeout thread while cleaning up SparkExe…
wangyum Aug 23, 2024
f4bcefb
[SPARK-49352][SQL][3.5] Avoid redundant array transform for identical…
viirya Aug 24, 2024
5eca953
[SPARK-49359][SQL] Allow StagedTableCatalog implementations to fall b…
cloud-fan Aug 23, 2024
870f1d8
[SPARK-49381][K8S][TESTS] Fix `SPARK-41388` test case to assert the r…
dongjoon-hyun Aug 26, 2024
2bc4c82
[SPARK-49385][K8S] Fix `getReusablePVCs` to use `podCreationTimeout` …
dongjoon-hyun Aug 26, 2024
c6f4dd7
[SPARK-49359][TESTS][FOLLOWUP][3.5] Fix compilation to rename `util.M…
dongjoon-hyun Aug 26, 2024
c979127
[SPARK-49402][PYTHON] Fix Binder integration in PySpark documentation
HyukjinKwon Aug 27, 2024
ddc99bc
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
2192437
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
ff83dd9
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
315d987
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
39f760c
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
51bfb88
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
e5a5d19
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
93fedc5
[SPARK-49402][PYTHON][FOLLOW-UP] Set upperfound for NumPy
HyukjinKwon Aug 27, 2024
dcfefd0
[SPARK-49402][PYTHON][FOLLOW-UP] Manually load ~/.profile in Spark Co…
HyukjinKwon Aug 28, 2024
2ad11b6
[SPARK-46037][SQL] Correctness fix for Shuffled Hash Join build left …
cloud-fan Aug 29, 2024
dce7c9a
[SPARK-43242][CORE][3.5] Fix throw 'Unexpected type of BlockId' in sh…
CavemanIV Aug 30, 2024
30a75d1
[MINOR] Add `artifacts` to `.gitignore`
zhengruifeng Aug 30, 2024
d5caaaa
[SPARK-49480][CORE] Fix NullPointerException from `SparkThrowableHelp…
jshmchenxi Sep 2, 2024
38ad0e7
[SPARK-49476][SQL] Fix nullability of base64 function
Kimahriman Sep 2, 2024
a1cd99d
Preparing Spark release v3.5.3-rc1
itholic Sep 3, 2024
0616ef2
Preparing development version 3.5.4-SNAPSHOT
itholic Sep 3, 2024
d8adf4b
[SPARK-49476][SQL][3.5][FOLLOWUP] Fix base64 proto test
Kimahriman Sep 3, 2024
7febde1
[SPARK-49275][SQL][3.5] Fix return type nullness of the xpath expression
chenhao-db Sep 4, 2024
7718777
[MINOR][DOCS] Fix site.SPARK_VERSION pattern in RDD Programming Guide
yaooqinn Sep 4, 2024
d83bf8c
[SPARK-49509][CORE] Use `Platform.allocateDirectBuffer` instead of `B…
cxzl25 Sep 4, 2024
e5ec16e
[SPARK-48965][SQL] Use the correct schema in `Dataset#toJSON`
bersprockets Sep 5, 2024
8fbeaf5
[SPARK-49408][SQL] Use IndexedSeq in ProjectingInternalRow
wzx140 Sep 5, 2024
d8f9d8d
[SPARK-49152][SQL][FOLLOWUP] DelegatingCatalogExtension should also u…
cloud-fan Sep 5, 2024
1f8c71f
Preparing Spark release v3.5.3-rc2
itholic Sep 6, 2024
46214da
Preparing development version 3.5.4-SNAPSHOT
itholic Sep 6, 2024
3f22ef1
[SPARK-49246][SQL][FOLLOW-UP] The behavior of SaveAsTable should not …
cloud-fan Sep 9, 2024
6292cfc
Preparing Spark release v3.5.3-rc3
itholic Sep 9, 2024
e923790
Preparing development version 3.5.4-SNAPSHOT
itholic Sep 9, 2024
ba374c6
fix import
cloud-fan Sep 9, 2024
32232e9
Preparing Spark release v3.5.3-rc3
itholic Sep 9, 2024
67421bb
Preparing development version 3.5.4-SNAPSHOT
itholic Sep 9, 2024
1206b52
[SPARK-49501][SQL] Fix double-escaping of table location
cstavr Sep 9, 2024
96eebeb
[SPARK-49595][CONNECT][SQL] Fix `DataFrame.unpivot/melt` in Spark Con…
xinrong-meng Sep 11, 2024
560efed
[SPARK-49261][SQL] Don't replace literals in aggregate expressions wi…
bersprockets Sep 12, 2024
e693e18
[SPARK-49628][SQL] ConstantFolding should copy stateful expression be…
cloud-fan Sep 13, 2024
e7ca790
[SPARK-49699][SS] Disable PruneFilters for streaming workloads
n-young-db Sep 20, 2024
dd76a82
[SPARK-49760][YARN] Correct handling of `SPARK_USER` env variable ove…
cnauroth Sep 24, 2024
b513297
[SPARK-49750][DOC] Mention delegation token support in K8s mode
pan3793 Sep 24, 2024
f1c69a5
[SPARK-49791][SQL] Make DelegatingCatalogExtension more extendable
cloud-fan Sep 26, 2024
b51db8b
[SPARK-49791][SQL][FOLLOWUP][3.5] Fix `import` statement
dongjoon-hyun Sep 26, 2024
1040657
[SPARK-49211][SQL][FOLLOW-UP] Support catalog in QualifiedTableName
amaliujia Sep 27, 2024
50c1783
[SPARK-49803][SQL][TESTS] Increase `spark.test.docker.connectionTimeo…
dongjoon-hyun Sep 27, 2024
be254c5
[SPARK-49804][K8S] Fix to use the exit code of executor container always
Sep 27, 2024
f888d57
[SPARK-46525][SQL][TESTS][3.5] Fix `docker-integration-tests` on Appl…
yaooqinn Sep 27, 2024
a49d6f4
[SPARK-47949][SQL][DOCKER][TESTS] MsSQLServer: Bump up mssql docker i…
yaooqinn Apr 23, 2024
16b8153
[SPARK-49819] Disable CollapseProject for correlated subqueries in pr…
n-young-db Sep 30, 2024
75860a3
[SPARK-49743][SQL] OptimizeCsvJsonExpr should not change schema field…
nikhilsheoran-db Oct 1, 2024
d9467b5
[SPARK-49816][SQL][3.5] Should only update out-going-ref-count for re…
cloud-fan Oct 1, 2024
b8f8456
[SPARK-49841][PYTHON][TESTS][3.5] Skip PySpark `test_cast_to_udt_with…
dongjoon-hyun Oct 1, 2024
ec28154
[SPARK-47702][CORE] Remove Shuffle service endpoint from the location…
Oct 4, 2024
aa6784c
Revert "[SPARK-47702][CORE] Remove Shuffle service endpoint from the …
LuciferYang Oct 4, 2024
d4b34d2
[SPARK-49843][SQL][3.5] Fix change comment on char/varchar columns
stefankandic Oct 4, 2024
9a5ae45
[SPARK-49836][SQL][SS] Fix possibly broken query when window is provi…
HeartSaVioR Oct 4, 2024
ca8407b
[SPARK-47702][CORE][3.5] Remove Shuffle service endpoint from the loc…
Oct 5, 2024
4d93a30
[SPARK-49806][PYTHON][TESTS][FOLLOW-UP] Skip newline difference in Sp…
HyukjinKwon Oct 7, 2024
4472fb2
[SPARK-49909][SQL][3.5] Fix the pretty name of some expressions
panbingkun Oct 9, 2024
5996c48
[SPARK-49782][SQL] ResolveDataFrameDropColumns rule resolves Unresolv…
LantaoJin Oct 10, 2024
36299b3
[SPARK-49959][SQL] Fix ColumnarArray.copy() to read nulls from the co…
chenhao-db Oct 15, 2024
92ec503
[SPARK-49981][CORE][TESTS] Fix `AsyncRDDActionsSuite.FutureAction res…
dongjoon-hyun Oct 16, 2024
d37a8b9
[SPARK-49983][CORE][TESTS] Fix `BarrierTaskContextSuite.successively …
dongjoon-hyun Oct 16, 2024
f5bc928
[SPARK-49979][SQL] Fix AQE hanging issue when collecting twice on a f…
liuzqt Oct 17, 2024
3d2d1c1
[SPARK-48155][SQL][3.5] AQEPropagateEmptyRelation for join should che…
AngersZhuuuu Oct 17, 2024
1fae1d7
[SPARK-49829][SS] Fix the bug on the optimization on adding input to …
HeartSaVioR Oct 18, 2024
6b9b3c0
Revert "[SPARK-49909][SQL][3.5] Fix the pretty name of some expressions"
panbingkun Oct 18, 2024
3a4ebae
[SPARK-50021][CORE][UI][3.5] Fix `ApplicationPage` to hide App UI lin…
dongjoon-hyun Oct 18, 2024
d24393b
[SPARK-50022][CORE][UI][3.5] Fix `MasterPage` to hide App UI links wh…
dongjoon-hyun Oct 18, 2024
e3e79e5
[SPARK-50123][TESTS] Move BitmapExpressionUtilsSuite & ExpressionImpl…
yaooqinn Oct 25, 2024
2f5e0fa
[SPARK-50150][BUILD][3.5] Upgrade Jetty to 9.4.56.v20240826
dongjoon-hyun Oct 29, 2024
175b5e9
[MINOR][BUILD] Skip `deepspeed` in requirements on MacOS
zhengruifeng Aug 10, 2023
4205b79
[SPARK-50155][3.5] Move scala and java files to their default folders
panbingkun Oct 30, 2024
0229c0e
[SPARK-50176][CONNECT][3.5] Disallow reattaching after the session is…
Nov 1, 2024
6df1966
[SPARK-50195][CORE] Fix `StandaloneRestServer` to propagate `spark.ap…
dongjoon-hyun Nov 1, 2024
9d47266
[SPARK-50199][PYTHON][TESTS] Use Spark 3.4.4 instead of 3.0.1 in `tes…
dongjoon-hyun Nov 1, 2024
08023c0
[SPARK-50176][CONNECT][FOLLOWUP][3.5] Fix ReattachableExecuteSuite fa…
Nov 4, 2024
1ddf4a9
[MINOR][DOCS][3.5] Fix specified java versions in `install.rst`
dvorst Nov 4, 2024
9b172de
[SPARK-50210][CORE] Fix `SparkSubmit` to show REST API `kill` respons…
dongjoon-hyun Nov 3, 2024
acccf53
[SPARK-50212][BUILD][3.5] Fix the conditional check for executing the…
LuciferYang Nov 5, 2024
d39f5ab
[SPARK-50235][SQL] Clean up ColumnVector resource after processing al…
viirya Nov 6, 2024
8da6987
[SPARK-50300][BUILD] Use mirror host instead of `archive.apache.org`
dongjoon-hyun Nov 13, 2024
8502a22
[SPARK-50304][INFRA] Remove `(any|empty).proto` from RAT exclusion
dongjoon-hyun Nov 14, 2024
e0bdfee
[SPARK-50316][BUILD][3.5] Upgrade ORC to 1.9.5
dongjoon-hyun Nov 14, 2024
242d333
[SPARK-50312][SQL] SparkThriftServer createServer parameter passing e…
Nov 15, 2024
08b195c
[MINOR][DOCS] Fix a HTML/Markdown syntax error in sql-migration-guide.md
yaooqinn Nov 20, 2024
df9b9de
[SPARK-50258][SQL] Fix output column order changed issue after AQE op…
wangyum Nov 20, 2024
5ff129a
[SPARK-50430][CORE] Use the standard Properties.clone instead of manu…
HyukjinKwon Nov 27, 2024
5e51e2c
[SPARK-49294][UI] Add width attribute for shuffle-write-time checkbox
xunxunmimi5577 Dec 2, 2024
1d6f7ad
[SPARK-50433][DOCS][TESTS][3.5] Fix configuring log4j2 guide docs for…
pan3793 Dec 3, 2024
5dc927b
[SPARK-50487][DOCS] Update broken jira link
huangxiaopingRD Dec 4, 2024
acedb15
[SPARK-50498][PYTHON] Avoid unnecessary py4j call in `listFunctions`
zhengruifeng Dec 5, 2024
86e29e9
[SPARK-50505][DOCS] Fix `spark.storage.replication.proactive` default…
dongjoon-hyun Dec 5, 2024
d01f34f
[SPARK-50492][SS] Fix java.util.NoSuchElementException when event tim…
liviazhu-db Dec 6, 2024
153cd9e
[SPARK-50492][SS][FOLLOWUP][3.5] Change `def references` to `lazy val…
LuciferYang Dec 6, 2024
bf29ab9
[SPARK-50421][CORE][3.5] Fix executor related memory config incorrect…
zjuwangg Dec 6, 2024
d8f3afa
[SPARK-50514][DOCS] Add `IDENTIFIER clause` page to `menu-sql.yaml`
dongjoon-hyun Dec 7, 2024
305d2a0
[SPARK-49695][SQL][3.5] Postgres fix xor push-down
andrej-db Dec 7, 2024
a57f3c2
[SPARK-50483][CORE][SQL][3.5] BlockMissingException should be thrown …
wangyum Dec 8, 2024
929a19f
Preparing Spark release v3.5.4-rc1
LuciferYang Dec 9, 2024
8e6507a
Preparing development version 3.5.5-SNAPSHOT
LuciferYang Dec 9, 2024
deabe49
[SPARK-49134][INFRA][3.5] Support retry for deploying artifacts to Ne…
yaooqinn Dec 9, 2024
a3cf28e
[SPARK-50463][SQL][3.5] Fix `ConstantColumnVector` with Columnar to R…
richardc-db Dec 10, 2024
bb953f9
[SPARK-50510][CONNECT][3.5] Fix sporadic ReattachableExecuteSuite fai…
Dec 11, 2024
e97580a
[SPARK-50087][SQL][3.5] Robust handling of boolean expressions in CAS…
andrej-db Dec 12, 2024
92e650c
[SPARK-50545][CORE][SQL][3.5] `AccessControlException` should be thro…
pan3793 Dec 12, 2024
91af6f9
Preparing Spark release v3.5.4-rc2
LuciferYang Dec 16, 2024
a764524
Preparing development version 3.5.5-SNAPSHOT
LuciferYang Dec 16, 2024
8168ea8
[SPARK-50430][CORE][FOLLOW-UP] Keep the logic of manual putting key a…
HyukjinKwon Dec 3, 2024
0fbe292
Revert "[SPARK-50430][CORE][FOLLOW-UP] Keep the logic of manual putti…
HyukjinKwon Dec 16, 2024
f7c48fe
Revert "[SPARK-50430][CORE] Use the standard Properties.clone instead…
HyukjinKwon Dec 16, 2024
b0a7d4d
[SPARK-50587][INFRA][3.5] Remove unsupported `curl` option `--retry-a…
LuciferYang Dec 16, 2024
a6f220d
Preparing Spark release v3.5.4-rc3
LuciferYang Dec 17, 2024
bcaa5a9
Preparing development version 3.5.5-SNAPSHOT
LuciferYang Dec 17, 2024
45349b6
[SPARK-50510][CONNECT][TEST][3.5] Fix flaky ReattachableExecuteSuite
Dec 17, 2024
a3d23fd
[MINOR][SS] Minor update to watermark propagation comments
neilramaswamy Dec 18, 2024
5a91172
[SPARK-50483][SPARK-50545][DOC][FOLLOWUP][3.5] Mention behavior chang…
pan3793 Dec 23, 2024
19ad7f7
[SPARK-50769][SQL] Fix ClassCastException in HistogramNumeric
linhongliu-db Jan 9, 2025
5a8483b
[SPARK-50784][TESTS] Fix `lint-scala` not to ignore `scalastyle` errors
dongjoon-hyun Jan 10, 2025
00b3833
[SPARK-50797][SQL][TESTS][3.5] Move `HiveCharVarcharTestSuite` from `…
dongjoon-hyun Jan 13, 2025
51fb84a
[SPARK-50624][SQL] Add TimestampNTZType to ColumnarRow/MutableColumna…
nastra Jan 13, 2025
5de80b7
[SPARK-50886][BUILD][3.5] Upgrade Avro to 1.11.4
dongjoon-hyun Jan 19, 2025
bdfa6ef
[SPARK-50870][SQL] Add the timezone when casting to timestamp in V2Sc…
Jan 21, 2025
e4e41b2
[SPARK-45013][CORE][TEST][3.5] Flaky Test with NPE: track allocated r…
yaooqinn Jan 21, 2025
7177891
[SPARK-50890][PYTHON][TESTS][CONNECT] Skip test_take in Spark Connect…
HyukjinKwon Jan 20, 2025
bd2b8e5
[SPARK-50946][CORE][TESTS] Add version check for Java 17.0.14 to make…
LuciferYang Jan 22, 2025
5abaeac
Revert "[SPARK-50946][CORE][TESTS] Add version check for Java 17.0.14…
dongjoon-hyun Jan 22, 2025
e608297
[SPARK-50946][CORE][TESTS][3.5] Add version check for Java 17.0.14 to…
dongjoon-hyun Jan 22, 2025
3912368
Revert "[SPARK-50890][PYTHON][TESTS][CONNECT] Skip test_take in Spark…
HyukjinKwon Jan 23, 2025
d061aad
Revert "Revert "[SPARK-50890][PYTHON][TESTS][CONNECT] Skip test_take …
HyukjinKwon Jan 23, 2025
d2f9b4c
[SPARK-50984][SQL][TESTS] Make `ExpressionImplUtilsSuite` robust by m…
dongjoon-hyun Jan 25, 2025
7fd9ced
[SPARK-50853][CORE][3.5] Close temp shuffle file writable channel
Jan 25, 2025
3e01d41
[SPARK-50091][SQL][3.5] Handle case of aggregates in left-hand operan…
bersprockets Jan 25, 2025
22118f9
[SPARK-50987][DOCS] Make `spark-connect-overview.md`s version strings…
dongjoon-hyun Jan 25, 2025
09a255a
[SPARK-50998][K8S][DOCS] Fix `spark.kubernetes.configMap.maxSize` def…
dongjoon-hyun Jan 27, 2025
1f81475
[SPARK-51033][CORE][TESTS] Fix `CryptoStreamUtilsSuite` to use `sameE…
dongjoon-hyun Jan 29, 2025
c339e50
[SPARK-51027][SQL] Prevent `HiveClient.runSqlHive` invocation in non-…
dongjoon-hyun Jan 29, 2025
1966d57
[SPARK-51039][BUILD] Fix `hive-llap-common` dependency to use `hive.l…
dongjoon-hyun Jan 30, 2025
fd86566
[SPARK-51061][CORE] Hide `Jetty` info in REST Submission API
dongjoon-hyun Feb 3, 2025
105bed7
[SPARK-51071][INFRA][3.5] Upgrade `upload-artifact` to v4 to recover …
dongjoon-hyun Feb 3, 2025
31fbf01
[SPARK-51104][DOC] Self-host JavaScript and CSS in Spark website
gengliangwang Feb 6, 2025
3b892bd
[SPARK-51104][DOC][FOLLOWUP] Self-host docsearch.min.css in Spark web…
gengliangwang Feb 6, 2025
e395f33
[SPARK-51128][DOC] Self host docsearch.min.css.map in Spark website
gengliangwang Feb 7, 2025
03f7200
[SPARK-51129][DOC] Fix code tab switching in Spark Website
gengliangwang Feb 7, 2025
4e024f4
[SPARK-51093][SQL][TESTS][3.5] Fix minor endianness issues in tests
jonathan-albrecht-ibm Feb 10, 2025
012df44
[MINOR][DOCS] Fix incorrect description of constraint on spark.sql.ad…
JoshRosen Feb 11, 2025
af45325
[SPARK-51164][CORE][TESTS] Fix `CallerContext` test by enabling `hado…
dongjoon-hyun Feb 11, 2025
2ea2c0b
[SPARK-51164][CORE][TESTS][FOLLOWUP] Add hadoop.caller.context.enable…
cnauroth Feb 12, 2025
6084757
[SPARK-51157][SQL][3.5] Add missing @VarArgs Scala annotation for Sca…
yaooqinn Feb 12, 2025
4962453
[SPARK-51042][SQL][3.5] Read and write the month and days fields of i…
jonathan-albrecht-ibm Feb 12, 2025
030e691
[SPARK-51219][SQL] Fix `ShowTablesExec.isTempView` to work with non-`…
ostronaut Feb 18, 2025
4d15f64
[SPARK-51219][SQL][TESTS][FOLLOWUP] ShowTablesExec` remove `ArrayImpl…
ostronaut Feb 20, 2025
23637fe
[SPARK-51185][CORE][3.5] Revert simplifications to PartitionedFileUti…
LukasRupprecht Feb 21, 2025
c0bfae6
[SPARK-51244][INFRA][3.5] Upgrade left Github Action image from `ubun…
wayneguow Feb 21, 2025
3ba36ec
[SPARK-51187][SQL][SS][3.5] Implement the graceful deprecation of inc…
HeartSaVioR Feb 23, 2025
eedb1aa
[SPARK-48899][K8S] Fix `ENV` key value format in K8s Dockerfiles
dongjoon-hyun Jul 15, 2024
d65ecc0
[SPARK-51173][TESTS] Add `configName` Scalastyle rule
dongjoon-hyun Feb 12, 2025
7c29c66
Preparing Spark release v3.5.5-rc1
dongjoon-hyun Feb 23, 2025
55049c5
Creating K8S only image
denglishcollibra Jun 4, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
11 changes: 5 additions & 6 deletions .github/workflows/benchmark.yml
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,7 @@ on:
jobs:
matrix-gen:
name: Generate matrix for job splits
runs-on: ubuntu-20.04
runs-on: ubuntu-22.04
outputs:
matrix: ${{ steps.set-matrix.outputs.matrix }}
env:
Expand All @@ -60,7 +60,7 @@ jobs:
tpcds-1g-gen:
name: "Generate an input dataset for TPCDSQueryBenchmark with SF=1"
if: contains(github.event.inputs.class, 'TPCDSQueryBenchmark') || contains(github.event.inputs.class, '*')
runs-on: ubuntu-20.04
runs-on: ubuntu-22.04
env:
SPARK_LOCAL_IP: localhost
steps:
Expand Down Expand Up @@ -98,7 +98,7 @@ jobs:
uses: actions/checkout@v3
with:
repository: databricks/tpcds-kit
ref: 2a5078a782192ddb6efbcead8de9973d6ab4f069
ref: 1b7fb7529edae091684201fab142d956d6afd881
path: ./tpcds-kit
- name: Build tpcds-kit
if: steps.cache-tpcds-sf-1.outputs.cache-hit != 'true'
Expand All @@ -117,8 +117,7 @@ jobs:
name: "Run benchmarks: ${{ github.event.inputs.class }} (JDK ${{ github.event.inputs.jdk }}, Scala ${{ github.event.inputs.scala }}, ${{ matrix.split }} out of ${{ github.event.inputs.num-splits }} splits)"
if: always()
needs: [matrix-gen, tpcds-1g-gen]
# Ubuntu 20.04 is the latest LTS. The next LTS is 22.04.
runs-on: ubuntu-20.04
runs-on: ubuntu-22.04
strategy:
fail-fast: false
matrix:
Expand Down Expand Up @@ -188,7 +187,7 @@ jobs:
echo "Preparing the benchmark results:"
tar -cvf benchmark-results-${{ github.event.inputs.jdk }}-${{ github.event.inputs.scala }}.tar `git diff --name-only` `git ls-files --others --exclude=tpcds-sf-1 --exclude-standard`
- name: Upload benchmark results
uses: actions/upload-artifact@v3
uses: actions/upload-artifact@v4
with:
name: benchmark-results-${{ github.event.inputs.jdk }}-${{ github.event.inputs.scala }}-${{ matrix.split }}
path: benchmark-results-${{ github.event.inputs.jdk }}-${{ github.event.inputs.scala }}.tar
Expand Down
101 changes: 69 additions & 32 deletions .github/workflows/build_and_test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -30,8 +30,7 @@ on:
description: Branch to run the build against
required: false
type: string
# Change 'master' to 'branch-3.5' in branch-3.5 branch after cutting it.
default: master
default: branch-3.5
hadoop:
description: Hadoop version to run with. HADOOP_PROFILE environment variable should accept it.
required: false
Expand Down Expand Up @@ -80,25 +79,34 @@ jobs:
id: set-outputs
run: |
if [ -z "${{ inputs.jobs }}" ]; then
pyspark=true; sparkr=true; tpcds=true; docker=true;
pyspark_modules=`cd dev && python -c "import sparktestsupport.modules as m; print(','.join(m.name for m in m.all_modules if m.name.startswith('pyspark')))"`
pyspark=`./dev/is-changed.py -m $pyspark_modules`
sparkr=`./dev/is-changed.py -m sparkr`
tpcds=`./dev/is-changed.py -m sql`
docker=`./dev/is-changed.py -m docker-integration-tests`
# 'build', 'scala-213', and 'java-11-17' are always true for now.
# It does not save significant time and most of PRs trigger the build.
if [[ "${{ github.repository }}" != 'apache/spark' ]]; then
pandas=$pyspark
kubernetes=`./dev/is-changed.py -m kubernetes`
sparkr=`./dev/is-changed.py -m sparkr`
tpcds=`./dev/is-changed.py -m sql`
docker=`./dev/is-changed.py -m docker-integration-tests`
else
pandas=false
kubernetes=false
sparkr=false
tpcds=false
docker=false
fi
build=`./dev/is-changed.py -m "core,unsafe,kvstore,avro,utils,network-common,network-shuffle,repl,launcher,examples,sketch,graphx,catalyst,hive-thriftserver,streaming,sql-kafka-0-10,streaming-kafka-0-10,mllib-local,mllib,yarn,mesos,kubernetes,hadoop-cloud,spark-ganglia-lgpl,sql,hive,connect,protobuf,api"`
precondition="
{
\"build\": \"true\",
\"build\": \"$build\",
\"pyspark\": \"$pyspark\",
\"pyspark-pandas\": \"$pandas\",
\"sparkr\": \"$sparkr\",
\"tpcds-1g\": \"$tpcds\",
\"docker-integration-tests\": \"$docker\",
\"scala-213\": \"true\",
\"java-11-17\": \"true\",
\"scala-213\": \"$build\",
\"java-11-17\": \"$build\",
\"lint\" : \"true\",
\"k8s-integration-tests\" : \"true\",
\"k8s-integration-tests\" : \"$kubernetes\",
\"breaking-changes-buf\" : \"true\",
}"
echo $precondition # For debugging
Expand Down Expand Up @@ -205,6 +213,9 @@ jobs:
HIVE_PROFILE: ${{ matrix.hive }}
GITHUB_PREV_SHA: ${{ github.event.before }}
SPARK_LOCAL_IP: localhost
SKIP_UNIDOC: true
SKIP_MIMA: true
SKIP_PACKAGING: true
steps:
- name: Checkout Spark repository
uses: actions/checkout@v3
Expand Down Expand Up @@ -256,7 +267,7 @@ jobs:
- name: Install Python packages (Python 3.8)
if: (contains(matrix.modules, 'sql') && !contains(matrix.modules, 'sql-'))
run: |
python3.8 -m pip install 'numpy>=1.20.0' pyarrow pandas scipy unittest-xml-reporting 'grpcio==1.56.0' 'protobuf==3.20.3'
python3.8 -m pip install 'numpy>=1.20.0' 'pyarrow==12.0.1' pandas scipy unittest-xml-reporting 'grpcio==1.56.0' 'protobuf==3.20.3'
python3.8 -m pip list
# Run the tests.
- name: Run tests
Expand All @@ -271,13 +282,13 @@ jobs:
./dev/run-tests --parallelism 1 --modules "$MODULES_TO_TEST" --included-tags "$INCLUDED_TAGS" --excluded-tags "$EXCLUDED_TAGS"
- name: Upload test results to report
if: always()
uses: actions/upload-artifact@v3
uses: actions/upload-artifact@v4
with:
name: test-results-${{ matrix.modules }}-${{ matrix.comment }}-${{ matrix.java }}-${{ matrix.hadoop }}-${{ matrix.hive }}
path: "**/target/test-reports/*.xml"
- name: Upload unit tests log files
if: failure()
uses: actions/upload-artifact@v3
uses: actions/upload-artifact@v4
with:
name: unit-tests-log-${{ matrix.modules }}-${{ matrix.comment }}-${{ matrix.java }}-${{ matrix.hadoop }}-${{ matrix.hive }}
path: "**/target/unit-tests.log"
Expand Down Expand Up @@ -344,6 +355,8 @@ jobs:
java:
- ${{ inputs.java }}
modules:
- >-
pyspark-errors
- >-
pyspark-sql, pyspark-mllib, pyspark-resource, pyspark-testing
- >-
Expand All @@ -353,11 +366,19 @@ jobs:
- >-
pyspark-pandas-slow
- >-
pyspark-connect, pyspark-errors
pyspark-connect
- >-
pyspark-pandas-connect
- >-
pyspark-pandas-slow-connect
exclude:
# Always run if pyspark-pandas == 'true', even infra-image is skip (such as non-master job)
# In practice, the build will run in individual PR, but not against the individual commit
# in Apache Spark repository.
- modules: ${{ fromJson(needs.precondition.outputs.required).pyspark-pandas != 'true' && 'pyspark-pandas' }}
- modules: ${{ fromJson(needs.precondition.outputs.required).pyspark-pandas != 'true' && 'pyspark-pandas-slow' }}
- modules: ${{ fromJson(needs.precondition.outputs.required).pyspark-pandas != 'true' && 'pyspark-pandas-connect' }}
- modules: ${{ fromJson(needs.precondition.outputs.required).pyspark-pandas != 'true' && 'pyspark-pandas-slow-connect' }}
env:
MODULES_TO_TEST: ${{ matrix.modules }}
HADOOP_PROFILE: ${{ inputs.hadoop }}
Expand All @@ -366,6 +387,7 @@ jobs:
SPARK_LOCAL_IP: localhost
SKIP_UNIDOC: true
SKIP_MIMA: true
SKIP_PACKAGING: true
METASPACE_SIZE: 1g
steps:
- name: Checkout Spark repository
Expand Down Expand Up @@ -404,6 +426,8 @@ jobs:
key: pyspark-coursier-${{ hashFiles('**/pom.xml', '**/plugins.sbt') }}
restore-keys: |
pyspark-coursier-
- name: Free up disk space
run: ./dev/free_disk_space_container
- name: Install Java ${{ matrix.java }}
uses: actions/setup-java@v3
with:
Expand All @@ -414,14 +438,20 @@ jobs:
python3.9 -m pip list
pypy3 -m pip list
- name: Install Conda for pip packaging test
if: ${{ matrix.modules == 'pyspark-errors' }}
run: |
curl -s https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh > miniconda.sh
bash miniconda.sh -b -p $HOME/miniconda
# Run the tests.
- name: Run tests
env: ${{ fromJSON(inputs.envs) }}
shell: 'script -q -e -c "bash {0}"'
run: |
export PATH=$PATH:$HOME/miniconda/bin
if [[ "$MODULES_TO_TEST" == "pyspark-errors" ]]; then
export PATH=$PATH:$HOME/miniconda/bin
export SKIP_PACKAGING=false
echo "Python Packaging Tests Enabled!"
fi
./dev/run-tests --parallelism 1 --modules "$MODULES_TO_TEST"
- name: Upload coverage to Codecov
if: fromJSON(inputs.envs).PYSPARK_CODECOV == 'true'
Expand All @@ -432,13 +462,13 @@ jobs:
name: PySpark
- name: Upload test results to report
if: always()
uses: actions/upload-artifact@v3
uses: actions/upload-artifact@v4
with:
name: test-results-${{ matrix.modules }}--8-${{ inputs.hadoop }}-hive2.3
path: "**/target/test-reports/*.xml"
- name: Upload unit tests log files
if: failure()
uses: actions/upload-artifact@v3
uses: actions/upload-artifact@v4
with:
name: unit-tests-log-${{ matrix.modules }}--8-${{ inputs.hadoop }}-hive2.3
path: "**/target/unit-tests.log"
Expand All @@ -457,6 +487,7 @@ jobs:
GITHUB_PREV_SHA: ${{ github.event.before }}
SPARK_LOCAL_IP: localhost
SKIP_MIMA: true
SKIP_PACKAGING: true
steps:
- name: Checkout Spark repository
uses: actions/checkout@v3
Expand Down Expand Up @@ -494,6 +525,8 @@ jobs:
key: sparkr-coursier-${{ hashFiles('**/pom.xml', '**/plugins.sbt') }}
restore-keys: |
sparkr-coursier-
- name: Free up disk space
run: ./dev/free_disk_space_container
- name: Install Java ${{ inputs.java }}
uses: actions/setup-java@v3
with:
Expand All @@ -509,7 +542,7 @@ jobs:
./dev/run-tests --parallelism 1 --modules sparkr
- name: Upload test results to report
if: always()
uses: actions/upload-artifact@v3
uses: actions/upload-artifact@v4
with:
name: test-results-sparkr--8-${{ inputs.hadoop }}-hive2.3
path: "**/target/test-reports/*.xml"
Expand Down Expand Up @@ -602,6 +635,8 @@ jobs:
key: docs-maven-${{ hashFiles('**/pom.xml') }}
restore-keys: |
docs-maven-
- name: Free up disk space
run: ./dev/free_disk_space_container
- name: Install Java 8
uses: actions/setup-java@v3
with:
Expand All @@ -611,6 +646,8 @@ jobs:
run: ./dev/check-license
- name: Dependencies test
run: ./dev/test-dependencies.sh
- name: MIMA test
run: ./dev/mima
- name: Scala linter
run: ./dev/lint-scala
- name: Java linter
Expand Down Expand Up @@ -662,16 +699,16 @@ jobs:
# See also https://issues.apache.org/jira/browse/SPARK-35375.
# Pin the MarkupSafe to 2.0.1 to resolve the CI error.
# See also https://issues.apache.org/jira/browse/SPARK-38279.
python3.9 -m pip install 'sphinx<3.1.0' mkdocs pydata_sphinx_theme nbsphinx numpydoc 'jinja2<3.0.0' 'markupsafe==2.0.1' 'pyzmq<24.0.0'
python3.9 -m pip install 'sphinx<3.1.0' mkdocs pydata_sphinx_theme 'sphinx-copybutton==0.5.2' 'nbsphinx==0.9.3' numpydoc 'jinja2<3.0.0' 'markupsafe==2.0.1' 'pyzmq<24.0.0' 'sphinxcontrib-applehelp==1.0.4' 'sphinxcontrib-devhelp==1.0.2' 'sphinxcontrib-htmlhelp==2.0.1' 'sphinxcontrib-qthelp==1.0.3' 'sphinxcontrib-serializinghtml==1.1.5' 'nest-asyncio==1.5.8' 'rpds-py==0.16.2' 'alabaster==0.7.13'
python3.9 -m pip install ipython_genutils # See SPARK-38517
python3.9 -m pip install sphinx_plotly_directive 'numpy>=1.20.0' pyarrow pandas 'plotly>=4.8'
python3.9 -m pip install sphinx_plotly_directive 'numpy>=1.20.0' 'pyarrow==12.0.1' pandas 'plotly>=4.8'
python3.9 -m pip install 'docutils<0.18.0' # See SPARK-39421
apt-get update -y
apt-get install -y ruby ruby-dev
Rscript -e "install.packages(c('devtools', 'testthat', 'knitr', 'rmarkdown', 'markdown', 'e1071', 'roxygen2', 'ggplot2', 'mvtnorm', 'statmod'), repos='https://cloud.r-project.org/')"
Rscript -e "devtools::install_version('pkgdown', version='2.0.1', repos='https://cloud.r-project.org')"
Rscript -e "devtools::install_version('preferably', version='0.4', repos='https://cloud.r-project.org')"
gem install bundler
gem install bundler -v 2.4.22
cd docs
bundle install
- name: R linter
Expand Down Expand Up @@ -794,8 +831,7 @@ jobs:
needs: precondition
if: fromJson(needs.precondition.outputs.required).tpcds-1g == 'true'
name: Run TPC-DS queries with SF=1
# Pin to 'Ubuntu 20.04' due to 'databricks/tpcds-kit' compilation
runs-on: ubuntu-20.04
runs-on: ubuntu-22.04
env:
SPARK_LOCAL_IP: localhost
steps:
Expand Down Expand Up @@ -845,7 +881,7 @@ jobs:
uses: actions/checkout@v3
with:
repository: databricks/tpcds-kit
ref: 2a5078a782192ddb6efbcead8de9973d6ab4f069
ref: 1b7fb7529edae091684201fab142d956d6afd881
path: ./tpcds-kit
- name: Build tpcds-kit
if: steps.cache-tpcds-sf-1.outputs.cache-hit != 'true'
Expand Down Expand Up @@ -878,13 +914,13 @@ jobs:
spark.sql.join.forceApplyShuffledHashJoin=true
- name: Upload test results to report
if: always()
uses: actions/upload-artifact@v3
uses: actions/upload-artifact@v4
with:
name: test-results-tpcds--8-${{ inputs.hadoop }}-hive2.3
path: "**/target/test-reports/*.xml"
- name: Upload unit tests log files
if: failure()
uses: actions/upload-artifact@v3
uses: actions/upload-artifact@v4
with:
name: unit-tests-log-tpcds--8-${{ inputs.hadoop }}-hive2.3
path: "**/target/unit-tests.log"
Expand All @@ -901,6 +937,7 @@ jobs:
SPARK_LOCAL_IP: localhost
ORACLE_DOCKER_IMAGE_NAME: gvenzl/oracle-xe:21.3.0
SKIP_MIMA: true
SKIP_PACKAGING: true
steps:
- name: Checkout Spark repository
uses: actions/checkout@v3
Expand Down Expand Up @@ -943,13 +980,13 @@ jobs:
./dev/run-tests --parallelism 1 --modules docker-integration-tests --included-tags org.apache.spark.tags.DockerTest
- name: Upload test results to report
if: always()
uses: actions/upload-artifact@v3
uses: actions/upload-artifact@v4
with:
name: test-results-docker-integration--8-${{ inputs.hadoop }}-hive2.3
path: "**/target/test-reports/*.xml"
- name: Upload unit tests log files
if: failure()
uses: actions/upload-artifact@v3
uses: actions/upload-artifact@v4
with:
name: unit-tests-log-docker-integration--8-${{ inputs.hadoop }}-hive2.3
path: "**/target/unit-tests.log"
Expand Down Expand Up @@ -1017,10 +1054,10 @@ jobs:
kubectl create clusterrolebinding serviceaccounts-cluster-admin --clusterrole=cluster-admin --group=system:serviceaccounts || true
kubectl apply -f https://raw.githubusercontent.com/volcano-sh/volcano/v1.7.0/installer/volcano-development.yaml || true
eval $(minikube docker-env)
build/sbt -Psparkr -Pkubernetes -Pvolcano -Pkubernetes-integration-tests -Dspark.kubernetes.test.driverRequestCores=0.5 -Dspark.kubernetes.test.executorRequestCores=0.2 -Dspark.kubernetes.test.volcanoMaxConcurrencyJobNum=1 -Dtest.exclude.tags=local "kubernetes-integration-tests/test"
build/sbt -Psparkr -Pkubernetes -Pvolcano -Pkubernetes-integration-tests -Dspark.kubernetes.test.volcanoMaxConcurrencyJobNum=1 -Dtest.exclude.tags=local "kubernetes-integration-tests/test"
- name: Upload Spark on K8S integration tests log files
if: failure()
uses: actions/upload-artifact@v3
uses: actions/upload-artifact@v4
with:
name: spark-on-kubernetes-it-log
path: "**/target/integration-tests.log"
2 changes: 1 addition & 1 deletion .github/workflows/notify_test_workflow.yml
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ on:
jobs:
notify:
name: Notify test workflow
runs-on: ubuntu-20.04
runs-on: ubuntu-22.04
permissions:
actions: read
checks: write
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/update_build_status.yml
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ on:
jobs:
update:
name: Update build status
runs-on: ubuntu-20.04
runs-on: ubuntu-22.04
permissions:
actions: read
checks: write
Expand Down
Loading