Skip to content

Commit 29f1436

Browse files
author
Marcelo Vanzin
committed
Merge branch 'master' into SPARK-24243
2 parents 9240b77 + 180f969 commit 29f1436

File tree

1,762 files changed

+70538
-54941
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

1,762 files changed

+70538
-54941
lines changed

.gitignore

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -77,7 +77,6 @@ target/
7777
unit-tests.log
7878
work/
7979
docs/.jekyll-metadata
80-
*.crc
8180

8281
# For Hive
8382
TempStatsStore/

.travis.yml

Lines changed: 0 additions & 50 deletions
This file was deleted.

R/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ export R_HOME=/home/username/R
1717

1818
#### Build Spark
1919

20-
Build Spark with [Maven](http://spark.apache.org/docs/latest/building-spark.html#building-with-buildmvn) and include the `-Psparkr` profile to build the R package. For example to use the default Hadoop versions you can run
20+
Build Spark with [Maven](http://spark.apache.org/docs/latest/building-spark.html#buildmvn) and include the `-Psparkr` profile to build the R package. For example to use the default Hadoop versions you can run
2121

2222
```bash
2323
build/mvn -DskipTests -Psparkr package

R/WINDOWS.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
To build SparkR on Windows, the following steps are required
44

55
1. Install R (>= 3.1) and [Rtools](http://cran.r-project.org/bin/windows/Rtools/). Make sure to
6-
include Rtools and R in `PATH`.
6+
include Rtools and R in `PATH`. Note that support for R prior to version 3.4 is deprecated as of Spark 3.0.0.
77

88
2. Install
99
[JDK8](http://www.oracle.com/technetwork/java/javase/downloads/jdk8-downloads-2133151.html) and set
@@ -14,7 +14,7 @@ directory in Maven in `PATH`.
1414

1515
4. Set `MAVEN_OPTS` as described in [Building Spark](http://spark.apache.org/docs/latest/building-spark.html).
1616

17-
5. Open a command shell (`cmd`) in the Spark directory and build Spark with [Maven](http://spark.apache.org/docs/latest/building-spark.html#building-with-buildmvn) and include the `-Psparkr` profile to build the R package. For example to use the default Hadoop versions you can run
17+
5. Open a command shell (`cmd`) in the Spark directory and build Spark with [Maven](http://spark.apache.org/docs/latest/building-spark.html#buildmvn) and include the `-Psparkr` profile to build the R package. For example to use the default Hadoop versions you can run
1818

1919
```bash
2020
mvn.cmd -DskipTests -Psparkr package

R/pkg/DESCRIPTION

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
Package: SparkR
22
Type: Package
3-
Version: 2.4.0
3+
Version: 3.0.0
44
Title: R Frontend for Apache Spark
55
Description: Provides an R Frontend for Apache Spark.
66
Authors@R: c(person("Shivaram", "Venkataraman", role = c("aut", "cre"),
@@ -15,7 +15,7 @@ URL: http://www.apache.org/ http://spark.apache.org/
1515
BugReports: http://spark.apache.org/contributing.html
1616
SystemRequirements: Java (== 8)
1717
Depends:
18-
R (>= 3.0),
18+
R (>= 3.1),
1919
methods
2020
Suggests:
2121
knitr,

R/pkg/NAMESPACE

Lines changed: 14 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -28,9 +28,8 @@ importFrom("utils", "download.file", "object.size", "packageVersion", "tail", "u
2828

2929
# S3 methods exported
3030
export("sparkR.session")
31-
export("sparkR.init")
32-
export("sparkR.stop")
3331
export("sparkR.session.stop")
32+
export("sparkR.stop")
3433
export("sparkR.conf")
3534
export("sparkR.version")
3635
export("sparkR.uiWebUrl")
@@ -42,9 +41,6 @@ export("sparkR.callJStatic")
4241

4342
export("install.spark")
4443

45-
export("sparkRSQL.init",
46-
"sparkRHive.init")
47-
4844
# MLlib integration
4945
exportMethods("glm",
5046
"spark.glm",
@@ -70,7 +66,8 @@ exportMethods("glm",
7066
"spark.svmLinear",
7167
"spark.fpGrowth",
7268
"spark.freqItemsets",
73-
"spark.associationRules")
69+
"spark.associationRules",
70+
"spark.findFrequentSequentialPatterns")
7471

7572
# Job group lifecycle management methods
7673
export("setJobGroup",
@@ -150,15 +147,13 @@ exportMethods("arrange",
150147
"printSchema",
151148
"randomSplit",
152149
"rbind",
153-
"registerTempTable",
154150
"rename",
155151
"repartition",
156152
"repartitionByRange",
157153
"rollup",
158154
"sample",
159155
"sample_frac",
160156
"sampleBy",
161-
"saveAsParquetFile",
162157
"saveAsTable",
163158
"saveDF",
164159
"schema",
@@ -200,10 +195,13 @@ exportMethods("%<=>%",
200195
"acos",
201196
"add_months",
202197
"alias",
198+
"approx_count_distinct",
203199
"approxCountDistinct",
204200
"approxQuantile",
205201
"array_contains",
206202
"array_distinct",
203+
"array_except",
204+
"array_intersect",
207205
"array_join",
208206
"array_max",
209207
"array_min",
@@ -212,6 +210,7 @@ exportMethods("%<=>%",
212210
"array_repeat",
213211
"array_sort",
214212
"arrays_overlap",
213+
"array_union",
215214
"arrays_zip",
216215
"asc",
217216
"ascii",
@@ -255,6 +254,7 @@ exportMethods("%<=>%",
255254
"dayofweek",
256255
"dayofyear",
257256
"decode",
257+
"degrees",
258258
"dense_rank",
259259
"desc",
260260
"element_at",
@@ -271,6 +271,7 @@ exportMethods("%<=>%",
271271
"floor",
272272
"format_number",
273273
"format_string",
274+
"from_csv",
274275
"from_json",
275276
"from_unixtime",
276277
"from_utc_timestamp",
@@ -336,6 +337,7 @@ exportMethods("%<=>%",
336337
"posexplode",
337338
"posexplode_outer",
338339
"quarter",
340+
"radians",
339341
"rand",
340342
"randn",
341343
"rank",
@@ -349,12 +351,15 @@ exportMethods("%<=>%",
349351
"row_number",
350352
"rpad",
351353
"rtrim",
354+
"schema_of_csv",
355+
"schema_of_json",
352356
"second",
353357
"sha1",
354358
"sha2",
355359
"shiftLeft",
356360
"shiftRight",
357361
"shiftRightUnsigned",
362+
"shuffle",
358363
"sd",
359364
"sign",
360365
"signum",
@@ -381,6 +386,7 @@ exportMethods("%<=>%",
381386
"tanh",
382387
"toDegrees",
383388
"toRadians",
389+
"to_csv",
384390
"to_date",
385391
"to_json",
386392
"to_timestamp",
@@ -409,18 +415,14 @@ export("as.DataFrame",
409415
"cacheTable",
410416
"clearCache",
411417
"createDataFrame",
412-
"createExternalTable",
413418
"createTable",
414419
"currentDatabase",
415-
"dropTempTable",
416420
"dropTempView",
417-
"jsonFile",
418421
"listColumns",
419422
"listDatabases",
420423
"listFunctions",
421424
"listTables",
422425
"loadDF",
423-
"parquetFile",
424426
"read.df",
425427
"read.jdbc",
426428
"read.json",

0 commit comments

Comments
 (0)