Skip to content

Commit

Permalink
Merge branch 'master' into branch-pz-smj
Browse files Browse the repository at this point in the history
  • Loading branch information
zecevicp committed Dec 10, 2018
2 parents 0a5c8de + 3bc83de commit 07ff4d3
Show file tree
Hide file tree
Showing 1,699 changed files with 68,660 additions and 53,847 deletions.
1 change: 0 additions & 1 deletion .gitignore
Expand Up @@ -77,7 +77,6 @@ target/
unit-tests.log
work/
docs/.jekyll-metadata
*.crc

# For Hive
TempStatsStore/
Expand Down
50 changes: 0 additions & 50 deletions .travis.yml

This file was deleted.

2 changes: 1 addition & 1 deletion R/WINDOWS.md
Expand Up @@ -3,7 +3,7 @@
To build SparkR on Windows, the following steps are required

1. Install R (>= 3.1) and [Rtools](http://cran.r-project.org/bin/windows/Rtools/). Make sure to
include Rtools and R in `PATH`.
include Rtools and R in `PATH`. Note that support for R prior to version 3.4 is deprecated as of Spark 3.0.0.

2. Install
[JDK8](http://www.oracle.com/technetwork/java/javase/downloads/jdk8-downloads-2133151.html) and set
Expand Down
4 changes: 2 additions & 2 deletions R/pkg/DESCRIPTION
@@ -1,6 +1,6 @@
Package: SparkR
Type: Package
Version: 2.4.0
Version: 3.0.0
Title: R Frontend for Apache Spark
Description: Provides an R Frontend for Apache Spark.
Authors@R: c(person("Shivaram", "Venkataraman", role = c("aut", "cre"),
Expand All @@ -15,7 +15,7 @@ URL: http://www.apache.org/ http://spark.apache.org/
BugReports: http://spark.apache.org/contributing.html
SystemRequirements: Java (== 8)
Depends:
R (>= 3.0),
R (>= 3.1),
methods
Suggests:
knitr,
Expand Down
22 changes: 10 additions & 12 deletions R/pkg/NAMESPACE
Expand Up @@ -28,9 +28,8 @@ importFrom("utils", "download.file", "object.size", "packageVersion", "tail", "u

# S3 methods exported
export("sparkR.session")
export("sparkR.init")
export("sparkR.stop")
export("sparkR.session.stop")
export("sparkR.stop")
export("sparkR.conf")
export("sparkR.version")
export("sparkR.uiWebUrl")
Expand All @@ -42,9 +41,6 @@ export("sparkR.callJStatic")

export("install.spark")

export("sparkRSQL.init",
"sparkRHive.init")

# MLlib integration
exportMethods("glm",
"spark.glm",
Expand All @@ -70,7 +66,8 @@ exportMethods("glm",
"spark.svmLinear",
"spark.fpGrowth",
"spark.freqItemsets",
"spark.associationRules")
"spark.associationRules",
"spark.findFrequentSequentialPatterns")

# Job group lifecycle management methods
export("setJobGroup",
Expand Down Expand Up @@ -150,15 +147,13 @@ exportMethods("arrange",
"printSchema",
"randomSplit",
"rbind",
"registerTempTable",
"rename",
"repartition",
"repartitionByRange",
"rollup",
"sample",
"sample_frac",
"sampleBy",
"saveAsParquetFile",
"saveAsTable",
"saveDF",
"schema",
Expand Down Expand Up @@ -200,6 +195,7 @@ exportMethods("%<=>%",
"acos",
"add_months",
"alias",
"approx_count_distinct",
"approxCountDistinct",
"approxQuantile",
"array_contains",
Expand Down Expand Up @@ -258,6 +254,7 @@ exportMethods("%<=>%",
"dayofweek",
"dayofyear",
"decode",
"degrees",
"dense_rank",
"desc",
"element_at",
Expand All @@ -274,6 +271,7 @@ exportMethods("%<=>%",
"floor",
"format_number",
"format_string",
"from_csv",
"from_json",
"from_unixtime",
"from_utc_timestamp",
Expand Down Expand Up @@ -339,6 +337,7 @@ exportMethods("%<=>%",
"posexplode",
"posexplode_outer",
"quarter",
"radians",
"rand",
"randn",
"rank",
Expand All @@ -352,6 +351,8 @@ exportMethods("%<=>%",
"row_number",
"rpad",
"rtrim",
"schema_of_csv",
"schema_of_json",
"second",
"sha1",
"sha2",
Expand Down Expand Up @@ -385,6 +386,7 @@ exportMethods("%<=>%",
"tanh",
"toDegrees",
"toRadians",
"to_csv",
"to_date",
"to_json",
"to_timestamp",
Expand Down Expand Up @@ -413,18 +415,14 @@ export("as.DataFrame",
"cacheTable",
"clearCache",
"createDataFrame",
"createExternalTable",
"createTable",
"currentDatabase",
"dropTempTable",
"dropTempView",
"jsonFile",
"listColumns",
"listDatabases",
"listFunctions",
"listTables",
"loadDF",
"parquetFile",
"read.df",
"read.jdbc",
"read.json",
Expand Down

0 comments on commit 07ff4d3

Please sign in to comment.