Skip to content

Commit 7d265f5

Browse files
committed
Merge remote-tracking branch 'upstream/master' into pyspark-update-cloudpickle-42-SPARK-23159
2 parents 2d19f0a + 4586ead commit 7d265f5

File tree

281 files changed

+6819
-4447
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

281 files changed

+6819
-4447
lines changed

R/pkg/R/DataFrame.R

Lines changed: 0 additions & 92 deletions
Large diffs are not rendered by default.

R/pkg/R/SQLContext.R

Lines changed: 0 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -123,7 +123,6 @@ infer_type <- function(x) {
123123
#' @return a list of config values with keys as their names
124124
#' @rdname sparkR.conf
125125
#' @name sparkR.conf
126-
#' @export
127126
#' @examples
128127
#'\dontrun{
129128
#' sparkR.session()
@@ -163,7 +162,6 @@ sparkR.conf <- function(key, defaultValue) {
163162
#' @return a character string of the Spark version
164163
#' @rdname sparkR.version
165164
#' @name sparkR.version
166-
#' @export
167165
#' @examples
168166
#'\dontrun{
169167
#' sparkR.session()
@@ -191,7 +189,6 @@ getDefaultSqlSource <- function() {
191189
#' limited by length of the list or number of rows of the data.frame
192190
#' @return A SparkDataFrame.
193191
#' @rdname createDataFrame
194-
#' @export
195192
#' @examples
196193
#'\dontrun{
197194
#' sparkR.session()
@@ -294,7 +291,6 @@ createDataFrame <- function(x, ...) {
294291

295292
#' @rdname createDataFrame
296293
#' @aliases createDataFrame
297-
#' @export
298294
#' @method as.DataFrame default
299295
#' @note as.DataFrame since 1.6.0
300296
as.DataFrame.default <- function(data, schema = NULL, samplingRatio = 1.0, numPartitions = NULL) {
@@ -304,7 +300,6 @@ as.DataFrame.default <- function(data, schema = NULL, samplingRatio = 1.0, numPa
304300
#' @param ... additional argument(s).
305301
#' @rdname createDataFrame
306302
#' @aliases as.DataFrame
307-
#' @export
308303
as.DataFrame <- function(data, ...) {
309304
dispatchFunc("as.DataFrame(data, schema = NULL)", data, ...)
310305
}
@@ -342,7 +337,6 @@ setMethod("toDF", signature(x = "RDD"),
342337
#' @param ... additional external data source specific named properties.
343338
#' @return SparkDataFrame
344339
#' @rdname read.json
345-
#' @export
346340
#' @examples
347341
#'\dontrun{
348342
#' sparkR.session()
@@ -371,7 +365,6 @@ read.json <- function(x, ...) {
371365

372366
#' @rdname read.json
373367
#' @name jsonFile
374-
#' @export
375368
#' @method jsonFile default
376369
#' @note jsonFile since 1.4.0
377370
jsonFile.default <- function(path) {
@@ -423,7 +416,6 @@ jsonRDD <- function(sqlContext, rdd, schema = NULL, samplingRatio = 1.0) {
423416
#' @param ... additional external data source specific named properties.
424417
#' @return SparkDataFrame
425418
#' @rdname read.orc
426-
#' @export
427419
#' @name read.orc
428420
#' @note read.orc since 2.0.0
429421
read.orc <- function(path, ...) {
@@ -444,7 +436,6 @@ read.orc <- function(path, ...) {
444436
#' @param path path of file to read. A vector of multiple paths is allowed.
445437
#' @return SparkDataFrame
446438
#' @rdname read.parquet
447-
#' @export
448439
#' @name read.parquet
449440
#' @method read.parquet default
450441
#' @note read.parquet since 1.6.0
@@ -466,7 +457,6 @@ read.parquet <- function(x, ...) {
466457
#' @param ... argument(s) passed to the method.
467458
#' @rdname read.parquet
468459
#' @name parquetFile
469-
#' @export
470460
#' @method parquetFile default
471461
#' @note parquetFile since 1.4.0
472462
parquetFile.default <- function(...) {
@@ -490,7 +480,6 @@ parquetFile <- function(x, ...) {
490480
#' @param ... additional external data source specific named properties.
491481
#' @return SparkDataFrame
492482
#' @rdname read.text
493-
#' @export
494483
#' @examples
495484
#'\dontrun{
496485
#' sparkR.session()
@@ -522,7 +511,6 @@ read.text <- function(x, ...) {
522511
#' @param sqlQuery A character vector containing the SQL query
523512
#' @return SparkDataFrame
524513
#' @rdname sql
525-
#' @export
526514
#' @examples
527515
#'\dontrun{
528516
#' sparkR.session()
@@ -556,7 +544,6 @@ sql <- function(x, ...) {
556544
#' @return SparkDataFrame
557545
#' @rdname tableToDF
558546
#' @name tableToDF
559-
#' @export
560547
#' @examples
561548
#'\dontrun{
562549
#' sparkR.session()
@@ -591,7 +578,6 @@ tableToDF <- function(tableName) {
591578
#' @rdname read.df
592579
#' @name read.df
593580
#' @seealso \link{read.json}
594-
#' @export
595581
#' @examples
596582
#'\dontrun{
597583
#' sparkR.session()
@@ -681,7 +667,6 @@ loadDF <- function(x = NULL, ...) {
681667
#' @return SparkDataFrame
682668
#' @rdname read.jdbc
683669
#' @name read.jdbc
684-
#' @export
685670
#' @examples
686671
#'\dontrun{
687672
#' sparkR.session()
@@ -734,7 +719,6 @@ read.jdbc <- function(url, tableName,
734719
#' @rdname read.stream
735720
#' @name read.stream
736721
#' @seealso \link{write.stream}
737-
#' @export
738722
#' @examples
739723
#'\dontrun{
740724
#' sparkR.session()

R/pkg/R/WindowSpec.R

Lines changed: 0 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,6 @@ NULL
2828
#' @seealso \link{windowPartitionBy}, \link{windowOrderBy}
2929
#'
3030
#' @param sws A Java object reference to the backing Scala WindowSpec
31-
#' @export
3231
#' @note WindowSpec since 2.0.0
3332
setClass("WindowSpec",
3433
slots = list(sws = "jobj"))
@@ -44,7 +43,6 @@ windowSpec <- function(sws) {
4443
}
4544

4645
#' @rdname show
47-
#' @export
4846
#' @note show(WindowSpec) since 2.0.0
4947
setMethod("show", "WindowSpec",
5048
function(object) {
@@ -63,7 +61,6 @@ setMethod("show", "WindowSpec",
6361
#' @name partitionBy
6462
#' @aliases partitionBy,WindowSpec-method
6563
#' @family windowspec_method
66-
#' @export
6764
#' @examples
6865
#' \dontrun{
6966
#' partitionBy(ws, "col1", "col2")
@@ -97,7 +94,6 @@ setMethod("partitionBy",
9794
#' @aliases orderBy,WindowSpec,character-method
9895
#' @family windowspec_method
9996
#' @seealso See \link{arrange} for use in sorting a SparkDataFrame
100-
#' @export
10197
#' @examples
10298
#' \dontrun{
10399
#' orderBy(ws, "col1", "col2")
@@ -113,7 +109,6 @@ setMethod("orderBy",
113109
#' @rdname orderBy
114110
#' @name orderBy
115111
#' @aliases orderBy,WindowSpec,Column-method
116-
#' @export
117112
#' @note orderBy(WindowSpec, Column) since 2.0.0
118113
setMethod("orderBy",
119114
signature(x = "WindowSpec", col = "Column"),
@@ -142,7 +137,6 @@ setMethod("orderBy",
142137
#' @aliases rowsBetween,WindowSpec,numeric,numeric-method
143138
#' @name rowsBetween
144139
#' @family windowspec_method
145-
#' @export
146140
#' @examples
147141
#' \dontrun{
148142
#' rowsBetween(ws, 0, 3)
@@ -174,7 +168,6 @@ setMethod("rowsBetween",
174168
#' @aliases rangeBetween,WindowSpec,numeric,numeric-method
175169
#' @name rangeBetween
176170
#' @family windowspec_method
177-
#' @export
178171
#' @examples
179172
#' \dontrun{
180173
#' rangeBetween(ws, 0, 3)
@@ -202,7 +195,6 @@ setMethod("rangeBetween",
202195
#' @name over
203196
#' @aliases over,Column,WindowSpec-method
204197
#' @family colum_func
205-
#' @export
206198
#' @examples
207199
#' \dontrun{
208200
#' df <- createDataFrame(mtcars)

R/pkg/R/broadcast.R

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -32,14 +32,12 @@
3232
# @seealso broadcast
3333
#
3434
# @param id Id of the backing Spark broadcast variable
35-
# @export
3635
setClass("Broadcast", slots = list(id = "character"))
3736

3837
# @rdname broadcast-class
3938
# @param value Value of the broadcast variable
4039
# @param jBroadcastRef reference to the backing Java broadcast object
4140
# @param objName name of broadcasted object
42-
# @export
4341
Broadcast <- function(id, value, jBroadcastRef, objName) {
4442
.broadcastValues[[id]] <- value
4543
.broadcastNames[[as.character(objName)]] <- jBroadcastRef
@@ -73,7 +71,6 @@ setMethod("value",
7371

7472
# @param bcastId The id of broadcast variable to set
7573
# @param value The value to be set
76-
# @export
7774
setBroadcastValue <- function(bcastId, value) {
7875
bcastIdStr <- as.character(bcastId)
7976
.broadcastValues[[bcastIdStr]] <- value

R/pkg/R/catalog.R

Lines changed: 0 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,6 @@
3434
#' @return A SparkDataFrame.
3535
#' @rdname createExternalTable-deprecated
3636
#' @seealso \link{createTable}
37-
#' @export
3837
#' @examples
3938
#'\dontrun{
4039
#' sparkR.session()
@@ -71,7 +70,6 @@ createExternalTable <- function(x, ...) {
7170
#' @return A SparkDataFrame.
7271
#' @rdname createTable
7372
#' @seealso \link{createExternalTable}
74-
#' @export
7573
#' @examples
7674
#'\dontrun{
7775
#' sparkR.session()
@@ -110,7 +108,6 @@ createTable <- function(tableName, path = NULL, source = NULL, schema = NULL, ..
110108
#' identifier is provided, it refers to a table in the current database.
111109
#' @return SparkDataFrame
112110
#' @rdname cacheTable
113-
#' @export
114111
#' @examples
115112
#'\dontrun{
116113
#' sparkR.session()
@@ -140,7 +137,6 @@ cacheTable <- function(x, ...) {
140137
#' identifier is provided, it refers to a table in the current database.
141138
#' @return SparkDataFrame
142139
#' @rdname uncacheTable
143-
#' @export
144140
#' @examples
145141
#'\dontrun{
146142
#' sparkR.session()
@@ -167,7 +163,6 @@ uncacheTable <- function(x, ...) {
167163
#' Removes all cached tables from the in-memory cache.
168164
#'
169165
#' @rdname clearCache
170-
#' @export
171166
#' @examples
172167
#' \dontrun{
173168
#' clearCache()
@@ -193,7 +188,6 @@ clearCache <- function() {
193188
#' @param tableName The name of the SparkSQL table to be dropped.
194189
#' @seealso \link{dropTempView}
195190
#' @rdname dropTempTable-deprecated
196-
#' @export
197191
#' @examples
198192
#' \dontrun{
199193
#' sparkR.session()
@@ -225,7 +219,6 @@ dropTempTable <- function(x, ...) {
225219
#' @return TRUE if the view is dropped successfully, FALSE otherwise.
226220
#' @rdname dropTempView
227221
#' @name dropTempView
228-
#' @export
229222
#' @examples
230223
#' \dontrun{
231224
#' sparkR.session()
@@ -251,7 +244,6 @@ dropTempView <- function(viewName) {
251244
#' @return a SparkDataFrame
252245
#' @rdname tables
253246
#' @seealso \link{listTables}
254-
#' @export
255247
#' @examples
256248
#'\dontrun{
257249
#' sparkR.session()
@@ -276,7 +268,6 @@ tables <- function(x, ...) {
276268
#' @param databaseName (optional) name of the database
277269
#' @return a list of table names
278270
#' @rdname tableNames
279-
#' @export
280271
#' @examples
281272
#'\dontrun{
282273
#' sparkR.session()
@@ -304,7 +295,6 @@ tableNames <- function(x, ...) {
304295
#' @return name of the current default database.
305296
#' @rdname currentDatabase
306297
#' @name currentDatabase
307-
#' @export
308298
#' @examples
309299
#' \dontrun{
310300
#' sparkR.session()
@@ -324,7 +314,6 @@ currentDatabase <- function() {
324314
#' @param databaseName name of the database
325315
#' @rdname setCurrentDatabase
326316
#' @name setCurrentDatabase
327-
#' @export
328317
#' @examples
329318
#' \dontrun{
330319
#' sparkR.session()
@@ -347,7 +336,6 @@ setCurrentDatabase <- function(databaseName) {
347336
#' @return a SparkDataFrame of the list of databases.
348337
#' @rdname listDatabases
349338
#' @name listDatabases
350-
#' @export
351339
#' @examples
352340
#' \dontrun{
353341
#' sparkR.session()
@@ -370,7 +358,6 @@ listDatabases <- function() {
370358
#' @rdname listTables
371359
#' @name listTables
372360
#' @seealso \link{tables}
373-
#' @export
374361
#' @examples
375362
#' \dontrun{
376363
#' sparkR.session()
@@ -403,7 +390,6 @@ listTables <- function(databaseName = NULL) {
403390
#' @return a SparkDataFrame of the list of column descriptions.
404391
#' @rdname listColumns
405392
#' @name listColumns
406-
#' @export
407393
#' @examples
408394
#' \dontrun{
409395
#' sparkR.session()
@@ -433,7 +419,6 @@ listColumns <- function(tableName, databaseName = NULL) {
433419
#' @return a SparkDataFrame of the list of function descriptions.
434420
#' @rdname listFunctions
435421
#' @name listFunctions
436-
#' @export
437422
#' @examples
438423
#' \dontrun{
439424
#' sparkR.session()
@@ -463,7 +448,6 @@ listFunctions <- function(databaseName = NULL) {
463448
#' identifier is provided, it refers to a table in the current database.
464449
#' @rdname recoverPartitions
465450
#' @name recoverPartitions
466-
#' @export
467451
#' @examples
468452
#' \dontrun{
469453
#' sparkR.session()
@@ -490,7 +474,6 @@ recoverPartitions <- function(tableName) {
490474
#' identifier is provided, it refers to a table in the current database.
491475
#' @rdname refreshTable
492476
#' @name refreshTable
493-
#' @export
494477
#' @examples
495478
#' \dontrun{
496479
#' sparkR.session()
@@ -512,7 +495,6 @@ refreshTable <- function(tableName) {
512495
#' @param path the path of the data source.
513496
#' @rdname refreshByPath
514497
#' @name refreshByPath
515-
#' @export
516498
#' @examples
517499
#' \dontrun{
518500
#' sparkR.session()

0 commit comments

Comments
 (0)