Skip to content

Commit af1e083

Browse files
committed
Merge branch 'master' into kafka-0.8.2-test-cleanup
2 parents 4298ac2 + 10ba188 commit af1e083

File tree

1,374 files changed

+86978
-32774
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

1,374 files changed

+86978
-32774
lines changed

.gitignore

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -65,6 +65,7 @@ scalastyle.txt
6565
scalastyle-output.xml
6666
R-unit-tests.log
6767
R/unit-tests.out
68+
python/lib/pyspark.zip
6869

6970
# For Hive
7071
metastore_db/

.rat-excludes

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,6 @@ graphlib-dot.min.js
3636
sorttable.js
3737
vis.min.js
3838
vis.min.css
39-
vis.map
4039
.*avsc
4140
.*txt
4241
.*json
@@ -80,6 +79,9 @@ local-1422981780767/*
8079
local-1425081759269/*
8180
local-1426533911241/*
8281
local-1426633911242/*
83-
local-1427397477963/*
82+
local-1430917381534/*
83+
local-1430917381535_1
84+
local-1430917381535_2
8485
DESCRIPTION
8586
NAMESPACE
87+
test_support/*

LICENSE

Lines changed: 63 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -836,6 +836,68 @@ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
836836
See the License for the specific language governing permissions and
837837
limitations under the License.
838838

839+
========================================================================
840+
For vis.js (core/src/main/resources/org/apache/spark/ui/static/vis.min.js):
841+
========================================================================
842+
Copyright (C) 2010-2015 Almende B.V.
843+
844+
Vis.js is dual licensed under both
845+
846+
* The Apache 2.0 License
847+
http://www.apache.org/licenses/LICENSE-2.0
848+
849+
and
850+
851+
* The MIT License
852+
http://opensource.org/licenses/MIT
853+
854+
Vis.js may be distributed under either license.
855+
856+
========================================================================
857+
For dagre-d3 (core/src/main/resources/org/apache/spark/ui/static/dagre-d3.min.js):
858+
========================================================================
859+
Copyright (c) 2013 Chris Pettitt
860+
861+
Permission is hereby granted, free of charge, to any person obtaining a copy
862+
of this software and associated documentation files (the "Software"), to deal
863+
in the Software without restriction, including without limitation the rights
864+
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
865+
copies of the Software, and to permit persons to whom the Software is
866+
furnished to do so, subject to the following conditions:
867+
868+
The above copyright notice and this permission notice shall be included in
869+
all copies or substantial portions of the Software.
870+
871+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
872+
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
873+
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
874+
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
875+
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
876+
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
877+
THE SOFTWARE.
878+
879+
========================================================================
880+
For graphlib-dot (core/src/main/resources/org/apache/spark/ui/static/graphlib-dot.min.js):
881+
========================================================================
882+
Copyright (c) 2012-2013 Chris Pettitt
883+
884+
Permission is hereby granted, free of charge, to any person obtaining a copy
885+
of this software and associated documentation files (the "Software"), to deal
886+
in the Software without restriction, including without limitation the rights
887+
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
888+
copies of the Software, and to permit persons to whom the Software is
889+
furnished to do so, subject to the following conditions:
890+
891+
The above copyright notice and this permission notice shall be included in
892+
all copies or substantial portions of the Software.
893+
894+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
895+
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
896+
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
897+
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
898+
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
899+
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
900+
THE SOFTWARE.
839901

840902
========================================================================
841903
BSD-style licenses
@@ -845,7 +907,7 @@ The following components are provided under a BSD-style license. See project lin
845907

846908
(BSD 3 Clause) core (com.github.fommil.netlib:core:1.1.2 - https://github.com/fommil/netlib-java/core)
847909
(BSD 3 Clause) JPMML-Model (org.jpmml:pmml-model:1.1.15 - https://github.com/jpmml/jpmml-model)
848-
(BSD 3-clause style license) jblas (org.jblas:jblas:1.2.3 - http://jblas.org/)
910+
(BSD 3-clause style license) jblas (org.jblas:jblas:1.2.4 - http://jblas.org/)
849911
(BSD License) AntLR Parser Generator (antlr:antlr:2.7.7 - http://www.antlr.org/)
850912
(BSD License) Javolution (javolution:javolution:5.5.1 - http://javolution.org)
851913
(BSD licence) ANTLR ST4 4.0.4 (org.antlr:ST4:4.0.4 - http://www.stringtemplate.org)

R/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ The SparkR documentation (Rd files and HTML files) are not a part of the source
5252
SparkR comes with several sample programs in the `examples/src/main/r` directory.
5353
To run one of them, use `./bin/sparkR <filename> <args>`. For example:
5454

55-
./bin/sparkR examples/src/main/r/pi.R local[2]
55+
./bin/sparkR examples/src/main/r/dataframe.R
5656

5757
You can also run the unit-tests for SparkR by running (you need to install the [testthat](http://cran.r-project.org/web/packages/testthat/index.html) package first):
5858

@@ -63,5 +63,5 @@ You can also run the unit-tests for SparkR by running (you need to install the [
6363
The `./bin/spark-submit` and `./bin/sparkR` can also be used to submit jobs to YARN clusters. You will need to set YARN conf dir before doing so. For example on CDH you can run
6464
```
6565
export YARN_CONF_DIR=/etc/hadoop/conf
66-
./bin/spark-submit --master yarn examples/src/main/r/pi.R 4
66+
./bin/spark-submit --master yarn examples/src/main/r/dataframe.R
6767
```

R/create-docs.sh

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,9 @@
2323
# After running this script the html docs can be found in
2424
# $SPARK_HOME/R/pkg/html
2525

26+
set -o pipefail
27+
set -e
28+
2629
# Figure out where the script is
2730
export FWDIR="$(cd "`dirname "$0"`"; pwd)"
2831
pushd $FWDIR

R/install-dev.sh

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -26,6 +26,8 @@
2626
# NOTE(shivaram): Right now we use $SPARK_HOME/R/lib to be the installation directory
2727
# to load the SparkR package on the worker nodes.
2828

29+
set -o pipefail
30+
set -e
2931

3032
FWDIR="$(cd `dirname $0`; pwd)"
3133
LIB_DIR="$FWDIR/lib"

R/pkg/DESCRIPTION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,11 +15,11 @@ Suggests:
1515
Description: R frontend for Spark
1616
License: Apache License (== 2.0)
1717
Collate:
18+
'schema.R'
1819
'generics.R'
1920
'jobj.R'
2021
'RDD.R'
2122
'pairRDD.R'
22-
'schema.R'
2323
'column.R'
2424
'group.R'
2525
'DataFrame.R'

R/pkg/NAMESPACE

Lines changed: 53 additions & 103 deletions
Original file line numberDiff line numberDiff line change
@@ -1,121 +1,49 @@
1-
#exportPattern("^[[:alpha:]]+")
2-
exportClasses("RDD")
3-
exportClasses("Broadcast")
4-
exportMethods(
5-
"aggregateByKey",
6-
"aggregateRDD",
7-
"cache",
8-
"cartesian",
9-
"checkpoint",
10-
"coalesce",
11-
"cogroup",
12-
"collect",
13-
"collectAsMap",
14-
"collectPartition",
15-
"combineByKey",
16-
"count",
17-
"countByKey",
18-
"countByValue",
19-
"distinct",
20-
"Filter",
21-
"filterRDD",
22-
"first",
23-
"flatMap",
24-
"flatMapValues",
25-
"fold",
26-
"foldByKey",
27-
"foreach",
28-
"foreachPartition",
29-
"fullOuterJoin",
30-
"glom",
31-
"groupByKey",
32-
"intersection",
33-
"join",
34-
"keyBy",
35-
"keys",
36-
"length",
37-
"lapply",
38-
"lapplyPartition",
39-
"lapplyPartitionsWithIndex",
40-
"leftOuterJoin",
41-
"lookup",
42-
"map",
43-
"mapPartitions",
44-
"mapPartitionsWithIndex",
45-
"mapValues",
46-
"maximum",
47-
"minimum",
48-
"numPartitions",
49-
"partitionBy",
50-
"persist",
51-
"pipeRDD",
52-
"reduce",
53-
"reduceByKey",
54-
"reduceByKeyLocally",
55-
"repartition",
56-
"rightOuterJoin",
57-
"sampleByKey",
58-
"sampleRDD",
59-
"saveAsTextFile",
60-
"saveAsObjectFile",
61-
"sortBy",
62-
"sortByKey",
63-
"subtract",
64-
"subtractByKey",
65-
"sumRDD",
66-
"take",
67-
"takeOrdered",
68-
"takeSample",
69-
"top",
70-
"unionRDD",
71-
"unpersist",
72-
"value",
73-
"values",
74-
"zipPartitions",
75-
"zipRDD",
76-
"zipWithIndex",
77-
"zipWithUniqueId"
78-
)
1+
# Imports from base R
2+
importFrom(methods, setGeneric, setMethod, setOldClass)
3+
4+
# Disable native libraries till we figure out how to package it
5+
# See SPARKR-7839
6+
#useDynLib(SparkR, stringHashCode)
797

808
# S3 methods exported
81-
export(
82-
"textFile",
83-
"objectFile",
84-
"parallelize",
85-
"hashCode",
86-
"includePackage",
87-
"broadcast",
88-
"setBroadcastValue",
89-
"setCheckpointDir"
90-
)
919
export("sparkR.init")
9210
export("sparkR.stop")
9311
export("print.jobj")
94-
useDynLib(SparkR, stringHashCode)
95-
importFrom(methods, setGeneric, setMethod, setOldClass)
96-
97-
# SparkRSQL
9812

9913
exportClasses("DataFrame")
10014

101-
exportMethods("columns",
15+
exportMethods("arrange",
16+
"cache",
17+
"collect",
18+
"columns",
19+
"count",
20+
"describe",
10221
"distinct",
22+
"dropna",
10323
"dtypes",
10424
"except",
10525
"explain",
26+
"fillna",
10627
"filter",
28+
"first",
29+
"group_by",
10730
"groupBy",
10831
"head",
10932
"insertInto",
11033
"intersect",
11134
"isLocal",
35+
"join",
11236
"limit",
11337
"orderBy",
38+
"mutate",
11439
"names",
40+
"persist",
11541
"printSchema",
11642
"registerTempTable",
43+
"rename",
11744
"repartition",
118-
"sampleDF",
45+
"sample",
46+
"sample_frac",
11947
"saveAsParquetFile",
12048
"saveAsTable",
12149
"saveDF",
@@ -124,42 +52,68 @@ exportMethods("columns",
12452
"selectExpr",
12553
"show",
12654
"showDF",
127-
"sortDF",
128-
"toJSON",
129-
"toRDD",
55+
"summarize",
56+
"take",
13057
"unionAll",
58+
"unpersist",
13159
"where",
13260
"withColumn",
133-
"withColumnRenamed")
61+
"withColumnRenamed",
62+
"write.df")
13463

13564
exportClasses("Column")
13665

13766
exportMethods("abs",
67+
"acos",
13868
"alias",
13969
"approxCountDistinct",
14070
"asc",
71+
"asin",
72+
"atan",
73+
"atan2",
14174
"avg",
14275
"cast",
76+
"cbrt",
77+
"ceiling",
14378
"contains",
79+
"cos",
80+
"cosh",
14481
"countDistinct",
14582
"desc",
14683
"endsWith",
84+
"exp",
85+
"expm1",
86+
"floor",
14787
"getField",
14888
"getItem",
89+
"hypot",
14990
"isNotNull",
15091
"isNull",
15192
"last",
15293
"like",
94+
"log",
95+
"log10",
96+
"log1p",
15397
"lower",
15498
"max",
15599
"mean",
156100
"min",
101+
"n",
102+
"n_distinct",
103+
"rint",
157104
"rlike",
105+
"sign",
106+
"sin",
107+
"sinh",
158108
"sqrt",
159109
"startsWith",
160110
"substr",
161111
"sum",
162112
"sumDistinct",
113+
"tan",
114+
"tanh",
115+
"toDegrees",
116+
"toRadians",
163117
"upper")
164118

165119
exportClasses("GroupedData")
@@ -174,19 +128,15 @@ export("cacheTable",
174128
"createExternalTable",
175129
"dropTempTable",
176130
"jsonFile",
177-
"jsonRDD",
178131
"loadDF",
179132
"parquetFile",
133+
"read.df",
180134
"sql",
181135
"table",
182136
"tableNames",
183137
"tables",
184-
"toDF",
185138
"uncacheTable")
186139

187-
export("sparkRSQL.init",
188-
"sparkRHive.init")
189-
190140
export("structField",
191141
"structField.jobj",
192142
"structField.character",

0 commit comments

Comments
 (0)