Skip to content

Commit c0fc0c7

Browse files
committed
Merge remote-tracking branch 'upstream/master' into qrDecomposition
2 parents 39b0b22 + 4f4ba8f commit c0fc0c7

File tree

877 files changed

+83116
-11557
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

877 files changed

+83116
-11557
lines changed

.gitignore

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -65,6 +65,7 @@ scalastyle.txt
6565
scalastyle-output.xml
6666
R-unit-tests.log
6767
R/unit-tests.out
68+
python/lib/pyspark.zip
6869

6970
# For Hive
7071
metastore_db/

.rat-excludes

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,6 @@ graphlib-dot.min.js
3636
sorttable.js
3737
vis.min.js
3838
vis.min.css
39-
vis.map
4039
.*avsc
4140
.*txt
4241
.*json
@@ -74,5 +73,12 @@ logs
7473
.*scalastyle-output.xml
7574
.*dependency-reduced-pom.xml
7675
known_translations
76+
json_expectation
77+
local-1422981759269/*
78+
local-1422981780767/*
79+
local-1425081759269/*
80+
local-1426533911241/*
81+
local-1426633911242/*
82+
local-1430917381534/*
7783
DESCRIPTION
7884
NAMESPACE

LICENSE

Lines changed: 47 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -643,6 +643,36 @@ LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
643643
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
644644
THE SOFTWARE.
645645

646+
========================================================================
647+
For d3 (core/src/main/resources/org/apache/spark/ui/static/d3.min.js):
648+
========================================================================
649+
650+
Copyright (c) 2010-2015, Michael Bostock
651+
All rights reserved.
652+
653+
Redistribution and use in source and binary forms, with or without
654+
modification, are permitted provided that the following conditions are met:
655+
656+
* Redistributions of source code must retain the above copyright notice, this
657+
list of conditions and the following disclaimer.
658+
659+
* Redistributions in binary form must reproduce the above copyright notice,
660+
this list of conditions and the following disclaimer in the documentation
661+
and/or other materials provided with the distribution.
662+
663+
* The name Michael Bostock may not be used to endorse or promote products
664+
derived from this software without specific prior written permission.
665+
666+
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
667+
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
668+
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
669+
DISCLAIMED. IN NO EVENT SHALL MICHAEL BOSTOCK BE LIABLE FOR ANY DIRECT,
670+
INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
671+
BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
672+
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
673+
OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
674+
NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE,
675+
EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
646676

647677
========================================================================
648678
For Scala Interpreter classes (all .scala files in repl/src/main/scala
@@ -806,6 +836,22 @@ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
806836
See the License for the specific language governing permissions and
807837
limitations under the License.
808838

839+
========================================================================
840+
For vis.js (core/src/main/resources/org/apache/spark/ui/static/vis.min.js):
841+
========================================================================
842+
Copyright (C) 2010-2015 Almende B.V.
843+
844+
Vis.js is dual licensed under both
845+
846+
* The Apache 2.0 License
847+
http://www.apache.org/licenses/LICENSE-2.0
848+
849+
and
850+
851+
* The MIT License
852+
http://opensource.org/licenses/MIT
853+
854+
Vis.js may be distributed under either license.
809855

810856
========================================================================
811857
BSD-style licenses
@@ -815,7 +861,7 @@ The following components are provided under a BSD-style license. See project lin
815861

816862
(BSD 3 Clause) core (com.github.fommil.netlib:core:1.1.2 - https://github.com/fommil/netlib-java/core)
817863
(BSD 3 Clause) JPMML-Model (org.jpmml:pmml-model:1.1.15 - https://github.com/jpmml/jpmml-model)
818-
(BSD 3-clause style license) jblas (org.jblas:jblas:1.2.3 - http://jblas.org/)
864+
(BSD 3-clause style license) jblas (org.jblas:jblas:1.2.4 - http://jblas.org/)
819865
(BSD License) AntLR Parser Generator (antlr:antlr:2.7.7 - http://www.antlr.org/)
820866
(BSD License) Javolution (javolution:javolution:5.5.1 - http://javolution.org)
821867
(BSD licence) ANTLR ST4 4.0.4 (org.antlr:ST4:4.0.4 - http://www.stringtemplate.org)

R/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ The SparkR documentation (Rd files and HTML files) are not a part of the source
5252
SparkR comes with several sample programs in the `examples/src/main/r` directory.
5353
To run one of them, use `./bin/sparkR <filename> <args>`. For example:
5454

55-
./bin/sparkR examples/src/main/r/pi.R local[2]
55+
./bin/sparkR examples/src/main/r/dataframe.R
5656

5757
You can also run the unit-tests for SparkR by running (you need to install the [testthat](http://cran.r-project.org/web/packages/testthat/index.html) package first):
5858

@@ -63,5 +63,5 @@ You can also run the unit-tests for SparkR by running (you need to install the [
6363
The `./bin/spark-submit` and `./bin/sparkR` can also be used to submit jobs to YARN clusters. You will need to set YARN conf dir before doing so. For example on CDH you can run
6464
```
6565
export YARN_CONF_DIR=/etc/hadoop/conf
66-
./bin/spark-submit --master yarn examples/src/main/r/pi.R 4
66+
./bin/spark-submit --master yarn examples/src/main/r/dataframe.R
6767
```

R/pkg/DESCRIPTION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,11 +15,11 @@ Suggests:
1515
Description: R frontend for Spark
1616
License: Apache License (== 2.0)
1717
Collate:
18+
'schema.R'
1819
'generics.R'
1920
'jobj.R'
2021
'RDD.R'
2122
'pairRDD.R'
22-
'schema.R'
2323
'column.R'
2424
'group.R'
2525
'DataFrame.R'

R/pkg/NAMESPACE

Lines changed: 51 additions & 103 deletions
Original file line numberDiff line numberDiff line change
@@ -1,121 +1,47 @@
1-
#exportPattern("^[[:alpha:]]+")
2-
exportClasses("RDD")
3-
exportClasses("Broadcast")
4-
exportMethods(
5-
"aggregateByKey",
6-
"aggregateRDD",
7-
"cache",
8-
"cartesian",
9-
"checkpoint",
10-
"coalesce",
11-
"cogroup",
12-
"collect",
13-
"collectAsMap",
14-
"collectPartition",
15-
"combineByKey",
16-
"count",
17-
"countByKey",
18-
"countByValue",
19-
"distinct",
20-
"Filter",
21-
"filterRDD",
22-
"first",
23-
"flatMap",
24-
"flatMapValues",
25-
"fold",
26-
"foldByKey",
27-
"foreach",
28-
"foreachPartition",
29-
"fullOuterJoin",
30-
"glom",
31-
"groupByKey",
32-
"intersection",
33-
"join",
34-
"keyBy",
35-
"keys",
36-
"length",
37-
"lapply",
38-
"lapplyPartition",
39-
"lapplyPartitionsWithIndex",
40-
"leftOuterJoin",
41-
"lookup",
42-
"map",
43-
"mapPartitions",
44-
"mapPartitionsWithIndex",
45-
"mapValues",
46-
"maximum",
47-
"minimum",
48-
"numPartitions",
49-
"partitionBy",
50-
"persist",
51-
"pipeRDD",
52-
"reduce",
53-
"reduceByKey",
54-
"reduceByKeyLocally",
55-
"repartition",
56-
"rightOuterJoin",
57-
"sampleByKey",
58-
"sampleRDD",
59-
"saveAsTextFile",
60-
"saveAsObjectFile",
61-
"sortBy",
62-
"sortByKey",
63-
"subtract",
64-
"subtractByKey",
65-
"sumRDD",
66-
"take",
67-
"takeOrdered",
68-
"takeSample",
69-
"top",
70-
"unionRDD",
71-
"unpersist",
72-
"value",
73-
"values",
74-
"zipPartitions",
75-
"zipRDD",
76-
"zipWithIndex",
77-
"zipWithUniqueId"
78-
)
1+
# Imports from base R
2+
importFrom(methods, setGeneric, setMethod, setOldClass)
3+
4+
# Disable native libraries till we figure out how to package it
5+
# See SPARKR-7839
6+
#useDynLib(SparkR, stringHashCode)
797

808
# S3 methods exported
81-
export(
82-
"textFile",
83-
"objectFile",
84-
"parallelize",
85-
"hashCode",
86-
"includePackage",
87-
"broadcast",
88-
"setBroadcastValue",
89-
"setCheckpointDir"
90-
)
919
export("sparkR.init")
9210
export("sparkR.stop")
9311
export("print.jobj")
94-
useDynLib(SparkR, stringHashCode)
95-
importFrom(methods, setGeneric, setMethod, setOldClass)
96-
97-
# SparkRSQL
9812

9913
exportClasses("DataFrame")
10014

101-
exportMethods("columns",
15+
exportMethods("arrange",
16+
"cache",
17+
"collect",
18+
"columns",
19+
"count",
20+
"describe",
10221
"distinct",
10322
"dtypes",
10423
"except",
10524
"explain",
10625
"filter",
26+
"first",
27+
"group_by",
10728
"groupBy",
10829
"head",
10930
"insertInto",
11031
"intersect",
11132
"isLocal",
33+
"join",
11234
"limit",
11335
"orderBy",
36+
"mutate",
11437
"names",
38+
"persist",
11539
"printSchema",
11640
"registerTempTable",
41+
"rename",
11742
"repartition",
118-
"sampleDF",
43+
"sample",
44+
"sample_frac",
11945
"saveAsParquetFile",
12046
"saveAsTable",
12147
"saveDF",
@@ -124,42 +50,68 @@ exportMethods("columns",
12450
"selectExpr",
12551
"show",
12652
"showDF",
127-
"sortDF",
128-
"toJSON",
129-
"toRDD",
53+
"summarize",
54+
"take",
13055
"unionAll",
56+
"unpersist",
13157
"where",
13258
"withColumn",
133-
"withColumnRenamed")
59+
"withColumnRenamed",
60+
"write.df")
13461

13562
exportClasses("Column")
13663

13764
exportMethods("abs",
65+
"acos",
13866
"alias",
13967
"approxCountDistinct",
14068
"asc",
69+
"asin",
70+
"atan",
71+
"atan2",
14172
"avg",
14273
"cast",
74+
"cbrt",
75+
"ceiling",
14376
"contains",
77+
"cos",
78+
"cosh",
14479
"countDistinct",
14580
"desc",
14681
"endsWith",
82+
"exp",
83+
"expm1",
84+
"floor",
14785
"getField",
14886
"getItem",
87+
"hypot",
14988
"isNotNull",
15089
"isNull",
15190
"last",
15291
"like",
92+
"log",
93+
"log10",
94+
"log1p",
15395
"lower",
15496
"max",
15597
"mean",
15698
"min",
99+
"n",
100+
"n_distinct",
101+
"rint",
157102
"rlike",
103+
"sign",
104+
"sin",
105+
"sinh",
158106
"sqrt",
159107
"startsWith",
160108
"substr",
161109
"sum",
162110
"sumDistinct",
111+
"tan",
112+
"tanh",
113+
"toDegrees",
114+
"toRadians",
163115
"upper")
164116

165117
exportClasses("GroupedData")
@@ -174,19 +126,15 @@ export("cacheTable",
174126
"createExternalTable",
175127
"dropTempTable",
176128
"jsonFile",
177-
"jsonRDD",
178129
"loadDF",
179130
"parquetFile",
131+
"read.df",
180132
"sql",
181133
"table",
182134
"tableNames",
183135
"tables",
184-
"toDF",
185136
"uncacheTable")
186137

187-
export("sparkRSQL.init",
188-
"sparkRHive.init")
189-
190138
export("structField",
191139
"structField.jobj",
192140
"structField.character",

0 commit comments

Comments
 (0)