Skip to content

Commit 1502d13

Browse files
committed
Resolv conflict
2 parents d6bd416 + e86fbdb commit 1502d13

File tree

1,257 files changed

+40166
-19894
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

1,257 files changed

+40166
-19894
lines changed

.rat-excludes

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -28,6 +28,7 @@ spark-env.sh
2828
spark-env.cmd
2929
spark-env.sh.template
3030
log4j-defaults.properties
31+
log4j-defaults-repl.properties
3132
bootstrap-tooltip.js
3233
jquery-1.11.1.min.js
3334
d3.min.js
@@ -80,5 +81,8 @@ local-1425081759269/*
8081
local-1426533911241/*
8182
local-1426633911242/*
8283
local-1430917381534/*
84+
local-1430917381535_1
85+
local-1430917381535_2
8386
DESCRIPTION
8487
NAMESPACE
88+
test_support/*

LICENSE

Lines changed: 47 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -853,6 +853,52 @@ and
853853

854854
Vis.js may be distributed under either license.
855855

856+
========================================================================
857+
For dagre-d3 (core/src/main/resources/org/apache/spark/ui/static/dagre-d3.min.js):
858+
========================================================================
859+
Copyright (c) 2013 Chris Pettitt
860+
861+
Permission is hereby granted, free of charge, to any person obtaining a copy
862+
of this software and associated documentation files (the "Software"), to deal
863+
in the Software without restriction, including without limitation the rights
864+
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
865+
copies of the Software, and to permit persons to whom the Software is
866+
furnished to do so, subject to the following conditions:
867+
868+
The above copyright notice and this permission notice shall be included in
869+
all copies or substantial portions of the Software.
870+
871+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
872+
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
873+
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
874+
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
875+
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
876+
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
877+
THE SOFTWARE.
878+
879+
========================================================================
880+
For graphlib-dot (core/src/main/resources/org/apache/spark/ui/static/graphlib-dot.min.js):
881+
========================================================================
882+
Copyright (c) 2012-2013 Chris Pettitt
883+
884+
Permission is hereby granted, free of charge, to any person obtaining a copy
885+
of this software and associated documentation files (the "Software"), to deal
886+
in the Software without restriction, including without limitation the rights
887+
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
888+
copies of the Software, and to permit persons to whom the Software is
889+
furnished to do so, subject to the following conditions:
890+
891+
The above copyright notice and this permission notice shall be included in
892+
all copies or substantial portions of the Software.
893+
894+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
895+
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
896+
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
897+
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
898+
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
899+
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
900+
THE SOFTWARE.
901+
856902
========================================================================
857903
BSD-style licenses
858904
========================================================================
@@ -861,7 +907,7 @@ The following components are provided under a BSD-style license. See project lin
861907

862908
(BSD 3 Clause) core (com.github.fommil.netlib:core:1.1.2 - https://github.com/fommil/netlib-java/core)
863909
(BSD 3 Clause) JPMML-Model (org.jpmml:pmml-model:1.1.15 - https://github.com/jpmml/jpmml-model)
864-
(BSD 3-clause style license) jblas (org.jblas:jblas:1.2.3 - http://jblas.org/)
910+
(BSD 3-clause style license) jblas (org.jblas:jblas:1.2.4 - http://jblas.org/)
865911
(BSD License) AntLR Parser Generator (antlr:antlr:2.7.7 - http://www.antlr.org/)
866912
(BSD License) Javolution (javolution:javolution:5.5.1 - http://javolution.org)
867913
(BSD licence) ANTLR ST4 4.0.4 (org.antlr:ST4:4.0.4 - http://www.stringtemplate.org)

R/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ The SparkR documentation (Rd files and HTML files) are not a part of the source
5252
SparkR comes with several sample programs in the `examples/src/main/r` directory.
5353
To run one of them, use `./bin/sparkR <filename> <args>`. For example:
5454

55-
./bin/sparkR examples/src/main/r/pi.R local[2]
55+
./bin/sparkR examples/src/main/r/dataframe.R
5656

5757
You can also run the unit-tests for SparkR by running (you need to install the [testthat](http://cran.r-project.org/web/packages/testthat/index.html) package first):
5858

@@ -63,5 +63,5 @@ You can also run the unit-tests for SparkR by running (you need to install the [
6363
The `./bin/spark-submit` and `./bin/sparkR` can also be used to submit jobs to YARN clusters. You will need to set YARN conf dir before doing so. For example on CDH you can run
6464
```
6565
export YARN_CONF_DIR=/etc/hadoop/conf
66-
./bin/spark-submit --master yarn examples/src/main/r/pi.R 4
66+
./bin/spark-submit --master yarn examples/src/main/r/dataframe.R
6767
```

R/create-docs.sh

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -23,14 +23,14 @@
2323
# After running this script the html docs can be found in
2424
# $SPARK_HOME/R/pkg/html
2525

26+
set -o pipefail
27+
set -e
28+
2629
# Figure out where the script is
2730
export FWDIR="$(cd "`dirname "$0"`"; pwd)"
2831
pushd $FWDIR
2932

30-
# Generate Rd file
31-
Rscript -e 'library(devtools); devtools::document(pkg="./pkg", roclets=c("rd"))'
32-
33-
# Install the package
33+
# Install the package (this will also generate the Rd files)
3434
./install-dev.sh
3535

3636
# Now create HTML files

R/install-dev.sh

Lines changed: 10 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,11 +26,20 @@
2626
# NOTE(shivaram): Right now we use $SPARK_HOME/R/lib to be the installation directory
2727
# to load the SparkR package on the worker nodes.
2828

29+
set -o pipefail
30+
set -e
2931

3032
FWDIR="$(cd `dirname $0`; pwd)"
3133
LIB_DIR="$FWDIR/lib"
3234

3335
mkdir -p $LIB_DIR
3436

35-
# Install R
37+
pushd $FWDIR
38+
39+
# Generate Rd files if devtools is installed
40+
Rscript -e ' if("devtools" %in% rownames(installed.packages())) { library(devtools); devtools::document(pkg="./pkg", roclets=c("rd")) }'
41+
42+
# Install SparkR to $LIB_DIR
3643
R CMD INSTALL --library=$LIB_DIR $FWDIR/pkg/
44+
45+
popd

R/log4j.properties

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@
1919
log4j.rootCategory=INFO, file
2020
log4j.appender.file=org.apache.log4j.FileAppender
2121
log4j.appender.file.append=true
22-
log4j.appender.file.file=R-unit-tests.log
22+
log4j.appender.file.file=R/target/unit-tests.log
2323
log4j.appender.file.layout=org.apache.log4j.PatternLayout
2424
log4j.appender.file.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss.SSS} %t %p %c{1}: %m%n
2525

R/pkg/NAMESPACE

Lines changed: 29 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,9 @@
11
# Imports from base R
22
importFrom(methods, setGeneric, setMethod, setOldClass)
3-
useDynLib(SparkR, stringHashCode)
3+
4+
# Disable native libraries till we figure out how to package it
5+
# See SPARKR-7839
6+
#useDynLib(SparkR, stringHashCode)
47

58
# S3 methods exported
69
export("sparkR.init")
@@ -16,9 +19,11 @@ exportMethods("arrange",
1619
"count",
1720
"describe",
1821
"distinct",
22+
"dropna",
1923
"dtypes",
2024
"except",
2125
"explain",
26+
"fillna",
2227
"filter",
2328
"first",
2429
"group_by",
@@ -59,33 +64,56 @@ exportMethods("arrange",
5964
exportClasses("Column")
6065

6166
exportMethods("abs",
67+
"acos",
6268
"alias",
6369
"approxCountDistinct",
6470
"asc",
71+
"asin",
72+
"atan",
73+
"atan2",
6574
"avg",
6675
"cast",
76+
"cbrt",
77+
"ceiling",
6778
"contains",
79+
"cos",
80+
"cosh",
6881
"countDistinct",
6982
"desc",
7083
"endsWith",
84+
"exp",
85+
"expm1",
86+
"floor",
7187
"getField",
7288
"getItem",
89+
"hypot",
7390
"isNotNull",
7491
"isNull",
7592
"last",
7693
"like",
94+
"log",
95+
"log10",
96+
"log1p",
7797
"lower",
7898
"max",
7999
"mean",
80100
"min",
81101
"n",
82102
"n_distinct",
103+
"rint",
83104
"rlike",
105+
"sign",
106+
"sin",
107+
"sinh",
84108
"sqrt",
85109
"startsWith",
86110
"substr",
87111
"sum",
88112
"sumDistinct",
113+
"tan",
114+
"tanh",
115+
"toDegrees",
116+
"toRadians",
89117
"upper")
90118

91119
exportClasses("GroupedData")

0 commit comments

Comments
 (0)