Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
217 commits
Select commit Hold shift + click to select a range
07923c4
support kafka in Python
Dec 16, 2014
75d485e
add mqtt
Dec 16, 2014
048dbe6
fix python style
Dec 16, 2014
5697a01
bypass decoder in scala
Dec 18, 2014
98c8d17
fix python style
Dec 18, 2014
815de54
[SPARK-4946] [CORE] Using AkkaUtils.askWithReply in MapOutputTracker.…
YanTangZhai Dec 29, 2014
8d72341
[Minor] Fix a typo of type parameter in JavaUtils.scala
sarutak Dec 29, 2014
02b55de
[SPARK-4409][MLlib] Additional Linear Algebra Utils
brkyvz Dec 29, 2014
9bc0df6
SPARK-4968: takeOrdered to skip reduce step in case mappers return no…
Dec 29, 2014
6cf6fdf
SPARK-4156 [MLLIB] EM algorithm for GMMs
tgaloppo Dec 29, 2014
343db39
Added setMinCount to Word2Vec.scala
ganonp Dec 29, 2014
040d6f2
[SPARK-4972][MLlib] Updated the scala doc for lasso and ridge regress…
Dec 30, 2014
9077e72
[SPARK-4920][UI] add version on master and worker page for standalone…
liyezhang556520 Dec 30, 2014
efa80a5
[SPARK-4882] Register PythonBroadcast with Kryo so that PySpark works…
JoshRosen Dec 30, 2014
480bd1d
[SPARK-4908][SQL] Prevent multiple concurrent hive native commands
marmbrus Dec 30, 2014
94d60b7
[SQL] enable view test
adrian-wang Dec 30, 2014
65357f1
[SPARK-4975][SQL] Fix HiveInspectorSuite test failure
scwf Dec 30, 2014
5595eaa
[SPARK-4959] [SQL] Attributes are case sensitive when using a select …
chenghao-intel Dec 30, 2014
63b84b7
[SPARK-4904] [SQL] Remove the unnecessary code change in Generic UDF
chenghao-intel Dec 30, 2014
daac221
[SPARK-5002][SQL] Using ascending by default when not specify order i…
scwf Dec 30, 2014
53f0a00
[Spark-4512] [SQL] Unresolved Attribute Exception in Sort By
chenghao-intel Dec 30, 2014
19a8802
[SPARK-4493][SQL] Tests for IsNull / IsNotNull in the ParquetFilterSuite
liancheng Dec 30, 2014
f7a41a0
[SPARK-4916][SQL][DOCS]Update SQL programming guide about cache section
luogankun Dec 30, 2014
2deac74
[SPARK-4930][SQL][DOCS]Update SQL programming guide, CACHE TABLE is e…
luogankun Dec 30, 2014
a75dd83
[SPARK-4928][SQL] Fix: Operator '>,<,>=,<=' with decimal between diff…
guowei2 Dec 30, 2014
61a99f6
[SPARK-4937][SQL] Normalizes conjunctions and disjunctions to elimina…
liancheng Dec 30, 2014
7425bec
[SPARK-4386] Improve performance when writing Parquet files
MickDavies Dec 30, 2014
8f29b7c
[SPARK-4935][SQL] When hive.cli.print.header configured, spark-sql ab…
scwf Dec 30, 2014
07fa191
[SPARK-4570][SQL]add BroadcastLeftSemiJoinHash
wangxiaojing Dec 30, 2014
b239ea1
SPARK-3955 part 2 [CORE] [HOTFIX] Different versions between jackson-…
srowen Dec 30, 2014
0f31992
[Spark-4995] Replace Vector.toBreeze.activeIterator with foreachActive
Dec 30, 2014
6a89782
[SPARK-4813][Streaming] Fix the issue that ContextWaiter didn't handl…
zsxwing Dec 30, 2014
035bac8
[SPARK-4998][MLlib]delete the "train" function
ljzzju Dec 30, 2014
352ed6b
[SPARK-1010] Clean up uses of System.setProperty in unit tests
JoshRosen Dec 31, 2014
06a9aa5
[SPARK-4797] Replace breezeSquaredDistance
viirya Dec 31, 2014
8e14c5e
[SPARK-4298][Core] - The spark-submit cannot read Main-Class from Man…
Dec 31, 2014
3d194cc
SPARK-4547 [MLLIB] OOM when making bins in BinaryClassificationMetrics
srowen Dec 31, 2014
e24d3a9
[HOTFIX] Disable Spark UI in SparkSubmitSuite tests
JoshRosen Dec 12, 2014
c88a3d7
[SPARK-5038][SQL] Add explicit return type for implicit functions in …
rxin Dec 31, 2014
3610d3c
[SPARK-4790][STREAMING] Fix ReceivedBlockTrackerSuite waits for old f…
harishreedharan Dec 31, 2014
fdc2aa4
[SPARK-5028][Streaming]Add total received and processed records metri…
jerryshao Dec 31, 2014
c4f0b4f
SPARK-5020 [MLlib] GaussianMixtureModel.predictMembership() should ta…
tgaloppo Dec 31, 2014
fe6efac
[SPARK-5035] [Streaming] ReceiverMessage trait should extend Serializ…
JoshRosen Jan 1, 2015
4bb1248
SPARK-2757 [BUILD] [STREAMING] Add Mima test for Spark Sink after 1.1…
srowen Jan 1, 2015
7749dd6
[SPARK-5038] Add explicit return type for implicit functions.
rxin Jan 1, 2015
0128398
[HOTFIX] Bind web UI to ephemeral port in DriverSuite
JoshRosen Jan 1, 2015
bd88b71
[SPARK-3325][Streaming] Add a parameter to the method print in class …
watermen Jan 2, 2015
cdccc26
Fixed typos in streaming-kafka-integration.md
Jan 2, 2015
342612b
[SPARK-5058] Updated broken links
sigmoidanalytics Jan 4, 2015
b96008d
[SPARK-794][Core] Remove sleep() in ClusterScheduler.stop
Jan 4, 2015
3fddc94
[SPARK-4787] Stop SparkContext if a DAGScheduler init error occurs
tigerquoll Jan 4, 2015
e767d7d
[SPARK-4631] unit test for MQTT
Jan 5, 2015
939ba1f
[SPARK-4835] Disable validateOutputSpecs for Spark Streaming jobs
JoshRosen Jan 5, 2015
7239652
[SPARK-5067][Core] Use '===' to compare well-defined case class
zsxwing Jan 5, 2015
6c726a3
[SPARK-5069][Core] Fix the race condition of TaskSchedulerImpl.dagSch…
zsxwing Jan 5, 2015
27e7f5a
[SPARK-5083][Core] Fix a flaky test in TaskResultGetterSuite
zsxwing Jan 5, 2015
5c506ce
[SPARK-5074][Core] Fix a non-deterministic test failure
zsxwing Jan 5, 2015
d3f07fd
[SPARK-4688] Have a single shared network timeout in Spark
varunsaxena Jan 5, 2015
ce39b34
[SPARK-5057] Log message in failed askWithReply attempts
WangTaoTheTonic Jan 5, 2015
1c0e7ce
[SPARK-4465] runAsSparkUser doesn't affect TaskRunner in Mesos enviro…
jongyoul Jan 5, 2015
6c6f325
[SPARK-5089][PYSPARK][MLLIB] Fix vector convert
freeman-lab Jan 5, 2015
bbcba3a
[SPARK-5093] Set spark.network.timeout to 120s consistently.
rxin Jan 5, 2015
04d55d8
[SPARK-5040][SQL] Support expressing unresolved attributes using $"at…
rxin Jan 5, 2015
451546a
SPARK-4843 [YARN] Squash ExecutorRunnableUtil and ExecutorRunnable
Jan 6, 2015
a6394bc
[SPARK-1600] Refactor FileInputStream tests to remove Thread.sleep() …
JoshRosen Jan 6, 2015
5e3ec11
[Minor] Fix comments for GraphX 2D partitioning strategy
Jan 6, 2015
4cba6eb
SPARK-4159 [CORE] Maven build doesn't run JUnit test suites
srowen Jan 6, 2015
4108e5f
SPARK-5017 [MLlib] - Use SVD to compute determinant and inverse of co…
tgaloppo Jan 6, 2015
bb38ebb
[SPARK-5050][Mllib] Add unit test for sqdist
viirya Jan 6, 2015
e21acc1
[SPARK-5099][Mllib] Simplify logistic loss function
viirya Jan 7, 2015
5fde661
[YARN][SPARK-4929] Bug fix: fix the yarn-client code to support HA
SaintBacchus Jan 7, 2015
8fdd489
[SPARK-2165][YARN]add support for setting maxAppAttempts in the Appli…
WangTaoTheTonic Jan 7, 2015
6e74ede
[SPARK-2458] Make failed application log visible on History Server
tsudukim Jan 7, 2015
60e2d9e
[SPARK-5128][MLLib] Add common used log1pExp API in MLUtils
Jan 7, 2015
d345ebe
[SPARK-5132][Core]Correct stage Attempt Id key in stageInfofromJson
suyanNone Jan 7, 2015
2b729d2
[SPARK-5126][Core] Verify Spark urls before creating Actors so that i…
zsxwing Jan 8, 2015
c66a976
[SPARK-5116][MLlib] Add extractor for SparseVector and DenseVector
coderxiang Jan 8, 2015
c082385
MAINTENANCE: Automated closing of pull requests.
pwendell Jan 8, 2015
8d45834
SPARK-5087. [YARN] Merge yarn.Client and yarn.ClientBase
sryza Jan 8, 2015
f825e19
[SPARK-4917] Add a function to convert into a graph with canonical ed…
maropu Jan 8, 2015
06dc4b5
[SPARK-4989][CORE] avoid wrong eventlog conf cause cluster down in st…
liyezhang556520 Jan 8, 2015
0a59727
[Minor] Fix the value represented by spark.executor.id for consistency.
sarutak Jan 8, 2015
0760787
[SPARK-5130][Deploy]Take yarn-cluster as cluster mode in spark-submit
WangTaoTheTonic Jan 8, 2015
538f221
Document that groupByKey will OOM for large keys
Jan 8, 2015
72df5a3
SPARK-5148 [MLlib] Make usersOut/productsOut storagelevel in ALS conf…
zeitos Jan 8, 2015
f6ce899
add example and fix bugs
Jan 8, 2015
eea16a7
refactor
Jan 8, 2015
a00af6b
[SPARK-4973][CORE] Local directory in the driver of client-mode conti…
sarutak Jan 8, 2015
c9c8b21
[SPARK-4891][PySpark][MLlib] Add gamma/log normal/exp dist sampling t…
rnowling Jan 8, 2015
48cecf6
[SPARK-4048] Enhance and extend hadoop-provided profile.
Jan 9, 2015
167a5ab
[SPARK-5122] Remove Shark from spark-ec2
nchammas Jan 9, 2015
f3da4bd
[SPARK-5169][YARN]fetch the correct max attempts
WangTaoTheTonic Jan 9, 2015
b4034c3
[Minor] Fix test RetryingBlockFetcherSuite after changed config name
aarondav Jan 9, 2015
547df97
SPARK-5136 [DOCS] Improve documentation around setting up Spark Intel…
srowen Jan 9, 2015
1790b38
HOTFIX: Minor improvements to make-distribution.sh
pwendell Jan 9, 2015
b6aa557
[SPARK-1143] Separate pool tests into their own suite.
kayousterhout Jan 9, 2015
e9ca16e
[SPARK-5145][Mllib] Add BLAS.dsyr and use it in GaussianMixtureEM
viirya Jan 9, 2015
454fe12
[SPARK-3619] Upgrade to Mesos 0.21 to work around MESOS-1688
jongyoul Jan 9, 2015
7e8e62a
[SPARK-5015] [mllib] Random seed for GMM + make test suite deterministic
jkbradley Jan 9, 2015
e966452
[SPARK-1953][YARN]yarn client mode Application Master memory size is …
WangTaoTheTonic Jan 9, 2015
e0f28e0
[SPARK-4737] Task set manager properly handles serialization errors
mccheah Jan 9, 2015
ae62872
[DOC] Fixed Mesos version in doc from 0.18.1 to 0.21.0
sarutak Jan 9, 2015
4e1f12d
[Minor] Fix import order and other coding style
Jan 9, 2015
8782eb9
[SPARK-4990][Deploy]to find default properties file, search SPARK_CON…
WangTaoTheTonic Jan 10, 2015
4554529
[SPARK-4406] [MLib] FIX: Validate k in SVD
MechCoder Jan 10, 2015
545dfcb
[SPARK-5141][SQL]CaseInsensitiveMap throws java.io.NotSerializableExc…
luogankun Jan 10, 2015
1e56eba
[SPARK-4925][SQL] Publish Spark SQL hive-thriftserver maven artifact
alexoss68 Jan 10, 2015
4b39fd1
[SPARK-4943][SQL] Allow table name having dot for db/catalog
alexoss68 Jan 10, 2015
693a323
[SPARK-4574][SQL] Adding support for defining schema in foreign DDL c…
scwf Jan 10, 2015
b3e86dc
[SPARK-4861][SQL] Refactory command in spark sql
scwf Jan 10, 2015
77106df
SPARK-4963 [SQL] Add copy to SQL's Sample operator
Jan 10, 2015
3684fd2
[SPARK-5187][SQL] Fix caching of tables with HiveUDFs in the WHERE cl…
marmbrus Jan 10, 2015
0ca51cc
[SPARK-4692] [SQL] Support ! boolean logic operator like NOT
YanTangZhai Jan 10, 2015
f0d558b
[SPARK-5181] do not print writing WAL log when WAL is disabled
CodingCat Jan 10, 2015
8a29dc7
[Minor]Resolve sbt warnings during build (MQTTStreamSuite.scala).
witgo Jan 10, 2015
92d9a70
[SPARK-4871][SQL] Show sql statement in spark ui when run sql with sp…
scwf Jan 11, 2015
d22a31f
[SPARK-5029][SQL] Enable from follow multiple brackets
scwf Jan 11, 2015
3313260
[SPARK-5032] [graphx] Remove GraphX MIMA exclude for 1.3
jkbradley Jan 11, 2015
1656aae
[SPARK-5073] spark.storage.memoryMapThreshold have two default value
Lewuathe Jan 11, 2015
6942b97
[SPARK-4951][Core] Fix the issue that a busy executor may be killed
zsxwing Jan 12, 2015
f38ef65
[SPARK-4033][Examples]Input of the SparkPi too big causes the emptio…
SaintBacchus Jan 12, 2015
2130de9
SPARK-5018 [MLlib] [WIP] Make MultivariateGaussian public
tgaloppo Jan 12, 2015
82fd38d
[SPARK-5200] Disable web UI in Hive ThriftServer tests
JoshRosen Jan 12, 2015
ef9224e
[SPARK-5102][Core]subclass of MapStatus needs to be registered with Kryo
lianhuiwang Jan 12, 2015
13e610b
SPARK-4159 [BUILD] Addendum: improve running of single test after ena…
srowen Jan 12, 2015
a3978f3
[SPARK-5078] Optionally read from SPARK_LOCAL_HOSTNAME
marmbrus Jan 12, 2015
aff49a3
SPARK-5172 [BUILD] spark-examples-***.jar shades a wrong Hadoop distr…
srowen Jan 12, 2015
3aed305
[SPARK-4999][Streaming] Change storeInBlockManager to false by default
jerryshao Jan 12, 2015
5d9fa55
[SPARK-5049][SQL] Fix ordering of partition columns in ParquetTableScan
marmbrus Jan 12, 2015
1e42e96
[SPARK-5138][SQL] Ensure schema can be inferred from a namedtuple
mulby Jan 13, 2015
f7741a9
[SPARK-5006][Deploy]spark.port.maxRetries doesn't work
WangTaoTheTonic Jan 13, 2015
9dea64e
[SPARK-4697][YARN]System properties should override environment varia…
WangTaoTheTonic Jan 13, 2015
39e333e
[SPARK-5131][Streaming][DOC]: There is a discrepancy in WAL implement…
uncleGen Jan 13, 2015
8ead999
[SPARK-5223] [MLlib] [PySpark] fix MapConverter and ListConverter in …
Jan 13, 2015
6463e0b
[SPARK-4912][SQL] Persistent tables for the Spark SQL data sources api
yhuai Jan 13, 2015
14e3f11
[SPARK-5168] Make SQLConf a field rather than mixin in SQLContext
rxin Jan 13, 2015
f996909
[SPARK-5123][SQL] Reconcile Java/Scala API for data types.
rxin Jan 14, 2015
d5eeb35
[SPARK-5167][SQL] Move Row into sql package and make it usable for Java.
rxin Jan 14, 2015
a3f7421
[SPARK-5248] [SQL] move sql.types.decimal.Decimal to sql.types.Decimal
adrian-wang Jan 14, 2015
81f72a0
[SPARK-5211][SQL]Restore HiveMetastoreTypes.toDataType
yhuai Jan 14, 2015
38bdc99
[SQL] some comments fix for GROUPING SETS
adrian-wang Jan 14, 2015
5840f54
[SPARK-2909] [MLlib] [PySpark] SparseVector in pyspark now supports i…
MechCoder Jan 14, 2015
9d4449c
[SPARK-5228][WebUI] Hide tables for "Active Jobs/Completed Jobs/Faile…
sarutak Jan 14, 2015
259936b
[SPARK-4014] Add TaskContext.attemptNumber and deprecate TaskContext.…
JoshRosen Jan 14, 2015
2fd7f72
[SPARK-5235] Make SQLConf Serializable
alexbaretta Jan 14, 2015
76389c5
[SPARK-5234][ml]examples for ml don't have sparkContext.stop
Jan 14, 2015
13d2406
[SPARK-5254][MLLIB] Update the user guide to position spark.ml better
mengxr Jan 15, 2015
cfa397c
[SPARK-5193][SQL] Tighten up SQLContext API
rxin Jan 15, 2015
6abc45e
[SPARK-5254][MLLIB] remove developers section from spark.ml guide
mengxr Jan 15, 2015
4b325c7
[SPARK-5193][SQL] Tighten up HiveContext API
rxin Jan 15, 2015
3c8650c
[SPARK-5224] [PySpark] improve performance of parallelize list/ndarray
Jan 15, 2015
1881431
[SPARK-5274][SQL] Reconcile Java and Scala UDFRegistration.
rxin Jan 16, 2015
65858ba
[Minor] Fix tiny typo in BlockManager
sarutak Jan 16, 2015
96c2c71
[SPARK-4857] [CORE] Adds Executor membership events to SparkListener
Jan 16, 2015
a79a9f9
[SPARK-4092] [CORE] Fix InputMetrics for coalesce'd Rdds
Jan 16, 2015
2be82b1
[SPARK-1507][YARN]specify # cores for ApplicationMaster
WangTaoTheTonic Jan 16, 2015
e200ac8
[SPARK-5201][CORE] deal with int overflow in the ParallelCollectionRD…
advancedxy Jan 16, 2015
f6b852a
[DOCS] Fix typo in return type of cogroup
srowen Jan 16, 2015
e8422c5
[SPARK-5231][WebUI] History Server shows wrong job submission time.
sarutak Jan 16, 2015
ecf943d
[WebUI] Fix collapse of WebUI layout
sarutak Jan 16, 2015
d05c9ee
[SPARK-4923][REPL] Add Developer API to REPL to allow re-publishing t…
Jan 16, 2015
fd3a8a1
[SPARK-733] Add documentation on use of accumulators in lazy transfor…
Jan 16, 2015
ee1c1f3
[SPARK-4937][SQL] Adding optimization to simplify the And, Or condit…
scwf Jan 16, 2015
61b427d
[SPARK-5193][SQL] Remove Spark SQL Java-specific API.
rxin Jan 17, 2015
f3bfc76
[SQL][minor] Improved Row documentation.
rxin Jan 17, 2015
c1f3c27
[SPARK-4937][SQL] Comment for the newly optimization rules in `Boolea…
scwf Jan 17, 2015
6999910
[SPARK-5096] Use sbt tasks instead of vals to get hadoop version
marmbrus Jan 18, 2015
e7884bc
[SQL][Minor] Added comments and examples to explain BooleanSimplifica…
rxin Jan 18, 2015
e12b5b6
MAINTENANCE: Automated closing of pull requests.
pwendell Jan 18, 2015
ad16da1
[HOTFIX]: Minor clean up regarding skipped artifacts in build files.
pwendell Jan 18, 2015
1727e08
[SPARK-5279][SQL] Use java.math.BigDecimal as the exposed Decimal type.
rxin Jan 18, 2015
1a200a3
[SQL][Minor] Update sql doc according to data type APIs changes
scwf Jan 18, 2015
1955645
[SQL][minor] Put DataTypes.java in java dir.
rxin Jan 19, 2015
7dbf1fd
[SQL] fix typo in class description
Jan 19, 2015
851b6a9
SPARK-5217 Spark UI should report pending stages during job execution…
ScrapCodes Jan 19, 2015
3453d57
[SPARK-3288] All fields in TaskMetrics should be private and use gett…
Jan 19, 2015
4a4f9cc
[SPARK-5088] Use spark-class for running executors directly
jongyoul Jan 19, 2015
1ac1c1d
MAINTENANCE: Automated closing of pull requests.
pwendell Jan 19, 2015
4432568
[SPARK-5282][mllib]: RowMatrix easily gets int overflow in the memory…
hhbyyh Jan 19, 2015
cd5da42
[SPARK-5284][SQL] Insert into Hive throws NPE when a inner complex ty…
yhuai Jan 19, 2015
2604bc3
[SPARK-5286][SQL] Fail to drop an invalid table when using the data s…
yhuai Jan 19, 2015
74de94e
[SPARK-4504][Examples] fix run-example failure if multiple assembly j…
gvramana Jan 19, 2015
e69fb8c
[SPARK-5214][Core] Add EventLoop and change DAGScheduler to an EventLoop
zsxwing Jan 20, 2015
306ff18
SPARK-5270 [CORE] Provide isEmpty() function in RDD API
srowen Jan 20, 2015
debc031
[SQL][minor] Add a log4j file for catalyst test.
rxin Jan 20, 2015
4afad9c
[SPARK-4803] [streaming] Remove duplicate RegisterReceiver message
ilayaperumalg Jan 20, 2015
9d9294a
[SPARK-5333][Mesos] MesosTaskLaunchData occurs BufferUnderflowException
jongyoul Jan 20, 2015
8140802
[SQL][Minor] Refactors deeply nested FP style code in BooleanSimplifi…
liancheng Jan 20, 2015
c93a57f
SPARK-4660: Use correct class loader in JavaSerializer (copy of PR #3…
jacek-lewandowski Jan 20, 2015
769aced
[SPARK-5329][WebUI] UIWorkloadGenerator should stop SparkContext.
sarutak Jan 20, 2015
23e2554
SPARK-5019 [MLlib] - GaussianMixtureModel exposes instances of Multiv…
tgaloppo Jan 20, 2015
bc20a52
[SPARK-5287][SQL] Add defaultSizeOf to every data type.
yhuai Jan 20, 2015
d181c2a
[SPARK-5323][SQL] Remove Row's Seq inheritance.
rxin Jan 20, 2015
2f82c84
[SPARK-5186] [MLLIB] Vector.equals and Vector.hashCode are very inef…
hhbyyh Jan 20, 2015
9a151ce
[SPARK-5294][WebUI] Hide tables in AllStagePages for "Active Stages, …
sarutak Jan 21, 2015
bad6c57
[SPARK-5275] [Streaming] include python source code
Jan 21, 2015
ec5b0f2
[HOTFIX] Update pom.xml to pull MapR's Hadoop version 2.4.1.
rkannan82 Jan 21, 2015
424d8c6
[SPARK-5297][Streaming] Fix Java file stream type erasure problem
jerryshao Jan 21, 2015
8c06a5f
[SPARK-5336][YARN]spark.executor.cores must not be less than spark.ta…
WangTaoTheTonic Jan 21, 2015
2eeada3
SPARK-1714. Take advantage of AMRMClient APIs to simplify logic in Ya…
sryza Jan 21, 2015
aa1e22b
[MLlib] [SPARK-5301] Missing conversions and operations on IndexedRow…
Jan 21, 2015
7450a99
[SPARK-4749] [mllib]: Allow initializing KMeans clusters using a seed
str-janus Jan 21, 2015
3ee3ab5
[SPARK-5064][GraphX] Add numEdges upperbound validation for R-MAT gra…
Jan 21, 2015
812d367
[SPARK-5244] [SQL] add coalesce() in sql parser
adrian-wang Jan 21, 2015
8361078
[SPARK-5009] [SQL] Long keyword support in SQL Parsers
chenghao-intel Jan 21, 2015
b328ac6
Revert "[SPARK-5244] [SQL] add coalesce() in sql parser"
JoshRosen Jan 21, 2015
ba19689
[SQL] [Minor] Remove deprecated parquet tests
liancheng Jan 21, 2015
3be2a88
[SPARK-4984][CORE][WEBUI] Adding a pop-up containing the full job des…
scwf Jan 21, 2015
9bad062
[SPARK-5355] make SparkConf thread-safe
Jan 22, 2015
aea8953
Kafka-assembly for Python API
tdas Jan 22, 2015
27bccc5
[SPARK-5202] [SQL] Add hql variable substitution support
chenghao-intel Jan 22, 2015
ca7910d
[SPARK-3424][MLLIB] cache point distances during k-means|| init
mengxr Jan 22, 2015
fcb3e18
[SPARK-5317]Set BoostingStrategy.defaultParams With Enumeration Algo.…
Peishen-Jia Jan 22, 2015
adeeb38
Merge pull request #3 from tdas/kafka-python-api
Jan 22, 2015
33730d1
Merge branch 'master' of github.com:apache/spark into kafka
Jan 22, 2015
2c567a5
update logging and comment
Jan 22, 2015
97386b3
address comment
Jan 22, 2015
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
20 changes: 20 additions & 0 deletions assembly/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -354,5 +354,25 @@
</dependency>
</dependencies>
</profile>

<!-- Profiles that disable inclusion of certain dependencies. -->
<profile>
<id>hadoop-provided</id>
<properties>
<hadoop.deps.scope>provided</hadoop.deps.scope>
</properties>
</profile>
<profile>
<id>hive-provided</id>
<properties>
<hive.deps.scope>provided</hive.deps.scope>
</properties>
</profile>
<profile>
<id>parquet-provided</id>
<properties>
<parquet.deps.scope>provided</parquet.deps.scope>
</properties>
</profile>
</profiles>
</project>
15 changes: 0 additions & 15 deletions bagel/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -40,15 +40,6 @@
<artifactId>spark-core_${scala.binary.version}</artifactId>
<version>${project.version}</version>
</dependency>
<dependency>
<groupId>org.eclipse.jetty</groupId>
<artifactId>jetty-server</artifactId>
</dependency>
<dependency>
<groupId>org.scalatest</groupId>
<artifactId>scalatest_${scala.binary.version}</artifactId>
<scope>test</scope>
</dependency>
<dependency>
<groupId>org.scalacheck</groupId>
<artifactId>scalacheck_${scala.binary.version}</artifactId>
Expand All @@ -58,11 +49,5 @@
<build>
<outputDirectory>target/scala-${scala.binary.version}/classes</outputDirectory>
<testOutputDirectory>target/scala-${scala.binary.version}/test-classes</testOutputDirectory>
<plugins>
<plugin>
<groupId>org.scalatest</groupId>
<artifactId>scalatest-maven-plugin</artifactId>
</plugin>
</plugins>
</build>
</project>
4 changes: 2 additions & 2 deletions bagel/src/test/resources/log4j.properties
Original file line number Diff line number Diff line change
Expand Up @@ -15,10 +15,10 @@
# limitations under the License.
#

# Set everything to be logged to the file bagel/target/unit-tests.log
# Set everything to be logged to the file target/unit-tests.log
log4j.rootCategory=INFO, file
log4j.appender.file=org.apache.log4j.FileAppender
log4j.appender.file.append=false
log4j.appender.file.append=true
log4j.appender.file.file=target/unit-tests.log
log4j.appender.file.layout=org.apache.log4j.PatternLayout
log4j.appender.file.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss.SSS} %t %p %c{1}: %m%n
Expand Down
7 changes: 7 additions & 0 deletions bin/compute-classpath.cmd
Original file line number Diff line number Diff line change
Expand Up @@ -109,6 +109,13 @@ if "x%YARN_CONF_DIR%"=="x" goto no_yarn_conf_dir
set CLASSPATH=%CLASSPATH%;%YARN_CONF_DIR%
:no_yarn_conf_dir

rem To allow for distributions to append needed libraries to the classpath (e.g. when
rem using the "hadoop-provided" profile to build Spark), check SPARK_DIST_CLASSPATH and
rem append it to tbe final classpath.
if not "x%$SPARK_DIST_CLASSPATH%"=="x" (
set CLASSPATH=%CLASSPATH%;%SPARK_DIST_CLASSPATH%
)

rem A bit of a hack to allow calling this script within run2.cmd without seeing output
if "%DONT_PRINT_CLASSPATH%"=="1" goto exit

Expand Down
34 changes: 22 additions & 12 deletions bin/compute-classpath.sh
Original file line number Diff line number Diff line change
Expand Up @@ -72,22 +72,25 @@ else
assembly_folder="$ASSEMBLY_DIR"
fi

num_jars="$(ls "$assembly_folder" | grep "spark-assembly.*hadoop.*\.jar$" | wc -l)"
if [ "$num_jars" -eq "0" ]; then
echo "Failed to find Spark assembly in $assembly_folder"
echo "You need to build Spark before running this program."
exit 1
fi
num_jars=0

for f in ${assembly_folder}/spark-assembly*hadoop*.jar; do
if [[ ! -e "$f" ]]; then
echo "Failed to find Spark assembly in $assembly_folder" 1>&2
echo "You need to build Spark before running this program." 1>&2
exit 1
fi
ASSEMBLY_JAR="$f"
num_jars=$((num_jars+1))
done

if [ "$num_jars" -gt "1" ]; then
jars_list=$(ls "$assembly_folder" | grep "spark-assembly.*hadoop.*.jar$")
echo "Found multiple Spark assembly jars in $assembly_folder:"
echo "$jars_list"
echo "Please remove all but one jar."
echo "Found multiple Spark assembly jars in $assembly_folder:" 1>&2
ls ${assembly_folder}/spark-assembly*hadoop*.jar 1>&2
echo "Please remove all but one jar." 1>&2
exit 1
fi

ASSEMBLY_JAR="$(ls "$assembly_folder"/spark-assembly*hadoop*.jar 2>/dev/null)"

# Verify that versions of java used to build the jars and run Spark are compatible
jar_error_check=$("$JAR_CMD" -tf "$ASSEMBLY_JAR" nonexistent/class/path 2>&1)
if [[ "$jar_error_check" =~ "invalid CEN header" ]]; then
Expand Down Expand Up @@ -146,4 +149,11 @@ if [ -n "$YARN_CONF_DIR" ]; then
CLASSPATH="$CLASSPATH:$YARN_CONF_DIR"
fi

# To allow for distributions to append needed libraries to the classpath (e.g. when
# using the "hadoop-provided" profile to build Spark), check SPARK_DIST_CLASSPATH and
# append it to tbe final classpath.
if [ -n "$SPARK_DIST_CLASSPATH" ]; then
CLASSPATH="$CLASSPATH:$SPARK_DIST_CLASSPATH"
fi

echo "$CLASSPATH"
27 changes: 21 additions & 6 deletions bin/run-example
Original file line number Diff line number Diff line change
Expand Up @@ -35,17 +35,32 @@ else
fi

if [ -f "$FWDIR/RELEASE" ]; then
export SPARK_EXAMPLES_JAR="`ls "$FWDIR"/lib/spark-examples-*hadoop*.jar`"
elif [ -e "$EXAMPLES_DIR"/target/scala-$SPARK_SCALA_VERSION/spark-examples-*hadoop*.jar ]; then
export SPARK_EXAMPLES_JAR="`ls "$EXAMPLES_DIR"/target/scala-$SPARK_SCALA_VERSION/spark-examples-*hadoop*.jar`"
JAR_PATH="${FWDIR}/lib"
else
JAR_PATH="${EXAMPLES_DIR}/target/scala-${SPARK_SCALA_VERSION}"
fi

if [[ -z "$SPARK_EXAMPLES_JAR" ]]; then
echo "Failed to find Spark examples assembly in $FWDIR/lib or $FWDIR/examples/target" 1>&2
echo "You need to build Spark before running this program" 1>&2
JAR_COUNT=0

for f in ${JAR_PATH}/spark-examples-*hadoop*.jar; do
if [[ ! -e "$f" ]]; then
echo "Failed to find Spark examples assembly in $FWDIR/lib or $FWDIR/examples/target" 1>&2
echo "You need to build Spark before running this program" 1>&2
exit 1
fi
SPARK_EXAMPLES_JAR="$f"
JAR_COUNT=$((JAR_COUNT+1))
done

if [ "$JAR_COUNT" -gt "1" ]; then
echo "Found multiple Spark examples assembly jars in ${JAR_PATH}" 1>&2
ls ${JAR_PATH}/spark-examples-*hadoop*.jar 1>&2
echo "Please remove all but one jar." 1>&2
exit 1
fi

export SPARK_EXAMPLES_JAR

EXAMPLE_MASTER=${MASTER:-"local[*]"}

if [[ ! $EXAMPLE_CLASS == org.apache.spark.examples* ]]; then
Expand Down
4 changes: 3 additions & 1 deletion bin/spark-class
Original file line number Diff line number Diff line change
Expand Up @@ -71,6 +71,8 @@ case "$1" in
'org.apache.spark.executor.MesosExecutorBackend')
OUR_JAVA_OPTS="$SPARK_JAVA_OPTS $SPARK_EXECUTOR_OPTS"
OUR_JAVA_MEM=${SPARK_EXECUTOR_MEMORY:-$DEFAULT_MEM}
export PYTHONPATH="$FWDIR/python:$PYTHONPATH"
export PYTHONPATH="$FWDIR/python/lib/py4j-0.8.2.1-src.zip:$PYTHONPATH"
;;

# Spark submit uses SPARK_JAVA_OPTS + SPARK_SUBMIT_OPTS +
Expand Down Expand Up @@ -148,7 +150,7 @@ fi
if [[ "$1" =~ org.apache.spark.tools.* ]]; then
if test -z "$SPARK_TOOLS_JAR"; then
echo "Failed to find Spark Tools Jar in $FWDIR/tools/target/scala-$SPARK_SCALA_VERSION/" 1>&2
echo "You need to build Spark before running $1." 1>&2
echo "You need to run \"build/sbt tools/package\" before running $1." 1>&2
exit 1
fi
CLASSPATH="$CLASSPATH:$SPARK_TOOLS_JAR"
Expand Down
10 changes: 9 additions & 1 deletion bin/spark-submit
Original file line number Diff line number Diff line change
Expand Up @@ -38,11 +38,19 @@ while (($#)); do
export SPARK_SUBMIT_CLASSPATH=$2
elif [ "$1" = "--driver-java-options" ]; then
export SPARK_SUBMIT_OPTS=$2
elif [ "$1" = "--master" ]; then
export MASTER=$2
fi
shift
done

DEFAULT_PROPERTIES_FILE="$SPARK_HOME/conf/spark-defaults.conf"
if [ -z "$SPARK_CONF_DIR" ]; then
export SPARK_CONF_DIR="$SPARK_HOME/conf"
fi
DEFAULT_PROPERTIES_FILE="$SPARK_CONF_DIR/spark-defaults.conf"
if [ "$MASTER" == "yarn-cluster" ]; then
SPARK_SUBMIT_DEPLOY_MODE=cluster
fi
export SPARK_SUBMIT_DEPLOY_MODE=${SPARK_SUBMIT_DEPLOY_MODE:-"client"}
export SPARK_SUBMIT_PROPERTIES_FILE=${SPARK_SUBMIT_PROPERTIES_FILE:-"$DEFAULT_PROPERTIES_FILE"}

Expand Down
12 changes: 11 additions & 1 deletion bin/spark-submit2.cmd
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,11 @@ set ORIG_ARGS=%*

rem Reset the values of all variables used
set SPARK_SUBMIT_DEPLOY_MODE=client
set SPARK_SUBMIT_PROPERTIES_FILE=%SPARK_HOME%\conf\spark-defaults.conf

if not defined %SPARK_CONF_DIR% (
set SPARK_CONF_DIR=%SPARK_HOME%\conf
)
set SPARK_SUBMIT_PROPERTIES_FILE=%SPARK_CONF_DIR%\spark-defaults.conf
set SPARK_SUBMIT_DRIVER_MEMORY=
set SPARK_SUBMIT_LIBRARY_PATH=
set SPARK_SUBMIT_CLASSPATH=
Expand All @@ -45,11 +49,17 @@ if [%1] == [] goto continue
set SPARK_SUBMIT_CLASSPATH=%2
) else if [%1] == [--driver-java-options] (
set SPARK_SUBMIT_OPTS=%2
) else if [%1] == [--master] (
set MASTER=%2
)
shift
goto loop
:continue

if [%MASTER%] == [yarn-cluster] (
set SPARK_SUBMIT_DEPLOY_MODE=cluster
)

rem For client mode, the driver will be launched in the same JVM that launches
rem SparkSubmit, so we may need to read the properties file for any extra class
rem paths, library paths, java options and memory early on. Otherwise, it will
Expand Down
18 changes: 0 additions & 18 deletions core/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -276,11 +276,6 @@
<artifactId>selenium-java</artifactId>
<scope>test</scope>
</dependency>
<dependency>
<groupId>org.scalatest</groupId>
<artifactId>scalatest_${scala.binary.version}</artifactId>
<scope>test</scope>
</dependency>
<dependency>
<groupId>org.mockito</groupId>
<artifactId>mockito-all</artifactId>
Expand Down Expand Up @@ -326,19 +321,6 @@
<outputDirectory>target/scala-${scala.binary.version}/classes</outputDirectory>
<testOutputDirectory>target/scala-${scala.binary.version}/test-classes</testOutputDirectory>
<plugins>
<plugin>
<groupId>org.scalatest</groupId>
<artifactId>scalatest-maven-plugin</artifactId>
<executions>
<execution>
<id>test</id>
<goals>
<goal>test</goal>
</goals>
</execution>
</executions>
</plugin>

<!-- Unzip py4j so we can include its files in the jar -->
<plugin>
<groupId>org.apache.maven.plugins</groupId>
Expand Down
97 changes: 97 additions & 0 deletions core/src/main/java/org/apache/spark/JavaSparkListener.java
Original file line number Diff line number Diff line change
@@ -0,0 +1,97 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/

package org.apache.spark;

import org.apache.spark.scheduler.SparkListener;
import org.apache.spark.scheduler.SparkListenerApplicationEnd;
import org.apache.spark.scheduler.SparkListenerApplicationStart;
import org.apache.spark.scheduler.SparkListenerBlockManagerAdded;
import org.apache.spark.scheduler.SparkListenerBlockManagerRemoved;
import org.apache.spark.scheduler.SparkListenerEnvironmentUpdate;
import org.apache.spark.scheduler.SparkListenerExecutorAdded;
import org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate;
import org.apache.spark.scheduler.SparkListenerExecutorRemoved;
import org.apache.spark.scheduler.SparkListenerJobEnd;
import org.apache.spark.scheduler.SparkListenerJobStart;
import org.apache.spark.scheduler.SparkListenerStageCompleted;
import org.apache.spark.scheduler.SparkListenerStageSubmitted;
import org.apache.spark.scheduler.SparkListenerTaskEnd;
import org.apache.spark.scheduler.SparkListenerTaskGettingResult;
import org.apache.spark.scheduler.SparkListenerTaskStart;
import org.apache.spark.scheduler.SparkListenerUnpersistRDD;

/**
* Java clients should extend this class instead of implementing
* SparkListener directly. This is to prevent java clients
* from breaking when new events are added to the SparkListener
* trait.
*
* This is a concrete class instead of abstract to enforce
* new events get added to both the SparkListener and this adapter
* in lockstep.
*/
public class JavaSparkListener implements SparkListener {

@Override
public void onStageCompleted(SparkListenerStageCompleted stageCompleted) { }

@Override
public void onStageSubmitted(SparkListenerStageSubmitted stageSubmitted) { }

@Override
public void onTaskStart(SparkListenerTaskStart taskStart) { }

@Override
public void onTaskGettingResult(SparkListenerTaskGettingResult taskGettingResult) { }

@Override
public void onTaskEnd(SparkListenerTaskEnd taskEnd) { }

@Override
public void onJobStart(SparkListenerJobStart jobStart) { }

@Override
public void onJobEnd(SparkListenerJobEnd jobEnd) { }

@Override
public void onEnvironmentUpdate(SparkListenerEnvironmentUpdate environmentUpdate) { }

@Override
public void onBlockManagerAdded(SparkListenerBlockManagerAdded blockManagerAdded) { }

@Override
public void onBlockManagerRemoved(SparkListenerBlockManagerRemoved blockManagerRemoved) { }

@Override
public void onUnpersistRDD(SparkListenerUnpersistRDD unpersistRDD) { }

@Override
public void onApplicationStart(SparkListenerApplicationStart applicationStart) { }

@Override
public void onApplicationEnd(SparkListenerApplicationEnd applicationEnd) { }

@Override
public void onExecutorMetricsUpdate(SparkListenerExecutorMetricsUpdate executorMetricsUpdate) { }

@Override
public void onExecutorAdded(SparkListenerExecutorAdded executorAdded) { }

@Override
public void onExecutorRemoved(SparkListenerExecutorRemoved executorRemoved) { }
}
Loading