Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
641 commits
Select commit Hold shift + click to select a range
e87ffca
Revert "[SPARK-21567][SQL] Dataset should work with type alias"
cloud-fan Aug 8, 2017
d023314
[SPARK-21503][UI] Spark UI shows incorrect task status for a killed E…
Aug 9, 2017
7446be3
[SPARK-21523][ML] update breeze to 0.13.2 for an emergency bugfix in …
WeichenXu123 Aug 9, 2017
f6d56d2
[SPARK-21596][SS] Ensure places calling HDFSMetadataLog.get check the…
zsxwing Aug 9, 2017
3ca55ea
[SPARK-21663][TESTS] test("remote fetch below max RPC message size") …
wangjiaochun Aug 9, 2017
c909496
[SPARK-21699][SQL] Remove unused getTableOption in ExternalCatalog
rxin Aug 11, 2017
406eb1c
[SPARK-21595] Separate thresholds for buffering and spilling in Exter…
tejasapatil Aug 11, 2017
7b98077
[SPARK-21563][CORE] Fix race condition when serializing TaskDescripti…
ash211 Aug 14, 2017
48bacd3
[SPARK-21696][SS] Fix a potential issue that may generate partial sna…
zsxwing Aug 14, 2017
d9c8e62
[SPARK-21721][SQL] Clear FileSystem deleteOnExit cache when paths are…
viirya Aug 15, 2017
f1accc8
[SPARK-21723][ML] Fix writing LibSVM (key not found: numFeatures)
Aug 16, 2017
f5ede0d
[SPARK-21656][CORE] spark dynamic allocation should not idle timeout …
Aug 16, 2017
2a96975
[SPARK-18464][SQL][BACKPORT] support old table which doesn't store sc…
cloud-fan Aug 16, 2017
fdea642
[SPARK-21739][SQL] Cast expression should initialize timezoneId when …
DonnyZone Aug 18, 2017
6c2a38a
[MINOR] Correct validateAndTransformSchema in GaussianMixture and AFT…
sharp-pixel Aug 20, 2017
0f640e9
[SPARK-21721][SQL][FOLLOWUP] Clear FileSystem deleteOnExit cache when…
viirya Aug 20, 2017
526087f
[SPARK-21617][SQL] Store correct table metadata when altering schema …
Aug 21, 2017
236b2f4
[SPARK-21805][SPARKR] Disable R vignettes code on Windows
felixcheung Aug 24, 2017
a585367
[SPARK-21826][SQL] outer broadcast hash join should not throw NPE
cloud-fan Aug 24, 2017
2b4bd79
[SPARK-21681][ML] fix bug of MLOR do not work correctly when featureS…
WeichenXu123 Aug 24, 2017
0d4ef2f
[SPARK-21818][ML][MLLIB] Fix bug of MultivariateOnlineSummarizer.vari…
WeichenXu123 Aug 28, 2017
59bb7eb
[SPARK-21798] No config to replace deprecated SPARK_CLASSPATH config …
Aug 28, 2017
59529b2
[SPARK-21714][CORE][BACKPORT-2.2] Avoiding re-uploading remote resour…
jerryshao Aug 29, 2017
917fe66
Revert "[SPARK-21714][CORE][BACKPORT-2.2] Avoiding re-uploading remot…
Aug 29, 2017
a6a9944
[SPARK-21254][WEBUI] History UI performance fixes
2ooom Aug 30, 2017
d10c9dc
[SPARK-21714][CORE][BACKPORT-2.2] Avoiding re-uploading remote resour…
jerryshao Aug 30, 2017
14054ff
[SPARK-21834] Incorrect executor request in case of dynamic allocation
Aug 30, 2017
50f86e1
[SPARK-21884][SPARK-21477][BACKPORT-2.2][SQL] Mark LocalTableScanExec…
gatorsmile Sep 1, 2017
fb1b5f0
[SPARK-21418][SQL] NoSuchElementException: None.get in DataSourceScan…
srowen Sep 4, 2017
1f7c486
[SPARK-21925] Update trigger interval documentation in docs with beha…
brkyvz Sep 5, 2017
7da8fbf
[MINOR][DOC] Update `Partition Discovery` section to enumerate all av…
dongjoon-hyun Sep 5, 2017
9afab9a
[SPARK-21924][DOCS] Update structured streaming programming guide doc
Sep 6, 2017
342cc2a
[SPARK-21901][SS] Define toString for StateOperatorProgress
jaceklaskowski Sep 6, 2017
49968de
Fixed pandoc dependency issue in python/setup.py
Sep 7, 2017
0848df1
[SPARK-21890] Credentials not being passed to add the tokens
Sep 7, 2017
4304d0b
[SPARK-21950][SQL][PYTHON][TEST] pyspark.sql.tests.SQLTests2 should s…
ueshin Sep 8, 2017
781a1f8
[SPARK-21915][ML][PYSPARK] Model 1 and Model 2 ParamMaps Missing
marktab Sep 8, 2017
08cb06a
[SPARK-21936][SQL][2.2] backward compatibility test framework for Hiv…
cloud-fan Sep 8, 2017
9ae7c96
[SPARK-21946][TEST] fix flaky test: "alter table: rename cached table…
kiszk Sep 8, 2017
9876821
[SPARK-21128][R][BACKPORT-2.2] Remove both "spark-warehouse" and "met…
HyukjinKwon Sep 8, 2017
182478e
[SPARK-21954][SQL] JacksonUtils should verify MapType's value type in…
viirya Sep 9, 2017
b1b5a7f
[SPARK-20098][PYSPARK] dataType's typeName fix
szalai1 Sep 10, 2017
10c6836
[SPARK-21976][DOC] Fix wrong documentation for Mean Absolute Error.
FavioVazquez Sep 12, 2017
63098dc
[DOCS] Fix unreachable links in the document
sarutak Sep 12, 2017
b606dc1
[SPARK-18608][ML] Fix double caching
zhengruifeng Sep 12, 2017
3a692e3
[SPARK-21980][SQL] References in grouping functions should be indexed…
DonnyZone Sep 13, 2017
51e5a82
[SPARK-18608][ML][FOLLOWUP] Fix double caching for PySpark OneVsRest.
yanboliang Sep 14, 2017
42852bb
[SPARK-21985][PYSPARK] PairDeserializer is broken for double-zipped RDDs
aray Sep 17, 2017
309c401
[SPARK-21953] Show both memory and disk bytes spilled if either is pr…
ash211 Sep 18, 2017
a86831d
[SPARK-22043][PYTHON] Improves error message for show_profiles and du…
HyukjinKwon Sep 18, 2017
48d6aef
[SPARK-22047][TEST] ignore HiveExternalCatalogVersionsSuite
cloud-fan Sep 18, 2017
d0234eb
[SPARK-22047][FLAKY TEST] HiveExternalCatalogVersionsSuite
cloud-fan Sep 19, 2017
6764408
[SPARK-22052] Incorrect Metric assigned in MetricsReporter.scala
Taaffy Sep 19, 2017
5d10586
[SPARK-22076][SQL] Expand.projections should not be a Stream
cloud-fan Sep 20, 2017
401ac20
[SPARK-21384][YARN] Spark + YARN fails with LocalFileSystem as defaul…
Sep 20, 2017
765fd92
[SPARK-21928][CORE] Set classloader on SerializerManager's private kryo
squito Sep 21, 2017
090b987
[SPARK-22094][SS] processAllAvailable should check the query state
zsxwing Sep 22, 2017
de6274a
[SPARK-22072][SPARK-22071][BUILD] Improve release build scripts
holdenk Sep 22, 2017
c0a34a9
[SPARK-18136] Fix SPARK_JARS_DIR for Python pip install on Windows
jsnowacki Sep 23, 2017
1a829df
[SPARK-22092] Reallocation in OffHeapColumnVector.reserveInternal cor…
ala Sep 23, 2017
211d81b
[SPARK-22109][SQL][BRANCH-2.2] Resolves type conflicts between string…
HyukjinKwon Sep 23, 2017
8acce00
[SPARK-22107] Change as to alias in python quickstart
Sep 25, 2017
9836ea1
[SPARK-22083][CORE] Release locks in MemoryStore.evictBlocksToFreeSpace
squito Sep 25, 2017
b0f30b5
[SPARK-22120][SQL] TestHiveSparkSession.reset() should clean out Hive…
Sep 25, 2017
a406473
[SPARK-22141][BACKPORT][SQL] Propagate empty relation before checking…
gengliangwang Sep 27, 2017
42e1727
[SPARK-22140] Add TPCDSQuerySuite
gatorsmile Sep 28, 2017
12a74e3
[SPARK-22135][MESOS] metrics in spark-dispatcher not being registered…
Sep 28, 2017
8c5ab4e
[SPARK-22143][SQL][BRANCH-2.2] Fix memory leak in OffHeapColumnVector
hvanhovell Sep 28, 2017
8b2d838
[SPARK-22129][SPARK-22138] Release script improvements
holdenk Sep 29, 2017
ac9a0f6
[SPARK-22161][SQL] Add Impala-modified TPC-DS queries
gatorsmile Sep 29, 2017
7bf25e0
[SPARK-22146] FileNotFoundException while reading ORC files containin…
mgaido91 Sep 29, 2017
b9adddb
[SPARK-22167][R][BUILD] sparkr packaging issue allow zinc
holdenk Oct 2, 2017
3c30be5
[SPARK-22158][SQL][BRANCH-2.2] convertMetastore should not ignore tab…
dongjoon-hyun Oct 3, 2017
5e3f254
[SPARK-22178][SQL] Refresh Persistent Views by REFRESH TABLE Command
gatorsmile Oct 3, 2017
81232ce
[SPARK-20466][CORE] HadoopRDD#addLocalConfiguration throws NPE
Oct 3, 2017
8a4e7dd
[SPARK-22206][SQL][SPARKR] gapply in R can't work on empty grouping c…
viirya Oct 5, 2017
0d3f166
[SPARK-21549][CORE] Respect OutputFormats with no output directory pr…
szhem Oct 7, 2017
c5889b5
[SPARK-22218] spark shuffle services fails to update secret on app re…
Oct 9, 2017
cd51e2c
[SPARK-21907][CORE][BACKPORT 2.2] oom during spill
Oct 12, 2017
cfc04e0
[SPARK-22217][SQL] ParquetFileFormat to support arbitrary OutputCommi…
steveloughran Oct 12, 2017
c9187db
[SPARK-22252][SQL][2.2] FileFormatWriter should respect the input que…
cloud-fan Oct 13, 2017
30d5c9f
[SPARK-14387][SPARK-16628][SPARK-18355][SQL] Use Spark schema to read…
dongjoon-hyun Oct 13, 2017
acbad83
[SPARK-22273][SQL] Fix key/value schema field names in HashMapGenerat…
ueshin Oct 14, 2017
6b6761e
[SPARK-21549][CORE] Respect OutputFormats with no/invalid output dire…
mridulm Oct 16, 2017
0f060a2
[SPARK-22223][SQL] ObjectHashAggregate should not introduce unnecessa…
viirya Oct 16, 2017
71d1cb6
[SPARK-22249][SQL] isin with empty list throws exception on cached Da…
mgaido91 Oct 17, 2017
c423091
[SPARK-22271][SQL] mean overflows and returns null for some decimal v…
huaxingao Oct 17, 2017
010b50c
[SPARK-22249][FOLLOWUP][SQL] Check if list of value for IN is empty i…
mgaido91 Oct 18, 2017
f8c83fd
[SPARK-21551][PYTHON] Increase timeout for PythonRDD.serveIterator
peay Oct 19, 2017
bf8163f
[SPARK-22319][CORE][BACKPORT-2.2] call loginUserFromKeytab before acc…
sjrand Oct 23, 2017
daa838b
[SPARK-21936][SQL][FOLLOW-UP] backward compatibility test framework f…
srowen Oct 24, 2017
4c1a868
[SPARK-21991][LAUNCHER] Fix race condition in LauncherServer#acceptCo…
Oct 25, 2017
9ed6404
[SPARK-22227][CORE] DiskBlockManager.getAllBlocks now tolerates temp …
Oct 25, 2017
35725f7
[SPARK-22332][ML][TEST] Fix NaiveBayes unit test occasionly fail (cau…
WeichenXu123 Oct 25, 2017
d2dc175
[SPARK-21991][LAUNCHER][FOLLOWUP] Fix java lint
ash211 Oct 25, 2017
24fe7cc
[SPARK-17902][R] Revive stringsAsFactors option for collect() in SparkR
HyukjinKwon Oct 26, 2017
a607ddc
[SPARK-22328][CORE] ClosureCleaner should not miss referenced supercl…
viirya Oct 26, 2017
2839280
[SPARK-22355][SQL] Dataset.collect is not threadsafe
cloud-fan Oct 27, 2017
cb54f29
[SPARK-22356][SQL] data source table should support overlapped column…
cloud-fan Oct 27, 2017
cac6506
[SPARK-19727][SQL][FOLLOWUP] Fix for round function that modifies ori…
cloud-fan Oct 29, 2017
f973587
[SPARK-22344][SPARKR] Set java.io.tmpdir for SparkR tests
shivaram Oct 30, 2017
7f8236c
[SPARK-22291][SQL] Conversion error when transforming array types of …
jmchung Oct 30, 2017
dd69ac6
[SPARK-19611][SQL][FOLLOWUP] set dataSchema correctly in HiveMetastor…
cloud-fan Oct 31, 2017
ab87a92
[SPARK-22333][SQL][BACKPORT-2.2] timeFunctionCall(CURRENT_DATE, CURRE…
DonnyZone Oct 31, 2017
c311c5e
[MINOR][DOC] automatic type inference supports also Date and Timestamp
mgaido91 Nov 2, 2017
4074ed2
[SPARK-22306][SQL][2.2] alter table schema should not erase the bucke…
cloud-fan Nov 2, 2017
5e38373
[SPARK-22211][SQL] Remove incorrect FOJ limit pushdown
Nov 5, 2017
e35c53a
[SPARK-22429][STREAMING] Streaming checkpointing code does not retry …
Nov 5, 2017
2695b92
[SPARK-22315][SPARKR] Warn if SparkR package version doesn't match Sp…
shivaram Nov 6, 2017
161ba18
[SPARK-22417][PYTHON] Fix for createDataFrame from pandas.DataFrame w…
BryanCutler Nov 7, 2017
5c9035b
[SPARK-22327][SPARKR][TEST][BACKPORT-2.2] check for version warning
felixcheung Nov 8, 2017
73a2ca0
[SPARK-22281][SPARKR] Handle R method breaking signature changes
felixcheung Nov 8, 2017
efaf73f
[SPARK-22211][SQL][FOLLOWUP] Fix bad merge for tests
henryr Nov 9, 2017
0e97c8e
[SPARK-22417][PYTHON][FOLLOWUP][BRANCH-2.2] Fix for createDataFrame f…
ueshin Nov 9, 2017
ede0e1a
[SPARK-22403][SS] Add optional checkpointLocation argument to Structu…
wypoon Nov 10, 2017
1b70c66
[SPARK-22287][MESOS] SPARK_DAEMON_MEMORY not honored by MesosClusterD…
Nov 10, 2017
7551524
[SPARK-22472][SQL] add null check for top-level primitive values
cloud-fan Nov 10, 2017
0568f28
[SPARK-22344][SPARKR] clean up install dir if running test as source …
felixcheung Nov 10, 2017
eb49c32
[SPARK-22243][DSTREAM] spark.yarn.jars should reload from config when…
Nov 2, 2017
371be22
[SPARK-22294][DEPLOY] Reset spark.driver.bindAddress when starting a …
ssaavedra Nov 10, 2017
6b4ec22
[SPARK-22284][SQL] Fix 64KB JVM bytecode limit problem in calculating…
kiszk Nov 10, 2017
8b7f72e
[SPARK-19644][SQL] Clean up Scala reflection garbage after creating E…
zsxwing Nov 10, 2017
4ef0bef
[SPARK-21667][STREAMING] ConsoleSink should not fail streaming query …
rekhajoshm Nov 10, 2017
f6ee3d9
[SPARK-19606][MESOS] Support constraints in spark-dispatcher
Nov 12, 2017
114dc42
[SPARK-21720][SQL] Fix 64KB JVM bytecode limit problem with AND or OR
kiszk Nov 12, 2017
00cb9d0
[SPARK-22488][BACKPORT-2.2][SQL] Fix the view resolution issue in the…
gatorsmile Nov 12, 2017
95981fa
[SPARK-22464][BACKPORT-2.2][SQL] No pushdown for Hive metastore parti…
gatorsmile Nov 12, 2017
2a04cfa
[SPARK-19606][BUILD][BACKPORT-2.2][MESOS] fix mesos break
felixcheung Nov 12, 2017
8acd02f
[SPARK-21694][R][ML] Reduce max iterations in Linear SVM test in R to…
HyukjinKwon Nov 12, 2017
f736377
[SPARK-22442][SQL][BRANCH-2.2] ScalaReflection should produce correct…
viirya Nov 13, 2017
2f6dece
[SPARK-22442][SQL][BRANCH-2.2][FOLLOWUP] ScalaReflection should produ…
viirya Nov 13, 2017
c68b4c5
[MINOR][CORE] Using bufferedInputStream for dataDeserializeStream
ConeyLiu Nov 13, 2017
41116ab
Preparing Spark release v2.2.1-rc1
Nov 13, 2017
af0b185
Preparing development version 2.2.2-SNAPSHOT
Nov 13, 2017
d905e85
[SPARK-22471][SQL] SQLListener consumes much memory causing OutOfMemo…
tashoyan Nov 13, 2017
3ea6fd0
[SPARK-22377][BUILD] Use /usr/sbin/lsof if lsof does not exists in re…
HyukjinKwon Nov 13, 2017
210f292
[SPARK-22511][BUILD] Update maven central repo address
srowen Nov 14, 2017
3cefdde
[SPARK-22490][DOC] Add PySpark doc for SparkSession.builder
dongjoon-hyun Nov 15, 2017
3ae187b
[SPARK-22469][SQL] Accuracy problem in comparison with string and num…
liutang123 Nov 15, 2017
b17ba35
[SPARK-22479][SQL][BRANCH-2.2] Exclude credentials from SaveintoDataS…
onursatici Nov 16, 2017
17ba7b9
[SPARK-22499][SQL] Fix 64KB JVM bytecode limit problem with least and…
kiszk Nov 16, 2017
52b05b6
[SPARK-22494][SQL] Fix 64KB limit exception with Coalesce and Atleast…
mgaido91 Nov 16, 2017
0b51fd3
[SPARK-22501][SQL] Fix 64KB JVM bytecode limit problem with in
kiszk Nov 16, 2017
be68f86
[SPARK-22535][PYSPARK] Sleep before killing the python worker in Pyth…
zsxwing Nov 16, 2017
ef7ccc1
[SPARK-22540][SQL] Ensure HighlyCompressedMapStatus calculates correc…
Nov 17, 2017
3bc37e5
[SPARK-22538][ML] SQLTransformer should not unpersist possibly cached…
viirya Nov 17, 2017
53a6076
[SPARK-22544][SS] FileStreamSource should use its own hadoop conf to …
zsxwing Nov 17, 2017
710d618
[SPARK-22498][SQL] Fix 64KB JVM bytecode limit problem with concat
kiszk Nov 18, 2017
ca02575
[SPARK-22549][SQL] Fix 64KB JVM bytecode limit problem with concat_ws
kiszk Nov 21, 2017
23eb4d7
[SPARK-22508][SQL] Fix 64KB JVM bytecode limit problem with GenerateU…
kiszk Nov 21, 2017
94f9227
[SPARK-22550][SQL] Fix 64KB JVM bytecode limit problem with elt
kiszk Nov 21, 2017
11a599b
[SPARK-22500][SQL] Fix 64KB JVM bytecode limit problem with cast
kiszk Nov 21, 2017
df9228b
[SPARK-22548][SQL] Incorrect nested AND expression pushed down to JDB…
jliwork Nov 22, 2017
b17f406
[SPARK-17920][SPARK-19580][SPARK-19878][SQL] Backport PR 19779 to bra…
vinodkc Nov 22, 2017
f8e73d0
[SPARK-17920][SQL] [FOLLOWUP] Backport PR 19779 to branch-2.2
vinodkc Nov 24, 2017
f4c457a
[SPARK-22591][SQL] GenerateOrdering shouldn't change CodegenContext.I…
viirya Nov 24, 2017
ad57141
[SPARK-22595][SQL] fix flaky test: CastSuite.SPARK-22500: cast for st…
kiszk Nov 24, 2017
b606cc2
[SPARK-22495] Fix setup of SPARK_HOME variable on Windows
jsnowacki Nov 24, 2017
c3b5df2
fix typo
Nov 24, 2017
e30e269
Preparing Spark release v2.2.1-rc2
Nov 24, 2017
455cea6
Preparing development version 2.2.2-SNAPSHOT
Nov 24, 2017
2cd4898
[SPARK-22607][BUILD] Set large stack size consistently for tests to a…
srowen Nov 26, 2017
eef72d3
[SPARK-22603][SQL] Fix 64KB JVM bytecode limit problem with FormatString
kiszk Nov 27, 2017
38a0532
[SPARK-22637][SQL] Only refresh a logical plan once.
hvanhovell Nov 29, 2017
d7b1474
[SPARK-22654][TESTS] Retry Spark tarball download if failed in HiveEx…
srowen Nov 30, 2017
0121ebc
[SPARK-22373] Bump Janino dependency version to fix thread safety issue…
Victsm Dec 1, 2017
af8a692
[SPARK-22653] executorAddress registered in CoarseGrainedSchedulerBac…
tgravescs Dec 1, 2017
ba00bd9
[SPARK-22601][SQL] Data load is getting displayed successful on provi…
sujith71955 Dec 1, 2017
f3f8c87
[SPARK-22635][SQL][ORC] FileNotFoundException while reading ORC files…
mgaido91 Nov 30, 2017
5b63000
[SPARK-22162][BRANCH-2.2] Executors and the driver should use consist…
Dec 5, 2017
7fd6d53
[SPARK-22686][SQL] DROP TABLE IF EXISTS should not show AnalysisExcep…
dongjoon-hyun Dec 6, 2017
2084675
[SPARK-22688][SQL] Upgrade Janino version to 3.0.8
kiszk Dec 7, 2017
9e2d96d
[SPARK-22688][SQL][HOTFIX] Upgrade Janino version to 3.0.8
srowen Dec 7, 2017
00cdb38
[SPARK-22289][ML] Add JSON support for Matrix parameters (LR with coe…
YY-OnCall Dec 12, 2017
728a45e
[SPARK-22574][MESOS][SUBMIT] Check submission request parameters
Gschiavon Dec 12, 2017
0230515
Revert "[SPARK-22574][MESOS][SUBMIT] Check submission request paramet…
gatorsmile Dec 12, 2017
b4f4be3
[SPARK-22574][MESOS][SUBMIT] Check submission request parameters
Gschiavon Dec 13, 2017
1e4cca0
[SPARK-22817][R] Use fixed testthat version for SparkR tests in AppVeyor
HyukjinKwon Dec 17, 2017
1cf3e3a
[SPARK-22862] Docs on lazy elimination of columns missing from an enc…
marmbrus Dec 22, 2017
7a97943
[SPARK-20694][EXAMPLES] Update SQLDataSourceExample.scala
CNRui Dec 23, 2017
41f705a
[SPARK-22889][SPARKR] Set overwrite=T when install SparkR in tests
shivaram Dec 23, 2017
7c30ae3
[SPARK-22983] Don't push filters beneath aggregates with empty groupi…
JoshRosen Jan 8, 2018
24f1f2a
[SPARK-22984] Fix incorrect bitmap copying and offset adjustment in G…
JoshRosen Jan 9, 2018
0d943d9
[SPARK-22972] Couldn't find corresponding Hive SerDe for data source …
xubo245 Jan 10, 2018
acab4e7
[SPARK-23001][SQL] Fix NullPointerException when DESC a database with…
gatorsmile Jan 11, 2018
20eea20
[SPARK-22982] Remove unsafe asynchronous close() call from FileDownlo…
JoshRosen Jan 10, 2018
105ae86
[SPARK-22975][SS] MetricsReporter should not throw exception when the…
mgaido91 Jan 12, 2018
7022ef8
[SPARK-23038][TEST] Update docker/spark-test (JDK/OS)
dongjoon-hyun Jan 14, 2018
d09eecc
[SPARK-23095][SQL] Decorrelation of scalar subquery fails with java.u…
dilipbiswal Jan 17, 2018
0e58fee
[DOCS] change to dataset for java code in structured-streaming-kafka-…
brandonJY Jan 19, 2018
5273cc7
[SPARK-23281][SQL] Query produces results in incorrect order when a c…
dilipbiswal Jan 31, 2018
cb73ecd
revert the removal of import in SPARK-23281
gatorsmile Jan 31, 2018
f65e653
[SPARK-23358][CORE] When the number of partitions is greater than 2^2…
10110346 Feb 9, 2018
1b4c6ab
[SPARK-23186][SQL][BRANCH-2.2] Initialize DriverManager first before …
dongjoon-hyun Feb 10, 2018
1694834
[SPARK-23376][SQL] creating UnsafeKVExternalSorter with BytesToBytesM…
cloud-fan Feb 11, 2018
14b5dbf
[SPARK-23391][CORE] It may lead to overflow for some integer multipli…
10110346 Feb 12, 2018
73263b2
[SPARK-23053][CORE] taskBinarySerialization and task partitions calcu…
Feb 13, 2018
a95c3e2
[SPARK-23230][SQL][BRANCH-2.2] When hive.default.fileformat is other …
cxzl25 Feb 14, 2018
1cc34f3
[SPARK-22700][ML] Bucketizer.transform incorrectly drops row containi…
zhengruifeng Feb 22, 2018
fa3667e
[SPARK-23438][DSTREAMS] Fix DStreams data loss with WAL when driver c…
gaborgsomogyi Feb 26, 2018
9bd25c9
[SPARK-23508][CORE] Fix BlockmanagerId in case blockManagerIdCache ca…
caneGuy Feb 28, 2018
4864d21
[SPARK-23434][SQL][BRANCH-2.2] Spark should not warn `metadata direct…
dongjoon-hyun Mar 5, 2018
175b221
[SPARK-23525][BACKPORT][SQL] Support ALTER TABLE CHANGE COLUMN COMMEN…
jiangxb1987 Mar 8, 2018
367a161
[SPARK-23649][SQL] Skipping chars disallowed in UTF-8
MaxGekk Mar 20, 2018
85ab72b
[SPARK-23759][UI] Unable to bind Spark UI to specific host name / IP
Mar 23, 2018
6b5f9c3
[SPARK-23788][SS] Fix race in StreamingQuerySuite
jose-torres Mar 25, 2018
e957c4e
[SPARK-23816][CORE] Killed tasks should ignore FetchFailures.
squito Apr 9, 2018
a902323
[SPARK-24007][SQL] EqualNullSafe for FloatType and DoubleType might g…
ueshin Apr 18, 2018
041aec4
[SPARK-23963][SQL] Properly handle large number of columns in query o…
bersprockets Apr 13, 2018
e77d62a
[MINOR][DOCS] Fix comments of SQLExecution#withExecutionId
seancxmao Apr 24, 2018
4f1ae3a
[SPARK-23941][MESOS] Mesos task failed on specific spark app name
BounkongK May 1, 2018
154bbc9
[SPARK-23433][CORE] Late zombie task completions update all tasksets
squito May 3, 2018
768d0b7
[SPARK-23489][SQL][TEST][BRANCH-2.2] HiveExternalCatalogVersionsSuite…
dongjoon-hyun May 4, 2018
866270e
[SPARK-23697][CORE] LegacyAccumulatorWrapper should define isZero cor…
cloud-fan May 4, 2018
f9d6a16
[SPARK-21278][PYSPARK] Upgrade to Py4J 0.10.6
dongjoon-hyun Jul 5, 2017
850b7d8
[PYSPARK] Update py4j to version 0.10.7.
Apr 13, 2018
f96d13d
[SPARKR] Match pyspark features in SparkR communication protocol.
Apr 17, 2018
8c223b6
[R][BACKPORT-2.2] backport lint fix
felixcheung May 16, 2018
6a55d8b
[SPARK-23850][SQL][BRANCH-2.2] Add separate config for SQL options re…
May 18, 2018
2379074
[SPARK-24257][SQL] LongToUnsafeRowMap calculate the new size may be w…
cxzl25 May 24, 2018
8abd0a7
fix compilation caused by SPARK-24257
cloud-fan May 24, 2018
fb86eb0
[MINOR] Add port SSL config in toString and scaladoc
mgaido91 May 25, 2018
c306a84
[WEBUI] Avoid possibility of script in query param keys
srowen May 31, 2018
bf0b212
[SPARK-24531][TESTS] Remove version 2.2.0 from testing versions in Hi…
mgaido91 Jun 12, 2018
1f81ade
[SPARK-24506][UI] Add UI filters to tabs added after binding
mgaido91 Jun 12, 2018
090b883
[SPARK-23732][DOCS] Fix source links in generated scaladoc.
Jun 12, 2018
8ce9e2a
Preparing Spark release v2.2.2-rc1
tgravescs Jun 18, 2018
e2e4d58
Preparing development version 2.2-3-SNAPSHOT
tgravescs Jun 18, 2018
7bfefc9
Preparing development version 2.2.3-SNAPSHOT
tgravescs Jun 18, 2018
751b008
[SPARK-24589][CORE] Correctly identify tasks in output commit coordin…
Jun 21, 2018
a600004
[SPARK-22897][CORE] Expose stageAttemptId in TaskContext
advancedxy Jun 22, 2018
72575d0
[SPARK-24552][CORE][BRANCH-2.2] Use unique id instead of attempt numb…
Jun 25, 2018
fc28ba3
Preparing Spark release v2.2.2-rc2
tgravescs Jun 27, 2018
4795827
Preparing development version 2.2.3-SNAPSHOT
tgravescs Jun 27, 2018
a8537a5
[SPARK-24603][SQL] Fix findTightestCommonType reference in comments
Jun 28, 2018
4bc4ccd
[SPARK-24813][TESTS][HIVE][HOTFIX][BRANCH-2.2] HiveExternalCatalogVer…
srowen Jul 16, 2018
17db572
Update for spark 2.2.2 release
tgravescs Jul 16, 2018
144426c
[SPARK-24677][CORE] Avoid NoSuchElementException from MedianHeap
cxzl25 Jul 18, 2018
f339e2f
[SPARK-22499][FOLLOWUP][SQL] Reduce input string expressions for Leas…
HyukjinKwon Jul 24, 2018
7376473
[SPARK-24927][BUILD][BRANCH-2.3] The scope of snappy-java cannot be "…
liancheng Jul 27, 2018
f52d0c4
[SPARK-24809][SQL] Serializing LongToUnsafeRowMap in executor may res…
liutang123 Jul 29, 2018
c4b3769
[MINOR][CORE][TEST] Fix afterEach() in TastSetManagerSuite and TaskSc…
jiangxb1987 Jul 30, 2018
22ce805
[SPARK-24957][SQL][BACKPORT-2.2] Average with decimal followed by agg…
mgaido91 Aug 1, 2018
a5624c7
[PYSPARK] Updates to Accumulators
LucaCanali Jul 18, 2018
53ac850
[SPARK-24948][SHS][BACKPORT-2.2] Delegate check access permissions to…
mgaido91 Aug 8, 2018
b283c1f
[SPARK-24950][SQL] DateTimeUtilsSuite daysToMillis and millisToDays f…
d80tb7 Jul 28, 2018
051ea3a
[MINOR][BUILD] Add ECCN notice required by http://www.apache.org/dev/…
srowen Aug 10, 2018
1e73ee2
[SPARK-25081][CORE] Nested spill in ShuffleExternalSorter should not …
zsxwing Aug 12, 2018
6641aa6
[MINOR][SQL][DOC] Fix `to_json` example in function description and doc
dongjoon-hyun Aug 14, 2018
124789b
[SPARK-25144][SQL][TEST][BRANCH-2.2] Free aggregate map when task ends
cloud-fan Aug 21, 2018
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
13 changes: 7 additions & 6 deletions LICENSE
Original file line number Diff line number Diff line change
Expand Up @@ -249,11 +249,11 @@ The text of each license is also included at licenses/LICENSE-[project].txt.
(Interpreter classes (all .scala files in repl/src/main/scala
except for Main.Scala, SparkHelper.scala and ExecutorClassLoader.scala),
and for SerializableMapWrapper in JavaUtils.scala)
(BSD-like) Scala Actors library (org.scala-lang:scala-actors:2.11.7 - http://www.scala-lang.org/)
(BSD-like) Scala Compiler (org.scala-lang:scala-compiler:2.11.7 - http://www.scala-lang.org/)
(BSD-like) Scala Compiler (org.scala-lang:scala-reflect:2.11.7 - http://www.scala-lang.org/)
(BSD-like) Scala Library (org.scala-lang:scala-library:2.11.7 - http://www.scala-lang.org/)
(BSD-like) Scalap (org.scala-lang:scalap:2.11.7 - http://www.scala-lang.org/)
(BSD-like) Scala Actors library (org.scala-lang:scala-actors:2.11.8 - http://www.scala-lang.org/)
(BSD-like) Scala Compiler (org.scala-lang:scala-compiler:2.11.8 - http://www.scala-lang.org/)
(BSD-like) Scala Compiler (org.scala-lang:scala-reflect:2.11.8 - http://www.scala-lang.org/)
(BSD-like) Scala Library (org.scala-lang:scala-library:2.11.8 - http://www.scala-lang.org/)
(BSD-like) Scalap (org.scala-lang:scalap:2.11.8 - http://www.scala-lang.org/)
(BSD-style) scalacheck (org.scalacheck:scalacheck_2.11:1.10.0 - http://www.scalacheck.org)
(BSD-style) spire (org.spire-math:spire_2.11:0.7.1 - http://spire-math.org)
(BSD-style) spire-macros (org.spire-math:spire-macros_2.11:0.7.1 - http://spire-math.org)
Expand All @@ -263,7 +263,7 @@ The text of each license is also included at licenses/LICENSE-[project].txt.
(New BSD license) Protocol Buffer Java API (org.spark-project.protobuf:protobuf-java:2.4.1-shaded - http://code.google.com/p/protobuf)
(The BSD License) Fortran to Java ARPACK (net.sourceforge.f2j:arpack_combined_all:0.1 - http://f2j.sourceforge.net)
(The BSD License) xmlenc Library (xmlenc:xmlenc:0.52 - http://xmlenc.sourceforge.net)
(The New BSD License) Py4J (net.sf.py4j:py4j:0.10.4 - http://py4j.sourceforge.net/)
(The New BSD License) Py4J (net.sf.py4j:py4j:0.10.7 - http://py4j.sourceforge.net/)
(Two-clause BSD-style license) JUnit-Interface (com.novocode:junit-interface:0.10 - http://github.com/szeiger/junit-interface/)
(BSD licence) sbt and sbt-launch-lib.bash
(BSD 3 Clause) d3.min.js (https://github.com/mbostock/d3/blob/master/LICENSE)
Expand Down Expand Up @@ -297,3 +297,4 @@ The text of each license is also included at licenses/LICENSE-[project].txt.
(MIT License) RowsGroup (http://datatables.net/license/mit)
(MIT License) jsonFormatter (http://www.jqueryscript.net/other/jQuery-Plugin-For-Pretty-JSON-Formatting-jsonFormatter.html)
(MIT License) modernizr (https://github.com/Modernizr/Modernizr/blob/master/LICENSE)
(MIT License) machinist (https://github.com/typelevel/machinist)
25 changes: 25 additions & 0 deletions NOTICE
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,31 @@ This product includes software developed at
The Apache Software Foundation (http://www.apache.org/).


Export Control Notice
---------------------

This distribution includes cryptographic software. The country in which you currently reside may have
restrictions on the import, possession, use, and/or re-export to another country, of encryption software.
BEFORE using any encryption software, please check your country's laws, regulations and policies concerning
the import, possession, or use, and re-export of encryption software, to see if this is permitted. See
<http://www.wassenaar.org/> for more information.

The U.S. Government Department of Commerce, Bureau of Industry and Security (BIS), has classified this
software as Export Commodity Control Number (ECCN) 5D002.C.1, which includes information security software
using or performing cryptographic functions with asymmetric algorithms. The form and manner of this Apache
Software Foundation distribution makes it eligible for export under the License Exception ENC Technology
Software Unrestricted (TSU) exception (see the BIS Export Administration Regulations, Section 740.13) for
both object code and source code.

The following provides more details on the included cryptographic software:

This software uses Apache Commons Crypto (https://commons.apache.org/proper/commons-crypto/) to
support authentication, and encryption and decryption of data sent across the network between
services.

This software includes Bouncy Castle (http://bouncycastle.org/) to support the jets3t library.


========================================================================
Common Development and Distribution License 1.0
========================================================================
Expand Down
6 changes: 1 addition & 5 deletions R/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -66,11 +66,7 @@ To run one of them, use `./bin/spark-submit <filename> <args>`. For example:
```bash
./bin/spark-submit examples/src/main/r/dataframe.R
```
You can also run the unit tests for SparkR by running. You need to install the [testthat](http://cran.r-project.org/web/packages/testthat/index.html) package first:
```bash
R -e 'install.packages("testthat", repos="http://cran.us.r-project.org")'
./R/run-tests.sh
```
You can run R unit tests by following the instructions under [Running R Tests](http://spark.apache.org/docs/latest/building-spark.html#running-r-tests).

### Running on YARN

Expand Down
3 changes: 1 addition & 2 deletions R/WINDOWS.md
Original file line number Diff line number Diff line change
Expand Up @@ -34,10 +34,9 @@ To run the SparkR unit tests on Windows, the following steps are required —ass

4. Set the environment variable `HADOOP_HOME` to the full path to the newly created `hadoop` directory.

5. Run unit tests for SparkR by running the command below. You need to install the [testthat](http://cran.r-project.org/web/packages/testthat/index.html) package first:
5. Run unit tests for SparkR by running the command below. You need to install the needed packages following the instructions under [Running R Tests](http://spark.apache.org/docs/latest/building-spark.html#running-r-tests) first:

```
R -e "install.packages('testthat', repos='http://cran.us.r-project.org')"
.\bin\spark-submit2.cmd --conf spark.hadoop.fs.defaultFS="file:///" R\pkg\tests\run-all.R
```

1 change: 1 addition & 0 deletions R/install-dev.sh
Original file line number Diff line number Diff line change
Expand Up @@ -28,6 +28,7 @@

set -o pipefail
set -e
set -x

FWDIR="$(cd "`dirname "${BASH_SOURCE[0]}"`"; pwd)"
LIB_DIR="$FWDIR/lib"
Expand Down
1 change: 1 addition & 0 deletions R/pkg/.Rbuildignore
Original file line number Diff line number Diff line change
Expand Up @@ -6,3 +6,4 @@
^README\.Rmd$
^src-native$
^html$
^tests/fulltests/*
4 changes: 2 additions & 2 deletions R/pkg/DESCRIPTION
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
Package: SparkR
Type: Package
Version: 2.2.0
Version: 2.2.3
Title: R Frontend for Apache Spark
Description: The SparkR package provides an R Frontend for Apache Spark.
Description: Provides an R Frontend for Apache Spark.
Authors@R: c(person("Shivaram", "Venkataraman", role = c("aut", "cre"),
email = "[email protected]"),
person("Xiangrui", "Meng", role = "aut",
Expand Down
1 change: 1 addition & 0 deletions R/pkg/NAMESPACE
Original file line number Diff line number Diff line change
Expand Up @@ -122,6 +122,7 @@ exportMethods("arrange",
"group_by",
"groupBy",
"head",
"hint",
"insertInto",
"intersect",
"isLocal",
Expand Down
47 changes: 42 additions & 5 deletions R/pkg/R/DataFrame.R
Original file line number Diff line number Diff line change
Expand Up @@ -591,7 +591,7 @@ setMethod("cache",
#'
#' Persist this SparkDataFrame with the specified storage level. For details of the
#' supported storage levels, refer to
#' \url{http://spark.apache.org/docs/latest/programming-guide.html#rdd-persistence}.
#' \url{http://spark.apache.org/docs/latest/rdd-programming-guide.html#rdd-persistence}.
#'
#' @param x the SparkDataFrame to persist.
#' @param newLevel storage level chosen for the persistance. See available options in
Expand Down Expand Up @@ -1174,6 +1174,9 @@ setMethod("collect",
vec <- do.call(c, col)
stopifnot(class(vec) != "list")
class(vec) <- PRIMITIVE_TYPES[[colType]]
if (is.character(vec) && stringsAsFactors) {
vec <- as.factor(vec)
}
df[[colIndex]] <- vec
} else {
df[[colIndex]] <- col
Expand Down Expand Up @@ -2642,6 +2645,7 @@ generateAliasesForIntersectedCols <- function (x, intersectedColNames, suffix) {
#' Input SparkDataFrames can have different schemas (names and data types).
#'
#' Note: This does not remove duplicate rows across the two SparkDataFrames.
#' Also as standard in SQL, this function resolves columns by position (not by name).
#'
#' @param x A SparkDataFrame
#' @param y A SparkDataFrame
Expand Down Expand Up @@ -3114,7 +3118,7 @@ setMethod("as.data.frame",
#'
#' @family SparkDataFrame functions
#' @rdname attach
#' @aliases attach,SparkDataFrame-method
#' @aliases attach attach,SparkDataFrame-method
#' @param what (SparkDataFrame) The SparkDataFrame to attach
#' @param pos (integer) Specify position in search() where to attach.
#' @param name (character) Name to use for the attached SparkDataFrame. Names
Expand All @@ -3130,9 +3134,12 @@ setMethod("as.data.frame",
#' @note attach since 1.6.0
setMethod("attach",
signature(what = "SparkDataFrame"),
function(what, pos = 2, name = deparse(substitute(what)), warn.conflicts = TRUE) {
newEnv <- assignNewEnv(what)
attach(newEnv, pos = pos, name = name, warn.conflicts = warn.conflicts)
function(what, pos = 2L, name = deparse(substitute(what), backtick = FALSE),
warn.conflicts = TRUE) {
args <- as.list(environment()) # capture all parameters - this must be the first line
newEnv <- assignNewEnv(args$what)
args$what <- newEnv
do.call(attach, args)
})

#' Evaluate a R expression in an environment constructed from a SparkDataFrame
Expand Down Expand Up @@ -3642,3 +3649,33 @@ setMethod("checkpoint",
df <- callJMethod(x@sdf, "checkpoint", as.logical(eager))
dataFrame(df)
})

#' hint
#'
#' Specifies execution plan hint and return a new SparkDataFrame.
#'
#' @param x a SparkDataFrame.
#' @param name a name of the hint.
#' @param ... optional parameters for the hint.
#' @return A SparkDataFrame.
#' @family SparkDataFrame functions
#' @aliases hint,SparkDataFrame,character-method
#' @rdname hint
#' @name hint
#' @export
#' @examples
#' \dontrun{
#' df <- createDataFrame(mtcars)
#' avg_mpg <- mean(groupBy(createDataFrame(mtcars), "cyl"), "mpg")
#'
#' head(join(df, hint(avg_mpg, "broadcast"), df$cyl == avg_mpg$cyl))
#' }
#' @note hint since 2.2.0
setMethod("hint",
signature(x = "SparkDataFrame", name = "character"),
function(x, name, ...) {
parameters <- list(...)
stopifnot(all(sapply(parameters, is.character)))
jdf <- callJMethod(x@sdf, "hint", name, parameters)
dataFrame(jdf)
})
2 changes: 1 addition & 1 deletion R/pkg/R/RDD.R
Original file line number Diff line number Diff line change
Expand Up @@ -227,7 +227,7 @@ setMethod("cacheRDD",
#'
#' Persist this RDD with the specified storage level. For details of the
#' supported storage levels, refer to
#'\url{http://spark.apache.org/docs/latest/programming-guide.html#rdd-persistence}.
#'\url{http://spark.apache.org/docs/latest/rdd-programming-guide.html#rdd-persistence}.
#'
#' @param x The RDD to persist
#' @param newLevel The new storage level to be assigned
Expand Down
6 changes: 3 additions & 3 deletions R/pkg/R/SQLContext.R
Original file line number Diff line number Diff line change
Expand Up @@ -334,7 +334,7 @@ setMethod("toDF", signature(x = "RDD"),
#'
#' Loads a JSON file, returning the result as a SparkDataFrame
#' By default, (\href{http://jsonlines.org/}{JSON Lines text format or newline-delimited JSON}
#' ) is supported. For JSON (one record per file), set a named property \code{wholeFile} to
#' ) is supported. For JSON (one record per file), set a named property \code{multiLine} to
#' \code{TRUE}.
#' It goes through the entire dataset once to determine the schema.
#'
Expand All @@ -348,7 +348,7 @@ setMethod("toDF", signature(x = "RDD"),
#' sparkR.session()
#' path <- "path/to/file.json"
#' df <- read.json(path)
#' df <- read.json(path, wholeFile = TRUE)
#' df <- read.json(path, multiLine = TRUE)
#' df <- jsonFile(path)
#' }
#' @name read.json
Expand Down Expand Up @@ -598,7 +598,7 @@ tableToDF <- function(tableName) {
#' df1 <- read.df("path/to/file.json", source = "json")
#' schema <- structType(structField("name", "string"),
#' structField("info", "map<string,double>"))
#' df2 <- read.df(mapTypeJsonPath, "json", schema, wholeFile = TRUE)
#' df2 <- read.df(mapTypeJsonPath, "json", schema, multiLine = TRUE)
#' df3 <- loadDF("data/test_table", "parquet", mergeSchema = "true")
#' }
#' @name read.df
Expand Down
4 changes: 2 additions & 2 deletions R/pkg/R/client.R
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@

# Creates a SparkR client connection object
# if one doesn't already exist
connectBackend <- function(hostname, port, timeout) {
connectBackend <- function(hostname, port, timeout, authSecret) {
if (exists(".sparkRcon", envir = .sparkREnv)) {
if (isOpen(.sparkREnv[[".sparkRCon"]])) {
cat("SparkRBackend client connection already exists\n")
Expand All @@ -29,7 +29,7 @@ connectBackend <- function(hostname, port, timeout) {

con <- socketConnection(host = hostname, port = port, server = FALSE,
blocking = TRUE, open = "wb", timeout = timeout)

doServerAuth(con, authSecret)
assign(".sparkRCon", con, envir = .sparkREnv)
con
}
Expand Down
10 changes: 7 additions & 3 deletions R/pkg/R/deserialize.R
Original file line number Diff line number Diff line change
Expand Up @@ -60,14 +60,18 @@ readTypedObject <- function(con, type) {
stop(paste("Unsupported type for deserialization", type)))
}

readString <- function(con) {
stringLen <- readInt(con)
raw <- readBin(con, raw(), stringLen, endian = "big")
readStringData <- function(con, len) {
raw <- readBin(con, raw(), len, endian = "big")
string <- rawToChar(raw)
Encoding(string) <- "UTF-8"
string
}

readString <- function(con) {
stringLen <- readInt(con)
readStringData(con, stringLen)
}

readInt <- function(con) {
readBin(con, integer(), n = 1, endian = "big")
}
Expand Down
16 changes: 9 additions & 7 deletions R/pkg/R/generics.R
Original file line number Diff line number Diff line change
Expand Up @@ -398,7 +398,8 @@ setGeneric("as.data.frame",
standardGeneric("as.data.frame")
})

#' @rdname attach
# Do not document the generic because of signature changes across R versions
#' @noRd
#' @export
setGeneric("attach")

Expand Down Expand Up @@ -572,6 +573,10 @@ setGeneric("group_by", function(x, ...) { standardGeneric("group_by") })
#' @export
setGeneric("groupBy", function(x, ...) { standardGeneric("groupBy") })

#' @rdname hint
#' @export
setGeneric("hint", function(x, name, ...) { standardGeneric("hint") })

#' @rdname insertInto
#' @export
setGeneric("insertInto", function(x, tableName, ...) { standardGeneric("insertInto") })
Expand Down Expand Up @@ -1356,12 +1361,9 @@ setGeneric("year", function(x) { standardGeneric("year") })
#' @export
setGeneric("fitted")

#' @param x,y For \code{glm}: logical values indicating whether the response vector
#' and model matrix used in the fitting process should be returned as
#' components of the returned value.
#' @inheritParams stats::glm
#' @rdname glm
# Do not carry stats::glm usage and param here, and do not document the generic
#' @export
#' @noRd
setGeneric("glm")

#' @param object a fitted ML model object.
Expand Down Expand Up @@ -1469,7 +1471,7 @@ setGeneric("write.ml", function(object, path, ...) { standardGeneric("write.ml")

#' @rdname awaitTermination
#' @export
setGeneric("awaitTermination", function(x, timeout) { standardGeneric("awaitTermination") })
setGeneric("awaitTermination", function(x, timeout = NULL) { standardGeneric("awaitTermination") })

#' @rdname isActive
#' @export
Expand Down
Loading