Skip to content

Commit 88a03cc

Browse files
committed
Add new lines between confs and return types
1 parent ce7c6c8 commit 88a03cc

File tree

1 file changed

+22
-3
lines changed

1 file changed

+22
-3
lines changed

sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala

Lines changed: 22 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -146,35 +146,44 @@ private[spark] object SQLConf {
146146
defaultValue = Some(true),
147147
doc = "When set to true Spark SQL will automatically select a compression codec for each " +
148148
"column based on statistics of the data.")
149+
149150
val COLUMN_BATCH_SIZE = intConf("spark.sql.inMemoryColumnarStorage.batchSize",
150151
defaultValue = Some(10000),
151152
doc = "Controls the size of batches for columnar caching. Larger batch sizes can improve " +
152153
"memory utilization and compression, but risk OOMs when caching data.")
154+
153155
val IN_MEMORY_PARTITION_PRUNING =
154156
booleanConf("spark.sql.inMemoryColumnarStorage.partitionPruning",
155157
defaultValue = Some(false),
156158
doc = "<TODO>")
159+
157160
val AUTO_BROADCASTJOIN_THRESHOLD = intConf("spark.sql.autoBroadcastJoinThreshold",
158161
defaultValue = Some(10 * 1024 * 1024),
159162
doc = "Configures the maximum size in bytes for a table that will be broadcast to all worker " +
160163
"nodes when performing a join. By setting this value to -1 broadcasting can be disabled. " +
161164
"Note that currently statistics are only supported for Hive Metastore tables where the " +
162165
"command<code>ANALYZE TABLE &lt;tableName&gt; COMPUTE STATISTICS noscan</code> has been run.")
166+
163167
val DEFAULT_SIZE_IN_BYTES = longConf("spark.sql.defaultSizeInBytes", isPublic = false)
168+
164169
val SHUFFLE_PARTITIONS = intConf("spark.sql.shuffle.partitions",
165170
defaultValue = Some(200),
166171
doc = "Configures the number of partitions to use when shuffling data for joins or " +
167172
"aggregations.")
173+
168174
val CODEGEN_ENABLED = booleanConf("spark.sql.codegen",
169175
defaultValue = Some(false),
170176
doc = "When true, code will be dynamically generated at runtime for expression evaluation in" +
171177
" a specific query. For some queries with complicated expression this option can lead to " +
172178
"significant speed-ups. However, for simple queries this can actually slow down query " +
173179
"execution.")
180+
174181
val UNSAFE_ENABLED = booleanConf("spark.sql.unsafe.enabled",
175182
defaultValue = Some(false),
176183
doc = "<TDDO>")
184+
177185
val DIALECT = stringConf("spark.sql.dialect", defaultValue = Some("sql"), doc = "<TODO>")
186+
178187
val CASE_SENSITIVE = booleanConf("spark.sql.caseSensitive",
179188
defaultValue = Some(true),
180189
doc = "<TODO>")
@@ -185,26 +194,31 @@ private[spark] object SQLConf {
185194
"Spark SQL, do not differentiate between binary data and strings when writing out the " +
186195
"Parquet schema. This flag tells Spark SQL to interpret binary data as a string to provide " +
187196
"compatibility with these systems.")
197+
188198
val PARQUET_INT96_AS_TIMESTAMP = booleanConf("spark.sql.parquet.int96AsTimestamp",
189199
defaultValue = Some(true),
190200
doc = "Some Parquet-producing systems, in particular Impala, store Timestamp into INT96. " +
191201
"Spark would also store Timestamp as INT96 because we need to avoid precision lost of the " +
192202
"nanoseconds field. This flag tells Spark SQL to interpret INT96 data as a timestamp to " +
193203
"provide compatibility with these systems.")
204+
194205
val PARQUET_CACHE_METADATA = booleanConf("spark.sql.parquet.cacheMetadata",
195206
defaultValue = Some(true),
196207
doc = "Turns on caching of Parquet schema metadata. Can speed up querying of static data.")
208+
197209
val PARQUET_COMPRESSION = stringConf("spark.sql.parquet.compression.codec",
198210
defaultValue = Some("gzip"),
199211
doc = "Sets the compression codec use when writing Parquet files. Acceptable values include: " +
200212
"uncompressed, snappy, gzip, lzo.")
213+
201214
val PARQUET_FILTER_PUSHDOWN_ENABLED = booleanConf("spark.sql.parquet.filterPushdown",
202215
defaultValue = Some(false),
203216
doc = "Turn on Parquet filter pushdown optimization. This feature is turned off by default" +
204217
" because of a known bug in Paruet 1.6.0rc3 " +
205218
"(<a href=\"https://issues.apache.org/jira/browse/PARQUET-136\">PARQUET-136</a>). However, " +
206219
"if your table doesn't contain any nullable string or binary columns, it's still safe to " +
207220
"turn this feature on.")
221+
208222
val PARQUET_USE_DATA_SOURCE_API = booleanConf("spark.sql.parquet.useDataSourceApi",
209223
defaultValue = Some(true),
210224
doc = "<TODO>")
@@ -220,6 +234,7 @@ private[spark] object SQLConf {
220234
val COLUMN_NAME_OF_CORRUPT_RECORD = stringConf("spark.sql.columnNameOfCorruptRecord",
221235
defaultValue = Some("_corrupt_record"),
222236
doc = "<TODO>")
237+
223238
val BROADCAST_TIMEOUT = intConf("spark.sql.broadcastTimeout",
224239
defaultValue = Some(5 * 60),
225240
doc = "<TODO>")
@@ -230,21 +245,25 @@ private[spark] object SQLConf {
230245
defaultValue = Some(false),
231246
doc = "When true, performs sorts spilling to disk as needed otherwise sort each partition in" +
232247
" memory.")
248+
233249
val SORTMERGE_JOIN = booleanConf("spark.sql.planner.sortMergeJoin",
234250
defaultValue = Some(false),
235251
doc = "<TODO>")
236252

237253
// This is only used for the thriftserver
238254
val THRIFTSERVER_POOL = stringConf("spark.sql.thriftserver.scheduler.pool", isPublic = false)
255+
239256
val THRIFTSERVER_UI_STATEMENT_LIMIT =
240257
intConf("spark.sql.thriftserver.ui.retainedStatements", isPublic = false)
258+
241259
val THRIFTSERVER_UI_SESSION_LIMIT =
242260
intConf("spark.sql.thriftserver.ui.retainedSessions", isPublic = false)
243261

244262
// This is used to set the default data source
245263
val DEFAULT_DATA_SOURCE_NAME = stringConf("spark.sql.sources.default",
246264
defaultValue = Some("org.apache.spark.sql.parquet"),
247265
doc = "<TODO>")
266+
248267
// This is used to control the when we will split a schema's JSON string to multiple pieces
249268
// in order to fit the JSON string in metastore's table property (by default, the value has
250269
// a length restriction of 4000 characters). We will split the JSON string of a schema
@@ -354,7 +373,7 @@ private[sql] class SQLConf extends Serializable with CatalystConf {
354373
private[spark] def orcFilterPushDown: Boolean = getConf(ORC_FILTER_PUSHDOWN_ENABLED)
355374

356375
/** When true uses verifyPartitionPath to prune the path which is not exists. */
357-
private[spark] def verifyPartitionPath = getConf(HIVE_VERIFY_PARTITIONPATH)
376+
private[spark] def verifyPartitionPath: Boolean = getConf(HIVE_VERIFY_PARTITIONPATH)
358377

359378
/** When true the planner will use the external sort, which may spill to disk. */
360379
private[spark] def externalSortEnabled: Boolean = getConf(EXTERNAL_SORT)
@@ -439,10 +458,10 @@ private[sql] class SQLConf extends Serializable with CatalystConf {
439458

440459
private[spark] def defaultDataSourceName: String = getConf(DEFAULT_DATA_SOURCE_NAME)
441460

442-
private[spark] def partitionDiscoveryEnabled() =
461+
private[spark] def partitionDiscoveryEnabled(): Boolean =
443462
getConf(SQLConf.PARTITION_DISCOVERY_ENABLED)
444463

445-
private[spark] def partitionColumnTypeInferenceEnabled() =
464+
private[spark] def partitionColumnTypeInferenceEnabled(): Boolean =
446465
getConf(SQLConf.PARTITION_COLUMN_TYPE_INFERENCE)
447466

448467
// Do not use a value larger than 4000 as the default value of this property.

0 commit comments

Comments
 (0)