@@ -146,35 +146,44 @@ private[spark] object SQLConf {
146146 defaultValue = Some (true ),
147147 doc = " When set to true Spark SQL will automatically select a compression codec for each " +
148148 " column based on statistics of the data." )
149+
149150 val COLUMN_BATCH_SIZE = intConf(" spark.sql.inMemoryColumnarStorage.batchSize" ,
150151 defaultValue = Some (10000 ),
151152 doc = " Controls the size of batches for columnar caching. Larger batch sizes can improve " +
152153 " memory utilization and compression, but risk OOMs when caching data." )
154+
153155 val IN_MEMORY_PARTITION_PRUNING =
154156 booleanConf(" spark.sql.inMemoryColumnarStorage.partitionPruning" ,
155157 defaultValue = Some (false ),
156158 doc = " <TODO>" )
159+
157160 val AUTO_BROADCASTJOIN_THRESHOLD = intConf(" spark.sql.autoBroadcastJoinThreshold" ,
158161 defaultValue = Some (10 * 1024 * 1024 ),
159162 doc = " Configures the maximum size in bytes for a table that will be broadcast to all worker " +
160163 " nodes when performing a join. By setting this value to -1 broadcasting can be disabled. " +
161164 " Note that currently statistics are only supported for Hive Metastore tables where the " +
162165 " command<code>ANALYZE TABLE <tableName> COMPUTE STATISTICS noscan</code> has been run." )
166+
163167 val DEFAULT_SIZE_IN_BYTES = longConf(" spark.sql.defaultSizeInBytes" , isPublic = false )
168+
164169 val SHUFFLE_PARTITIONS = intConf(" spark.sql.shuffle.partitions" ,
165170 defaultValue = Some (200 ),
166171 doc = " Configures the number of partitions to use when shuffling data for joins or " +
167172 " aggregations." )
173+
168174 val CODEGEN_ENABLED = booleanConf(" spark.sql.codegen" ,
169175 defaultValue = Some (false ),
170176 doc = " When true, code will be dynamically generated at runtime for expression evaluation in" +
171177 " a specific query. For some queries with complicated expression this option can lead to " +
172178 " significant speed-ups. However, for simple queries this can actually slow down query " +
173179 " execution." )
180+
174181 val UNSAFE_ENABLED = booleanConf(" spark.sql.unsafe.enabled" ,
175182 defaultValue = Some (false ),
176183 doc = " <TDDO>" )
184+
177185 val DIALECT = stringConf(" spark.sql.dialect" , defaultValue = Some (" sql" ), doc = " <TODO>" )
186+
178187 val CASE_SENSITIVE = booleanConf(" spark.sql.caseSensitive" ,
179188 defaultValue = Some (true ),
180189 doc = " <TODO>" )
@@ -185,26 +194,31 @@ private[spark] object SQLConf {
185194 " Spark SQL, do not differentiate between binary data and strings when writing out the " +
186195 " Parquet schema. This flag tells Spark SQL to interpret binary data as a string to provide " +
187196 " compatibility with these systems." )
197+
188198 val PARQUET_INT96_AS_TIMESTAMP = booleanConf(" spark.sql.parquet.int96AsTimestamp" ,
189199 defaultValue = Some (true ),
190200 doc = " Some Parquet-producing systems, in particular Impala, store Timestamp into INT96. " +
191201 " Spark would also store Timestamp as INT96 because we need to avoid precision lost of the " +
192202 " nanoseconds field. This flag tells Spark SQL to interpret INT96 data as a timestamp to " +
193203 " provide compatibility with these systems." )
204+
194205 val PARQUET_CACHE_METADATA = booleanConf(" spark.sql.parquet.cacheMetadata" ,
195206 defaultValue = Some (true ),
196207 doc = " Turns on caching of Parquet schema metadata. Can speed up querying of static data." )
208+
197209 val PARQUET_COMPRESSION = stringConf(" spark.sql.parquet.compression.codec" ,
198210 defaultValue = Some (" gzip" ),
199211 doc = " Sets the compression codec use when writing Parquet files. Acceptable values include: " +
200212 " uncompressed, snappy, gzip, lzo." )
213+
201214 val PARQUET_FILTER_PUSHDOWN_ENABLED = booleanConf(" spark.sql.parquet.filterPushdown" ,
202215 defaultValue = Some (false ),
203216 doc = " Turn on Parquet filter pushdown optimization. This feature is turned off by default" +
204217 " because of a known bug in Paruet 1.6.0rc3 " +
205218 " (<a href=\" https://issues.apache.org/jira/browse/PARQUET-136\" >PARQUET-136</a>). However, " +
206219 " if your table doesn't contain any nullable string or binary columns, it's still safe to " +
207220 " turn this feature on." )
221+
208222 val PARQUET_USE_DATA_SOURCE_API = booleanConf(" spark.sql.parquet.useDataSourceApi" ,
209223 defaultValue = Some (true ),
210224 doc = " <TODO>" )
@@ -220,6 +234,7 @@ private[spark] object SQLConf {
220234 val COLUMN_NAME_OF_CORRUPT_RECORD = stringConf(" spark.sql.columnNameOfCorruptRecord" ,
221235 defaultValue = Some (" _corrupt_record" ),
222236 doc = " <TODO>" )
237+
223238 val BROADCAST_TIMEOUT = intConf(" spark.sql.broadcastTimeout" ,
224239 defaultValue = Some (5 * 60 ),
225240 doc = " <TODO>" )
@@ -230,21 +245,25 @@ private[spark] object SQLConf {
230245 defaultValue = Some (false ),
231246 doc = " When true, performs sorts spilling to disk as needed otherwise sort each partition in" +
232247 " memory." )
248+
233249 val SORTMERGE_JOIN = booleanConf(" spark.sql.planner.sortMergeJoin" ,
234250 defaultValue = Some (false ),
235251 doc = " <TODO>" )
236252
237253 // This is only used for the thriftserver
238254 val THRIFTSERVER_POOL = stringConf(" spark.sql.thriftserver.scheduler.pool" , isPublic = false )
255+
239256 val THRIFTSERVER_UI_STATEMENT_LIMIT =
240257 intConf(" spark.sql.thriftserver.ui.retainedStatements" , isPublic = false )
258+
241259 val THRIFTSERVER_UI_SESSION_LIMIT =
242260 intConf(" spark.sql.thriftserver.ui.retainedSessions" , isPublic = false )
243261
244262 // This is used to set the default data source
245263 val DEFAULT_DATA_SOURCE_NAME = stringConf(" spark.sql.sources.default" ,
246264 defaultValue = Some (" org.apache.spark.sql.parquet" ),
247265 doc = " <TODO>" )
266+
248267 // This is used to control the when we will split a schema's JSON string to multiple pieces
249268 // in order to fit the JSON string in metastore's table property (by default, the value has
250269 // a length restriction of 4000 characters). We will split the JSON string of a schema
@@ -354,7 +373,7 @@ private[sql] class SQLConf extends Serializable with CatalystConf {
354373 private [spark] def orcFilterPushDown : Boolean = getConf(ORC_FILTER_PUSHDOWN_ENABLED )
355374
356375 /** When true uses verifyPartitionPath to prune the path which is not exists. */
357- private [spark] def verifyPartitionPath = getConf(HIVE_VERIFY_PARTITIONPATH )
376+ private [spark] def verifyPartitionPath : Boolean = getConf(HIVE_VERIFY_PARTITIONPATH )
358377
359378 /** When true the planner will use the external sort, which may spill to disk. */
360379 private [spark] def externalSortEnabled : Boolean = getConf(EXTERNAL_SORT )
@@ -439,10 +458,10 @@ private[sql] class SQLConf extends Serializable with CatalystConf {
439458
440459 private [spark] def defaultDataSourceName : String = getConf(DEFAULT_DATA_SOURCE_NAME )
441460
442- private [spark] def partitionDiscoveryEnabled () =
461+ private [spark] def partitionDiscoveryEnabled (): Boolean =
443462 getConf(SQLConf .PARTITION_DISCOVERY_ENABLED )
444463
445- private [spark] def partitionColumnTypeInferenceEnabled () =
464+ private [spark] def partitionColumnTypeInferenceEnabled (): Boolean =
446465 getConf(SQLConf .PARTITION_COLUMN_TYPE_INFERENCE )
447466
448467 // Do not use a value larger than 4000 as the default value of this property.
0 commit comments