Skip to content

Commit e51b9ac

Browse files
committed
renamed ExtractPythonUdfs to ExtractPythonUDFs
1 parent 8c756f1 commit e51b9ac

File tree

3 files changed

+4
-4
lines changed

3 files changed

+4
-4
lines changed

sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -146,7 +146,7 @@ class SQLContext(@transient val sparkContext: SparkContext)
146146
protected[sql] lazy val analyzer: Analyzer =
147147
new Analyzer(catalog, functionRegistry, conf) {
148148
override val extendedResolutionRules =
149-
ExtractPythonUdfs ::
149+
ExtractPythonUDFs ::
150150
sources.PreInsertCastAndRename ::
151151
Nil
152152

sql/core/src/main/scala/org/apache/spark/sql/execution/pythonUDFs.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@ private[spark] case class PythonUDF(
6969
* This has the limitation that the input to the Python UDF is not allowed include attributes from
7070
* multiple child operators.
7171
*/
72-
private[spark] object ExtractPythonUdfs extends Rule[LogicalPlan] {
72+
private[spark] object ExtractPythonUDFs extends Rule[LogicalPlan] {
7373
def apply(plan: LogicalPlan): LogicalPlan = plan transform {
7474
// Skip EvaluatePython nodes.
7575
case plan: EvaluatePython => plan

sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ import org.apache.spark.sql.SQLConf.SQLConfEntry._
4242
import org.apache.spark.sql.catalyst.ParserDialect
4343
import org.apache.spark.sql.catalyst.analysis._
4444
import org.apache.spark.sql.catalyst.plans.logical._
45-
import org.apache.spark.sql.execution.{ExecutedCommand, ExtractPythonUdfs, SetCommand}
45+
import org.apache.spark.sql.execution.{ExecutedCommand, ExtractPythonUDFs, SetCommand}
4646
import org.apache.spark.sql.hive.client._
4747
import org.apache.spark.sql.hive.execution.{DescribeHiveTableCommand, HiveNativeCommand}
4848
import org.apache.spark.sql.sources.DataSourceStrategy
@@ -381,7 +381,7 @@ class HiveContext(sc: SparkContext) extends SQLContext(sc) {
381381
catalog.ParquetConversions ::
382382
catalog.CreateTables ::
383383
catalog.PreInsertionCasts ::
384-
ExtractPythonUdfs ::
384+
ExtractPythonUDFs ::
385385
ResolveHiveWindowFunction ::
386386
sources.PreInsertCastAndRename ::
387387
Nil

0 commit comments

Comments
 (0)