From cb144a102f9be1a93c8500afca8513bf01451363 Mon Sep 17 00:00:00 2001 From: yangjie01 Date: Mon, 13 Nov 2023 14:58:27 +0800 Subject: [PATCH] init --- .../execution/datasources/DataSource.scala | 25 +------------------ 1 file changed, 1 insertion(+), 24 deletions(-) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala index b3784dbf8137..cd295f3b17bd 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala @@ -29,9 +29,8 @@ import org.apache.spark.SparkException import org.apache.spark.deploy.SparkHadoopUtil import org.apache.spark.internal.Logging import org.apache.spark.sql._ -import org.apache.spark.sql.catalyst.analysis.{Resolver, UnresolvedAttribute} +import org.apache.spark.sql.catalyst.analysis.UnresolvedAttribute import org.apache.spark.sql.catalyst.catalog.{BucketSpec, CatalogStorageFormat, CatalogTable, CatalogUtils} -import org.apache.spark.sql.catalyst.expressions.Attribute import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan import org.apache.spark.sql.catalyst.util.{CaseInsensitiveMap, TypeUtils} import org.apache.spark.sql.connector.catalog.TableProvider @@ -822,26 +821,4 @@ object DataSource extends Logging { throw QueryCompilationErrors.writeEmptySchemasUnsupportedByDataSourceError() } } - - /** - * Resolve partition columns using output columns of the query plan. - */ - def resolvePartitionColumns( - partitionColumns: Seq[Attribute], - outputColumns: Seq[Attribute], - plan: LogicalPlan, - resolver: Resolver): Seq[Attribute] = { - partitionColumns.map { col => - // The partition columns created in `planForWritingFileFormat` should always be - // `UnresolvedAttribute` with a single name part. - assert(col.isInstanceOf[UnresolvedAttribute]) - val unresolved = col.asInstanceOf[UnresolvedAttribute] - assert(unresolved.nameParts.length == 1) - val name = unresolved.nameParts.head - outputColumns.find(a => resolver(a.name, name)).getOrElse { - throw QueryCompilationErrors.cannotResolveAttributeError( - name, plan.output.map(_.name).mkString(", ")) - } - } - } }