@@ -37,7 +37,7 @@ import org.apache.spark.sql.catalyst.analysis.{ResolvedStar, UnresolvedAttribute
3737import org .apache .spark .sql .catalyst .expressions ._
3838import org .apache .spark .sql .catalyst .plans .logical .{Filter , _ }
3939import org .apache .spark .sql .catalyst .plans .{Inner , JoinType }
40- import org .apache .spark .sql .catalyst .{CatalystTypeConverters , ScalaReflection , SqlParser }
40+ import org .apache .spark .sql .catalyst .{expressions , CatalystTypeConverters , ScalaReflection , SqlParser }
4141import org .apache .spark .sql .execution .{EvaluatePython , ExplainCommand , LogicalRDD }
4242import org .apache .spark .sql .jdbc .JDBCWriteDetails
4343import org .apache .spark .sql .json .JacksonGenerator
@@ -400,7 +400,9 @@ class DataFrame private[sql](
400400 joined.left,
401401 joined.right,
402402 joinType = Inner ,
403- Some (EqualTo (joined.left.resolve(usingColumn), joined.right.resolve(usingColumn))))
403+ Some (expressions.EqualTo (
404+ joined.left.resolve(usingColumn),
405+ joined.right.resolve(usingColumn))))
404406 )
405407 }
406408
@@ -1343,7 +1345,7 @@ class DataFrame private[sql](
13431345 mode : SaveMode ,
13441346 options : java.util.Map [String , String ],
13451347 partitionColumns : java.util.List [String ]): Unit = {
1346- ???
1348+ saveAsTable(tableName, source, mode, options.toMap, partitionColumns)
13471349 }
13481350
13491351 /**
@@ -1399,7 +1401,15 @@ class DataFrame private[sql](
13991401 mode : SaveMode ,
14001402 options : Map [String , String ],
14011403 partitionColumns : Seq [String ]): Unit = {
1402- ???
1404+ sqlContext.executePlan(
1405+ CreateTableUsingAsSelect (
1406+ tableName,
1407+ source,
1408+ temporary = false ,
1409+ partitionColumns.toArray,
1410+ mode,
1411+ options,
1412+ logicalPlan)).toRdd
14031413 }
14041414
14051415 /**
0 commit comments