|
17 | 17 |
|
18 | 18 | package org.apache.spark.sql |
19 | 19 |
|
20 | | -import org.apache.spark.sql.catalyst.expressions.NamedExpression |
| 20 | +import org.apache.hadoop.io.{LongWritable, Text} |
| 21 | +import org.apache.hadoop.mapreduce.lib.input.{TextInputFormat => NewTextInputFormat} |
21 | 22 | import org.scalatest.Matchers._ |
22 | 23 |
|
| 24 | +import org.apache.spark.sql.catalyst.expressions.NamedExpression |
23 | 25 | import org.apache.spark.sql.execution.Project |
24 | 26 | import org.apache.spark.sql.functions._ |
25 | 27 | import org.apache.spark.sql.test.SharedSQLContext |
@@ -591,15 +593,44 @@ class ColumnExpressionSuite extends QueryTest with SharedSQLContext { |
591 | 593 | ) |
592 | 594 | } |
593 | 595 |
|
594 | | - test("InputFileName") { |
| 596 | + test("InputFileName - SqlNewHadoopRDD") { |
595 | 597 | withTempPath { dir => |
596 | 598 | val data = sparkContext.parallelize(0 to 10).toDF("id") |
597 | 599 | data.write.parquet(dir.getCanonicalPath) |
598 | | - val answer = sqlContext.read.parquet(dir.getCanonicalPath).select(inputFileName()) |
| 600 | + val answer = sqlContext.read.parquet(dir.getCanonicalPath).select(input_file_name()) |
599 | 601 | .head.getString(0) |
600 | 602 | assert(answer.contains(dir.getCanonicalPath)) |
601 | 603 |
|
602 | | - checkAnswer(data.select(inputFileName()).limit(1), Row("")) |
| 604 | + checkAnswer(data.select(input_file_name()).limit(1), Row("")) |
| 605 | + } |
| 606 | + } |
| 607 | + |
| 608 | + test("input_file_name - HadoopRDD") { |
| 609 | + withTempPath { dir => |
| 610 | + val data = sparkContext.parallelize((0 to 10).map(_.toString)).toDF() |
| 611 | + data.write.text(dir.getCanonicalPath) |
| 612 | + val df = sparkContext.textFile(dir.getCanonicalPath).toDF() |
| 613 | + val answer = df.select(input_file_name()).head.getString(0) |
| 614 | + assert(answer.contains(dir.getCanonicalPath)) |
| 615 | + |
| 616 | + checkAnswer(data.select(input_file_name()).limit(1), Row("")) |
| 617 | + } |
| 618 | + } |
| 619 | + |
| 620 | + test("input_file_name - NewHadoopRDD") { |
| 621 | + withTempPath { dir => |
| 622 | + val data = sparkContext.parallelize((0 to 10).map(_.toString)).toDF() |
| 623 | + data.write.text(dir.getCanonicalPath) |
| 624 | + val rdd = sparkContext.newAPIHadoopFile( |
| 625 | + dir.getCanonicalPath, |
| 626 | + classOf[NewTextInputFormat], |
| 627 | + classOf[LongWritable], |
| 628 | + classOf[Text]) |
| 629 | + val df = rdd.map(pair => pair._2.toString).toDF() |
| 630 | + val answer = df.select(input_file_name()).head.getString(0) |
| 631 | + assert(answer.contains(dir.getCanonicalPath)) |
| 632 | + |
| 633 | + checkAnswer(data.select(input_file_name()).limit(1), Row("")) |
603 | 634 | } |
604 | 635 | } |
605 | 636 |
|
|
0 commit comments