Skip to content

Commit 795a763

Browse files
committed
Change name to WriteInputFormatTestDataGenerator. Cleanup some var names. Use SPARK_HOME in path for writing test sequencefile data.
1 parent 174f520 commit 795a763

File tree

1 file changed

+16
-16
lines changed

1 file changed

+16
-16
lines changed

core/src/main/scala/org/apache/spark/api/python/WriteInputFormatTests.scala renamed to core/src/main/scala/org/apache/spark/api/python/WriteInputFormatTestDataGenerator.scala

Lines changed: 16 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -5,32 +5,33 @@ import org.apache.hadoop.io._
55
import scala.Array
66
import java.io.{DataOutput, DataInput}
77

8-
case class TestWritable(var str: String, var numi: Int, var numd: Double) extends Writable {
8+
case class TestWritable(var str: String, var int: Int, var double: Double) extends Writable {
99
def this() = this("", 0, 0.0)
1010

11-
def write(p1: DataOutput) = {
12-
p1.writeUTF(str)
13-
p1.writeInt(numi)
14-
p1.writeDouble(numd)
11+
def write(out: DataOutput) = {
12+
out.writeUTF(str)
13+
out.writeInt(int)
14+
out.writeDouble(double)
1515
}
1616

17-
def readFields(p1: DataInput) = {
18-
str = p1.readUTF()
19-
numi = p1.readInt()
20-
numd = p1.readDouble()
17+
def readFields(in: DataInput) = {
18+
str = in.readUTF()
19+
int = in.readInt()
20+
double = in.readDouble()
2121
}
2222
}
2323

24-
object WriteInputFormatTests extends App {
24+
object WriteInputFormatTestDataGenerator extends App {
2525
import SparkContext._
2626

2727
val sc = new SparkContext("local[2]", "test")
2828

29-
val textPath = "../python/test_support/data/sftext/"
30-
val intPath = "../python/test_support/data/sfint/"
31-
val doublePath = "../python/test_support/data/sfdouble/"
32-
val arrPath = "../python/test_support/data/sfarray/"
33-
val classPath = "../python/test_support/data/sfclass/"
29+
val sparkHome = sys.env.get("SPARK_HOME").orElse(sys.props.get("spark.home")).get
30+
val textPath = s"$sparkHome/python/test_support/data/sftext/"
31+
val intPath = s"$sparkHome/python/test_support/data/sfint/"
32+
val doublePath = s"$sparkHome/python/test_support/data/sfdouble/"
33+
val arrPath = s"$sparkHome/python/test_support/data/sfarray/"
34+
val classPath = s"$sparkHome/python/test_support/data/sfclass/"
3435

3536
val intKeys = Seq((1.0, "aa"), (2.0, "bb"), (2.0, "aa"), (3.0, "cc"), (2.0, "bb"), (1.0, "aa"))
3637
sc.parallelize(intKeys).saveAsSequenceFile(intPath)
@@ -60,5 +61,4 @@ object WriteInputFormatTests extends App {
6061
classOf[Text], classOf[TestWritable],
6162
classOf[org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat[Text, TestWritable]])
6263

63-
6464
}

0 commit comments

Comments
 (0)