@@ -57,7 +57,7 @@ object SparkHadoopWriter extends Logging {
5757 */
5858 def write [K , V : ClassTag ](
5959 rdd : RDD [(K , V )],
60- config : SparkHadoopWriterConfig [K , V ]): Unit = {
60+ config : HadoopWriteConfigUtil [K , V ]): Unit = {
6161 // Extract context and configuration from RDD.
6262 val sparkContext = rdd.context
6363 val stageId = rdd.id
@@ -119,7 +119,7 @@ object SparkHadoopWriter extends Logging {
119119 /** Write a RDD partition out in a single Spark task. */
120120 private def executeTask [K , V : ClassTag ](
121121 context : TaskContext ,
122- config : SparkHadoopWriterConfig [K , V ],
122+ config : HadoopWriteConfigUtil [K , V ],
123123 jobTrackerId : String ,
124124 sparkStageId : Int ,
125125 sparkPartitionId : Int ,
@@ -175,8 +175,8 @@ object SparkHadoopWriter extends Logging {
175175 * A helper class that reads JobConf from older mapred API, creates output Format/Committer/Writer.
176176 */
177177private [spark]
178- class SparkHadoopMapRedWriterConfig [K , V : ClassTag ](conf : SerializableJobConf )
179- extends SparkHadoopWriterConfig [K , V ] with Logging {
178+ class HadoopMapRedWriteConfigUtil [K , V : ClassTag ](conf : SerializableJobConf )
179+ extends HadoopWriteConfigUtil [K , V ] with Logging {
180180
181181 private var outputFormat : Class [_ <: OutputFormat [K , V ]] = null
182182 private var writer : RecordWriter [K , V ] = null
@@ -308,8 +308,8 @@ class SparkHadoopMapRedWriterConfig[K, V: ClassTag](conf: SerializableJobConf)
308308 * Format/Committer/Writer.
309309 */
310310private [spark]
311- class SparkHadoopMapReduceWriterConfig [K , V : ClassTag ](conf : SerializableConfiguration )
312- extends SparkHadoopWriterConfig [K , V ] with Logging {
311+ class HadoopMapReduceWriteConfigUtil [K , V : ClassTag ](conf : SerializableConfiguration )
312+ extends HadoopWriteConfigUtil [K , V ] with Logging {
313313
314314 private var outputFormat : Class [_ <: NewOutputFormat [K , V ]] = null
315315 private var writer : NewRecordWriter [K , V ] = null
0 commit comments