|
17 | 17 |
|
18 | 18 | package org.apache.spark.deploy.yarn |
19 | 19 |
|
20 | | -import java.net.{InetAddress, UnknownHostException, URI} |
21 | | -import java.nio.ByteBuffer |
| 20 | +import java.net.{InetAddress, URI, UnknownHostException} |
22 | 21 |
|
23 | 22 | import scala.collection.JavaConversions._ |
24 | | -import scala.collection.mutable.HashMap |
25 | | -import scala.collection.mutable.Map |
| 23 | +import scala.collection.mutable.{HashMap, Map} |
26 | 24 |
|
27 | 25 | import org.apache.hadoop.conf.Configuration |
28 | 26 | import org.apache.hadoop.fs._ |
29 | | -import org.apache.hadoop.fs.permission.FsPermission; |
30 | | -import org.apache.hadoop.io.DataOutputBuffer |
| 27 | +import org.apache.hadoop.fs.permission.FsPermission |
31 | 28 | import org.apache.hadoop.mapred.Master |
32 | 29 | import org.apache.hadoop.mapreduce.MRJobConfig |
33 | | -import org.apache.hadoop.net.NetUtils |
34 | 30 | import org.apache.hadoop.security.UserGroupInformation |
35 | 31 | import org.apache.hadoop.util.StringUtils |
36 | 32 | import org.apache.hadoop.yarn.api._ |
37 | 33 | import org.apache.hadoop.yarn.api.ApplicationConstants.Environment |
38 | 34 | import org.apache.hadoop.yarn.api.protocolrecords._ |
39 | 35 | import org.apache.hadoop.yarn.api.records._ |
40 | 36 | import org.apache.hadoop.yarn.conf.YarnConfiguration |
41 | | -import org.apache.hadoop.yarn.ipc.YarnRPC |
42 | | -import org.apache.hadoop.yarn.util.{Records, Apps} |
43 | | - |
| 37 | +import org.apache.hadoop.yarn.util.{Apps, Records} |
44 | 38 | import org.apache.spark.{Logging, SparkConf} |
45 | | -import org.apache.spark.util.Utils |
46 | | -import org.apache.spark.deploy.SparkHadoopUtil |
47 | | -import org.apache.spark.deploy.ExecutorLauncher |
48 | | -import org.apache.hadoop.yarn.api.ApplicationConstants.Environment |
49 | | - |
50 | 39 |
|
51 | 40 | /** |
52 | 41 | * The entry point (starting in Client#main() and Client#run()) for launching Spark on YARN. The |
@@ -355,8 +344,8 @@ trait ClientBase extends Logging { |
355 | 344 | JAVA_OPTS += s"-D$k=$v" |
356 | 345 | } |
357 | 346 | // TODO: honor driver classpath here: sys.props.get("spark.driver.classPath") |
358 | | - sys.props.get("spark.driver.javaOpts").map(opts => JAVA_OPTS += opts) |
359 | | - sys.props.get("spark.driver.libraryPath").map(p => JAVA_OPTS + s"-Djava.library.path=$p") |
| 347 | + sys.props.get("spark.driver.javaOpts").foreach(opts => JAVA_OPTS += opts) |
| 348 | + sys.props.get("spark.driver.libraryPath").foreach(p => JAVA_OPTS += s"-Djava.library.path=$p") |
360 | 349 | } |
361 | 350 |
|
362 | 351 | if (!localResources.contains(ClientBase.LOG4J_PROP)) { |
|
0 commit comments