File tree Expand file tree Collapse file tree 2 files changed +0
-18
lines changed
core/src/main/scala/org/apache/spark/internal/config Expand file tree Collapse file tree 2 files changed +0
-18
lines changed Original file line number Diff line number Diff line change @@ -253,11 +253,6 @@ package object config {
253253 private [spark] val EXECUTOR_ID =
254254 ConfigBuilder (" spark.executor.id" ).version(" 1.2.0" ).stringConf.createOptional
255255
256- private [spark] val EXECUTOR_BIND_ADDRESS = ConfigBuilder (" spark.executor.bindAddress" )
257- .doc(" Address where to bind network listen sockets on the executor." )
258- .stringConf
259- .createWithDefault(Utils .localHostName())
260-
261256 private [spark] val EXECUTOR_CLASS_PATH =
262257 ConfigBuilder (SparkLauncher .EXECUTOR_EXTRA_CLASSPATH )
263258 .version(" 1.0.0" )
Original file line number Diff line number Diff line change @@ -349,19 +349,6 @@ of the most common options to set are:
349349 </td >
350350 <td >3.0.0</td >
351351</tr >
352- <tr >
353- <td ><code >spark.executor.bindAddress</code ></td >
354- <td >(local hostname)</td >
355- <td >
356- Hostname or IP address where to bind listening sockets. This config overrides the SPARK_LOCAL_IP
357- environment variable (see below).
358- <br />It also allows a different address from the local one to be advertised to other
359- executors or external systems. This is useful, for example, when running containers with bridged networking.
360- For this to properly work, the different ports used by the driver (RPC, block manager and UI) need to be
361- forwarded from the container's host.
362- </td >
363- <td >4.0.0</td >
364- </tr >
365352<tr >
366353 <td ><code >spark.extraListeners</code ></td >
367354 <td >(none)</td >
You can’t perform that action at this time.
0 commit comments