@@ -138,8 +138,8 @@ private[spark] object PythonRDD extends Logging {
138138 * or to enable local execution.
139139 *
140140 * @return 3-tuple (as a Java array) with the port number of a local socket which serves the
141- * data collected from this job, the secret for authentication, and a server object
142- * that can be used to sync the JVM serving thread in Python.
141+ * data collected from this job, the secret for authentication, and a socket auth
142+ * server object that can be used to join the JVM serving thread in Python.
143143 */
144144 def runJob (
145145 sc : SparkContext ,
@@ -158,8 +158,8 @@ private[spark] object PythonRDD extends Logging {
158158 * A helper function to collect an RDD as an iterator, then serve it via socket.
159159 *
160160 * @return 3-tuple (as a Java array) with the port number of a local socket which serves the
161- * data collected from this job, the secret for authentication, and a server object
162- * that can be used to sync the JVM serving thread in Python.
161+ * data collected from this job, the secret for authentication, and a socket auth
162+ * server object that can be used to join the JVM serving thread in Python.
163163 */
164164 def collectAndServe [T ](rdd : RDD [T ]): Array [Any ] = {
165165 serveIterator(rdd.collect().iterator, s " serve RDD ${rdd.id}" )
@@ -170,12 +170,12 @@ private[spark] object PythonRDD extends Logging {
170170 * are collected as separate jobs, by order of index. Partition data is first requested by a
171171 * non-zero integer to start a collection job. The response is prefaced by an integer with 1
172172 * meaning partition data will be served, 0 meaning the local iterator has been consumed,
173- * and -1 meaining an error occurred during collection. This function is used by
173+ * and -1 meaning an error occurred during collection. This function is used by
174174 * pyspark.rdd._local_iterator_from_socket().
175175 *
176176 * @return 3-tuple (as a Java array) with the port number of a local socket which serves the
177- * data collected from this job, the secret for authentication, and a server object
178- * that can be used to sync the JVM serving thread in Python.
177+ * data collected from this job, the secret for authentication, and a socket auth
178+ * server object that can be used to join the JVM serving thread in Python.
179179 */
180180 def toLocalIteratorAndServe [T ](rdd : RDD [T ]): Array [Any ] = {
181181 val handleFunc = (sock : Socket ) => {
@@ -447,8 +447,8 @@ private[spark] object PythonRDD extends Logging {
447447 * The thread will terminate after all the data are sent or any exceptions happen.
448448 *
449449 * @return 3-tuple (as a Java array) with the port number of a local socket which serves the
450- * data collected from this job, the secret for authentication, and a server object
451- * that can be used to sync the JVM serving thread in Python.
450+ * data collected from this job, the secret for authentication, and a socket auth
451+ * server object that can be used to join the JVM serving thread in Python.
452452 */
453453 def serveIterator (items : Iterator [_], threadName : String ): Array [Any ] = {
454454 serveToStream(threadName) { out =>
@@ -470,8 +470,8 @@ private[spark] object PythonRDD extends Logging {
470470 * exceptions happen.
471471 *
472472 * @return 3-tuple (as a Java array) with the port number of a local socket which serves the
473- * data collected from this job, the secret for authentication, and a server object
474- * that can be used to sync the JVM serving thread in Python.
473+ * data collected from this job, the secret for authentication, and a socket auth
474+ * server object that can be used to join the JVM serving thread in Python.
475475 */
476476 private [spark] def serveToStream (
477477 threadName : String )(writeFunc : OutputStream => Unit ): Array [Any ] = {
0 commit comments