-
Notifications
You must be signed in to change notification settings - Fork 28.9k
[SPARK-11340][SPARKR] Support setting driver properties when starting Spark from R programmatically or from RStudio #9290
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
I checked, the user could also set SPARK_DRIVER_MEMORY before running |
|
Test build #44405 has finished for PR 9290 at commit
|
|
Test build #44409 has finished for PR 9290 at commit
|
R/pkg/R/sparkR.R
Outdated
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Lets add the support for other options of the same limitation, according to http://spark.apache.org/docs/latest/configuration.html, we also have:
spark.driver.extraClassPath
spark.driver.extraJavaOptions
spark.driver.extraLibraryPath
…aLibraryPath from feedback add quote " around parameter values
|
Test build #44490 has finished for PR 9290 at commit
|
|
Test build #44494 has finished for PR 9290 at commit
|
|
Test build #44509 has finished for PR 9290 at commit
|
R/pkg/R/sparkR.R
Outdated
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
wording:
A few Spark properties can not take effect after the driver JVM has started, as documented in
http://spark.apache.org/docs/latest/configuration.html#application-properties
When starting SparkR without using spark-submit, for example, in Rstudio, add them to spark-submit commandline if not already set in SPARKR_SUBMIT_ARGS so that they can take effect when launching the JVM.
|
code looks good. Could you update the SparkR documentation to document this support? |
|
Test build #44547 has finished for PR 9290 at commit
|
|
Test build #44552 has finished for PR 9290 at commit
|
R/pkg/R/sparkR.R
Outdated
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Adding example here may imply that "spark.driver.memory" always works. Maybe move this line to be near with the code change?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I could take this out - this is more for API doc/roxygen. If we are moving it out it would be visible only when someone is reading the code. Maybe pointing to the SparkR programming guide is better.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah we should take this out of here. I think we should add a new section in the Programming guide called 'Running SparkR from RStudio' and include this option there. Here we could just have a link to the latest programming guide with something like 'For more options on how to initialize and use SparkR see our programming guide'
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
updated. created JIRA for additional programming guide change: https://issues.apache.org/jira/browse/SPARK-11407
I could take a shot at the doc change if you'd like
|
Test build #44624 has finished for PR 9290 at commit
|
|
LGTM. Thanks @felixcheung -- Merging this. |


Mapping spark.driver.memory from sparkEnvir to spark-submit commandline arguments.
@shivaram suggested that we possibly add other spark.driver.* properties - do we want to add all of those? I thought those could be set in SparkConf?
@sun-rui