-
Notifications
You must be signed in to change notification settings - Fork 28.9k
[SPARK-11998] [SQL] [test-hadoop2.0] When downloading Hadoop artifacts from maven, we need to try to download the version that is used by Spark #9979
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@marmbrus This should be a val, right?
|
cc @marmbrus @JoshRosen for review. |
|
what is the pic? |
|
test this please |
|
Test build #46716 has finished for PR 9979 at commit
|
|
Test build #46717 has finished for PR 9979 at commit
|
|
https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/46722/consoleFull is the hadoop 2.2 run. |
|
LGTM pending tests, per our offline discussion; this seems fine given that this auto-downloading of Hive classes isn't the recommended approach for production deployments, so the risks here don't seem huge. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't think you need this. It's a partial function
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Done
|
Test build #46722 has finished for PR 9979 at commit
|
…p that matches the Hadoop used by Spark. If the Hadoop artifact cannot be resolved (e.g. Hadoop version is a vendor specific version like 2.0.0-cdh4.1.1), we will use Hadoop 2.4.0 (we used to hard code this version as the hadoop that we will download from maven) and we will not share Hadoop classes.
|
https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/46778/consoleFull is for hadoop 2.0 test. |
|
Test build #46778 has finished for PR 9979 at commit
|
|
OK. I am merging this to master and branch 1.6. I will watch the builds and see if there is any new issues. |
|
hmm... not sure why https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/46778/consoleFull used hadoop 2.3 profile. |
|
Looks like there is a typo at https://github.com/apache/spark/blob/master/dev/run-tests-jenkins.py#L167-L170 |
…from maven, we need to try to download the version that is used by Spark If we need to download Hive/Hadoop artifacts, try to download a Hadoop that matches the Hadoop used by Spark. If the Hadoop artifact cannot be resolved (e.g. Hadoop version is a vendor specific version like 2.0.0-cdh4.1.1), we will use Hadoop 2.4.0 (we used to hard code this version as the hadoop that we will download from maven) and we will not share Hadoop classes. I tested this match in my laptop with the following confs (these confs are used by our builds). All tests are good. ``` build/sbt -Phadoop-1 -Dhadoop.version=1.2.1 -Pkinesis-asl -Phive-thriftserver -Phive build/sbt -Phadoop-1 -Dhadoop.version=2.0.0-mr1-cdh4.1.1 -Pkinesis-asl -Phive-thriftserver -Phive build/sbt -Pyarn -Phadoop-2.2 -Pkinesis-asl -Phive-thriftserver -Phive build/sbt -Pyarn -Phadoop-2.3 -Dhadoop.version=2.3.0 -Pkinesis-asl -Phive-thriftserver -Phive ``` Author: Yin Huai <[email protected]> Closes #9979 from yhuai/versionsSuite. (cherry picked from commit ad76562) Signed-off-by: Yin Huai <[email protected]>
If we need to download Hive/Hadoop artifacts, try to download a Hadoop that matches the Hadoop used by Spark. If the Hadoop artifact cannot be resolved (e.g. Hadoop version is a vendor specific version like 2.0.0-cdh4.1.1), we will use Hadoop 2.4.0 (we used to hard code this version as the hadoop that we will download from maven) and we will not share Hadoop classes.
I tested this match in my laptop with the following confs (these confs are used by our builds). All tests are good.