From f628b4dc9aef3a6c17cc4b478a491ed67082282c Mon Sep 17 00:00:00 2001 From: Michael Armbrust Date: Fri, 18 Apr 2014 19:35:27 -0700 Subject: [PATCH] Use scala deprecation instead of java. --- .../main/scala/org/apache/spark/api/java/JavaSparkContext.scala | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala b/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala index cf30523ab523e..bda9272b43393 100644 --- a/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala +++ b/core/src/main/scala/org/apache/spark/api/java/JavaSparkContext.scala @@ -114,7 +114,7 @@ class JavaSparkContext(val sc: SparkContext) extends JavaSparkContextVarargsWork * @deprecated As of Spark 1.0.0, defaultMinSplits is deprecated, use * {@link #defaultMinPartitions()} instead */ - @Deprecated + @deprecated("use defaultMinPartitions", "1.0.0") def defaultMinSplits: java.lang.Integer = sc.defaultMinSplits /** Default min number of partitions for Hadoop RDDs when not given by user */