Skip to content

Commit 42c1f09

Browse files
committed
Using better labels
1 parent 9d48cbf commit 42c1f09

20 files changed

+48
-48
lines changed

core/src/main/scala/org/apache/spark/Aggregator.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ package org.apache.spark
2020
import org.apache.spark.util.collection.{AppendOnlyMap, ExternalAppendOnlyMap}
2121

2222
/**
23-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
23+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
2424
*
2525
* A set of functions used to aggregate data.
2626
*

core/src/main/scala/org/apache/spark/Dependency.scala

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -21,15 +21,15 @@ import org.apache.spark.rdd.RDD
2121
import org.apache.spark.serializer.Serializer
2222

2323
/**
24-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
24+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
2525
*
2626
* Base class for dependencies.
2727
*/
2828
abstract class Dependency[T](val rdd: RDD[T]) extends Serializable
2929

3030

3131
/**
32-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
32+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
3333
*
3434
* Base class for dependencies where each partition of the parent RDD is used by at most one
3535
* partition of the child RDD. Narrow dependencies allow for pipelined execution.
@@ -45,7 +45,7 @@ abstract class NarrowDependency[T](rdd: RDD[T]) extends Dependency(rdd) {
4545

4646

4747
/**
48-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
48+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
4949
*
5050
* Represents a dependency on the output of a shuffle stage.
5151
* @param rdd the parent RDD
@@ -65,7 +65,7 @@ class ShuffleDependency[K, V](
6565

6666

6767
/**
68-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
68+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
6969
*
7070
* Represents a one-to-one dependency between partitions of the parent and child RDDs.
7171
*/
@@ -75,7 +75,7 @@ class OneToOneDependency[T](rdd: RDD[T]) extends NarrowDependency[T](rdd) {
7575

7676

7777
/**
78-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
78+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
7979
*
8080
* Represents a one-to-one dependency between ranges of partitions in the parent and child RDDs.
8181
* @param rdd the parent RDD

core/src/main/scala/org/apache/spark/FutureAction.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ import org.apache.spark.rdd.RDD
2525
import org.apache.spark.scheduler.{JobFailed, JobSucceeded, JobWaiter}
2626

2727
/**
28-
* <span class="badge badge-red" style="float: right;">EXPERIMENTAL</span>
28+
* <span class="badge badge-red" style="float: right;">EXPERIMENTAL API</span>
2929
*
3030
* A future for the result of an action to support cancellation. This is an extension of the
3131
* Scala Future interface to support cancellation.
@@ -150,7 +150,7 @@ class SimpleFutureAction[T] private[spark](jobWaiter: JobWaiter[_], resultFunc:
150150

151151

152152
/**
153-
* <span class="badge badge-red" style="float: right;">EXPERIMENTAL</span>
153+
* <span class="badge badge-red" style="float: right;">EXPERIMENTAL API</span>
154154
*
155155
* A [[FutureAction]] for actions that could trigger multiple Spark jobs. Examples include take,
156156
* takeSample. Cancellation works by setting the cancelled flag to true and interrupting the

core/src/main/scala/org/apache/spark/TaskContext.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ import scala.collection.mutable.ArrayBuffer
2222
import org.apache.spark.executor.TaskMetrics
2323

2424
/**
25-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
25+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
2626
*
2727
* Contextual information about a task which can be read or mutated during execution.
2828
*/

core/src/main/scala/org/apache/spark/executor/TaskMetrics.scala

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ package org.apache.spark.executor
2020
import org.apache.spark.storage.{BlockId, BlockStatus}
2121

2222
/**
23-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
23+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
2424
*
2525
* Metrics tracked during the execution of a task.
2626
*/
@@ -88,7 +88,7 @@ object TaskMetrics {
8888

8989

9090
/**
91-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
91+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
9292
*
9393
* Metrics pertaining to shuffle data read in a given task.
9494
*/
@@ -127,7 +127,7 @@ class ShuffleReadMetrics extends Serializable {
127127
}
128128

129129
/**
130-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
130+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
131131
*
132132
* Metrics pertaining to shuffle data written in a given task.
133133
*/

core/src/main/scala/org/apache/spark/io/CompressionCodec.scala

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ import org.xerial.snappy.{SnappyInputStream, SnappyOutputStream}
2525
import org.apache.spark.SparkConf
2626

2727
/**
28-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
28+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
2929
*
3030
* CompressionCodec allows the customization of choosing different compression implementations
3131
* to be used in block storage.
@@ -58,7 +58,7 @@ private[spark] object CompressionCodec {
5858

5959

6060
/**
61-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
61+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
6262
*
6363
* LZF implementation of [[org.apache.spark.io.CompressionCodec]].
6464
*
@@ -77,7 +77,7 @@ class LZFCompressionCodec(conf: SparkConf) extends CompressionCodec {
7777

7878

7979
/**
80-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
80+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
8181
*
8282
* Snappy implementation of [[org.apache.spark.io.CompressionCodec]].
8383
* Block size can be configured by spark.io.compression.snappy.block.size.

core/src/main/scala/org/apache/spark/partial/BoundedDouble.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@
1818
package org.apache.spark.partial
1919

2020
/**
21-
* <span class="badge badge-red" style="float: right;">EXPERIMENTAL</span>
21+
* <span class="badge badge-red" style="float: right;">EXPERIMENTAL API</span>
2222
*
2323
* A Double value with error bars and associated confidence.
2424
*/

core/src/main/scala/org/apache/spark/partial/PartialResult.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@
1818
package org.apache.spark.partial
1919

2020
/**
21-
* <span class="badge badge-red" style="float: right;">EXPERIMENTAL</span>
21+
* <span class="badge badge-red" style="float: right;">EXPERIMENTAL API</span>
2222
*/
2323
class PartialResult[R](initialVal: R, isFinal: Boolean) {
2424
private var finalValue: Option[R] = if (isFinal) Some(initialVal) else None

core/src/main/scala/org/apache/spark/rdd/AsyncRDDActions.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ import scala.reflect.ClassTag
2626
import org.apache.spark.{ComplexFutureAction, FutureAction, Logging}
2727

2828
/**
29-
* <span class="badge badge-red" style="float: right;">EXPERIMENTAL</span>
29+
* <span class="badge badge-red" style="float: right;">EXPERIMENTAL API</span>
3030
*
3131
* A set of asynchronous RDD actions available through an implicit conversion.
3232
* Import `org.apache.spark.SparkContext._` at the top of your program to use these functions.

core/src/main/scala/org/apache/spark/rdd/HadoopRDD.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -70,7 +70,7 @@ private[spark] class HadoopPartition(rddId: Int, idx: Int, @transient s: InputSp
7070
}
7171

7272
/**
73-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
73+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
7474
*
7575
* An RDD that provides core functionality for reading data stored in Hadoop (e.g., files in HDFS,
7676
* sources in HBase, or S3), using the older MapReduce API (`org.apache.hadoop.mapred`).

core/src/main/scala/org/apache/spark/rdd/NewHadoopRDD.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ class NewHadoopPartition(rddId: Int, val index: Int, @transient rawSplit: InputS
3636
}
3737

3838
/**
39-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
39+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
4040
*
4141
* An RDD that provides core functionality for reading data stored in Hadoop (e.g., files in HDFS,
4242
* sources in HBase, or S3), using the new MapReduce API (`org.apache.hadoop.mapreduce`).

core/src/main/scala/org/apache/spark/rdd/PartitionPruningRDD.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@ private[spark] class PruneDependency[T](rdd: RDD[T], @transient partitionFilterF
4646

4747

4848
/**
49-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
49+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
5050
*
5151
* A RDD used to prune RDD partitions/partitions so we can avoid launching tasks on
5252
* all partitions. An example use case: If we know the RDD is partitioned by range,
@@ -67,7 +67,7 @@ class PartitionPruningRDD[T: ClassTag](
6767

6868

6969
/**
70-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
70+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
7171
*/
7272
object PartitionPruningRDD {
7373

core/src/main/scala/org/apache/spark/rdd/ShuffledRDD.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ private[spark] class ShuffledRDDPartition(val idx: Int) extends Partition {
2828
}
2929

3030
/**
31-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
31+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
3232
*
3333
* The resulting RDD from a shuffle (e.g. repartitioning of data).
3434
* @param prev the parent RDD.

core/src/main/scala/org/apache/spark/scheduler/SparkListener.scala

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -27,23 +27,23 @@ import org.apache.spark.executor.TaskMetrics
2727
import org.apache.spark.storage.BlockManagerId
2828
import org.apache.spark.util.{Distribution, Utils}
2929

30-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
30+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
3131
sealed trait SparkListenerEvent
3232

33-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
33+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
3434
case class SparkListenerStageSubmitted(stageInfo: StageInfo, properties: Properties = null)
3535
extends SparkListenerEvent
3636

37-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
37+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
3838
case class SparkListenerStageCompleted(stageInfo: StageInfo) extends SparkListenerEvent
3939

40-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
40+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
4141
case class SparkListenerTaskStart(stageId: Int, taskInfo: TaskInfo) extends SparkListenerEvent
4242

43-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
43+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
4444
case class SparkListenerTaskGettingResult(taskInfo: TaskInfo) extends SparkListenerEvent
4545

46-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
46+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
4747
case class SparkListenerTaskEnd(
4848
stageId: Int,
4949
taskType: String,
@@ -52,34 +52,34 @@ case class SparkListenerTaskEnd(
5252
taskMetrics: TaskMetrics)
5353
extends SparkListenerEvent
5454

55-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
55+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
5656
case class SparkListenerJobStart(jobId: Int, stageIds: Seq[Int], properties: Properties = null)
5757
extends SparkListenerEvent
5858

59-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
59+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
6060
case class SparkListenerJobEnd(jobId: Int, jobResult: JobResult) extends SparkListenerEvent
6161

62-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
62+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
6363
case class SparkListenerEnvironmentUpdate(environmentDetails: Map[String, Seq[(String, String)]])
6464
extends SparkListenerEvent
6565

66-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
66+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
6767
case class SparkListenerBlockManagerAdded(blockManagerId: BlockManagerId, maxMem: Long)
6868
extends SparkListenerEvent
6969

70-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
70+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
7171
case class SparkListenerBlockManagerRemoved(blockManagerId: BlockManagerId)
7272
extends SparkListenerEvent
7373

74-
/** <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span> */
74+
/** <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span> */
7575
case class SparkListenerUnpersistRDD(rddId: Int) extends SparkListenerEvent
7676

7777
/** An event used in the listener to shutdown the listener daemon thread. */
7878
private[spark] case object SparkListenerShutdown extends SparkListenerEvent
7979

8080

8181
/**
82-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
82+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
8383
*
8484
* Interface for listening to events from the Spark scheduler. Note that this is an internal
8585
* interface which might change in different Spark releases.
@@ -143,7 +143,7 @@ trait SparkListener {
143143
}
144144

145145
/**
146-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
146+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
147147
*
148148
* Simple SparkListener that logs a few summary statistics when each stage completes
149149
*/

core/src/main/scala/org/apache/spark/serializer/JavaSerializer.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -94,7 +94,7 @@ private[spark] class JavaSerializerInstance(counterReset: Int) extends Serialize
9494
}
9595

9696
/**
97-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
97+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
9898
*
9999
* A Spark serializer that uses Java's built-in serialization.
100100
*

core/src/main/scala/org/apache/spark/serializer/Serializer.scala

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ import org.apache.spark.util.{ByteBufferInputStream, NextIterator}
2626
import org.apache.spark.SparkEnv
2727

2828
/**
29-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
29+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
3030
*
3131
* A serializer. Because some serialization libraries are not thread safe, this class is used to
3232
* create [[org.apache.spark.serializer.SerializerInstance]] objects that do the actual
@@ -55,7 +55,7 @@ object Serializer {
5555

5656

5757
/**
58-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
58+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
5959
*
6060
* An instance of a serializer, for use by one thread at a time.
6161
*/
@@ -89,7 +89,7 @@ trait SerializerInstance {
8989

9090

9191
/**
92-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
92+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
9393
*
9494
* A stream for writing serialized objects.
9595
*/
@@ -108,7 +108,7 @@ trait SerializationStream {
108108

109109

110110
/**
111-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
111+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
112112
*
113113
* A stream for reading serialized objects.
114114
*/

core/src/main/scala/org/apache/spark/util/BoundedPriorityQueue.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ import scala.collection.JavaConverters._
2424
import scala.collection.generic.Growable
2525

2626
/**
27-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
27+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
2828
*
2929
* Bounded priority queue. This class wraps the original PriorityQueue
3030
* class and modifies it such that only the top K elements are retained.

core/src/main/scala/org/apache/spark/util/MutablePair.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@
1818
package org.apache.spark.util
1919

2020
/**
21-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
21+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
2222
*
2323
* A tuple of 2 elements. This can be used as an alternative to Scala's Tuple2 when we want to
2424
* minimize object allocation.

core/src/main/scala/org/apache/spark/util/random/Pseudorandom.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@
1818
package org.apache.spark.util.random
1919

2020
/**
21-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
21+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
2222
*
2323
* A class with pseudorandom behavior.
2424
*/

core/src/main/scala/org/apache/spark/util/random/RandomSampler.scala

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ import cern.jet.random.Poisson
2323
import cern.jet.random.engine.DRand
2424

2525
/**
26-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
26+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
2727
*
2828
* A pseudorandom sampler. It is possible to change the sampled item type. For example, we might
2929
* want to add weights for stratified sampling or importance sampling. Should only use
@@ -42,7 +42,7 @@ trait RandomSampler[T, U] extends Pseudorandom with Cloneable with Serializable
4242
}
4343

4444
/**
45-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
45+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
4646
*
4747
* A sampler based on Bernoulli trials.
4848
*
@@ -71,7 +71,7 @@ class BernoulliSampler[T](lb: Double, ub: Double, complement: Boolean = false)
7171
}
7272

7373
/** Ps
74-
* <span class="badge badge-red" style="float: right;">SEMI-PRIVATE</span>
74+
* <span class="badge badge-red" style="float: right;">UNSTABLE DEVELOPER API</span>
7575
*
7676
* A sampler based on values drawn from Poisson distribution.
7777
*

0 commit comments

Comments
 (0)