Interface | Description |
---|---|
AccumulableParam<R,T> |
Helper object defining how to accumulate values of a particular type.
|
AccumulatorParam<T> |
A simpler version of
AccumulableParam where the only data type you can add
in is the same type as the accumulated value. |
FutureAction<T> |
:: Experimental ::
A future for the result of an action to support cancellation.
|
Logging |
:: DeveloperApi ::
Utility trait for classes that want to log data.
|
Partition |
A partition of an RDD.
|
TaskEndReason |
:: DeveloperApi ::
Various possible reasons why a task ended.
|
TaskFailedReason |
:: DeveloperApi ::
Various possible reasons why a task failed.
|
Class | Description |
---|---|
Accumulable<R,T> |
A data type that can be accumulated, ie has an commutative and associative "add" operation,
but where the result type,
R , may be different from the element type being added, T . |
Accumulator<T> |
A simpler value of
Accumulable where the result type being accumulated is the same
as the types of elements being merged, i.e. |
Aggregator<K,V,C> |
:: DeveloperApi ::
A set of functions used to aggregate data.
|
ComplexFutureAction<T> |
:: Experimental ::
A
FutureAction for actions that could trigger multiple Spark jobs. |
Dependency<T> |
:: DeveloperApi ::
Base class for dependencies.
|
ExceptionFailure |
:: DeveloperApi ::
Task failed due to a runtime exception.
|
ExecutorLostFailure |
:: DeveloperApi ::
The task failed because the executor that it was running on was lost.
|
FetchFailed |
:: DeveloperApi ::
Task failed to fetch shuffle data from a remote node.
|
HashPartitioner |
A
Partitioner that implements hash-based partitioning using
Java's Object.hashCode . |
InterruptibleIterator<T> |
:: DeveloperApi ::
An iterator that wraps around an existing iterator to provide task killing functionality.
|
NarrowDependency<T> |
:: DeveloperApi ::
Base class for dependencies where each partition of the parent RDD is used by at most one
partition of the child RDD.
|
OneToOneDependency<T> |
:: DeveloperApi ::
Represents a one-to-one dependency between partitions of the parent and child RDDs.
|
Partitioner |
An object that defines how the elements in a key-value pair RDD are partitioned by key.
|
RangeDependency<T> |
:: DeveloperApi ::
Represents a one-to-one dependency between ranges of partitions in the parent and child RDDs.
|
RangePartitioner<K,V> |
A
Partitioner that partitions sortable records by range into roughly
equal ranges. |
Resubmitted |
:: DeveloperApi ::
A
ShuffleMapTask that completed successfully earlier, but we
lost the executor before the stage completed. |
SerializableWritable<T extends org.apache.hadoop.io.Writable> | |
ShuffleDependency<K,V> |
:: DeveloperApi ::
Represents a dependency on the output of a shuffle stage.
|
SimpleFutureAction<T> |
:: Experimental ::
A
FutureAction holding the result of an action that triggers a single job. |
SparkConf |
Configuration for a Spark application.
|
SparkContext |
Main entry point for Spark functionality.
|
SparkContext.DoubleAccumulatorParam$ | |
SparkContext.FloatAccumulatorParam$ | |
SparkContext.IntAccumulatorParam$ | |
SparkContext.LongAccumulatorParam$ | |
SparkEnv |
:: DeveloperApi ::
Holds all the runtime environment objects for a running Spark instance (either master or worker),
including the serializer, Akka actor system, block manager, map output tracker, etc.
|
SparkFiles |
Resolves paths to files added through
SparkContext.addFile() . |
Success |
:: DeveloperApi ::
Task succeeded.
|
TaskContext |
:: DeveloperApi ::
Contextual information about a task which can be read or mutated during execution.
|
TaskKilled |
:: DeveloperApi ::
Task was killed intentionally and needs to be rescheduled.
|
TaskResultLost |
:: DeveloperApi ::
The task finished successfully, but the result was lost from the executor's block manager before
it was fetched.
|
UnknownReason |
:: DeveloperApi ::
We don't know why the task ended -- for example, because of a ClassNotFound exception when
deserializing the task result.
|
Exception | Description |
---|---|
SparkException | |
TaskKilledException |
:: DeveloperApi ::
Exception thrown when a task is explicitly killed (i.e., task failure is expected).
|
Accumulator
and StorageLevel
, are also used in Java, but the
org.apache.spark.api.java
package contains the main Java API.