public class ShuffledRDD<K,V,C> extends RDD<scala.Tuple2<K,C>>
Constructor and Description |
---|
ShuffledRDD(RDD<? extends scala.Product2<K,V>> prev,
Partitioner part,
scala.reflect.ClassTag<K> evidence$1,
scala.reflect.ClassTag<V> evidence$2,
scala.reflect.ClassTag<C> evidence$3) |
Modifier and Type | Method and Description |
---|---|
void |
clearDependencies()
Clears the dependencies of this RDD.
|
scala.collection.Iterator<scala.Tuple2<K,C>> |
compute(Partition split,
TaskContext context)
:: DeveloperApi ::
Implemented by subclasses to compute a given partition.
|
scala.collection.Seq<Dependency<?>> |
getDependencies()
Implemented by subclasses to return how this RDD depends on parent RDDs.
|
Partition[] |
getPartitions()
Implemented by subclasses to return the set of partitions in this RDD.
|
protected scala.collection.Seq<java.lang.String> |
getPreferredLocations(Partition partition)
Optionally overridden by subclasses to specify placement preferences.
|
scala.Some<Partitioner> |
partitioner()
Optionally overridden by subclasses to specify how they are partitioned.
|
java.lang.Object |
prev() |
ShuffledRDD<K,V,C> |
setAggregator(Aggregator<K,V,C> aggregator)
Set aggregator for RDD's shuffle.
|
ShuffledRDD<K,V,C> |
setKeyOrdering(scala.math.Ordering<K> keyOrdering)
Set key ordering for RDD's shuffle.
|
ShuffledRDD<K,V,C> |
setMapSideCombine(boolean mapSideCombine)
Set mapSideCombine flag for RDD's shuffle.
|
ShuffledRDD<K,V,C> |
setSerializer(Serializer serializer)
Set a serializer for this RDD's shuffle, or null to use the default (spark.serializer)
|
aggregate, cache, cartesian, checkpoint, checkpointData, coalesce, collect, collect, context, count, countApprox, countApproxDistinct, countApproxDistinct, countByValue, countByValueApprox, creationSite, dependencies, distinct, distinct, doubleRDDToDoubleRDDFunctions, filter, filterWith, first, firstParent, flatMap, flatMapWith, fold, foreach, foreachPartition, foreachWith, getCheckpointFile, getNumPartitions, getStorageLevel, glom, groupBy, groupBy, groupBy, id, intersection, intersection, intersection, isCheckpointed, isEmpty, iterator, keyBy, localCheckpoint, map, mapPartitions, mapPartitionsWithContext, mapPartitionsWithIndex, mapPartitionsWithSplit, mapWith, max, min, name, numericRDDToDoubleRDDFunctions, parent, partitions, persist, persist, pipe, pipe, pipe, preferredLocations, randomSplit, rddToAsyncRDDActions, rddToOrderedRDDFunctions, rddToPairRDDFunctions, rddToSequenceFileRDDFunctions, reduce, repartition, sample, saveAsObjectFile, saveAsTextFile, saveAsTextFile, scope, setName, sortBy, sparkContext, subtract, subtract, subtract, take, takeOrdered, takeSample, toArray, toDebugString, toJavaRDD, toLocalIterator, top, toString, treeAggregate, treeReduce, union, unpersist, zip, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipWithIndex, zipWithUniqueId
clone, equals, finalize, getClass, hashCode, notify, notifyAll, wait, wait, wait
initializeIfNecessary, initializeLogging, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
public java.lang.Object prev()
public ShuffledRDD<K,V,C> setSerializer(Serializer serializer)
public ShuffledRDD<K,V,C> setKeyOrdering(scala.math.Ordering<K> keyOrdering)
public ShuffledRDD<K,V,C> setAggregator(Aggregator<K,V,C> aggregator)
public ShuffledRDD<K,V,C> setMapSideCombine(boolean mapSideCombine)
public scala.collection.Seq<Dependency<?>> getDependencies()
RDD
getDependencies
in class RDD<scala.Tuple2<K,C>>
public scala.Some<Partitioner> partitioner()
RDD
partitioner
in class RDD<scala.Tuple2<K,C>>
public Partition[] getPartitions()
RDD
getPartitions
in class RDD<scala.Tuple2<K,C>>
protected scala.collection.Seq<java.lang.String> getPreferredLocations(Partition partition)
RDD
getPreferredLocations
in class RDD<scala.Tuple2<K,C>>
partition
- (undocumented)public scala.collection.Iterator<scala.Tuple2<K,C>> compute(Partition split, TaskContext context)
RDD
public void clearDependencies()
RDD
UnionRDD
for an example.clearDependencies
in class RDD<scala.Tuple2<K,C>>