public class PipedRDD<T> extends RDD<String>
Modifier and Type | Class and Description |
---|---|
class |
PipedRDD.NotEqualsFileNameFilter
A FilenameFilter that accepts anything that isn't equal to the name passed in.
|
Constructor and Description |
---|
PipedRDD(RDD<T> prev,
scala.collection.Seq<String> command,
scala.collection.Map<String,String> envVars,
scala.Function1<scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit> printPipeContext,
scala.Function2<T,scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit> printRDDElement,
boolean separateWorkingDir,
scala.reflect.ClassTag<T> evidence$1) |
PipedRDD(RDD<T> prev,
String command,
scala.collection.Map<String,String> envVars,
scala.Function1<scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit> printPipeContext,
scala.Function2<T,scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit> printRDDElement,
boolean separateWorkingDir,
scala.reflect.ClassTag<T> evidence$2) |
Modifier and Type | Method and Description |
---|---|
scala.collection.Iterator<String> |
compute(Partition split,
TaskContext context)
:: DeveloperApi ::
Implemented by subclasses to compute a given partition.
|
Partition[] |
getPartitions()
Implemented by subclasses to return the set of partitions in this RDD.
|
static scala.collection.Seq<String> |
tokenize(String command) |
aggregate, cache, cartesian, checkpoint, checkpointData, coalesce, collect, collect, collectPartitions, computeOrReadCheckpoint, conf, context, count, countApprox, countApproxDistinct, countApproxDistinct, countByValue, countByValueApprox, creationSite, dependencies, distinct, distinct, doCheckpoint, elementClassTag, filter, filterWith, first, flatMap, flatMapWith, fold, foreach, foreachPartition, foreachWith, getCheckpointFile, getCreationSite, getNarrowAncestors, getStorageLevel, glom, groupBy, groupBy, groupBy, id, intersection, intersection, intersection, isCheckpointed, iterator, keyBy, map, mapPartitions, mapPartitionsWithContext, mapPartitionsWithIndex, mapPartitionsWithSplit, mapWith, markCheckpointed, max, min, name, partitioner, partitions, persist, persist, pipe, pipe, pipe, preferredLocations, randomSplit, reduce, repartition, retag, retag, sample, saveAsObjectFile, saveAsTextFile, saveAsTextFile, setName, sortBy, sparkContext, subtract, subtract, subtract, take, takeOrdered, takeSample, toArray, toDebugString, toJavaRDD, toLocalIterator, top, toString, union, unpersist, zip, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipWithIndex, zipWithUniqueId
initializeIfNecessary, initializeLogging, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
public PipedRDD(RDD<T> prev, scala.collection.Seq<String> command, scala.collection.Map<String,String> envVars, scala.Function1<scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit> printPipeContext, scala.Function2<T,scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit> printRDDElement, boolean separateWorkingDir, scala.reflect.ClassTag<T> evidence$1)
public PipedRDD(RDD<T> prev, String command, scala.collection.Map<String,String> envVars, scala.Function1<scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit> printPipeContext, scala.Function2<T,scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit> printRDDElement, boolean separateWorkingDir, scala.reflect.ClassTag<T> evidence$2)
public static scala.collection.Seq<String> tokenize(String command)
public Partition[] getPartitions()
RDD
public scala.collection.Iterator<String> compute(Partition split, TaskContext context)
RDD