public class InMemoryColumnarTableScan extends SparkPlan implements LeafNode, scala.Product, scala.Serializable
Constructor and Description |
---|
InMemoryColumnarTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> predicates,
InMemoryRelation relation) |
Modifier and Type | Method and Description |
---|---|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
attributes() |
scala.PartialFunction<org.apache.spark.sql.catalyst.expressions.Expression,org.apache.spark.sql.catalyst.expressions.Expression> |
buildFilter() |
RDD<org.apache.spark.sql.catalyst.expressions.Row> |
execute()
Runs this query returning the result as an RDD.
|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
output() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> |
partitionFilters() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> |
predicates() |
Accumulator<Object> |
readBatches() |
Accumulator<Object> |
readPartitions() |
InMemoryRelation |
relation() |
SQLContext |
sqlContext()
A handle to the SQL Context that was used to create this plan.
|
codegenEnabled, executeCollect, makeCopy, outputPartitioning, requiredChildDistribution
expressions, inputSet, missingInput, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, references, schema, schemaString, simpleString, statePrefix, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, generateTreeString, getNodeNumbered, map, mapChildren, nodeName, numberedTreeString, otherCopyArgs, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
productArity, productElement, productIterator, productPrefix
initializeIfNecessary, initializeLogging, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
public InMemoryColumnarTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> predicates, InMemoryRelation relation)
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> predicates()
public InMemoryRelation relation()
public SQLContext sqlContext()
SparkPlan
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output()
output
in class org.apache.spark.sql.catalyst.plans.QueryPlan<SparkPlan>
public scala.PartialFunction<org.apache.spark.sql.catalyst.expressions.Expression,org.apache.spark.sql.catalyst.expressions.Expression> buildFilter()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> partitionFilters()
public Accumulator<Object> readPartitions()
public Accumulator<Object> readBatches()