public class InMemoryColumnarTableScan extends SparkPlan implements LeafNode, scala.Product, scala.Serializable
| Constructor and Description |
|---|
InMemoryColumnarTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> predicates,
InMemoryRelation relation) |
| Modifier and Type | Method and Description |
|---|---|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
attributes() |
scala.PartialFunction<org.apache.spark.sql.catalyst.expressions.Expression,org.apache.spark.sql.catalyst.expressions.Expression> |
buildFilter() |
RDD<org.apache.spark.sql.catalyst.expressions.Row> |
execute()
Runs this query returning the result as an RDD.
|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
output() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> |
partitionFilters() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> |
predicates() |
Accumulator<Object> |
readBatches() |
Accumulator<Object> |
readPartitions() |
InMemoryRelation |
relation() |
SQLContext |
sqlContext()
A handle to the SQL Context that was used to create this plan.
|
codegenEnabled, executeCollect, makeCopy, outputPartitioning, requiredChildDistributionexpressions, inputSet, missingInput, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, references, schema, schemaString, simpleString, statePrefix, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUpapply, argString, asCode, children, collect, fastEquals, flatMap, foreach, generateTreeString, getNodeNumbered, map, mapChildren, nodeName, numberedTreeString, otherCopyArgs, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildrenproductArity, productElement, productIterator, productPrefixinitializeIfNecessary, initializeLogging, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarningpublic InMemoryColumnarTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> predicates,
InMemoryRelation relation)
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> predicates()
public InMemoryRelation relation()
public SQLContext sqlContext()
SparkPlanpublic scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output()
output in class org.apache.spark.sql.catalyst.plans.QueryPlan<SparkPlan>public scala.PartialFunction<org.apache.spark.sql.catalyst.expressions.Expression,org.apache.spark.sql.catalyst.expressions.Expression> buildFilter()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> partitionFilters()
public Accumulator<Object> readPartitions()
public Accumulator<Object> readBatches()