public class HiveTableScan
extends org.apache.spark.sql.execution.SparkPlan
implements org.apache.spark.sql.execution.LeafNode, scala.Product, scala.Serializable
Constructor and Description |
---|
HiveTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> requestedAttributes,
MetastoreRelation relation,
scala.Option<org.apache.spark.sql.catalyst.expressions.Expression> partitionPruningPred,
HiveContext context) |
Modifier and Type | Method and Description |
---|---|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> |
attributes() |
HiveContext |
context() |
RDD<org.apache.spark.sql.Row> |
execute() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> |
output() |
scala.Option<org.apache.spark.sql.catalyst.expressions.Expression> |
partitionPruningPred() |
scala.collection.Seq<org.apache.hadoop.hive.ql.metadata.Partition> |
prunePartitions(scala.collection.Seq<org.apache.hadoop.hive.ql.metadata.Partition> partitions)
Prunes partitions not involve the query plan.
|
MetastoreRelation |
relation() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
requestedAttributes() |
codegenEnabled, executeCollect, executeTake, isTraceEnabled, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning, makeCopy, newMutableProjection, newOrdering, newPredicate, newProjection, org$apache$spark$Logging$$log__$eq, org$apache$spark$Logging$$log_, outputPartitioning, requiredChildDistribution, sparkContext, sqlContext
expressions, inputSet, missingInput, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, references, schema, schemaString, simpleString, statePrefix, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, foreachUp, generateTreeString, getNodeNumbered, map, mapChildren, nodeName, numberedTreeString, origin, otherCopyArgs, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
productArity, productElement, productIterator, productPrefix
initializeIfNecessary, initializeLogging, log_
public HiveTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> requestedAttributes, MetastoreRelation relation, scala.Option<org.apache.spark.sql.catalyst.expressions.Expression> partitionPruningPred, HiveContext context)
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> requestedAttributes()
public MetastoreRelation relation()
public scala.Option<org.apache.spark.sql.catalyst.expressions.Expression> partitionPruningPred()
public HiveContext context()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> attributes()
public scala.collection.Seq<org.apache.hadoop.hive.ql.metadata.Partition> prunePartitions(scala.collection.Seq<org.apache.hadoop.hive.ql.metadata.Partition> partitions)
partitions
- All partitions of the relation.public RDD<org.apache.spark.sql.Row> execute()
execute
in class org.apache.spark.sql.execution.SparkPlan
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> output()