public class ParquetTableScan extends SparkPlan implements LeafNode, scala.Product, scala.Serializable
ParquetRelation
as a
RDD[Row]
.Constructor and Description |
---|
ParquetTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes,
ParquetRelation relation,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> columnPruningPred) |
Modifier and Type | Method and Description |
---|---|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
attributes() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> |
columnPruningPred() |
RDD<org.apache.spark.sql.catalyst.expressions.Row> |
execute()
Runs this query returning the result as an RDD.
|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
output() |
ParquetTableScan |
pruneColumns(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> prunedAttributes)
Applies a (candidate) projection.
|
ParquetRelation |
relation() |
scala.Tuple2<Object,Object>[] |
requestedPartitionOrdinals() |
codegenEnabled, executeCollect, makeCopy, outputPartitioning, requiredChildDistribution
expressions, inputSet, missingInput, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, references, schema, schemaString, simpleString, statePrefix, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, generateTreeString, getNodeNumbered, map, mapChildren, nodeName, numberedTreeString, otherCopyArgs, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
productArity, productElement, productIterator, productPrefix
initializeIfNecessary, initializeLogging, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
public ParquetTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes, ParquetRelation relation, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> columnPruningPred)
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes()
public ParquetRelation relation()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> columnPruningPred()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output()
output
in class org.apache.spark.sql.catalyst.plans.QueryPlan<SparkPlan>
public scala.Tuple2<Object,Object>[] requestedPartitionOrdinals()
public RDD<org.apache.spark.sql.catalyst.expressions.Row> execute()
SparkPlan
public ParquetTableScan pruneColumns(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> prunedAttributes)
prunedAttributes
- The list of attributes to be used in the projection.