public abstract class SparkPlan extends org.apache.spark.sql.catalyst.plans.QueryPlan<SparkPlan> implements Logging, scala.Serializable
Constructor and Description |
---|
SparkPlan() |
Modifier and Type | Method and Description |
---|---|
boolean |
codegenEnabled() |
abstract RDD<org.apache.spark.sql.catalyst.expressions.Row> |
execute()
Runs this query returning the result as an RDD.
|
org.apache.spark.sql.catalyst.expressions.Row[] |
executeCollect()
Runs this query returning the result as an array.
|
SparkPlan |
makeCopy(Object[] newArgs)
Overridden make copy also propogates sqlContext to copied plan.
|
org.apache.spark.sql.catalyst.plans.physical.Partitioning |
outputPartitioning()
Specifies how data is partitioned across different nodes in the cluster.
|
scala.collection.Seq<org.apache.spark.sql.catalyst.plans.physical.Distribution> |
requiredChildDistribution()
Specifies any partition requirements on the input data for this operator.
|
expressions, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, output, outputSet, printSchema, schema, schemaString, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, generateTreeString, getNodeNumbered, id, map, mapChildren, nextId, nodeName, numberedTreeString, otherCopyArgs, sameInstance, simpleString, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
initialized, initializeIfNecessary, initializeLogging, initLock, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
public boolean codegenEnabled()
public SparkPlan makeCopy(Object[] newArgs)
makeCopy
in class org.apache.spark.sql.catalyst.trees.TreeNode<SparkPlan>
public org.apache.spark.sql.catalyst.plans.physical.Partitioning outputPartitioning()
public scala.collection.Seq<org.apache.spark.sql.catalyst.plans.physical.Distribution> requiredChildDistribution()
public abstract RDD<org.apache.spark.sql.catalyst.expressions.Row> execute()
public org.apache.spark.sql.catalyst.expressions.Row[] executeCollect()