public class Aggregate extends SparkPlan implements org.apache.spark.sql.catalyst.expressions.NoBind, scala.Product, scala.Serializable
groupingExpressions
and computes the aggregateExpressions
for each
group.
Modifier and Type | Class and Description |
---|---|
class |
Aggregate.ComputedAggregate
An aggregate that needs to be computed for each row in a group.
|
class |
Aggregate.ComputedAggregate$ |
Constructor and Description |
---|
Aggregate(boolean partial,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> groupingExpressions,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression> aggregateExpressions,
SparkPlan child,
SQLContext sqlContext) |
Modifier and Type | Method and Description |
---|---|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression> |
aggregateExpressions() |
SparkPlan |
child() |
RDD<org.apache.spark.sql.catalyst.expressions.Row> |
execute()
Substituted version of aggregateExpressions expressions which are used to compute final
output rows given a group and the result of all aggregate computations.
|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> |
groupingExpressions() |
scala.collection.immutable.List<SQLContext> |
otherCopyArgs() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
output() |
org.apache.spark.sql.catalyst.plans.physical.Partitioning |
outputPartitioning() |
boolean |
partial() |
scala.collection.Seq<org.apache.spark.sql.catalyst.plans.physical.Distribution> |
requiredChildDistribution()
Specifies any partition requirements on the input data for this operator.
|
executeCollect, outputPartitioning
expressions, generateSchemaString, generateSchemaString, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, schemaString, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, generateTreeString, getNodeNumbered, id, makeCopy, map, mapChildren, nextId, nodeName, numberedTreeString, sameInstance, simpleString, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
public Aggregate(boolean partial, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> groupingExpressions, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression> aggregateExpressions, SparkPlan child, SQLContext sqlContext)
public boolean partial()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> groupingExpressions()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression> aggregateExpressions()
public SparkPlan child()
public scala.collection.Seq<org.apache.spark.sql.catalyst.plans.physical.Distribution> requiredChildDistribution()
SparkPlan
requiredChildDistribution
in class SparkPlan
public scala.collection.immutable.List<SQLContext> otherCopyArgs()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output()
output
in class org.apache.spark.sql.catalyst.plans.QueryPlan<SparkPlan>
public RDD<org.apache.spark.sql.catalyst.expressions.Row> execute()
public org.apache.spark.sql.catalyst.plans.physical.Partitioning outputPartitioning()