public class Filter extends SparkPlan implements scala.Product, scala.Serializable
Constructor and Description |
---|
Filter(org.apache.spark.sql.catalyst.expressions.Expression condition,
SparkPlan child) |
Modifier and Type | Method and Description |
---|---|
SparkPlan |
child() |
org.apache.spark.sql.catalyst.expressions.Expression |
condition() |
scala.Function1<org.apache.spark.sql.catalyst.expressions.Row,Object> |
conditionEvaluator() |
RDD<org.apache.spark.sql.catalyst.expressions.Row> |
execute()
Runs this query returning the result as an RDD.
|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
output() |
org.apache.spark.sql.catalyst.plans.physical.Partitioning |
outputPartitioning() |
codegenEnabled, executeCollect, makeCopy, outputPartitioning, requiredChildDistribution
expressions, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, schema, schemaString, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, generateTreeString, getNodeNumbered, id, map, mapChildren, nextId, nodeName, numberedTreeString, otherCopyArgs, sameInstance, simpleString, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
productArity, productElement, productIterator, productPrefix
initialized, initializeIfNecessary, initializeLogging, initLock, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
public Filter(org.apache.spark.sql.catalyst.expressions.Expression condition, SparkPlan child)
public org.apache.spark.sql.catalyst.expressions.Expression condition()
public SparkPlan child()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output()
output
in class org.apache.spark.sql.catalyst.plans.QueryPlan<SparkPlan>
public scala.Function1<org.apache.spark.sql.catalyst.expressions.Row,Object> conditionEvaluator()
public RDD<org.apache.spark.sql.catalyst.expressions.Row> execute()
SparkPlan
public org.apache.spark.sql.catalyst.plans.physical.Partitioning outputPartitioning()