public class BroadcastNestedLoopJoin extends SparkPlan implements BinaryNode, scala.Product, scala.Serializable
Constructor and Description |
---|
BroadcastNestedLoopJoin(SparkPlan left,
SparkPlan right,
org.apache.spark.sql.execution.joins.BuildSide buildSide,
org.apache.spark.sql.catalyst.plans.JoinType joinType,
scala.Option<org.apache.spark.sql.catalyst.expressions.Expression> condition) |
Modifier and Type | Method and Description |
---|---|
org.apache.spark.sql.execution.joins.BuildSide |
buildSide() |
scala.Option<org.apache.spark.sql.catalyst.expressions.Expression> |
condition() |
RDD<org.apache.spark.sql.catalyst.expressions.Row> |
execute()
Runs this query returning the result as an RDD.
|
org.apache.spark.sql.catalyst.plans.JoinType |
joinType() |
SparkPlan |
left() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
output() |
org.apache.spark.sql.catalyst.plans.physical.Partitioning |
outputPartitioning()
Specifies how data is partitioned across different nodes in the cluster.
|
SparkPlan |
right() |
codegenEnabled, executeCollect, makeCopy, requiredChildDistribution
expressions, inputSet, missingInput, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, references, schema, schemaString, simpleString, statePrefix, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, generateTreeString, getNodeNumbered, map, mapChildren, nodeName, numberedTreeString, otherCopyArgs, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
productArity, productElement, productIterator, productPrefix
initializeIfNecessary, initializeLogging, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
public SparkPlan left()
left
in interface org.apache.spark.sql.catalyst.trees.BinaryNode<SparkPlan>
public SparkPlan right()
right
in interface org.apache.spark.sql.catalyst.trees.BinaryNode<SparkPlan>
public org.apache.spark.sql.execution.joins.BuildSide buildSide()
public org.apache.spark.sql.catalyst.plans.JoinType joinType()
public scala.Option<org.apache.spark.sql.catalyst.expressions.Expression> condition()
public org.apache.spark.sql.catalyst.plans.physical.Partitioning outputPartitioning()
SparkPlan
outputPartitioning
in class SparkPlan
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output()
output
in class org.apache.spark.sql.catalyst.plans.QueryPlan<SparkPlan>