public class ScriptTransformationExec
extends org.apache.spark.sql.execution.SparkPlan
implements org.apache.spark.sql.execution.UnaryExecNode, scala.Product, scala.Serializable
param: input the set of expression that should be passed to the script. param: script the command that should be executed. param: output the attributes that are produced by the script.
Constructor and Description |
---|
ScriptTransformationExec(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> input,
String script,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output,
org.apache.spark.sql.execution.SparkPlan child,
HiveScriptIOSchema ioschema) |
Modifier and Type | Method and Description |
---|---|
abstract static R |
apply(T1 v1,
T2 v2,
T3 v3,
T4 v4,
T5 v5) |
org.apache.spark.sql.execution.SparkPlan |
child() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> |
input() |
HiveScriptIOSchema |
ioschema() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
output() |
org.apache.spark.sql.catalyst.plans.physical.Partitioning |
outputPartitioning() |
org.apache.spark.sql.catalyst.expressions.AttributeSet |
producedAttributes() |
String |
script() |
static String |
toString() |
cleanupResources, doExecuteBroadcast, doPrepare, execute, executeBroadcast, executeCollect, executeCollectIterator, executeCollectPublic, executeQuery, executeTake, executeToIterator, initializeLogIfNecessary, initializeLogIfNecessary, initializeLogIfNecessary$default$2, isTraceEnabled, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning, longMetric, makeCopy, metrics, newMutableProjection, newMutableProjection$default$3, newNaturalAscendingOrdering, newOrdering, newPredicate, org$apache$spark$internal$Logging$$log__$eq, org$apache$spark$internal$Logging$$log_, org$apache$spark$sql$execution$SparkPlan$$decodeUnsafeRows, org$apache$spark$sql$execution$SparkPlan$$runningSubqueries, outputOrdering, prepare, prepareSubqueries, requiredChildDistribution, requiredChildOrdering, resetMetrics, sparkContext, sqlContext, subexpressionEliminationEnabled, waitForSubqueries
allAttributes, canEvaluate, canEvaluateWithinJoin, canonicalized, conf, doCanonicalize, expressions, innerChildren, inputSet, isCanonicalizedPlan, mapExpressions, missingInput, normalizeExprId, normalizePredicates, org$apache$spark$sql$catalyst$plans$QueryPlan$$recursiveTransform$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$seqToExpressions$1, outputSet, printSchema, references, replaceAlias, sameResult, schema, schemaString, semanticHash, simpleString, splitConjunctivePredicates, splitDisjunctivePredicates, statePrefix, subqueries, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp, verboseString
apply, argString, asCode, children, collect, collectFirst, collectLeaves, containsChild, fastEquals, find, flatMap, foreach, foreachUp, generateTreeString, generateTreeString$default$5, generateTreeString$default$6, hashCode, jsonFields, map, mapChildren, mapProductIterator, nodeName, numberedTreeString, org$apache$spark$sql$catalyst$trees$TreeNode$$allChildren, org$apache$spark$sql$catalyst$trees$TreeNode$$collectJsonValue$1, org$apache$spark$sql$catalyst$trees$TreeNode$$getNodeNumbered, org$apache$spark$sql$catalyst$trees$TreeNode$$mapChild$1, org$apache$spark$sql$catalyst$trees$TreeNode$$mapChild$2, org$apache$spark$sql$catalyst$trees$TreeNode$$mapTreeNode$1, org$apache$spark$sql$catalyst$trees$TreeNode$$parseToJson, origin, otherCopyArgs, p, prettyJson, productIterator, productPrefix, stringArgs, toJSON, toString, transform, transformDown, transformUp, treeString, treeString, treeString$default$2, verboseStringWithSuffix, withNewChildren
productArity, productElement, productIterator, productPrefix
initializeLogging, log_
public ScriptTransformationExec(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> input, String script, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output, org.apache.spark.sql.execution.SparkPlan child, HiveScriptIOSchema ioschema)
public abstract static R apply(T1 v1, T2 v2, T3 v3, T4 v4, T5 v5)
public static String toString()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> input()
public String script()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output()
output
in class org.apache.spark.sql.catalyst.plans.QueryPlan<org.apache.spark.sql.execution.SparkPlan>
public org.apache.spark.sql.execution.SparkPlan child()
child
in interface org.apache.spark.sql.execution.UnaryExecNode
public HiveScriptIOSchema ioschema()
public org.apache.spark.sql.catalyst.expressions.AttributeSet producedAttributes()
producedAttributes
in class org.apache.spark.sql.catalyst.plans.QueryPlan<org.apache.spark.sql.execution.SparkPlan>
public org.apache.spark.sql.catalyst.plans.physical.Partitioning outputPartitioning()
outputPartitioning
in class org.apache.spark.sql.execution.SparkPlan