public abstract class SparkPlan extends org.apache.spark.sql.catalyst.plans.QueryPlan<SparkPlan> implements Logging, scala.Serializable
| Constructor and Description |
|---|
SparkPlan() |
| Modifier and Type | Method and Description |
|---|---|
boolean |
codegenEnabled() |
abstract RDD<org.apache.spark.sql.catalyst.expressions.Row> |
execute()
Runs this query returning the result as an RDD.
|
org.apache.spark.sql.catalyst.expressions.Row[] |
executeCollect()
Runs this query returning the result as an array.
|
SparkPlan |
makeCopy(Object[] newArgs)
Overridden make copy also propogates sqlContext to copied plan.
|
org.apache.spark.sql.catalyst.plans.physical.Partitioning |
outputPartitioning()
Specifies how data is partitioned across different nodes in the cluster.
|
scala.collection.Seq<org.apache.spark.sql.catalyst.plans.physical.Distribution> |
requiredChildDistribution()
Specifies any partition requirements on the input data for this operator.
|
expressions, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, output, outputSet, printSchema, schema, schemaString, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUpapply, argString, asCode, children, collect, fastEquals, flatMap, foreach, generateTreeString, getNodeNumbered, id, map, mapChildren, nextId, nodeName, numberedTreeString, otherCopyArgs, sameInstance, simpleString, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildreninitialized, initializeIfNecessary, initializeLogging, initLock, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarningpublic boolean codegenEnabled()
public SparkPlan makeCopy(Object[] newArgs)
makeCopy in class org.apache.spark.sql.catalyst.trees.TreeNode<SparkPlan>public org.apache.spark.sql.catalyst.plans.physical.Partitioning outputPartitioning()
public scala.collection.Seq<org.apache.spark.sql.catalyst.plans.physical.Distribution> requiredChildDistribution()
public abstract RDD<org.apache.spark.sql.catalyst.expressions.Row> execute()
public org.apache.spark.sql.catalyst.expressions.Row[] executeCollect()