public abstract class SparkPlan extends org.apache.spark.sql.catalyst.plans.QueryPlan<SparkPlan> implements com.typesafe.scalalogging.slf4j.Logging
Constructor and Description |
---|
SparkPlan() |
Modifier and Type | Method and Description |
---|---|
abstract RDD<org.apache.spark.sql.catalyst.expressions.Row> |
execute()
Runs this query returning the result as an RDD.
|
org.apache.spark.sql.catalyst.expressions.Row[] |
executeCollect()
Runs this query returning the result as an array.
|
org.apache.spark.sql.catalyst.plans.physical.Partitioning |
outputPartitioning()
Specifies how data is partitioned across different nodes in the cluster.
|
scala.collection.Seq<org.apache.spark.sql.catalyst.plans.physical.Distribution> |
requiredChildDistribution()
Specifies any partition requirements on the input data for this operator.
|
expressions, generateSchemaString, generateSchemaString, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, output, outputSet, printSchema, schemaString, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, generateTreeString, getNodeNumbered, id, makeCopy, map, mapChildren, nextId, nodeName, numberedTreeString, otherCopyArgs, sameInstance, simpleString, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
public org.apache.spark.sql.catalyst.plans.physical.Partitioning outputPartitioning()
public scala.collection.Seq<org.apache.spark.sql.catalyst.plans.physical.Distribution> requiredChildDistribution()
public abstract RDD<org.apache.spark.sql.catalyst.expressions.Row> execute()
public org.apache.spark.sql.catalyst.expressions.Row[] executeCollect()