public class Limit extends SparkPlan implements UnaryNode, scala.Product, scala.Serializable
Modifier and Type | Method and Description |
---|---|
SparkPlan |
child() |
RDD<org.apache.spark.sql.catalyst.expressions.Row> |
execute()
Runs this query returning the result as an RDD.
|
org.apache.spark.sql.catalyst.expressions.Row[] |
executeCollect()
A custom implementation modeled after the take function on RDDs but which never runs any job
locally.
|
int |
limit() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
output() |
org.apache.spark.sql.catalyst.plans.physical.SinglePartition$ |
outputPartitioning()
Specifies how data is partitioned across different nodes in the cluster.
|
codegenEnabled, makeCopy, requiredChildDistribution
expressions, inputSet, missingInput, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, references, schema, schemaString, simpleString, statePrefix, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, generateTreeString, getNodeNumbered, map, mapChildren, nodeName, numberedTreeString, otherCopyArgs, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
productArity, productElement, productIterator, productPrefix
initializeIfNecessary, initializeLogging, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
public Limit(int limit, SparkPlan child)
public int limit()
public SparkPlan child()
child
in interface org.apache.spark.sql.catalyst.trees.UnaryNode<SparkPlan>
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output()
output
in class org.apache.spark.sql.catalyst.plans.QueryPlan<SparkPlan>
public org.apache.spark.sql.catalyst.plans.physical.SinglePartition$ outputPartitioning()
SparkPlan
outputPartitioning
in class SparkPlan
public org.apache.spark.sql.catalyst.expressions.Row[] executeCollect()
executeCollect
in class SparkPlan