public class CreateTableAsSelect
extends org.apache.spark.sql.catalyst.plans.logical.Command
implements org.apache.spark.sql.execution.RunnableCommand, scala.Product, scala.Serializable
Constructor and Description |
---|
CreateTableAsSelect(String database,
String tableName,
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan query,
boolean allowExisting,
scala.Option<org.apache.hadoop.hive.ql.plan.CreateTableDesc> desc) |
Modifier and Type | Method and Description |
---|---|
boolean |
allowExisting() |
String |
argString() |
String |
database() |
scala.Option<org.apache.hadoop.hive.ql.plan.CreateTableDesc> |
desc() |
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan |
query() |
scala.collection.Seq<Row> |
run(SQLContext sqlContext) |
String |
tableName() |
childrenResolved, cleanArgs, isTraceEnabled, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning, org$apache$spark$Logging$$log__$eq, org$apache$spark$Logging$$log_, org$apache$spark$sql$catalyst$plans$logical$LogicalPlan$$resolveAsColumn, org$apache$spark$sql$catalyst$plans$logical$LogicalPlan$$resolveAsTableColumn, resolve, resolve, resolve$default$3, resolveChildren, resolveChildren$default$3, resolved, resolveGetField, sameResult, statePrefix, statistics
expressions, inputSet, missingInput, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, references, schema, schemaString, simpleString, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, asCode, collect, fastEquals, flatMap, foreach, foreachUp, generateTreeString, getNodeNumbered, makeCopy, map, mapChildren, nodeName, numberedTreeString, origin, otherCopyArgs, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
productArity, productElement, productIterator, productPrefix
initializeIfNecessary, initializeLogging, log_
public CreateTableAsSelect(String database, String tableName, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan query, boolean allowExisting, scala.Option<org.apache.hadoop.hive.ql.plan.CreateTableDesc> desc)
public String database()
public String tableName()
public org.apache.spark.sql.catalyst.plans.logical.LogicalPlan query()
public boolean allowExisting()
public scala.Option<org.apache.hadoop.hive.ql.plan.CreateTableDesc> desc()
public scala.collection.Seq<Row> run(SQLContext sqlContext)
run
in interface org.apache.spark.sql.execution.RunnableCommand
public String argString()
argString
in class org.apache.spark.sql.catalyst.trees.TreeNode<org.apache.spark.sql.catalyst.plans.logical.LogicalPlan>