public class HiveTableScan
extends org.apache.spark.sql.execution.SparkPlan
implements org.apache.spark.sql.execution.LeafNode, scala.Product, scala.Serializable
| Constructor and Description |
|---|
HiveTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> requestedAttributes,
MetastoreRelation relation,
scala.Option<org.apache.spark.sql.catalyst.expressions.Expression> partitionPruningPred,
HiveContext context) |
| Modifier and Type | Method and Description |
|---|---|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> |
attributes() |
HiveContext |
context() |
RDD<org.apache.spark.sql.Row> |
execute() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> |
output() |
scala.Option<org.apache.spark.sql.catalyst.expressions.Expression> |
partitionPruningPred() |
scala.collection.Seq<org.apache.hadoop.hive.ql.metadata.Partition> |
prunePartitions(scala.collection.Seq<org.apache.hadoop.hive.ql.metadata.Partition> partitions)
Prunes partitions not involve the query plan.
|
MetastoreRelation |
relation() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
requestedAttributes() |
codegenEnabled, executeCollect, executeTake, isTraceEnabled, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning, makeCopy, newMutableProjection, newOrdering, newPredicate, newProjection, org$apache$spark$Logging$$log__$eq, org$apache$spark$Logging$$log_, outputPartitioning, requiredChildDistribution, sparkContext, sqlContextexpressions, inputSet, missingInput, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, references, schema, schemaString, simpleString, statePrefix, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUpapply, argString, asCode, children, collect, fastEquals, flatMap, foreach, foreachUp, generateTreeString, getNodeNumbered, map, mapChildren, nodeName, numberedTreeString, origin, otherCopyArgs, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildrenproductArity, productElement, productIterator, productPrefixinitializeIfNecessary, initializeLogging, log_public HiveTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> requestedAttributes,
MetastoreRelation relation,
scala.Option<org.apache.spark.sql.catalyst.expressions.Expression> partitionPruningPred,
HiveContext context)
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> requestedAttributes()
public MetastoreRelation relation()
public scala.Option<org.apache.spark.sql.catalyst.expressions.Expression> partitionPruningPred()
public HiveContext context()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> attributes()
public scala.collection.Seq<org.apache.hadoop.hive.ql.metadata.Partition> prunePartitions(scala.collection.Seq<org.apache.hadoop.hive.ql.metadata.Partition> partitions)
partitions - All partitions of the relation.public RDD<org.apache.spark.sql.Row> execute()
execute in class org.apache.spark.sql.execution.SparkPlanpublic scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> output()