public class BinaryFileRDD<T> extends NewHadoopRDD<String,T>
NewHadoopRDD.NewHadoopMapPartitionsWithSplitRDD<U,T>, NewHadoopRDD.NewHadoopMapPartitionsWithSplitRDD$| Constructor and Description |
|---|
BinaryFileRDD(SparkContext sc,
Class<? extends StreamFileInputFormat<T>> inputFormatClass,
Class<String> keyClass,
Class<T> valueClass,
org.apache.hadoop.conf.Configuration conf,
int minPartitions) |
| Modifier and Type | Method and Description |
|---|---|
Partition[] |
getPartitions()
Implemented by subclasses to return the set of partitions in this RDD.
|
compute, getConf, getPreferredLocations, mapPartitionsWithInputSplitaggregate, cache, cartesian, checkpoint, checkpointData, coalesce, collect, collect, collectPartitions, computeOrReadCheckpoint, conf, context, count, countApprox, countApproxDistinct, countApproxDistinct, countByValue, countByValueApprox, creationSite, dependencies, distinct, distinct, doCheckpoint, elementClassTag, filter, filterWith, first, flatMap, flatMapWith, fold, foreach, foreachPartition, foreachWith, getCheckpointFile, getCreationSite, getNarrowAncestors, getStorageLevel, glom, groupBy, groupBy, groupBy, id, intersection, intersection, intersection, isCheckpointed, iterator, keyBy, map, mapPartitions, mapPartitionsWithContext, mapPartitionsWithIndex, mapPartitionsWithSplit, mapWith, markCheckpointed, max, min, name, partitioner, partitions, persist, persist, pipe, pipe, pipe, preferredLocations, randomSplit, reduce, repartition, retag, retag, sample, saveAsObjectFile, saveAsTextFile, saveAsTextFile, setName, sortBy, sparkContext, subtract, subtract, subtract, take, takeOrdered, takeSample, toArray, toDebugString, toJavaRDD, toLocalIterator, top, toString, union, unpersist, zip, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipWithIndex, zipWithUniqueIdfirstAvailableClass, newJobContext, newTaskAttemptContext, newTaskAttemptIDinitializeIfNecessary, initializeLogging, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarningpublic BinaryFileRDD(SparkContext sc, Class<? extends StreamFileInputFormat<T>> inputFormatClass, Class<String> keyClass, Class<T> valueClass, org.apache.hadoop.conf.Configuration conf, int minPartitions)
public Partition[] getPartitions()
RDDgetPartitions in class NewHadoopRDD<String,T>