Package org.apache.spark.rdd
Class ShuffledRDD<K,V,C>  
Object
org.apache.spark.rdd.RDD<scala.Tuple2<K,C>>
 
org.apache.spark.rdd.ShuffledRDD<K,V,C>  
- Type Parameters:
- K- the key class.
- V- the value class.
- C- the combiner class.
- All Implemented Interfaces:
- Serializable,- org.apache.spark.internal.Logging
:: DeveloperApi ::
 The resulting RDD from a shuffle (e.g. repartitioning of data).
 param:  prev the parent RDD.
 param:  part the partitioner used to partition the RDD
- See Also:
- 
Nested Class SummaryNested classes/interfaces inherited from interface org.apache.spark.internal.Loggingorg.apache.spark.internal.Logging.LogStringContext, org.apache.spark.internal.Logging.SparkShellLoggingFilter
- 
Constructor SummaryConstructorsConstructorDescriptionShuffledRDD(RDD<? extends scala.Product2<K, V>> prev, Partitioner part, scala.reflect.ClassTag<K> evidence$1, scala.reflect.ClassTag<V> evidence$2, scala.reflect.ClassTag<C> evidence$3) 
- 
Method SummaryModifier and TypeMethodDescriptionvoidcompute(Partition split, TaskContext context) :: DeveloperApi :: Implemented by subclasses to compute a given partition.scala.collection.immutable.Seq<Dependency<?>>scala.Some<Partitioner>Optionally overridden by subclasses to specify how they are partitioned.prev()ShuffledRDD<K,V, C> setAggregator(Aggregator<K, V, C> aggregator) Set aggregator for RDD's shuffle.ShuffledRDD<K,V, C> setKeyOrdering(scala.math.Ordering<K> keyOrdering) Set key ordering for RDD's shuffle.ShuffledRDD<K,V, C> setMapSideCombine(boolean mapSideCombine) Set mapSideCombine flag for RDD's shuffle.ShuffledRDD<K,V, C> setSerializer(Serializer serializer) Set a serializer for this RDD's shuffle, or null to use the default (spark.serializer)Methods inherited from class org.apache.spark.rdd.RDDaggregate, barrier, cache, cartesian, checkpoint, cleanShuffleDependencies, coalesce, collect, collect, context, count, countApprox, countApproxDistinct, countApproxDistinct, countByValue, countByValueApprox, dependencies, distinct, distinct, doubleRDDToDoubleRDDFunctions, filter, first, flatMap, fold, foreach, foreachPartition, getCheckpointFile, getNumPartitions, getResourceProfile, getStorageLevel, glom, groupBy, groupBy, groupBy, id, intersection, intersection, intersection, isCheckpointed, isEmpty, iterator, keyBy, localCheckpoint, map, mapPartitions, mapPartitionsWithEvaluator, mapPartitionsWithIndex, max, min, name, numericRDDToDoubleRDDFunctions, partitions, persist, persist, pipe, pipe, pipe, preferredLocations, randomSplit, rddToAsyncRDDActions, rddToOrderedRDDFunctions, rddToPairRDDFunctions, rddToSequenceFileRDDFunctions, reduce, repartition, sample, saveAsObjectFile, saveAsTextFile, saveAsTextFile, setName, sortBy, sparkContext, subtract, subtract, subtract, take, takeOrdered, takeSample, toDebugString, toJavaRDD, toLocalIterator, top, toString, treeAggregate, treeAggregate, treeReduce, union, unpersist, withResources, zip, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitionsWithEvaluator, zipWithIndex, zipWithUniqueIdMethods inherited from class java.lang.Objectequals, getClass, hashCode, notify, notifyAll, wait, wait, waitMethods inherited from interface org.apache.spark.internal.LogginginitializeForcefully, initializeLogIfNecessary, initializeLogIfNecessary, initializeLogIfNecessary$default$2, isTraceEnabled, log, logBasedOnLevel, logDebug, logDebug, logDebug, logDebug, logError, logError, logError, logError, logInfo, logInfo, logInfo, logInfo, logName, LogStringContext, logTrace, logTrace, logTrace, logTrace, logWarning, logWarning, logWarning, logWarning, MDC, org$apache$spark$internal$Logging$$log_, org$apache$spark$internal$Logging$$log__$eq, withLogContext
- 
Constructor Details- 
ShuffledRDD
 
- 
- 
Method Details- 
clearDependenciespublic void clearDependencies()
- 
computeDescription copied from class:RDD:: DeveloperApi :: Implemented by subclasses to compute a given partition.
- 
getDependencies
- 
getPartitions
- 
partitionerDescription copied from class:RDDOptionally overridden by subclasses to specify how they are partitioned.- Overrides:
- partitionerin class- RDD<scala.Tuple2<K,- C>> 
 
- 
prev
- 
setAggregatorSet aggregator for RDD's shuffle.
- 
setKeyOrderingSet key ordering for RDD's shuffle.
- 
setMapSideCombineSet mapSideCombine flag for RDD's shuffle.
- 
setSerializerSet a serializer for this RDD's shuffle, or null to use the default (spark.serializer)
 
-