org.apache.spark.api.r
Class BaseRRDD<T,U>

Object
  extended by org.apache.spark.rdd.RDD<U>
      extended by org.apache.spark.api.r.BaseRRDD<T,U>
All Implemented Interfaces:
java.io.Serializable, Logging
Direct Known Subclasses:
PairwiseRRDD, RRDD, StringRRDD

public abstract class BaseRRDD<T,U>
extends RDD<U>
implements Logging

See Also:
Serialized Form

Constructor Summary
BaseRRDD(RDD<T> parent, int numPartitions, byte[] func, String deserializer, String serializer, byte[] packageNames, String rLibDir, Broadcast<Object>[] broadcastVars, scala.reflect.ClassTag<T> evidence$1, scala.reflect.ClassTag<U> evidence$2)
           
 
Method Summary
 scala.collection.Iterator<U> compute(Partition partition, TaskContext context)
          :: DeveloperApi :: Implemented by subclasses to compute a given partition.
 Partition[] getPartitions()
          Implemented by subclasses to return the set of partitions in this RDD.
 
Methods inherited from class org.apache.spark.rdd.RDD
aggregate, cache, cartesian, checkpoint, checkpointData, coalesce, collect, collect, context, count, countApprox, countApproxDistinct, countApproxDistinct, countByValue, countByValueApprox, creationSite, dependencies, distinct, distinct, doubleRDDToDoubleRDDFunctions, filter, filterWith, first, flatMap, flatMapWith, fold, foreach, foreachPartition, foreachWith, getCheckpointFile, getStorageLevel, glom, groupBy, groupBy, groupBy, id, intersection, intersection, intersection, isCheckpointed, isEmpty, iterator, keyBy, map, mapPartitions, mapPartitionsWithContext, mapPartitionsWithIndex, mapPartitionsWithSplit, mapWith, max, min, name, numericRDDToDoubleRDDFunctions, partitioner, partitions, persist, persist, pipe, pipe, pipe, preferredLocations, randomSplit, rddToAsyncRDDActions, rddToOrderedRDDFunctions, rddToPairRDDFunctions, rddToSequenceFileRDDFunctions, reduce, repartition, sample, saveAsObjectFile, saveAsTextFile, saveAsTextFile, scope, setName, sortBy, sparkContext, subtract, subtract, subtract, take, takeOrdered, takeSample, toArray, toDebugString, toJavaRDD, toLocalIterator, top, toString, treeAggregate, treeReduce, union, unpersist, zip, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipWithIndex, zipWithUniqueId
 
Methods inherited from class Object
equals, getClass, hashCode, notify, notifyAll, wait, wait, wait
 
Methods inherited from interface org.apache.spark.Logging
initializeIfNecessary, initializeLogging, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
 

Constructor Detail

BaseRRDD

public BaseRRDD(RDD<T> parent,
                int numPartitions,
                byte[] func,
                String deserializer,
                String serializer,
                byte[] packageNames,
                String rLibDir,
                Broadcast<Object>[] broadcastVars,
                scala.reflect.ClassTag<T> evidence$1,
                scala.reflect.ClassTag<U> evidence$2)
Method Detail

getPartitions

public Partition[] getPartitions()
Description copied from class: RDD
Implemented by subclasses to return the set of partitions in this RDD. This method will only be called once, so it is safe to implement a time-consuming computation in it.

Returns:
(undocumented)

compute

public scala.collection.Iterator<U> compute(Partition partition,
                                            TaskContext context)
Description copied from class: RDD
:: DeveloperApi :: Implemented by subclasses to compute a given partition.

Specified by:
compute in class RDD<U>
Parameters:
partition - (undocumented)
context - (undocumented)
Returns:
(undocumented)