org.apache.spark.graphx.impl
Class EdgeRDDImpl<ED,VD>

Object
  extended by org.apache.spark.rdd.RDD<Edge<ED>>
      extended by org.apache.spark.graphx.EdgeRDD<ED>
          extended by org.apache.spark.graphx.impl.EdgeRDDImpl<ED,VD>
All Implemented Interfaces:
java.io.Serializable, Logging

public class EdgeRDDImpl<ED,VD>
extends EdgeRDD<ED>

See Also:
Serialized Form

Method Summary
 EdgeRDDImpl<ED,VD> cache()
          Persists the edge partitions using `targetStorageLevel`, which defaults to MEMORY_ONLY.
 void checkpoint()
          Mark this RDD for checkpointing.
 Edge<ED>[] collect()
          Return an array that contains all of the elements in this RDD.
 long count()
          The number of edges in the RDD.
 EdgeRDDImpl<ED,VD> filter(scala.Function1<EdgeTriplet<VD,ED>,Object> epred, scala.Function2<Object,VD,Object> vpred)
           
 scala.Option<String> getCheckpointFile()
          Gets the name of the file to which this RDD was checkpointed
 StorageLevel getStorageLevel()
          Get the RDD's current storage level, or StorageLevel.NONE if none is set.
<ED2,ED3> EdgeRDDImpl<ED3,VD>
innerJoin(EdgeRDD<ED2> other, scala.Function4<Object,Object,ED,ED2,ED3> f, scala.reflect.ClassTag<ED2> evidence$4, scala.reflect.ClassTag<ED3> evidence$5)
          Inner joins this EdgeRDD with another EdgeRDD, assuming both are partitioned using the same PartitionStrategy.
 boolean isCheckpointed()
          Return whether this RDD has been checkpointed or not
<ED2,VD2> EdgeRDDImpl<ED2,VD2>
mapEdgePartitions(scala.Function2<Object,org.apache.spark.graphx.impl.EdgePartition<ED,VD>,org.apache.spark.graphx.impl.EdgePartition<ED2,VD2>> f, scala.reflect.ClassTag<ED2> evidence$6, scala.reflect.ClassTag<VD2> evidence$7)
           
<ED2> EdgeRDDImpl<ED2,VD>
mapValues(scala.Function1<Edge<ED>,ED2> f, scala.reflect.ClassTag<ED2> evidence$3)
          Map the values in an edge partitioning preserving the structure but changing the values.
 scala.Option<Partitioner> partitioner()
          If partitionsRDD already has a partitioner, use it.
 RDD<scala.Tuple2<Object,org.apache.spark.graphx.impl.EdgePartition<ED,VD>>> partitionsRDD()
           
 EdgeRDDImpl<ED,VD> persist(StorageLevel newLevel)
          Persists the edge partitions at the specified storage level, ignoring any existing target storage level.
 EdgeRDDImpl<ED,VD> reverse()
          Reverse all the edges in this RDD.
 EdgeRDDImpl<ED,VD> setName(String _name)
          Assign a name to this RDD
 StorageLevel targetStorageLevel()
           
 EdgeRDDImpl<ED,VD> unpersist(boolean blocking)
          Mark the RDD as non-persistent, and remove all blocks for it from memory and disk.
 
Methods inherited from class org.apache.spark.graphx.EdgeRDD
compute, fromEdges
 
Methods inherited from class org.apache.spark.rdd.RDD
aggregate, cartesian, checkpointData, coalesce, collect, context, countApprox, countApproxDistinct, countApproxDistinct, countByValue, countByValueApprox, creationSite, dependencies, distinct, distinct, doubleRDDToDoubleRDDFunctions, filter, filterWith, first, flatMap, flatMapWith, fold, foreach, foreachPartition, foreachWith, glom, groupBy, groupBy, groupBy, id, intersection, intersection, intersection, isEmpty, iterator, keyBy, map, mapPartitions, mapPartitionsWithContext, mapPartitionsWithIndex, mapPartitionsWithSplit, mapWith, max, min, name, numericRDDToDoubleRDDFunctions, partitions, persist, pipe, pipe, pipe, preferredLocations, randomSplit, rddToAsyncRDDActions, rddToOrderedRDDFunctions, rddToPairRDDFunctions, rddToSequenceFileRDDFunctions, reduce, repartition, sample, saveAsObjectFile, saveAsTextFile, saveAsTextFile, scope, sortBy, sparkContext, subtract, subtract, subtract, take, takeOrdered, takeSample, toArray, toDebugString, toJavaRDD, toLocalIterator, top, toString, treeAggregate, treeReduce, union, zip, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipPartitions, zipWithIndex, zipWithUniqueId
 
Methods inherited from class Object
equals, getClass, hashCode, notify, notifyAll, wait, wait, wait
 
Methods inherited from interface org.apache.spark.Logging
initializeIfNecessary, initializeLogging, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
 

Method Detail

partitionsRDD

public RDD<scala.Tuple2<Object,org.apache.spark.graphx.impl.EdgePartition<ED,VD>>> partitionsRDD()

targetStorageLevel

public StorageLevel targetStorageLevel()

setName

public EdgeRDDImpl<ED,VD> setName(String _name)
Description copied from class: RDD
Assign a name to this RDD

Overrides:
setName in class RDD<Edge<ED>>

partitioner

public scala.Option<Partitioner> partitioner()
If partitionsRDD already has a partitioner, use it. Otherwise assume that the PartitionIDs in partitionsRDD correspond to the actual partitions and create a new partitioner that allows co-partitioning with partitionsRDD.

Overrides:
partitioner in class RDD<Edge<ED>>
Returns:
(undocumented)

collect

public Edge<ED>[] collect()
Description copied from class: RDD
Return an array that contains all of the elements in this RDD.

Overrides:
collect in class RDD<Edge<ED>>
Returns:
(undocumented)

persist

public EdgeRDDImpl<ED,VD> persist(StorageLevel newLevel)
Persists the edge partitions at the specified storage level, ignoring any existing target storage level.

Overrides:
persist in class RDD<Edge<ED>>
Parameters:
newLevel - (undocumented)
Returns:
(undocumented)

unpersist

public EdgeRDDImpl<ED,VD> unpersist(boolean blocking)
Description copied from class: RDD
Mark the RDD as non-persistent, and remove all blocks for it from memory and disk.

Overrides:
unpersist in class RDD<Edge<ED>>
Parameters:
blocking - Whether to block until all blocks are deleted.
Returns:
This RDD.

cache

public EdgeRDDImpl<ED,VD> cache()
Persists the edge partitions using `targetStorageLevel`, which defaults to MEMORY_ONLY.

Overrides:
cache in class RDD<Edge<ED>>

getStorageLevel

public StorageLevel getStorageLevel()
Description copied from class: RDD
Get the RDD's current storage level, or StorageLevel.NONE if none is set.

Overrides:
getStorageLevel in class RDD<Edge<ED>>

checkpoint

public void checkpoint()
Description copied from class: RDD
Mark this RDD for checkpointing. It will be saved to a file inside the checkpoint directory set with SparkContext.setCheckpointDir() and all references to its parent RDDs will be removed. This function must be called before any job has been executed on this RDD. It is strongly recommended that this RDD is persisted in memory, otherwise saving it on a file will require recomputation.

Overrides:
checkpoint in class RDD<Edge<ED>>

isCheckpointed

public boolean isCheckpointed()
Description copied from class: RDD
Return whether this RDD has been checkpointed or not

Overrides:
isCheckpointed in class RDD<Edge<ED>>
Returns:
(undocumented)

getCheckpointFile

public scala.Option<String> getCheckpointFile()
Description copied from class: RDD
Gets the name of the file to which this RDD was checkpointed

Overrides:
getCheckpointFile in class RDD<Edge<ED>>
Returns:
(undocumented)

count

public long count()
The number of edges in the RDD.

Overrides:
count in class RDD<Edge<ED>>
Returns:
(undocumented)

mapValues

public <ED2> EdgeRDDImpl<ED2,VD> mapValues(scala.Function1<Edge<ED>,ED2> f,
                                           scala.reflect.ClassTag<ED2> evidence$3)
Description copied from class: EdgeRDD
Map the values in an edge partitioning preserving the structure but changing the values.

Specified by:
mapValues in class EdgeRDD<ED>
Parameters:
f - the function from an edge to a new edge value
evidence$3 - (undocumented)
Returns:
a new EdgeRDD containing the new edge values

reverse

public EdgeRDDImpl<ED,VD> reverse()
Description copied from class: EdgeRDD
Reverse all the edges in this RDD.

Specified by:
reverse in class EdgeRDD<ED>
Returns:
a new EdgeRDD containing all the edges reversed

filter

public EdgeRDDImpl<ED,VD> filter(scala.Function1<EdgeTriplet<VD,ED>,Object> epred,
                                 scala.Function2<Object,VD,Object> vpred)

innerJoin

public <ED2,ED3> EdgeRDDImpl<ED3,VD> innerJoin(EdgeRDD<ED2> other,
                                               scala.Function4<Object,Object,ED,ED2,ED3> f,
                                               scala.reflect.ClassTag<ED2> evidence$4,
                                               scala.reflect.ClassTag<ED3> evidence$5)
Description copied from class: EdgeRDD
Inner joins this EdgeRDD with another EdgeRDD, assuming both are partitioned using the same PartitionStrategy.

Specified by:
innerJoin in class EdgeRDD<ED>
Parameters:
other - the EdgeRDD to join with
f - the join function applied to corresponding values of this and other
evidence$4 - (undocumented)
evidence$5 - (undocumented)
Returns:
a new EdgeRDD containing only edges that appear in both this and other, with values supplied by f

mapEdgePartitions

public <ED2,VD2> EdgeRDDImpl<ED2,VD2> mapEdgePartitions(scala.Function2<Object,org.apache.spark.graphx.impl.EdgePartition<ED,VD>,org.apache.spark.graphx.impl.EdgePartition<ED2,VD2>> f,
                                                        scala.reflect.ClassTag<ED2> evidence$6,
                                                        scala.reflect.ClassTag<VD2> evidence$7)