Interface | Description |
---|---|
JdbcRDD.ConnectionFactory | |
PartitionCoalescer |
::DeveloperApi::
A PartitionCoalescer defines how to coalesce the partitions of a given RDD.
|
Class | Description |
---|---|
AsyncRDDActions<T> |
A set of asynchronous RDD actions available through an implicit conversion.
|
CheckpointState |
Enumeration to manage state transitions of an RDD through checkpointing
|
CoGroupedRDD<K> |
:: DeveloperApi ::
An RDD that cogroups its parents.
|
DefaultPartitionCoalescer |
Coalesce the partitions of a parent RDD (
prev ) into fewer partitions, so that each partition of
this RDD computes one or more of the parent ones. |
DeterministicLevel |
The deterministic level of RDD's output (i.e.
|
DoubleRDDFunctions |
Extra functions available on RDDs of Doubles through an implicit conversion.
|
HadoopRDD<K,V> |
:: DeveloperApi ::
An RDD that provides core functionality for reading data stored in Hadoop (e.g., files in HDFS,
sources in HBase, or S3), using the older MapReduce API (
org.apache.hadoop.mapred ). |
HadoopRDD.HadoopMapPartitionsWithSplitRDD$ | |
InputFileBlockHolder |
This holds file names of the current Spark task.
|
JdbcRDD<T> |
An RDD that executes a SQL query on a JDBC connection and reads results.
|
NewHadoopRDD<K,V> |
:: DeveloperApi ::
An RDD that provides core functionality for reading data stored in Hadoop (e.g., files in HDFS,
sources in HBase, or S3), using the new MapReduce API (
org.apache.hadoop.mapreduce ). |
NewHadoopRDD.NewHadoopMapPartitionsWithSplitRDD$ | |
OrderedRDDFunctions<K,V,P extends scala.Product2<K,V>> |
Extra functions available on RDDs of (key, value) pairs where the key is sortable through
an implicit conversion.
|
PairRDDFunctions<K,V> |
Extra functions available on RDDs of (key, value) pairs through an implicit conversion.
|
PartitionGroup |
::DeveloperApi::
A group of
Partition s
param: prefLoc preferred location for the partition group |
PartitionPruningRDD<T> |
:: DeveloperApi ::
An RDD used to prune RDD partitions/partitions so we can avoid launching tasks on
all partitions.
|
RDD<T> |
A Resilient Distributed Dataset (RDD), the basic abstraction in Spark.
|
RDDBarrier<T> |
:: Experimental ::
Wraps an RDD in a barrier stage, which forces Spark to launch tasks of this stage together.
|
SequenceFileRDDFunctions<K,V> |
Extra functions available on RDDs of (key, value) pairs to create a Hadoop SequenceFile,
through an implicit conversion.
|
ShuffledRDD<K,V,C> |
:: DeveloperApi ::
The resulting RDD from a shuffle (e.g.
|
UnionRDD<T> |