org.apache

spark

package spark

Core Spark functionality. org.apache.spark.SparkContext serves as the main entry point to Spark, while org.apache.spark.rdd.RDD is the data type representing a distributed collection, and provides most parallel operations.

In addition, org.apache.spark.rdd.PairRDDFunctions contains operations available only on RDDs of key-value pairs, such as groupByKey and join; org.apache.spark.rdd.DoubleRDDFunctions contains operations available only on RDDs of Doubles; and org.apache.spark.rdd.SequenceFileRDDFunctions contains operations available on RDDs that can be saved as SequenceFiles. These operations are automatically available on any RDD of the right type (e.g. RDD[(Int, Int)] through implicit conversions when you import org.apache.spark.SparkContext._.

Java programmers should reference the org.apache.spark.api.java package for Spark programming APIs in Java.

Classes and methods marked with Experimental are user-facing features which have not been officially adopted by the Spark project. These are subject to change or removal in minor releases.

Classes and methods marked with Developer API are intended for advanced users want to extend Spark through lower level interfaces. These are subject to changes or removal in minor releases.

Linear Supertypes
AnyRef, Any
Ordering
  1. Alphabetic
  2. By inheritance
Inherited
  1. spark
  2. AnyRef
  3. Any
  1. Hide All
  2. Show all
Learn more about member selection
Visibility
  1. Public
  2. All

Type Members

  1. class Accumulable[R, T] extends Serializable

    A data type that can be accumulated, ie has an commutative and associative "add" operation, but where the result type, R, may be different from the element type being added, T.

  2. trait AccumulableParam[R, T] extends Serializable

    Helper object defining how to accumulate values of a particular type.

  3. class Accumulator[T] extends Accumulable[T, T]

    A simpler value of Accumulable where the result type being accumulated is the same as the types of elements being merged, i.

  4. trait AccumulatorParam[T] extends AccumulableParam[T, T]

    A simpler version of org.apache.spark.AccumulableParam where the only data type you can add in is the same type as the accumulated value.

  5. case class Aggregator[K, V, C](createCombiner: (V) ⇒ C, mergeValue: (C, V) ⇒ C, mergeCombiners: (C, C) ⇒ C) extends Product with Serializable

    :: DeveloperApi :: A set of functions used to aggregate data.

  6. class ComplexFutureAction[T] extends FutureAction[T]

    :: Experimental :: A FutureAction for actions that could trigger multiple Spark jobs.

  7. abstract class Dependency[T] extends Serializable

    :: DeveloperApi :: Base class for dependencies.

  8. case class ExceptionFailure(className: String, description: String, stackTrace: Array[StackTraceElement], metrics: Option[TaskMetrics]) extends TaskFailedReason with Product with Serializable

    :: DeveloperApi :: Task failed due to a runtime exception.

  9. case class FetchFailed(bmAddress: BlockManagerId, shuffleId: Int, mapId: Int, reduceId: Int) extends TaskFailedReason with Product with Serializable

    :: DeveloperApi :: Task failed to fetch shuffle data from a remote node.

  10. trait FutureAction[T] extends Future[T]

    :: Experimental :: A future for the result of an action to support cancellation.

  11. class HashPartitioner extends Partitioner

    A org.apache.spark.Partitioner that implements hash-based partitioning using Java's Object.hashCode.

  12. class InterruptibleIterator[+T] extends Iterator[T]

    :: DeveloperApi :: An iterator that wraps around an existing iterator to provide task killing functionality.

  13. trait Logging extends AnyRef

    :: DeveloperApi :: Utility trait for classes that want to log data.

  14. abstract class NarrowDependency[T] extends Dependency[T]

    :: DeveloperApi :: Base class for dependencies where each partition of the child RDD depends on a small number of partitions of the parent RDD.

  15. class OneToOneDependency[T] extends NarrowDependency[T]

    :: DeveloperApi :: Represents a one-to-one dependency between partitions of the parent and child RDDs.

  16. trait Partition extends Serializable

    A partition of an RDD.

  17. abstract class Partitioner extends Serializable

    An object that defines how the elements in a key-value pair RDD are partitioned by key.

  18. class RangeDependency[T] extends NarrowDependency[T]

    :: DeveloperApi :: Represents a one-to-one dependency between ranges of partitions in the parent and child RDDs.

  19. class RangePartitioner[K, V] extends Partitioner

    A org.apache.spark.Partitioner that partitions sortable records by range into roughly equal ranges.

  20. class SerializableWritable[T <: Writable] extends Serializable

    Annotations
    @DeveloperApi()
  21. class ShuffleDependency[K, V, C] extends Dependency[Product2[K, V]]

    :: DeveloperApi :: Represents a dependency on the output of a shuffle stage.

  22. class SimpleFutureAction[T] extends FutureAction[T]

    :: Experimental :: A FutureAction holding the result of an action that triggers a single job.

  23. class SparkConf extends Cloneable with Logging

    Configuration for a Spark application.

  24. class SparkContext extends Logging

    Main entry point for Spark functionality.

  25. class SparkEnv extends Logging

    :: DeveloperApi :: Holds all the runtime environment objects for a running Spark instance (either master or worker), including the serializer, Akka actor system, block manager, map output tracker, etc.

  26. class SparkException extends Exception

  27. class TaskContext extends Serializable

    :: DeveloperApi :: Contextual information about a task which can be read or mutated during execution.

  28. sealed trait TaskEndReason extends AnyRef

    :: DeveloperApi :: Various possible reasons why a task ended.

  29. sealed trait TaskFailedReason extends TaskEndReason

    :: DeveloperApi :: Various possible reasons why a task failed.

  30. class TaskKilledException extends RuntimeException

    :: DeveloperApi :: Exception thrown when a task is explicitly killed (i.

Value Members

  1. object ExecutorLostFailure extends TaskFailedReason with Product with Serializable

    :: DeveloperApi :: The task failed because the executor that it was running on was lost.

  2. object Partitioner extends Serializable

  3. object Resubmitted extends TaskFailedReason with Product with Serializable

    :: DeveloperApi :: A org.apache.spark.scheduler.ShuffleMapTask that completed successfully earlier, but we lost the executor before the stage completed.

  4. object SparkContext extends Logging

    The SparkContext object contains a number of implicit conversions and parameters for use with various Spark features.

  5. object SparkEnv extends Logging

  6. object SparkFiles

    Resolves paths to files added through SparkContext.addFile().

  7. object Success extends TaskEndReason with Product with Serializable

    :: DeveloperApi :: Task succeeded.

  8. object TaskKilled extends TaskFailedReason with Product with Serializable

    :: DeveloperApi :: Task was killed intentionally and needs to be rescheduled.

  9. object TaskResultLost extends TaskFailedReason with Product with Serializable

    :: DeveloperApi :: The task finished successfully, but the result was lost from the executor's block manager before it was fetched.

  10. object UnknownReason extends TaskFailedReason with Product with Serializable

    :: DeveloperApi :: We don't know why the task ended -- for example, because of a ClassNotFound exception when deserializing the task result.

  11. package annotation

    Spark annotations to mark an API experimental or intended only for advanced usages by developers.

  12. package api

  13. package bagel

    Bagel: An implementation of Pregel in Spark.

  14. package broadcast

    Spark's broadcast variables, used to broadcast immutable datasets to all nodes.

  15. package deploy

  16. package examples

  17. package executor

    Executor components used with various cluster managers.

  18. package graphx

    ALPHA COMPONENT GraphX is a graph processing framework built on top of Spark.

  19. package io

    IO codecs used for compression.

  20. package mllib

    Spark's machine learning library.

  21. package network

  22. package partial

    :: Experimental ::

  23. package rdd

    Provides several RDD implementations.

  24. package scheduler

    Spark's scheduling components.

  25. package serializer

    Pluggable serializers for RDD and shuffle data.

  26. package sql

    Allows the execution of relational queries, including those expressed in SQL using Spark.

  27. package storage

  28. package streaming

    Spark Streaming functionality.

  29. package ui

  30. package util

    Spark utilities.

Inherited from AnyRef

Inherited from Any

Ungrouped