Re-create a StreamingContext from a checkpoint file.
Re-create a StreamingContext from a checkpoint file.
Path either to the directory that was specified as the checkpoint directory, or to the checkpoint file 'graph' or 'graph.bk'.
Create a StreamingContext by providing the details necessary for creating a new SparkContext.
Create a StreamingContext by providing the details necessary for creating a new SparkContext.
Cluster URL to connect to (e.g. mesos://host:port, spark://host:port, local[4]).
A name for your job, to display on the cluster web UI
The time interval at which streaming data will be divided into batches
Create a StreamingContext using an existing SparkContext.
Create a StreamingContext using an existing SparkContext.
Existing SparkContext
The time interval at which streaming data will be divided into batches
Create an input stream with any arbitrary user implemented actor receiver.
Create an input stream with any arbitrary user implemented actor receiver.
Props object defining creation of the actor
Name of the actor
RDD storage level. Defaults to memory-only.
An important point to note: Since Actor may exist outside the spark framework, It is thus user's responsibility to ensure the type safety, i.e parametrized type of data received and actorStream should be same.
Set the context to periodically checkpoint the DStream operations for master fault-tolerance.
Set the context to periodically checkpoint the DStream operations for master fault-tolerance. The graph will be checkpointed every batch interval.
HDFS-compatible directory where the checkpoint data will be reliably stored
Create a input stream that monitors a Hadoop-compatible filesystem for new files and reads them using the given key-value types and input format.
Create a input stream that monitors a Hadoop-compatible filesystem for new files and reads them using the given key-value types and input format.
Key type for reading HDFS file
Value type for reading HDFS file
Input format for reading HDFS file
HDFS directory to monitor for new file
Function to filter paths to process
Should process only new files and ignore existing files in the directory
Create a input stream that monitors a Hadoop-compatible filesystem for new files and reads them using the given key-value types and input format.
Create a input stream that monitors a Hadoop-compatible filesystem for new files and reads them using the given key-value types and input format. File names starting with . are ignored.
Key type for reading HDFS file
Value type for reading HDFS file
Input format for reading HDFS file
HDFS directory to monitor for new file
Create a input stream from a Flume source.
Create a input stream from a Flume source.
Hostname of the slave machine to which the flume data will be sent
Port of the slave machine to which the flume data will be sent
Storage level to use for storing the received objects
Create an input stream with any arbitrary user implemented network receiver.
Create an input stream with any arbitrary user implemented network receiver.
Custom implementation of NetworkReceiver
Create an input stream from a queue of RDDs.
Create an input stream from a queue of RDDs. In each batch, it will process either one or all of the RDDs returned by the queue.
Type of objects in the RDD
Queue of RDDs
Whether only one RDD should be consumed from the queue in every interval
Default RDD is returned by the DStream when the queue is empty. Set as null if no RDD should be returned when empty
Create an input stream from a queue of RDDs.
Create an input stream from a queue of RDDs. In each batch, it will process either one or all of the RDDs returned by the queue.
Type of objects in the RDD
Queue of RDDs
Whether only one RDD should be consumed from the queue in every interval
Create a input stream from network source hostname:port, where data is received as serialized blocks (serialized using the Spark's serializer) that can be directly pushed into the block manager without deserializing them.
Create a input stream from network source hostname:port, where data is received as serialized blocks (serialized using the Spark's serializer) that can be directly pushed into the block manager without deserializing them. This is the most efficient way to receive data.
Type of the objects in the received blocks
Hostname to connect to for receiving data
Port to connect to for receiving data
Storage level to use for storing the received objects
Register an input stream that will be started (InputDStream.
Register an input stream that will be started (InputDStream.start() called) to get the input data.
Register an output stream that will be computed every interval
Set each DStreams in this context to remember RDDs it generated in the last given duration.
Set each DStreams in this context to remember RDDs it generated in the last given duration. DStreams remember RDDs only for a limited duration of time and releases them for garbage collection. This method allows the developer to specify how to long to remember the RDDs ( if the developer wishes to query old data outside the DStream computation).
Minimum duration that each DStream should remember its RDDs
Create a input stream from TCP source hostname:port.
Create a input stream from TCP source hostname:port. Data is received using a TCP socket and the receive bytes it interepreted as object using the given converter.
Type of the objects received (after converting bytes to objects)
Hostname to connect to for receiving data
Port to connect to for receiving data
Function to convert the byte stream to objects
Storage level to use for storing the received objects
Create a input stream from TCP source hostname:port.
Create a input stream from TCP source hostname:port. Data is received using
a TCP socket and the receive bytes is interpreted as UTF8 encoded \n
delimited
lines.
Hostname to connect to for receiving data
Port to connect to for receiving data
Storage level to use for storing the received objects (default: StorageLevel.MEMORY_AND_DISK_SER_2)
Return the associated Spark context
Start the execution of the streams.
Stop the execution of the streams.
Create a input stream that monitors a Hadoop-compatible filesystem for new files and reads them as text files (using key as LongWritable, value as Text and input format as TextInputFormat).
Create a input stream that monitors a Hadoop-compatible filesystem for new files and reads them as text files (using key as LongWritable, value as Text and input format as TextInputFormat). File names starting with . are ignored.
HDFS directory to monitor for new file
Create a input stream that returns tweets received from Twitter.
Create a input stream that returns tweets received from Twitter.
Twitter4J authentication, or None to use Twitter4J's default OAuth authorization; this uses the system properties twitter4j.oauth.consumerKey, .consumerSecret, .accessToken and .accessTokenSecret.
Set of filter strings to get only those tweets that match them
Storage level to use for storing the received objects
Create a unified DStream from multiple DStreams of the same type and same interval
Create an input stream that receives messages pushed by a zeromq publisher.
Create an input stream that receives messages pushed by a zeromq publisher.
Url of remote zeromq publisher
topic to subscribe to
A zeroMQ stream publishes sequence of frames for each topic and each frame has sequence of byte thus it needs the converter (which might be deserializer of bytes) to translate from sequence of sequence of bytes, where sequence refer to a frame and sub sequence refer to its payload.
RDD storage level. Defaults to memory-only.
A StreamingContext is the main entry point for Spark Streaming functionality. Besides the basic information (such as, cluster URL and job name) to internally create a SparkContext, it provides methods used to create DStream from various input sources.