class StreamingTableWrite extends StreamingFlowExecution
A StreamingFlowExecution
that writes a streaming DataFrame
to a Table
.
- Source
- FlowExecution.scala
- Alphabetic
- By Inheritance
- StreamingTableWrite
- StreamingFlowExecution
- Logging
- FlowExecution
- AnyRef
- Any
- Hide All
- Show All
- Public
- Protected
Instance Constructors
- new StreamingTableWrite(identifier: TableIdentifier, flow: ResolvedFlow, graph: DataflowGraph, updateContext: PipelineUpdateContext, checkpointPath: String, trigger: Trigger, destination: Table, sqlConf: Map[String, String])
Type Members
- implicit class LogStringContext extends AnyRef
- Definition Classes
- Logging
Value Members
- final def !=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
- final def ##: Int
- Definition Classes
- AnyRef → Any
- final def ==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
- final def asInstanceOf[T0]: T0
- Definition Classes
- Any
- val checkpointPath: String
Structured Streaming checkpoint.
Structured Streaming checkpoint.
- Definition Classes
- StreamingTableWrite → StreamingFlowExecution
- def clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.CloneNotSupportedException]) @IntrinsicCandidate() @native()
- val destination: Table
The destination that this
FlowExecution
is writing to.The destination that this
FlowExecution
is writing to.- Definition Classes
- StreamingTableWrite → FlowExecution
- final def displayName: String
Returns a user-visible name for the flow.
Returns a user-visible name for the flow.
- Definition Classes
- FlowExecution
- final def eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
- def equals(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef → Any
- def exception: Option[Throwable]
Returns an optional exception that occurred during execution, if any.
Returns an optional exception that occurred during execution, if any.
- Definition Classes
- FlowExecution
- final def executeAsync(): Unit
Executes this FlowExecution asynchronously to perform its intended update.
Executes this FlowExecution asynchronously to perform its intended update. A future that can be used to track execution status is saved, and can be retrieved with
getFuture
.- Definition Classes
- FlowExecution
- final def executeInternal(): Future[Unit]
Executes this
StreamingFlowExecution
by starting its stream with the correct scheduling pool and confs.Executes this
StreamingFlowExecution
by starting its stream with the correct scheduling pool and confs.- returns
a Future that completes when the execution is finished or stopped.
- Definition Classes
- StreamingFlowExecution → FlowExecution
- implicit val executionContext: ExecutionContext
The thread execution context for the current
FlowExecution
.The thread execution context for the current
FlowExecution
.- Definition Classes
- FlowExecution
- val flow: ResolvedFlow
The
ResolvedFlow
that thisStreamingFlowExecution
is executing.The
ResolvedFlow
that thisStreamingFlowExecution
is executing.- Definition Classes
- StreamingTableWrite → StreamingFlowExecution
- final def getClass(): Class[_ <: AnyRef]
- Definition Classes
- AnyRef → Any
- Annotations
- @IntrinsicCandidate() @native()
- def getFuture: Future[ExecutionResult]
Retrieves the future that can be used to track execution status.
Retrieves the future that can be used to track execution status.
- Definition Classes
- FlowExecution
- def getOrigin: QueryOrigin
Origin to use when recording events for this flow.
Origin to use when recording events for this flow.
- Definition Classes
- StreamingTableWrite → FlowExecution
- val graph: DataflowGraph
- def hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @IntrinsicCandidate() @native()
- val identifier: TableIdentifier
Identifier of this physical flow
Identifier of this physical flow
- Definition Classes
- StreamingTableWrite → FlowExecution
- def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- Attributes
- protected
- Definition Classes
- Logging
- def initializeLogIfNecessary(isInterpreter: Boolean): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def isCompleted: Boolean
Returns true if and only if this
FlowExecution
has been completed with either success or an exception.Returns true if and only if this
FlowExecution
has been completed with either success or an exception.- Definition Classes
- FlowExecution
- final def isInstanceOf[T0]: Boolean
- Definition Classes
- Any
- def isStreaming: Boolean
Returns true iff this
FlowExecution
executes using Spark Structured Streaming.Returns true iff this
FlowExecution
executes using Spark Structured Streaming.- Definition Classes
- StreamingFlowExecution → FlowExecution
- def isTraceEnabled(): Boolean
- Attributes
- protected
- Definition Classes
- Logging
- def log: Logger
- Attributes
- protected
- Definition Classes
- Logging
- def logBasedOnLevel(level: Level)(f: => MessageWithContext): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logDebug(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logDebug(entry: LogEntry, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logDebug(entry: LogEntry): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logDebug(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logError(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logError(entry: LogEntry, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logError(entry: LogEntry): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logError(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logInfo(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logInfo(entry: LogEntry, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logInfo(entry: LogEntry): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logInfo(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logName: String
- Attributes
- protected
- Definition Classes
- Logging
- def logTrace(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logTrace(entry: LogEntry, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logTrace(entry: LogEntry): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logTrace(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logWarning(msg: => String, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logWarning(entry: LogEntry, throwable: Throwable): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logWarning(entry: LogEntry): Unit
- Attributes
- protected
- Definition Classes
- Logging
- def logWarning(msg: => String): Unit
- Attributes
- protected
- Definition Classes
- Logging
- final def ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
- final def notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @IntrinsicCandidate() @native()
- final def notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @IntrinsicCandidate() @native()
- def spark: classic.SparkSession
SparkSession to execute this physical flow with.
SparkSession to execute this physical flow with.
The default value for streaming flows is the pipeline's spark session because the source dataframe is resolved using the pipeline's spark session, and a new session will be started implicitly by the streaming query.
The default value for batch flows is a cloned spark session from the pipeline's spark session.
Please make sure that the execution thread runs in a different spark session than the pipeline's spark session.
- Attributes
- protected
- Definition Classes
- FlowExecution
- val sqlConf: Map[String, String]
Spark confs that must be set when starting this flow.
Spark confs that must be set when starting this flow.
- Definition Classes
- StreamingTableWrite → StreamingFlowExecution
- def startStream(): StreamingQuery
Starts a stream and returns its streaming query.
Starts a stream and returns its streaming query.
- Definition Classes
- StreamingTableWrite → StreamingFlowExecution
- def stop(): Unit
Stops execution of this
FlowExecution
.Stops execution of this
FlowExecution
. If you override this, please be sure to callsuper.stop()
at the beginning of your method, so we can properly handle errors when a user tries to stop a flow.- Definition Classes
- FlowExecution
- final def synchronized[T0](arg0: => T0): T0
- Definition Classes
- AnyRef
- def toString(): String
- Definition Classes
- AnyRef → Any
- val trigger: Trigger
Structured Streaming trigger.
Structured Streaming trigger.
- Definition Classes
- StreamingTableWrite → StreamingFlowExecution
- val updateContext: PipelineUpdateContext
Context about this pipeline update.
Context about this pipeline update.
- Definition Classes
- StreamingTableWrite → FlowExecution
- final def wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException])
- final def wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException]) @native()
- final def wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.InterruptedException])
- def withLogContext(context: Map[String, String])(body: => Unit): Unit
- Attributes
- protected
- Definition Classes
- Logging
Deprecated Value Members
- def finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws(classOf[java.lang.Throwable]) @Deprecated
- Deprecated
(Since version 9)