Waits for the termination of this
query, either by query.stop()
or by an exception.
Waits for the termination of this
query, either by query.stop()
or by an exception.
If the query has terminated with an exception, then the exception will be thrown.
Otherwise, it returns whether the query has terminated or not within the timeoutMs
milliseconds.
If the query has terminated, then all subsequent calls to this method will either return
true
immediately (if the query was terminated by stop()
), or throw the exception
immediately (if the query has terminated with exception).
2.0.0
if this
query has terminated with an exception
Waits for the termination of this
query, either by query.stop()
or by an exception.
Waits for the termination of this
query, either by query.stop()
or by an exception.
If the query has terminated with an exception, then the exception will be thrown.
If the query has terminated, then all subsequent calls to this method will either return
immediately (if the query was terminated by stop()
), or throw the exception
immediately (if the query has terminated with exception).
2.0.0
if this
query has terminated with an exception.
Tracks how much data we have processed and committed to the sink or state store from each input source.
Tracks how much data we have processed and committed to the sink or state store from each input source. Only the scheduler thread should modify this field, and only in atomic steps. Other threads should make a shallow copy if they are going to access this field more than once, since the field's value may change at any time.
Returns the StreamingQueryException if the query was terminated by an exception.
Returns the StreamingQueryException if the query was terminated by an exception.
Prints the physical plan to the console for debugging purposes.
Prints the physical plan to the console for debugging purposes.
2.0.0
Prints the physical plan to the console for debugging purposes.
Prints the physical plan to the console for debugging purposes.
whether to do extended explain or not
2.0.0
Expose for tests
Returns the unique id of this query.
Returns the unique id of this query. This id is automatically generated and is unique across all queries that have been started in the current process.
2.0.0
Whether the query is currently active or not
Whether the query is currently active or not
The thread that runs the micro-batches of this stream.
The thread that runs the micro-batches of this stream. Note that this thread must be org.apache.spark.util.UninterruptibleThread to avoid potential deadlocks in using HDFSMetadataLog. See SPARK-14131 for more details.
Returns the name of the query.
Returns the name of the query. This name is unique across all active queries. This can be
set in the DataStreamWriter as
dataframe.writeStream.queryName("query").start()
.
2.0.0
A write-ahead-log that records the offsets that are present in each batch.
A write-ahead-log that records the offsets that are present in each batch. In order to ensure that a given batch will always consist of the same data, we write to this log *before* any processing is done. Thus, the Nth record in this log indicated data that is currently being processed and the N-1th entry indicates which offsets have been durably committed to the sink.
Blocks until all available data in the source has been processed and committed to the sink.
Blocks until all available data in the source has been processed and committed to the sink.
This method is intended for testing. Note that in the case of continually arriving data, this
method may block forever. Additionally, this method is only guaranteed to block until data that
has been synchronously appended data to a org.apache.spark.sql.execution.streaming.Source
prior to invocation. (i.e. getOffset
must immediately reflect the addition).
2.0.0
Returns current status of the sink.
Returns current status of the sink.
Returns current status of all the sources.
Returns current status of all the sources.
Returns the SparkSession associated with this
.
Returns the SparkSession associated with this
.
2.0.0
Starts the execution.
Starts the execution. This returns only after the thread has started and QueryStartedEvent has been posted to all the listeners.
Returns the current status of the query.
Returns the current status of the query.
Signals to the thread executing micro-batches that it should stop running after the next batch.
Signals to the thread executing micro-batches that it should stop running after the next batch. This method blocks until the thread stops running.
Manages the execution of a streaming Spark SQL query that is occurring in a separate thread. Unlike a standard query, a streaming query executes repeatedly each time new data arrives at any Source present in the query plan. Whenever new data arrives, a QueryExecution is created and the results are committed transactionally to the given Sink.