Package org.apache.spark.sql.connector.read.streaming
package org.apache.spark.sql.connector.read.streaming
-
ClassDescriptionIndicates that the source accepts the latest seen offset, which requires streaming execution to provide the latest seen offset when restarting the streaming query from checkpoint./** Represents a
ReadLimit
where theMicroBatchStream
should scan approximately given maximum number of rows with at least the given minimum number of rows.A variation onPartitionReader
for use with continuous streaming processing.A variation onPartitionReaderFactory
that returnsContinuousPartitionReader
instead ofPartitionReader
.ASparkDataStream
for streaming queries with continuous mode.ASparkDataStream
for streaming queries with micro-batch mode.An abstract representation of progress through aMicroBatchStream
orContinuousStream
.Used for per-partition offsets in continuous processing.Represents aReadLimit
where theMicroBatchStream
must scan all the data available at the streaming source.Interface representing limits on how much to read from aMicroBatchStream
when it implementsSupportsAdmissionControl
.Represents aReadLimit
where theMicroBatchStream
should scan approximately the given maximum number of files.Represents aReadLimit
where theMicroBatchStream
should scan approximately the given maximum number of rows.Represents aReadLimit
where theMicroBatchStream
should scan approximately at least the given minimum number of rows.A mix-in interface for streaming sinks to signal that they can report metrics.A mix-in interface forSparkDataStream
streaming sources to signal that they can report metrics.The base interface representing a readable data stream in a Spark streaming query.A mix-in interface forSparkDataStream
streaming sources to signal that they can control the rate of data ingested into the system.An interface for streaming sources that supports running in Trigger.AvailableNow mode, which will process all the available data at the beginning of the query in (possibly) multiple batches.