package streaming
Type Members
-
trait
ContinuousPartitionReader[T] extends PartitionReader[T]
A variation on
PartitionReaderfor use with continuous streaming processing.A variation on
PartitionReaderfor use with continuous streaming processing.- Since
3.0.0
-
trait
ContinuousPartitionReaderFactory extends PartitionReaderFactory
A variation on
PartitionReaderFactorythat returnsContinuousPartitionReaderinstead ofPartitionReader.A variation on
PartitionReaderFactorythat returnsContinuousPartitionReaderinstead ofPartitionReader. It's used for continuous streaming processing.- Since
3.0.0
-
trait
ContinuousStream extends SparkDataStream
A
SparkDataStreamfor streaming queries with continuous mode.A
SparkDataStreamfor streaming queries with continuous mode.- Since
3.0.0
-
trait
MicroBatchStream extends SparkDataStream
A
SparkDataStreamfor streaming queries with micro-batch mode.A
SparkDataStreamfor streaming queries with micro-batch mode.- Since
3.0.0
-
abstract
class
Offset extends AnyRef
An abstract representation of progress through a
MicroBatchStreamorContinuousStream.An abstract representation of progress through a
MicroBatchStreamorContinuousStream. During execution, offsets provided by the data source implementation will be logged and used as restart checkpoints. Each source should provide an offset implementation which the source can use to reconstruct a position in the stream up to which data has been seen/processed.- Since
3.0.0
-
trait
PartitionOffset extends Serializable
Used for per-partition offsets in continuous processing.
Used for per-partition offsets in continuous processing. ContinuousReader implementations will provide a method to merge these into a global Offset.
These offsets must be serializable.
- Since
3.0.0
-
final
class
ReadAllAvailable extends ReadLimit
Represents a
ReadLimitwhere theMicroBatchStreammust scan all the data available at the streaming source.Represents a
ReadLimitwhere theMicroBatchStreammust scan all the data available at the streaming source. This is meant to be a hard specification as being able to return all available data is necessary for Trigger.Once() to work correctly. If a source is unable to scan all available data, then it must throw an error.- Since
3.0.0
- See also
SupportsAdmissionControl#latestOffset(Offset, ReadLimit)
-
trait
ReadLimit extends AnyRef
Interface representing limits on how much to read from a
MicroBatchStreamwhen it implementsSupportsAdmissionControl.Interface representing limits on how much to read from a
MicroBatchStreamwhen it implementsSupportsAdmissionControl. There are several child interfaces representing various kinds of limits.- Since
3.0.0
- See also
SupportsAdmissionControl#latestOffset(Offset, ReadLimit)
ReadAllAvailable
ReadMaxRows
-
class
ReadMaxFiles extends ReadLimit
Represents a
ReadLimitwhere theMicroBatchStreamshould scan approximately the given maximum number of files.Represents a
ReadLimitwhere theMicroBatchStreamshould scan approximately the given maximum number of files.- Since
3.0.0
- See also
SupportsAdmissionControl#latestOffset(Offset, ReadLimit)
-
final
class
ReadMaxRows extends ReadLimit
Represents a
ReadLimitwhere theMicroBatchStreamshould scan approximately the given maximum number of rows.Represents a
ReadLimitwhere theMicroBatchStreamshould scan approximately the given maximum number of rows.- Since
3.0.0
- See also
SupportsAdmissionControl#latestOffset(Offset, ReadLimit)
-
trait
SparkDataStream extends AnyRef
The base interface representing a readable data stream in a Spark streaming query.
The base interface representing a readable data stream in a Spark streaming query. It's responsible to manage the offsets of the streaming source in the streaming query.
Data sources should implement concrete data stream interfaces:
MicroBatchStreamandContinuousStream.- Since
3.0.0
-
trait
SupportsAdmissionControl extends SparkDataStream
A mix-in interface for
SparkDataStreamstreaming sources to signal that they can control the rate of data ingested into the system.A mix-in interface for
SparkDataStreamstreaming sources to signal that they can control the rate of data ingested into the system. These rate limits can come implicitly from the contract of triggers, e.g. Trigger.Once() requires that a micro-batch process all data available to the system at the start of the micro-batch. Alternatively, sources can decide to limit ingest through data source options.Through this interface, a MicroBatchStream should be able to return the next offset that it will process until given a
ReadLimit.- Since
3.0.0