DataGen Connector #
The DataGen connector provides a
Source implementation that allows for generating input data for
It is useful when developing locally or demoing without access to external systems such as Kafka.
The DataGen connector is built-in, no additional dependencies are required.
DataGeneratorSource produces N data points in parallel. The source splits the sequence
into as many parallel sub-sequences as there are parallel source subtasks. It drives the data
generation process by supplying “index” values of type
Long to the user-provided
GeneratorFunction is then used for mapping the (sub-)sequences of
into the generated events of an arbitrary data type. For instance, the following code will produce the sequence of
["Number: 0", "Number: 1", ... , "Number: 999"] records.
GeneratorFunction<Long, String> generatorFunction = index -> "Number: " + index; long numberOfRecords = 1000; DataGeneratorSource<String> source = new DataGeneratorSource<>(generatorFunction, numberOfRecords, Types.STRING); DataStreamSource<String> stream = env.fromSource(source, WatermarkStrategy.noWatermarks(), "Generator Source");
The order of elements depends on the parallelism. Each sub-sequence will be produced in order.
Consequently, if the parallelism is limited to one, this will produce one sequence in order from
"Number: 0" to
Rate Limiting #
DataGeneratorSource has built-in support for rate limiting. The following code will produce a stream of
Long values at the overall source rate (across all source subtasks) not exceeding 100 events per second.
GeneratorFunction<Long, Long> generatorFunction = index -> index; double recordsPerSecond = 100; DataGeneratorSource<String> source = new DataGeneratorSource<>( generatorFunction, Long.MAX_VALUE, RateLimiterStrategy.perSecond(recordsPerSecond), Types.STRING);
Additional rate limiting strategies, such as limiting the number of records emitted per checkpoint, can be found in RateLimiterStrategy .
This source is always bounded. From a practical perspective, however, setting the number of records
Long.MAX_VALUE turns it into an effectively unbounded source (the end will never be reached). For finite sequences users may want to consider running the application in
BATCH execution mode
DataGeneratorSourcecan be used to implement Flink jobs with at-least-once and end-to-end exactly-once processing guarantees under the condition that the output of the
GeneratorFunctionis deterministic with respect to its input, in other words supplying the same
Longnumber always leads to generating the same output.
Note: it is possible to also produce deterministic watermarks right at the source based on the generated events and a custom WatermarkStrategy .