How it works
is a critical part of Artie’s data pipeline that determines when and how data gets written to your destination.Data buffering
- Artie’s reading process will read changes from your source database and publish them to Kafka
- Artie’s writing process will read messages from Kafka and write them to your destination
- Messages are temporarily stored in memory and deduplicated based on primary key(s) or unique index
- Multiple changes to the same record are merged to reduce write volume
Flush trigger evaluation
- Artie continuously monitors three flush conditions
- When any condition is met, a flush is triggered
- Reading from Kafka pauses during the flush operation
Conditions
Artie evaluates three conditions to determine when to flush data. Any one of these conditions will trigger a flush:Time elapsed
Maximum time in seconds — Ensures data freshness even during low-volume periods
Message count
Number of deduplicated messages — Based on unique primary keys or unique index.
Byte size
Total bytes of deduplicated data — Actual payload size after deduplication