Status
Current state: Under Discussion
Discussion thread: here
JIRA: KAFKA-9983
Please keep the discussion on the mailing list rather than commenting on the wiki (wiki discussions get unwieldy fast).
Motivation
Currently the actual end-to-end latency of a record flowing through Streams is difficult to gauge at best. It's also difficult to build real-time applications without some sense of the latency you can expect between the time that an event occurs and when this event is processed and reflected in the output results. Being able to bound this latency is an important requisite for many apps, and exposing this through metrics should go a long way towards enabling users to make the right design choices.
For example, application builders may be producing events in reaction to user actions. Without a sense of how long it will take for this new event to be reflected in the results, how do you know what to do next – should they show a brief loading screen before issuing an interactive query, or will they need to wait hours for this record to be processed? (Hopefully not). Without end-to-end latency metrics this question is difficult to answer.
Public Interfaces
The following metrics would be added:
- record-staleness-max [ms]
- record-staleness-p99 [ms] (99th percentile)
- record-staleness-p75 [ms] (75th percentile)
- record-staleness-avg [ms] (mean)
- record-staleness-min [ms]
These will be exposed on the task-level at the recording level INFO with the following tags
- type = stream-task-metrics
- thread-id=[threadId]
- task-id=[taskId]
We will also expose these metrics on the operator-level at the recording level DEBUG with the following tags:
- type = stream-processor-node-metrics
- thread-id=[threadId]
- task-id=[taskId]
- processor-node-id=[processorNodeId]
Proposed Changes
Imagine a simple 3-node subtopology with source node O, filter node F, and sink node I. For any record flowing through this with record timestamp t, let tO be the system (wallclock) time when it is read from the source topic, tF be the time when it is processed by the filter, and tI be the time when it reaches the sink node. The staleness at operator O for a given record is defined as
SO (t) = tO - t
and likewise for the other operator-level end-to-end latencies. This represents the age of the record at the time is was received by operator O. The task-level staleness S will be computed based on the source node, ie S = SO. The source nodes reading from the user input topics therefore represent the consumption latency, the time it took for a newly-created event to be read by Streams. This can be especially interesting in cases where some records may be severely delayed: for example by a IoT device with unstable network connections, or when a user's smartphone reconnects to the internet after a flight and pushes all the latest updates.
Note that for a given record, SO <= SF <= SI . This holds true within and across subtopologies. A downstream subtopology will always have a task-level end-to-end latency staleness greater than or equal to that of an upstream subtopology for a single task, which in turn implies the same holds true for the statistical measures exposed via the new metrics. Comparing the staleness across tasks (or across operators) will also be of interest as this represents the processing delay: the amount of time it took for Streams to actually process the record from point A to point B within the topology. Seeing a large processing delay indicates a possible bottleneck in the topology and may help users debug their application performance. (Debugging topology bottlenecks is not the primary motivation of this KIP, but it is a nice side effect.)
Late arriving records will be included in this metric, even if they are otherwise dropped due to the grace period having passed. Although we already expose a metric for the number of late dropped records, there is no way for a user to find out how late the record was. Including them in the staleness metrics may for one thing help users to set a reasonable grace period if they see that a large number of records are being dropped. Another slight difference with the concept of late records is that they are dropped based on stream time, whereas this metric is always reported with respect to the current time. The stream-time may lag the system time in low traffic, for example, or when all records are considerably delayed. This might mean the user sees no dropped records even though the staleness is large.
Compatibility, Deprecation, and Migration Plan
N/A
Rejected Alternatives
Using stream time
We define the staleness in terms of individual record timestamps, but we could have instead it as the difference between the system time and the stream time, ie SO = tO - st where st is the stream-time. This approach has some drawbacks; first and foremost, we are losing information in the case of out-of-order data, and may not notice records with extreme delays. In the example above, an IoT device that regularly disconnects for long periods of time would push a lot of out-of-order data when it reconnects. Since these records would not advance or effect the stream-time, this delay would not be reflected as an increase in the processing latency metric. But the end-to-end latency of these records is of course quite high