Latency Monitoring #
Redict is often used for demanding use cases, where it serves a large number of queries per second per instance, but also has strict latency requirements for the average response time and the worst-case latency.
While Redict is an in-memory system, it deals with the operating system in different ways, for example, in the context of persisting to disk. Moreover Redict implements a rich set of commands. Certain commands are fast and run in constant or logarithmic time. Other commands are slower O(N) commands that can cause latency spikes.
Finally, Redict is single threaded. This is usually an advantage from the point of view of the amount of work it can perform per core, and in the latency figures it is able to provide. However, it poses a challenge for latency, since the single thread must be able to perform certain tasks incrementally, for example key expiration, in a way that does not impact the other clients that are served.
Latency Monitoring helps the user to check and troubleshoot possible latency problems. Latency monitoring is composed of the following conceptual parts:
- Latency hooks that sample different latency-sensitive code paths.
- Time series recording of latency spikes, split by different events.
- Reporting engine to fetch raw data from the time series.
- Analysis engine to provide human-readable reports and hints according to the measurements.
The rest of this document covers the latency monitoring subsystem details. For more information about the general topic of Redict and latency, see Redict latency problems troubleshooting.
Events and time series #
Different monitored code paths have different names and are called events.
For example, command
is an event that measures latency spikes of possibly slow
command executions, while fast-command
is the event name for the monitoring
of the O(1) and O(log N) commands. Other events are less generic and monitor
specific operations performed by Redict. For example, the fork
event
only monitors the time taken by Redict to execute the fork(2)
system call.
A latency spike is an event that takes more time to run than the configured latency threshold. There is a separate time series associated with every monitored event. This is how the time series work:
- Every time a latency spike happens, it is logged in the appropriate time series.
- Every time series is composed of 160 elements.
- Each element is a pair made of a Unix timestamp of the time the latency spike was measured and the number of milliseconds the event took to execute.
- Latency spikes for the same event that occur in the same second are merged by taking the maximum latency. Even if continuous latency spikes are measured for a given event, which could happen with a low threshold, at least 160 seconds of history are available.
- Records the all-time maximum latency for every element.
The framework monitors and logs latency spikes in the execution time of these events:
command
: regular commands.fast-command
: O(1) and O(log N) commands.fork
: thefork(2)
system call.rdb-unlink-temp-file
: theunlink(2)
system call.aof-fsync-always
: thefsync(2)
system call when invoked by theappendfsync allways
policy.aof-write
: writing to the AOF - a catchall event forwrite(2)
system calls.aof-write-pending-fsync
: thewrite(2)
system call when there is a pending fsync.aof-write-active-child
: thewrite(2)
system call when there are active child processes.aof-write-alone
: thewrite(2)
system call when no pending fsync and no active child process.aof-fstat
: thefstat(2)
system call.aof-rename
: therename(2)
system call for renaming the temporary file after completingBGREWRITEAOF
.aof-rewrite-diff-write
: writing the differences accumulated while performingBGREWRITEAOF
.active-defrag-cycle
: the active defragmentation cycle.expire-cycle
: the expiration cycle.eviction-cycle
: the eviction cycle.eviction-del
: deletes during the eviction cycle.
How to enable latency monitoring #
What is high latency for one use case may not be considered high latency for another. Some applications may require that all queries be served in less than 1 millisecond. For other applications, it may be acceptable for a small amount of clients to experience a 2 second latency on occasion.
The first step to enable the latency monitor is to set a latency threshold in milliseconds. Only events that take longer than the specified threshold will be logged as latency spikes. The user should set the threshold according to their needs. For example, if the application requires a maximum acceptable latency of 100 milliseconds, the threshold should be set to log all the events blocking the server for a time equal or greater to 100 milliseconds.
Enable the latency monitor at runtime in a production server with the following command:
CONFIG SET latency-monitor-threshold 100
Monitoring is turned off by default (threshold set to 0), even if the actual cost of latency monitoring is near zero. While the memory requirements of latency monitoring are very small, there is no good reason to raise the baseline memory usage of a Redict instance that is working well.
Report information with the LATENCY command #
The user interface to the latency monitoring subsystem is the LATENCY
command.
Like many other Redict commands, LATENCY
accepts subcommands that modify its behavior. These subcommands are:
LATENCY LATEST
- returns the latest latency samples for all events.LATENCY HISTORY
- returns latency time series for a given event.LATENCY RESET
- resets latency time series data for one or more events.LATENCY GRAPH
- renders an ASCII-art graph of an event’s latency samples.LATENCY DOCTOR
- replies with a human-readable latency analysis report.
Refer to each subcommand’s documentation page for further information.