Last week I wrote about the so-called “knee” in the M/M/m queueing theory response time curve. In that post I examined one definition of the knee; here is my analysis of the others, including the idea that there is no such thing as the knee.
There are potentially several ways to think about the “knee” in the queueing curve. In the previous post I dug into Cary Millsap’s definition: the knee is the point where a line tangent to the queueing curve passes through the origin:
Here are a few others to consider:
- The knee is where the curve has a slope of 1, that is, it’s growing as fast upwards as it is rightwards.
- The knee is the utilization at which you can expect requests to spend as much time waiting as being serviced, i.e. \( 2S \).
- The knee is related to tail latency, for example where the 99th percentile of the latency is double the median.
- The knee is where the curve’s derivative starts to change quickly, i.e. the wait time changes greatly with only a small change in utilization.
Let’s look at each of these in turn.
The Curve Has Slope=1
The first definition is the point at which the queueing curve has a slope of 1, at which point response times are increasing directly proportional to utilization \( \rho \). Past this point, response times increase faster than utilization, so this is a reasonable point to say the system “becomes superlinear” in the sense that linearity is where slope is 1.
Using the heuristic formula for the queueing theory response time curve, which again is exact for \( m=1 \) and \( m=2 \) service channels, and slightly underestimates qeueuing wait time for greater values of \( m \) but still serves as a reasonable approximation, we proceed as follows.
The derivative of that equation is
So the solution for \( \rho \) as a function of \( m \) is the real root between 0 and 1 of
Solving this for \( \rho \) turned out to be an interesting challenge which I did not complete. Among other things, this equation has very different behavior at \( m=1 \). In a single-server queueing system, the “knee” where the slope is 1 is at utilization of 0.
If this is a surprise to you, take another look at the queueing “hockey stick” curve. You’ll usually see it plotted with a stretched horizontal axis, because \( 0<\rho<1 \), and compressed vertical axis. That fools your eye! But when plotted without distortion, it’s quite clear that the queueing curve is on a diagonal right away when \( m=1 \). Here’s the queuing curve for 1, 2, and 8 servers (red, blue, and green lines).
For \( m=2 \) and greater, the solution is a lot of algebra. I resorted to using Wolfram Alpha. For 2 service channels, the equation reduces to
The solution of this is
Or \( \rho \approx 0.371507 \). This gets worse for \( m=3 \) and above, and there’s no single solution for all \( m \). It’s not too bad to plot, though; here’s an interactive Desmos calculator where you can see the queueing curve in blue, and its derivative in orange, as you vary \( m \). Where the derivative crosses the line \( y=1 \) is the “knee” by the definition I’m using in this section.
The behavior, although I can’t plot it with an equation, is not different from what you’ve seen previously: as you increase the number of service channels, the knee moves to the right. Notice how the orange line jumps when you set \( m=1 \).
I’d conclude this section by saying that I think this isn’t the right direction to take this discussion; the “knee” is supposed to be a kind of rule of thumb. It’s the point above which performance becomes “bad” by some definition. Clearly, “performance is bad above utilization of 0” isn’t useful.
Requests Spend As Much Time Waiting As In Service
The second definition is simply the point where \( R=2S \), or \( R/S=1 \); the solution of
The solution in terms of \( \rho \) turns out to be not too difficult:
Which is not too different, in fact, than the solution to Cary Millsap’s definition. Here are both plotted together, showing similar behavior. Blue is “wait equals service time” and red is Cary Millsap’s definition.
This definition is reasonable, in some sense, if you think “bad performance” is roughly equivalent to “my wait is longer than my job, that’s a waste.” But it’s still arbitrary. (More on this later.)
Define The Knee In Terms Of Tail Latency
As you can see, Cary Millsap’s definition is much more pessimistic, and the “wait equals service time” definition is more optimistic. But as I often say, tail latency is much more important than average latency, and the M/M/m queueing curve is in terms of averages. What if we define the knee as something like “the utilization at which the 99th percentile latency is twice the average service time?”
Although this would be possible to do, it would be tedious, because latency quantile calculations for M/M/m queueing systems are a bit intricate. Neil Gunther’s book delves into that a bit if you’re curious. Solving for \( \rho \) when quantile latency equals some multiple of service time, as a function of the number of service channels, would probably be harder than solving for the slope of the queueing curve.
Note also that the behavior in general would remain the same as the curves we’ve been seeing—as service channels increase, so too does the maximum utilization the system can bear before causing bad performance.
Define The Knee As Rapid Change In Residence Time
The final definition I proposed is that the knee might be where the curve’s derivative starts to change rapidly, showing that the wait time starts to be very sensitive to changes in utilization. The change in the curve’s derivative, of course, is its second derivative. When does this become large?
To answer this formally, I’d have to define “large” and I’d also have to solve the second derivative of the queueing curve for \( \rho \) as a function of \( m \) as usual. Here’s the second derivative:
This looks hard to solve, and I’m not sure what value I’d pick for “large;” it would necessarily be arbitrary. It turns out it is easy to avoid finding this answer by simply plotting the derivatives and observing that they all increase monotonically as utilization approaches 1. In other words, there’s no “turning point” where you can point at the graph and say “beyond this, things really get even more worser even more faster.”
This supports Neil Gunther’s view that there is no queueing knee. And if you really want to dig deeper into math and see even more potential definitions of the queueing knee, you should read that article. It goes into different types of considerations than I’ve done here.
This two-part series was in some sense an effort to point out that there’s no specific definition of knee in the M/M/m response time curve. Why was this worth doing? Put simply, because it continues to persist even after being analyzed and discussed by many performance experts. Not long ago a self-styled performance expert told me “as you know, past the knee at 75% utilization, the system starts to perform badly” and I had to bite my tongue. (To rub salt into the wound, this was a server with many CPU cores. What a waste.)
All knee definitions turn out to be arbitrary, but examining different definitions, inasmuch as they lend themselves to the algebra required, consistently points out that the “knee” grows larger as the number of service channels increases. (The exact shape and rate of that growth varies, but it always grows. This follows directly from Erlang’s formulas.)
In the end, that principle is the only takeaway from this. There’s no knee, but the more CPUs (or servers) you have, the more fully utilized you should expect to run that system. You’re foolish to do otherwise.