This is a very useful way of thinking about responses and latency, and a great example of a domain with a lot of very specialized knowledge (industrial engineering and logistics) that probably doesn't overall have enough cross-talk into CS given how much we talk about queues, latency, and workers.<p>I'm going to have to brush up on my IE one of these days to try to figure out if there are some useful insights to potentially be gleaned from models that account for server-level concurrency more directly, along the lines of <a href="https://en.wikipedia.org/wiki/M/M/c_queue" rel="nofollow">https://en.wikipedia.org/wiki/M/M/c_queue</a>
Instead of giving the entire dataset to the model (which in this case already identifies a performance limit visually), give a partial set and then see if these algorithms can predict the remaining known values. That should be the real test of whether prediction works.<p>Here's what I found last time I did it: <a href="https://github.com/brendangregg/PerfModels/blob/master/scale.pdf" rel="nofollow">https://github.com/brendangregg/PerfModels/blob/master/scale...</a><p>ie, I'm splitting the dataset into "model input" and "extra measurements". USL was not really working.
Everyone interested in this should have a read of the Neil Gunther's own books: <a href="http://www.perfdynamics.com/books.html" rel="nofollow">http://www.perfdynamics.com/books.html</a>