This may seem a bit ‘back to basics’, but it’s a topic of discussions I’ve been having recently.
When presenting performance metrics, folks sometimes use mean, sometimes median, sometimes both, sometimes they include standard deviation, percentiles, etc. I’ve been looking for some concrete guidance on what metrics to use, and in what contexts.
In my search, I came across a fantastic book (ironically, written by an old college professor or mine), called The Art of Computer
Systems Performance Analysis.
I specifically want to call out chapter 12, on Summarizing Measured Data (slides can be download at the bottom of the page). Don’t get too caught up in all the math – you can still get a lot out of it without understanding every equation.
Here’s my interpretation/summarization of the chapter.
When attempting to summarize data with a single number, we want to show both indices of central tendencies, and indices of dispersion. In other words, when measuring Web page performance, what’s the most common experience, and how variable is the data.
Web performance data is not normally distributed, but is positively skewed (i.e. has a long tail), as is common with computer response times. Because the Mean is so impacted by the tail, it can be far from the central tendency. The Median, while not perfect, is closer to the central tendency for skewed distributions, so is the preferred metric. This is demonstrated on slide 15 and 16, with more details in surrounding slides.
Correspondingly, the suggestion for indicating data variability (dispersion) is either percentiles or SIQR (Semi-Interquartile Range, defined on slide 41). Again, this is due to response time data having a skewed distribution. Refer to slide 45, with more details in surrounding slides.