3 Key Performance Testing Metrics Every Tester Should Know

Dive into the core of performance testing metrics and discover the importance of accurate analysis in ensuring optimal system performance. Through simple explanations, you’ll be guided toward making informed decisions in your testing endeavors.

Embarking on performance testing unveils three crucial metrics: Average, Standard Deviation, and Percentiles. Each offers unique insights, painting a comprehensive picture of system performance. Through a thoughtful analysis of these metrics, we lay the foundation for informed decision-making and enhanced system responsiveness.

Find out why Performance Testing Services are the Master Key to Improving Your Software in this article.

Making Sense of The Average, Standard Deviation, and Percentiles in Performance Testing Reports

There are certain performance testing metrics that are essential to understand properly in order to draw the right conclusions from your performance tests. These performance metrics require some basic understanding of math and statistics, but nothing too complicated.

The issue is that if you don’t understand well what each one means or what they represent, you’ll come to some very wrong conclusions. 

In this post, I want to focus on average response time, standard deviation, and percentiles. Without going into a lot of math, we’ll discuss test metrics and their usefulness when analyzing performance results.

The Importance of Analyzing Data as a Graph

The first time I thought about this subject was during a course that Scott Barber gave in 2008 (when we were just starting up Abstracta), on his visit to Uruguay. He showed us a table with values like this:

Performance testing data chart

He asked us which data set we thought had the best performance, which is not quite as easy to discern as when you display the data in a graph:

Performance testing metrics: Data Set A

In Set A, you can tell there was a peak, but then it recovers.

Performance testing metrics: Data Set B

In Set B, it seems that it started out with a very poor response time, and probably 20 seconds into testing, the system collapsed and began to respond to an error page, which then got resolved in a second.

Performance testing metrics: Data Set C

Finally, in Set C, it’s clear that as time passed, the system performance continued to degrade.

Barber’s aim with this exercise was to show that it’s much easier to analyze information when it’s presented in a graph. In addition, in the table, the information is summarized, but in the graphs, you can see all the points. Thus, with more data points, we can gain a clearer picture of what is going on.

Understanding Key Performance Testing Metrics

Okay, now let’s see what each of the metrics for performance testing means, as a key part of your performance testing process. Let’s do it one by one, checking their importance for analysis purposes.

Average Response Time

To calculate the average, simply add up all the values of the samples and then divide that number by the quantity of samples.

Let’s say I do this and my resulting average peak response time is 3 seconds. The problem with this is that, at face value, it gives you a false sense that all response times are about three seconds, some a little more and some a little less, but that might not be the case.

Imagine I had three samples, the first two with a response time of one second, the third with a response time of seven:

1 + 1 + 7 = 9

9/3 = 3

This is a very simple example that shows that three very different values could result in an average of three, yet the individual values may not be anywhere close to 3.

My colleague, Fabian Baptista, made a funny comment related to this:

“If I were to put one hand in a bucket of water at -100 degrees Fahrenheit and another hand in a bucket of burning lava, on average, my hand temperature would be fine, but I’d lose both of my hands.” 

So, when analyzing average response time, it’s possible to have a result that’s within the acceptable level, but be careful with the conclusions you reach.

That’s why it is not recommended to define service level agreements (SLAs) using averages; instead, have something like “The service must respond in less than 1 second for 99% of cases.” We’ll see more about this later with the percentile metric.

Don’t miss this Quality Sense episode about why observability is such relevant in software testing, with Federico Toledo and Lisa Crispin.

Standard Deviation

Standard deviation is a measure of dispersion with respect to the average, how much the values vary with respect to their average, or how far apart they are.

If the value of the standard deviation is small, this indicates that all the values of the samples are close to the average, but if it’s large, then they are far apart and have a greater range.

To understand how to interpret this value, let’s look at a couple of examples.

If all the values are equal, then the standard deviation is 0. If there are very scattered values, for example, consider 9 samples with values from 1 to 9 (1, 2, 3, 4, 5, 6, 7, 8, 9), the standard deviation is ~ 2.6 (you can use this online calculator to calculate it).

Although the value of the average as a metric can be greatly improved by also including the standard deviation, what’s more useful yet are the percentile values.


A percentile is a very useful performance testing metric that gives a measure under which a percentage of the sample is found.

For instance, the 90th percentile (abbreviated as p90) indicates that 90% of the sample is below that value and the rest of the values (that is, the other 10%) are above it.

It’s practical to analyze more than one percentile value (JMeter and Gatling do so in their reports), showing how many samples are p90, p95, and p99.

If these are complemented with the minimum, maximum, and average, then it’s possible to have a much more complete view of the data and thus understand how the system behaves.

Some percentiles have particular names, such as p100 which is the maximum (100% of the data is below this value), and p50, which is the median (half of the data is below it and half are above).

The percentile is typically used to establish acceptance criteria, indicating that 90% of the sample should be below a certain value. This is done to rule out outliers.

Careful with Performance Testing Metrics

So, before you go analyzing your next performance test’s results, make sure to remember these key considerations:

  • Never consider the average as “the” value to pay attention to, since it can be deceiving, as it often hides important information.
  • Consider the standard deviation to know just how useful the average is, the higher the standard deviation, the less meaningful it is.
  • Observe the percentile values and define acceptance criteria based on that, keeping in mind that if you select the 90th percentile, you’re basically saying, “I don’t care if 10% of my users experience bad response times”

If you are interested in learning about the best continuous performance testing practices for improving your system’s performance, we invite you to read this article.

What other considerations and performance issues do you have when analyzing performance testing metrics? Let us know!

Reaching for open-source software, or a free performance load testing tool? Get to know JMeter .Net DSL, one of the leading open-source performance testing tools, bridging JMeter and . NET. It revolutionizes performance testing and open-source tools, enhancing software quality, efficiency, and reliability.

Looking for a Quality Partner for Performance Testing?

Embrace agility and cost-effectiveness through our Performance Testing Services, hand in hand with our performance test engineers.

Are you unsure of which type of performance tests you need? 

We are quality partners! Learn more about our solutions here and understand why we have received such praise. Contact us to discuss how we can help you grow your business.

Follow us on Linkedin & X to be part of our community!

379 / 446