When performance testing practises mature, recognising the connection between response time and system capacity, insight into the nature of the correlation becomes important. As often as there is a significant gap between a performance test scenario and potential live scenarios, the value of performance testing takes a blow. Filling that gap cost-effectively requires insight, which can only be gathered by correlating the workload, response time and system utilization. At this point, creating a performance model is the only technique that fits the gap.
But performance modelling can be a complex process, filled with uncertainty and concerns about time investments. Often, it's better err on the side of simplicity rather than drive into the details, however tempting it may be. Leadership in this field is essential to derive cost/efficiencies out of this maturity transformation; there are many complexities lurking for the unwary.
In my company HyPerformix, one recent customer success story came at PepsiCo. Doug Taylor, Enterprise Test Centre Architect, recently said "if I can shrink the amount of work I do on performance testing, it not only saves hardware and software dollars, because the machines are smaller, but it also decreases the amount of time that the performance team has to build an environment, run and environment, and test it".
Doug was able to make significant efficiency savings by moving from a performance testing model to a performance engineering model, and reduce the amount of time taken in one case from 5 months to just 1 day. Don't believe it? Review Doug’s webinar online here