Skip navigation

Standard Performance Evaluation Corporation

 
 

 

Benchmarking Breeds Controversy

Editor's Forum

By its very nature, graphics performance benchmarking is controversial; especially when you are doing it across different hardware architectures and APIs. How couldn't there be controversy? You are taking different systems with different performance assets and trying to apply a single criteria to measure performance. And not just any kind of performance, but graphics performance, arguably one of the most difficult characteristics of a computer system to measure.

The situation becomes even more heated when you try to take a series of performance numbers and boil them down to a single number that is supposed to be the final word on performance measurement. In actuality, everyone knows what the composite number is -- it's the performance equivalent of a sound bite (I'll leave it to the reader to supply the bite/byte pun).

The reason more than a dozen prominent vendors and researchers on the GPC Group take on this seemingly thankless task is simple: standardized performance measurement is desperately needed in the industry, by both vendors and users. But, just because it's a good cause doesn't mean everyone is going to agree about how to go about it. Good intentions don't necessarily lead to equanimity when it gets down to the dirty details.

Certainly the GPC Group has had its share of controversy. Compromises are reached, but the controversy never really goes away; it's endemic to performance benchmarking. As evidenced by the performance benchmarking results found in this publication, plenty of progress has been made. And, no doubt there will be plenty of progress to report over the upcoming months. But, as always, it's likely to be ushered in by controversy.