The AnTuTu Benchmark, a benchmarking tool for Android smartphones and tablets, has been revised following some discrepancy over whether the latest Intel Atom processor outperformed ARM-based chips from several vendors in key aspects of the benchmark. Under the revised benchmark, overall scores for the Atom Z2580 dropped by about 20 percent.
As I indicated in a post on EE Times earlier this week, there appeared to be some discrepancies with the AnTuTu benchmark in relation to older versions of the benchmark and other benchmarks. (See: Has Intel Really Beaten ARM?) Technical consulting firm BDTI pointed out that the compiled code for the Intel processor was not executing all instructions that were intended for the RAM test. This artificially improved the results for the Lenovo K900 smartphone and the Intel Atom processors. The problem appears to arise from the ICC compiler introduced around version 2.9.4 of the AnTuTu benchmark and used just for the Intel processors. Code for all other processors uses a different compiler, called the GCC.
To rectify the situation, AnTuTu issued revision 3.2.2 to the benchmark Wednesday evening. The revision still uses the ICC compiler, but the resulting scores are drastically different for the Intel processor. The AnTuTu CPU and overall scores dropped by approximately 20 percent, while the AnTuTu RAM score plummeted by approximately 50 percent, as shown in the figure below. All other scores for the Intel processor remain relatively the same. Similarly, scores for the Samsung S4 Oct and Qualcomm Snapdragon 600, both used in the Samsung Galaxy S4, remain relatively unchanged.

As a result of the revised scores, the AnTuTu benchmark is no longer an outlier from the other benchmarks and paints a competitive picture similar to the other benchmarks, with the Samsung Exynos 5 Octa processor outperforming the Intel Atom Z2580 processor (see the figure below). But, are these numbers valid? It is unclear what changes were made to revision 3.2.2 the benchmark.

However, AnTuTu did indicate that new testing standards would be issued in August, presumably with the next major revision of the benchmark. At that point, it will be interesting to see the new “standards” and the resulting changes in the benchmark scores.
Suddenly, the battle between smartphone processors looks much different than all the hype that was floating around because of the AnTuTu benchmark. It’s clear that the ARM processors still hold a significant advantage over the Intel processors, whether you include the new AnTuTu scores or just eliminate them completely from the evaluation.
The current and upcoming revisions to the AnTuTu benchmark will also drastically alter the scores of Intel’s upcoming Bay Trail Atom processor. As with the Atom Z2580, many in the press prematurely proclaimed it the victor over the next-generation ARM processors. Now that also appears to be highly questionable.
So, will all the sensationalistic bloggers retract their stories about Intel beating ARM? Likely not, but this definitely questions their credibility just as it does benchmarks. The moral of the story is that you have to question all benchmark data and use many data points from different sources before drawing a conclusion. It’s clear that Intel still has an uphill climb to catch the ARM camp, much less surpass it. It is also clear that we need better benchmarks for mobile devices that test for platform efficiency and usage models.
— Jim McGregor is founder and principal analyst at Tirias Research.
Related posts:




There are two target audiences for benchmarks: engineers and end-users.
Engineers have price/performance targets they want to meet when designing systems, and benchmarks are metrics that tell them where they are on meeting the targets.
End-users will see benchmarks as ways of comparing systems.
But in the smartphone market, where the AnTuTu bechmarks seem to be popular, the question is what decisions are made on the basis of them.
Smartphones are better seen as fashion accessories than as tech. Most buyers are interested in cool. Smartphones are status markers, and the incentive will be "My phone is cooler than yours!"
What makes a phone cool? Brand will be critical. iPhone buyers aren't buying iOS, they're buying Apple. There are scads of Android phones, so while Android is cool, it's a common denominator, and just running Android is not a deciding feature. Windows Phone is not cool, and that may be the biggest challenge Microsoft and Nokia have in getting a share of the market.
The smartphone market reminds me of the movies. In the movie business, you are as good as your last hit picture, and if your studio hits a dry patch and doesn't have hit pictures for a while, you may go out of business. The smartphone market is similar. Motorola had a hit with the Razr, didn't have a followup hit, and was rumored for a bit to be looking at getting out of the smartphone business.
I'd bet that most folks who run the AnTuTu benchmarks are measuring performance on the phone they already bought for other reasons, and aren't making a decision on which phone to buy based on them.
Given that, how much should anyone care what the AnTuTu benchmarks say?