BenchmarkXPRT Blog banner

Tag Archives: HDXPRT

Gearing up for a busy year ahead

We hope everyone’s 2018 kicked off on a happy note, and you’re starting the year rested, refreshed, and inspired. Here at the XPRTs, we already have a busy slate of activity lined up, and we want to share a quick preview of what community members can expect in the coming months.

Next week, I’ll be travelling to CES 2018 in Las Vegas. CES provides us with a great opportunity to survey emerging tech and industry trends, and I look forward to sharing my thoughts and impressions from the show. If you’re attending this year, and would like to meet and discuss any aspect of the XPRTs, let me know.

There’s also more WebXPRT news to come. We’re working on several new features for the WebXPRT Processor Comparison Chart that we think will prove to be useful, and we hope to take the updated chart live very soon. We’re also getting closer to the much-anticipated WebXPRT 3 general release! If you’ve been testing the WebXPRT 3 Community Preview, be sure to send in your feedback soon.

Work on the next version of HDXPRT is progressing as well, and we’ll share more details about UI and workload updates as we get closer to a community preview build.

Last but not least, we’re considering the prospect of updating TouchXPRT and MobileXPRT later in the year. We look forward to working with the community on improved versions of each of those benchmarks.

Justin

Our community’s goal

Computer system performance evaluation has a long and complex history. Many of the earliest tests were simple, short code snippets, such as Whetstone, that did little more than give an indication of how fast a particular computer subsystem was able to operate. Unfortunately, such simple benchmarks quickly lost their value, in part because they were very crude measures, and in part because software tools on the things they were measuring could easily optimize for them. In some cases, a compiler could even recognize a test and “optimize” the code by simply producing the final result!

Over time, though, benchmarks have become more complex and more relevant. Whole organizations exist and have existed to build benchmarks. Notable ones include the Ziff-Davis Benchmark Operation (ZDBOp), which the Ziff-Davis computer magazines funded in the 1990s and which Mark and I ran; the Standard Performance Evaluation Corporation (SPEC), which its member companies fund and of which PT is a member; and the Business Applications Performance Corporation (BAPCo), which its member companies fund. Each of these organizations has developed widely used products, such as Winstone (ZDBOp), SPEC CPU (SPEC), and SYSmark (BAPCo). Each organization has also always faced challenges. In the case of ZDBOp, for example, Ziff Davis could no longer support the costs of developing its benchmarks, so they discontinued the group. SPEC continues to develop good benchmarks, but its process can sometimes yield years between versions.

The goal with HDXPRT and the HDXPRT Development Community (HDC) is to explore a new way to develop benchmarks. By utilizing the expertise and experience of a community of interested people, we hope to be able develop benchmarks in an open and collaborative environment while keeping them timely.

HDXPRT 2011 is the first test of this approach. We believe that it and subsequent versions of it, as well as other benchmarks, will give the industry a new model for creating world-class performance measurement tools.

If you’re not a member of the HDC, please consider joining us and helping define the future of performance evaluation.

Bill

Comment on this post in the forums

Check out the other XPRTs:

Forgot your password?