The Alberta Workloads for the SPEC CPU 2017 Benchmark Suite
Citations Over TimeTop 14% of 2018 papers
Abstract
A proper evaluation of techniques that require multiple training and evaluation executions of a benchmark, such as Feedback-Directed Optimization (FDO), requires multiple workloads that can be used to characterize variations on the behaviour of a program based on the workload. This paper aims to improve the performance evaluation of computer systems - including compilers, computer architecture simulation, and operating-system prototypes - that rely on the industrystandard SPEC CPU benchmark suite. A main concern with the use of this suite in research is that it is distributed with a very small number of workloads. This paper describes the process to create additional workloads for this suite and offers useful insights in many of its benchmarks. The set of additional workloads created, named the Alberta Workloads for the SPEC CPU 2017 Benchmark Suite1 is made freely available with the goal of providing additional data points for the exploration of learning in computing systems. These workloads should also contribute to ameliorate the hidden learning problem where a researcher sets parameters to a system during development based on a set of benchmarks and then evaluates the system using the very same set of benchmarks with the very same workloads.
Related Papers
- → Reflections on the creation of a real-time parallel benchmark suite(1999)1 cited
- → New CPU benchmark suites from SPEC(2003)39 cited
- → Perspector: Benchmarking Benchmark Suites(2023)1 cited
- → An evaluation of the Meiko CS-2 using the GENESIS Benchmark Suite(1994)2 cited
- Discussion on SPEC Benchmark Suites and Evaluation Indicators(2010)