Understanding the Memory Performance of Data-Mining Workloads on Small, Medium, and Large-Scale CMPs Using Hardware-Software Co-simulation
Citations Over TimeTop 10% of 2007 papers
Abstract
With the amount of data continuing to grow, extracting "data of interest" is becoming popular, pervasive, and more important than ever. Data mining, as this process is known as, seeks to draw meaningful conclusions, extract knowledge, and acquire models from vast amounts of data. These compute-intensive data-mining applications, where thread-level parallelism can be effectively exploited, are the design targets of future multi-core systems. As a result, future multi-core systems will be required to process terabyte-level workloads. To understand the memory system performance of data-mining applications, this paper presents the use of hardware-software co-simulation to explore the cache design space of several multi-threaded data mining applications. Our study reveals that the workloads are memory intensive, have large working-set sizes, and exhibit good data locality. We find that large DRAM caches can be useful to address their large working-set sizes
Related Papers
- → Clarification of the Stretching Process of Newly-Developed Open-End Rotor-Spun Hollow Yarn(2012)2 cited
- → <title>Challenge of terabyte optical recording</title>(2001)1 cited
- Dublin City University at the TREC 2005 Terabyte Track(2005)
- → Optical Mass Storage For A 10 Terabyte Archive: System Integration Problems And User Interface(1987)
- The Influence of Spun Yarn Fittings on Yarn Quality(2002)