ADAPT: Algorithmic Differentiation Applied to Floating-Point Precision Tuning
Citations Over TimeTop 10% of 2018 papers
Abstract
HPC applications use floating point arithmetic operations extensively to solve computational problems. Mixed-precision computing seeks to use the lowest precision data type that is sufficient to achieve a desired accuracy, improving performance and reducing power consumption. Manually optimizing a program to use mixed precision is challenging as it not only requires extensive knowledge about the numerical behavior of the algorithm but also estimates of the rounding errors. In this work, we present ADAPT, a scalable approach for mixed-precision analysis on HPC workloads using algorithmic differentiation to provide accurate estimates about the final output error. ADAPT provides a floating-point precision sensitivity profile while incurring an overhead of only a constant multiple of the original computation irrespective of the number of variables analyzed. The sensitivity profile can be used to make algorithmic choices and to develop mixed-precision configurations of a program. We evaluate ADAPT on six benchmarks and a proxy application (LULESH) and show that we are able to achieve a speedup of 1.2× on the proxy application.
Related Papers
- → 2. Basic Numerical Methods(1996)3 cited
- → Converting integer to floating point number(2003)1 cited
- → Computing Integer Powers in Floating-Point Arithmetic(2007)2 cited
- → ROUND-OFF ERRORS IN FLOATING-POINT ADDITIONS(1985)1 cited
- → Functional Analysis and Synthesis of Floating Point Arithmetic Devices(2012)