DocumentCode :
1954618
Title :
Machine learning for predictive auto-tuning with boosted regression trees
Author :
Bergstra, James ; Pinto, Nicolas ; Cox, David
Author_Institution :
Rowland Inst. at Harvard, Harvard Univ., Cambridge, MA, USA
fYear :
2012
fDate :
13-14 May 2012
Firstpage :
1
Lastpage :
9
Abstract :
The rapidly evolving landscape of multicore architectures makes the construction of efficient libraries a daunting task. A family of methods known collectively as “auto-tuning” has emerged to address this challenge. Two major approaches to auto-tuning are empirical and model-based: empirical autotuning is a generic but slow approach that works by measuring runtimes of candidate implementations, model-based auto-tuning predicts those runtimes using simplified abstractions designed by hand. We show that machine learning methods for non-linear regression can be used to estimate timing models from data, capturing the best of both approaches. A statistically-derived model offers the speed of a model-based approach, with the generality and simplicity of empirical auto-tuning. We validate our approach using the filterbank correlation kernel described in Pinto and Cox [2012], where we find that 0.1 seconds of hill climbing on the regression model (“predictive auto-tuning”) can achieve almost the same speed-up as is brought by minutes of empirical auto-tuning. Our approach is not specific to filterbank correlation, nor even to GPU kernel auto-tuning, and can be applied to almost any templated-code optimization problem, spanning a wide variety of problem types, kernel types, and platforms.
Keywords :
channel bank filters; learning (artificial intelligence); multiprocessing systems; regression analysis; trees (mathematics); boosted regression trees; candidate implementation runtime measurement; empirical autotuning; filterbank correlation kernel; machine learning; model-based approach; model-based autotuning; multicore architectures; nonlinear regression; predictive autotuning; statistically-derived model; templated-code optimization problem; time 0.1 s; timing model estimation; Correlation; Graphics processing unit; Instruction sets; Kernel; Libraries; Optimization; Regression tree analysis;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Innovative Parallel Computing (InPar), 2012
Conference_Location :
San Jose, CA
Print_ISBN :
978-1-4673-2632-2
Electronic_ISBN :
978-1-4673-2631-5
Type :
conf
DOI :
10.1109/InPar.2012.6339587
Filename :
6339587
Link To Document :
بازگشت