Friday July 9, 2021
Vers des méthodes transparentes et parcimonieuses pour l’optimisation automatique des performances
Abstract:
 
The  end  of Moore's  Law  and  the breakdown  of  Dennard's  scaling mean  that increasing hardware complexity and optimizing code efficiently are indispensable to maintain the exponential performance  improvements of the past decades.  Hand optimizing code is not suited to the sheer number of configurations of many code optimization  problems,  but  fitting   these  problems  into  the  mathematical optimization and learning frameworks enables applying methods from these domains to automatically  optimize code  for performance,  a process  called autotuning. Commonly  used  autotuning  methods  are either  not  conducive  to  statistical analysis, such as genetic algorithms, or reliant on restrictive hypotheses about the target  search space, such as  gradient descent.  In this  thesis we develop and evaluate  the performance  of an  autotuning method based  on the  Design of Experiments, a  branch of statistics  that is not  widely studied or  applied in autotuning  problems,  and   which  aids  in  the   parsimonious  production  of statistically interpretable and accurate surrogate models.

We  present a  series of  descriptions and  discussions of  various optimizationn methods, from  the perspective  of performance  tuning.  We  describe heuristics from  mathematical optimization,  and parametric  and nonparametric  statistical modeling methods, describing how these surrogate  models can be used to minimize an unknown  function.  We  then discuss  how the  Design of  Experiments enables managing  the   compromise  between  experimental  budget   and  model  quality, establishing  a  link  with  Online Learning  methods,  focusing  on  parsimony,
progressive model improvement, uncertainty,  and robustness, the properties that are most relevant for a method's applicability to autotuning problems.

The key  contribution of  this thesis  is the development  of a  transparent and parsimonious autotuning  approach based on  the Design of Experiments,  which we apply to  diverse problems such as  optimizing the configuration of  GPU and CPU kernels  and  finding  mixed-precision  bit  quantization  policies  for  neural networks.  We also present a series of empirical evaluations of other methods on autotuning problems from  different High Performance Computing  domains, such as search  heuristics   coordinated  by   a  bandit   algorithm  to   optimize  the
configuration  of compilers  for several  GPU and  FPGA kernels.   Although some experimental scenarios  eluded the  detection and  exploitation of  search space structure,  regardless  of the  chosen  method,  we demonstrate  how  autotuning methods based on the Design of  Experiments can aid in interpretable, efficient, and effective code optimization.
Mis à jour le 1 July 2021