Home > Research > Publications & Outputs > End-to-end Deep Learning of Optimization Heuris...

Electronic data

  • pact17-paper47

    Accepted author manuscript, 1 MB, PDF document

    Available under license: CC BY-NC: Creative Commons Attribution-NonCommercial 4.0 International License

Links

Text available via DOI:

View graph of relations

End-to-end Deep Learning of Optimization Heuristics

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
  • Chris Cummins
  • Pavlos Petoumenos
  • Zheng Wang
  • Hugh Leather
Close
Publication date9/09/2017
Host publicationThe 26th International Conference on Parallel Architectures and Compilation Techniques (PACT) 2017
PublisherIEEE
Pages219-232
Number of pages14
ISBN (electronic)9781509067640
ISBN (print)9781509067657
<mark>Original language</mark>English

Abstract

Accurate automatic optimization heuristics are necessary for dealing with the complexity and diversity of modern hardware and software. Machine learning is a proven technique for learning such heuristics, but its success is bound by the quality of the features used. These features must be hand crafted by developers through a combination of expert domain knowledge and trial and error. This makes the quality of the final model directly dependent on the skill and available time of the system architect.

Our work introduces a better way for building heuristics. We develop a deep neural network that learns heuristics over raw code, entirely without using code features. The neural network simultaneously constructs appropriate representations of the code and learns how best to optimize, removing the need for manual feature creation. Further, we show that our neural nets can transfer learning from one optimization problem to another, improving the accuracy of new models, without the help of human experts.

We compare the effectiveness of our automatically generated heuristics against ones with features hand-picked by experts. We examine two challenging tasks: predicting optimal mapping for
heterogeneous parallelism and GPU thread coarsening factors. In 89% of the cases, the quality of our fully automatic heuristics matches or surpasses that of state-of-the-art predictive models
using hand-crafted features, providing on average 14% and 12% more performance with no human effort expended on designing features.