Reinforcement Learning and Adaptive Sampling for Optimized DNN Compilation
Abstract
Achieving faster execution with shorter compilation time can enable further diversity and innovation in neural networks. However, the current paradigm of executing neural networks either relies on hand-optimized libraries, traditional compilation heuristics, or very recently, simulated annealing and genetic algorithms. Our work takes a unique approach by formulating compiler optimizations for neural networks as a reinforcement learning problem, whose solution takes fewer steps to converge. This solution, dubbed ReLeASE, comes with a sampling algorithm that leverages clustering to focus the costly samples (real hardware measurements) on representative points, subsuming an entire subspace. Our adaptive sampling not only reduces the number of samples, but also improves the quality of samples for better exploration in shorter time. As such, experimentation with real hardware shows that reinforcement learning with adaptive sampling provides 4.45x speed up in optimization time over AutoTVM, while also improving inference time of the modern deep networks by 5.6%. Further experiments also confirm that our adaptive sampling can even improve AutoTVM’s simulated annealing by 4.00x.
Cite
Text
Ahn et al. "Reinforcement Learning and Adaptive Sampling for Optimized DNN Compilation." ICML 2019 Workshops: RL4RealLife, 2019.Markdown
[Ahn et al. "Reinforcement Learning and Adaptive Sampling for Optimized DNN Compilation." ICML 2019 Workshops: RL4RealLife, 2019.](https://mlanthology.org/icmlw/2019/ahn2019icmlw-reinforcement/)BibTeX
@inproceedings{ahn2019icmlw-reinforcement,
title = {{Reinforcement Learning and Adaptive Sampling for Optimized DNN Compilation}},
author = {Ahn, Byung Hoon and Pilligundla, Prannoy and Esmaeilzadeh, Hadi},
booktitle = {ICML 2019 Workshops: RL4RealLife},
year = {2019},
url = {https://mlanthology.org/icmlw/2019/ahn2019icmlw-reinforcement/}
}