A Graph Neural Network-based performance model for Deep Learning Applications
Tue 14 Jun 2022 04:45 - 05:00 at Boardroom - Evening
The unprecedented proliferation of machine learning based software brings an ever-increasing need to optimize the implementation of such applications. State-of-the-art compilers for neural networks, such as Halide and TVM, incorporate a machine learning-based performance model to search the space of valid implementations of a given deep learning algorithm. For a given application, the model predicts the value of performance metrics such as the run time without executing the application on hardware. Such models speed up the compilation process by obviating the need to benchmark an enormous number of candidate implementations, referred to as schedules, on hardware. Existing performance models employ feed-forward networks, recurrent networks, or decision tree ensembles to estimate the performance of different implementations of a neural network. Graphs present a natural and intuitive way to model deep-learning networks where each node represents a computational stage or operation. Incorporating the inherent graph structure of these workloads in the performance model can enable a better representation and learning of inter-stage interactions. The accuracy of the performance model has direct implications on the efficiency of the search strategy, making it a crucial component of this class of deep-learning compilers. In this work, we develop a novel performance model that adopts a graph representation. In our model, each stage of computation represents a node characterized by features that capture the operations performed by the stage. The interaction between nodes is achieved using graph convolutions. Experimental evaluation shows a 7.75x and 12x reduction in prediction error compared to the existing Halide and TVM models, respectively.
Mon 13 JunDisplayed time zone: Pacific Time (US & Canada) change
15:30 - 17:00 | |||
15:30 45mKeynote | Unsupervised Program Synthesis: Hierarchy and Perception MAPS Kevin Ellis Cornell University | ||
16:15 15mTalk | ExeBench: An ML-scale dataset of executable C functions MAPS Jordi Armengol-Estapé University of Edinburgh, Jackson Woodruff University of Edinburgh, Alexander Brauckmann University of Edinburgh, José Wesley de Souza Magalhães University of Edinburgh, Michael F. P. O'Boyle University of Edinburgh | ||
16:30 15mTalk | Automatically Debugging AutoML Pipelines Using Maro: ML Automated Remediation Oracle MAPS | ||
16:45 15mTalk | A Graph Neural Network-based performance model for Deep Learning Applications MAPS Shikhar Singh University of Texas, James Hegarty Facebook, Hugh Leather University of Edinburgh, UK, Benoit Steiner Facebook |
Tue 14 JunDisplayed time zone: Pacific Time (US & Canada) change
03:30 - 05:00 | |||
03:30 45mKeynote | Unsupervised Program Synthesis: Hierarchy and Perception MAPS Kevin Ellis Cornell University | ||
04:15 15mTalk | ExeBench: An ML-scale dataset of executable C functions MAPS Jordi Armengol-Estapé University of Edinburgh, Jackson Woodruff University of Edinburgh, Alexander Brauckmann University of Edinburgh, José Wesley de Souza Magalhães University of Edinburgh, Michael F. P. O'Boyle University of Edinburgh | ||
04:30 15mTalk | Automatically Debugging AutoML Pipelines Using Maro: ML Automated Remediation Oracle MAPS | ||
04:45 15mTalk | A Graph Neural Network-based performance model for Deep Learning Applications MAPS Shikhar Singh University of Texas, James Hegarty Facebook, Hugh Leather University of Edinburgh, UK, Benoit Steiner Facebook |