Mon 13 Jun 2022 16:45 - 17:00 at Boardroom - Evening Chair(s): Swarat Chaudhuri
Tue 14 Jun 2022 04:45 - 05:00 at Boardroom - Evening

The unprecedented proliferation of machine learning based software brings an ever-increasing need to optimize the implementation of such applications. State-of-the-art compilers for neural networks, such as Halide and TVM, incorporate a machine learning-based performance model to search the space of valid implementations of a given deep learning algorithm. For a given application, the model predicts the value of performance metrics such as the run time without executing the application on hardware. Such models speed up the compilation process by obviating the need to benchmark an enormous number of candidate implementations, referred to as schedules, on hardware. Existing performance models employ feed-forward networks, recurrent networks, or decision tree ensembles to estimate the performance of different implementations of a neural network. Graphs present a natural and intuitive way to model deep-learning networks where each node represents a computational stage or operation. Incorporating the inherent graph structure of these workloads in the performance model can enable a better representation and learning of inter-stage interactions. The accuracy of the performance model has direct implications on the efficiency of the search strategy, making it a crucial component of this class of deep-learning compilers. In this work, we develop a novel performance model that adopts a graph representation. In our model, each stage of computation represents a node characterized by features that capture the operations performed by the stage. The interaction between nodes is achieved using graph convolutions. Experimental evaluation shows a 7.75x and 12x reduction in prediction error compared to the existing Halide and TVM models, respectively.

Mon 13 Jun

Displayed time zone: Pacific Time (US & Canada) change

15:30 - 17:00
EveningMAPS at Boardroom +12h
Chair(s): Swarat Chaudhuri University of Texas at Austin
15:30
45m
Keynote
Unsupervised Program Synthesis: Hierarchy and Perception
MAPS
Kevin Ellis Cornell University
16:15
15m
Talk
ExeBench: An ML-scale dataset of executable C functions
MAPS
Jordi Armengol-Estapé University of Edinburgh, Jackson Woodruff University of Edinburgh, Alexander Brauckmann University of Edinburgh, José Wesley de Souza Magalhães University of Edinburgh, Michael F. P. O'Boyle University of Edinburgh
16:30
15m
Talk
Automatically Debugging AutoML Pipelines Using Maro: ML Automated Remediation Oracle
MAPS
Julian Dolby IBM Research, USA, Jason Tsay IBM Research, Martin Hirzel IBM Research
16:45
15m
Talk
A Graph Neural Network-based performance model for Deep Learning Applications
MAPS
Shikhar Singh University of Texas, James Hegarty Facebook, Hugh Leather University of Edinburgh, UK, Benoit Steiner Facebook

Tue 14 Jun

Displayed time zone: Pacific Time (US & Canada) change

03:30 - 05:00
EveningMAPS at Boardroom
03:30
45m
Keynote
Unsupervised Program Synthesis: Hierarchy and Perception
MAPS
Kevin Ellis Cornell University
04:15
15m
Talk
ExeBench: An ML-scale dataset of executable C functions
MAPS
Jordi Armengol-Estapé University of Edinburgh, Jackson Woodruff University of Edinburgh, Alexander Brauckmann University of Edinburgh, José Wesley de Souza Magalhães University of Edinburgh, Michael F. P. O'Boyle University of Edinburgh
04:30
15m
Talk
Automatically Debugging AutoML Pipelines Using Maro: ML Automated Remediation Oracle
MAPS
Julian Dolby IBM Research, USA, Jason Tsay IBM Research, Martin Hirzel IBM Research
04:45
15m
Talk
A Graph Neural Network-based performance model for Deep Learning Applications
MAPS
Shikhar Singh University of Texas, James Hegarty Facebook, Hugh Leather University of Edinburgh, UK, Benoit Steiner Facebook