%0 Conference Paper
%B IEEE High Performance Extreme Computing Conference (HPEC)
%D 2019
%T Application of Approximate Matrix Multiplication to Neural Networks and Distributed SLAM
%A Brian Plancher
%A Camelia Brumar
%A Iulian Brumar
%A Lillian Pentecost
%A Saketh Rama
%A David Brooks
%K deep learning
%K profiling
%X Computational efficiency is a critical constraint for a variety of cutting-edge real-time applications. In this work, we identify an opportunity to speed up the end-to-end runtime of two such compute bound applications by incorporating approximate linear algebra techniques. Particularly, we apply approximate matrix multiplication to artificial Neural Networks (NNs) for image classification and to the robotics problem of Distributed Simultaneous Localization and Mapping (DSLAM). Expanding upon recent sampling-based Monte Carlo approximation strategies for matrix multiplication, we develop updated theoretical bounds, and an adaptive error prediction strategy. We then apply these techniques in the context of NNs and DSLAM increasing the speed of both applications by 15-20% while maintaining a 97% classification accuracy for NNs running on the MNIST dataset and keeping the average robot position error under 1 meter (vs 0.32 meters for the exact solution). However, both applications experience variance in their results. This suggests that Monte Carlo matrix multiplication may be an effective technique to reduce the memory and computational burden of certain algorithms when used carefully, but more research is needed before these techniques can be widely used in practice.
%B IEEE High Performance Extreme Computing Conference (HPEC)
%C Waltham, MA, USA
%G eng
%U https://doi.org/10.1109/HPEC.2019.8916468