Repository logo
 

Sparsification and Separation of Deep Learning Layers for Constrained Resource Inference on Wearables

Accepted version
Peer-reviewed

Type

Conference Object

Change log

Authors

Bhattacharya, Sourav 
Lane, Nicholas D 

Abstract

Deep learning has revolutionized the way sensor data are analyzed and interpreted. The accuracy gains these approaches o↵er make them attractive for the next generation of mobile, wearable and embedded sensory applications. However, state-of-the-art deep learning algorithms typically require a significant amount of device and processor resources, even just for the inference stages that are used to discriminate high-level classes from low-level data. The limited availability of memory, computation, and energy on mobile and embedded platforms thus pose a significant challenge to the adoption of these powerful learning techniques. In this paper, we propose SparseSep, a new approach that leverages the sparsification of fully connected layers and separation of convolutional kernels to reduce the resource requirements of popular deep learning algorithms. As a result, SparseSep allows large-scale DNNs and CNNs to run eciently on mobile and embedded hardware with only minimal impact on inference accuracy. We experiment using SparseSep across a variety of common processors such as the Qualcomm Snapdragon 400, ARM Cortex M0 and M3, and Nvidia Tegra K1, and show that it allows inference for various deep models to execute more eciently; for example, on average requiring 11.3 times less memory and running 13.3 times faster on these representative platforms.

Description

Keywords

46 Information and Computing Sciences, 4611 Machine Learning, 7 Affordable and Clean Energy

Journal Title

Proceedings of the 14th ACM Conference on Embedded Network Sensor Systems CD-ROM

Conference Name

SenSys '16: The 14th ACM Conference on Embedded Network Sensor Systems

Journal ISSN

Volume Title

Publisher

ACM

Rights

All rights reserved