Improving Interpretability and Regularization in Deep Learning


Type
Article
Change log
Authors
Gales, MJF 
Ragni, A 
Karanasou, P 
Sim, KC 
Abstract

IEEE Deep learning approaches yield state-of-the-art performance in a range of tasks, including automatic speech recognition. However, the highly distributed representation in a deep neural network (DNN) or other network variations are difficult to analyse, making further parameter interpretation and regularisation challenging. This paper presents a regularisation scheme acting on the activation function output to improve the network interpretability and regularisation. The proposed approach, referred to as activation regularisation, encourages activation function outputs to satisfy a target pattern. By defining appropriate target patterns, different learning concepts can be imposed on the network. This method can aid network interpretability and also has the potential to reduce over-fitting. The scheme is evaluated on several continuous speech recognition tasks: the Wall Street Journal continuous speech recognition task, eight conversational telephone speech tasks from the IARPA Babel program and a U.S. English broadcast news task. On all the tasks, the activation regularisation achieved consistent performance gains over the standard DNN baselines.

Description
Keywords
Activation regularisation, interpretability, visualisation, neural network, deep learning
Journal Title
IEEE/ACM Transactions on Audio Speech and Language Processing
Conference Name
Journal ISSN
2329-9290
2329-9304
Volume Title
26
Publisher
Institute of Electrical and Electronics Engineers (IEEE)
Sponsorship
IARPA (4912046943)
Cambridge Assessment (unknown)