Repository logo
 

Improving Interpretability and Regularization in Deep Learning

Accepted version
Peer-reviewed

Type

Article

Change log

Authors

Gales, MJF 
Ragni, A 
Karanasou, P 
Sim, KC 

Abstract

IEEE Deep learning approaches yield state-of-the-art performance in a range of tasks, including automatic speech recognition. However, the highly distributed representation in a deep neural network (DNN) or other network variations are difficult to analyse, making further parameter interpretation and regularisation challenging. This paper presents a regularisation scheme acting on the activation function output to improve the network interpretability and regularisation. The proposed approach, referred to as activation regularisation, encourages activation function outputs to satisfy a target pattern. By defining appropriate target patterns, different learning concepts can be imposed on the network. This method can aid network interpretability and also has the potential to reduce over-fitting. The scheme is evaluated on several continuous speech recognition tasks: the Wall Street Journal continuous speech recognition task, eight conversational telephone speech tasks from the IARPA Babel program and a U.S. English broadcast news task. On all the tasks, the activation regularisation achieved consistent performance gains over the standard DNN baselines.

Description

Keywords

Activation regularisation, interpretability, visualisation, neural network, deep learning

Journal Title

IEEE/ACM Transactions on Audio Speech and Language Processing

Conference Name

Journal ISSN

2329-9290
2329-9304

Volume Title

26

Publisher

Institute of Electrical and Electronics Engineers (IEEE)
Sponsorship
IARPA (4912046943)
Cambridge Assessment (unknown)