Repository logo
 

Multi-Task Learning for Coherence Modeling.

Published version
Peer-reviewed

Type

Conference Object

Change log

Authors

Farag, Youmna 
Giannakoudaki, Eleni 

Abstract

We address the task of assessing discourse coherence, an aspect of text quality that is essential for many NLP tasks, such as summarization and language assessment. We propose a hierarchical neural network trained in a multi-task fashion that learns to predict a document-level coherence score (at the network's top layers) along with word-level grammatical roles (at the bottom layers), taking advantage of inductive transfer between the two tasks. We assess the extent to which our framework generalizes to different domains and prediction tasks, and demonstrate its effectiveness not only on standard binary evaluation coherence tasks, but also on real-world tasks involving the prediction of varying degrees of coherence, achieving a new state of the art.

Description

Keywords

Journal Title

In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.

Conference Name

57th Annual Meeting of the Association for Computational Linguistics.

Journal ISSN

Volume Title

Publisher

ACL
Sponsorship
Cambridge Assessment (unknown)