arXiv Analytics

Sign in

arXiv:1907.02427 [cs.CL]AbstractReferencesReviewsResources

Multi-Task Learning for Coherence Modeling

Youmna Farag, Helen Yannakoudakis

Published 2019-07-04Version 1

We address the task of assessing discourse coherence, an aspect of text quality that is essential for many NLP tasks, such as summarization and language assessment. We propose a hierarchical neural network trained in a multi-task fashion that learns to predict a document-level coherence score (at the network's top layers) along with word-level grammatical roles (at the bottom layers), taking advantage of inductive transfer between the two tasks. We assess the extent to which our framework generalizes to different domains and prediction tasks, and demonstrate its effectiveness not only on standard binary evaluation coherence tasks, but also on real-world tasks involving the prediction of varying degrees of coherence, achieving a new state of the art.

Comments: 11 pages, 3 figures, Accepted at ACL 2019
Journal: THE 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019)
Categories: cs.CL, cs.LG
Related articles: Most relevant | Search more
arXiv:2205.12701 [cs.CL] (Published 2022-05-25)
Eliciting Transferability in Multi-task Learning with Task-level Mixture-of-Experts
arXiv:2011.00425 [cs.CL] (Published 2020-11-01)
Analyzing the Effect of Multi-task Learning for Biomedical Named Entity Recognition
arXiv:1905.08511 [cs.CL] (Published 2019-05-21)
Answering while Summarizing: Multi-task Learning for Multi-hop QA with Evidence Extraction