Using Explicit Discourse Connectives in Translation for Implicit Discourse Relation Classification
Citations Over Time
Abstract
Implicit discourse relation recognition is an extremely challenging task due to the lack of indicative connectives. Various neural network architectures have been proposed for this task recently, but most of them suffer from the shortage of labeled data. In this paper, we address this problem by procuring additional training data from parallel corpora: When humans translate a text, they sometimes add connectives (a process known as explicitation). We automatically back-translate it into an English connective and use it to infer a label with high confidence. We show that a training set several times larger than the original training set can be generated this way. With the extra labeled instances, we show that even a simple bidirectional Long Short-Term Memory Network can outperform the current state-of-the-art.
Related Papers
- → Multi-task Attention-based Neural Networks for Implicit Discourse Relationship Representation and Identification(2017)107 cited
- Deep Enhanced Representation for Implicit Discourse Relation Recognition(2018)
- Adapting BERT to Implicit Discourse Relation Classification with a Focus on Discourse Connectives(2020)
- → Let’s be explicit about that: Distant supervision for implicit discourse relation classification via connective prediction(2021)1 cited
- → Implicit Discourse Relation Classification via Multi-Task Neural Networks(2016)75 cited