- Abstract: We propose an effective multitask learning setup for reducing distant supervision noise by leveraging sentence-level supervision. We show how sentence-level supervision can be used to improve the encoding of individual sentences, and to learn which input sentences are more likely to express the relationship between a pair of entities. We also introduce a novel neural architecture for collecting signals from multiple input sentences, which combines the benefits of attention and maxpooling. The proposed method increases AUC by 10% (from 0.261 to 0.284), and outperforms recently published results on the FB-NYT dataset.
- Keywords: relation extraction, distant supervision, maxpooled attention, multitask learning
- TL;DR: A new form of attention that works well for the distant supervision setting, and a multitask learning approach to add sentence-level annotations.