Abstract: Labeled data for the task of Coreference Resolution is a scarce resource, requiring significant human effort. While state-of-the-art coreference models rely on such data, we propose an approach that leverages an end-to-end neural model in settings where labeled data is unavailable. Specifically, using weak supervision, we transfer the linguistic knowledge encoded by Stanford’s rule-based coreference system to the end-to-end model, which jointly learns rich, contextualized span representations and coreference chains. Our experiments on the English OntoNotes corpus demonstrate that our approach effectively benefits from the noisy coreference supervision, producing an improvement over Stanford’s rule-based system (+3.7 F$_1$) and outperforming the previous best unsupervised model (+0.9 F$_1$). Additionally, we validate the efficacy of our method on two other datasets: PreCo and Litbank (+2.5 and +4 F$_1$ on Stanford's system, respectively).
0 Replies
Loading