Transitioning Representations between Languages for Cross-lingual Event Detection via Langevin Dynamics

Published: 07 Oct 2023, Last Modified: 01 Dec 2023EMNLP 2023 FindingsEveryoneRevisionsBibTeX
Submission Type: Regular Short Paper
Submission Track: Information Extraction
Submission Track 2: Multilinguality and Linguistic Diversity
Keywords: Event Detection, Information Extraction, Cross-lingual Transfer Learning, Langevin Dynamics
TL;DR: We introduce a novel method for cross-lingual transfer learning for event detection that utilizes Langevin Dynamics to transition target language representations into the source language space.
Abstract: Cross-lingual transfer learning (CLTL) for event detection (ED) aims to develop models in high-resource source languages that can be directly applied to produce effective performance for lower-resource target languages. Previous research in this area has focused on representation matching methods to develop a language-universal representation space into which source- and target-language example representations can be mapped to achieve cross-lingual transfer. However, as this approach modifies the representations for the source-language examples, the models might lose discriminative features for ED that are learned over training data of the source language to prevent effective predictions. To this end, our work introduces a novel approach for cross-lingual ED where we only aim to transition the representations for the target-language examples into the source-language space, thus preserving the representations in the source language and their discriminative information. Our method introduces Langevin Dynamics to perform representation transition and a semantic preservation framework to retain event type features during the transition process. Extensive experiments over three languages demonstrate the state-of-the-art performance for ED in CLTL.
Submission Number: 4765
Loading