Long text analysis using sliced recurrent neural networks with breaking point information enrichmentOpen Website

23 Feb 2022OpenReview Archive Direct UploadReaders: Everyone
Abstract: Sliced recurrent neural networks (SRNNs) are the state-ofthe-art efficient solution for long text analysis tasks; however, their slicing operations inevitably result in long-term dependency loss in lower-level networks and thus limit their accuracy. Therefore, we propose a breaking point information enrichment mechanism to strengthen dependencies between sliced subsequences without hindering parallelization. Then, the resulting BPIE-SRNN model is further extended to a bidirectional model, BPIE-BiSRNN, to utilize the dependency information in not only the previous but also the following contexts. Experiments on four large public real-world datasets demonstrate that the BPIE-SRNN and BPIE-BiSRNN models always achieve a much better accuracy than SRNNs and BiSRNNs, while maintaining a superior training efficiency
0 Replies

Loading