Cutting Recursive Autoencoder TreesDownload PDF

18 Apr 2024 (modified: 20 Jan 2013)ICLR 2013 conference submissionReaders: Everyone
Decision: conferencePoster-iclr2013-conference
Abstract: Deep Learning models enjoy considerable success in Natural Language Processing. While deep architectures produce useful representations that lead to improvements in various tasks, they are often difficult to interpret. This makes the analysis of learned structures particularly difficult. We therefore have to rely on empirical tests to see whether a particular structure makes sense. In this paper, we present an analysis of a well-received model that produces structural representations of text: the Semi-Supervised Recursive Autoencoder. We show that for certain tasks, the structure of the autoencoder may be significantly reduced and we evaluate the produced structures through human judgment.
7 Replies

Loading