Treepiece: Faster Semantic Parsing via Tree Tokenization

Published: 07 Oct 2023, Last Modified: 01 Dec 2023EMNLP 2023 FindingsEveryoneRevisionsBibTeX
Submission Type: Regular Short Paper
Submission Track: Semantics: Lexical, Sentence level, Document Level, Textual Inference, etc.
Submission Track 2: Syntax, Parsing and their Applications
Keywords: semantic parsing, decoding, tokenization algorithm, parse tree
TL;DR: We introduce TreePiece, a new technique that accelerates semantic parsing modeling by tokenizing parse trees into subtrees.
Abstract: \emph{Autoregressive} (AR) encoder-decoder neural networks have proved successful in many NLP problems, including \emph{Semantic Parsing} -- a task that translates natural language to machine-readable \emph{parse trees}. However, the sequential prediction process of AR models can be slow. To accelerate AR for semantic parsing, we introduce a new technique called \emph{TreePiece} that tokenizes a parse tree into subtrees and generates one subtree per decoding step. On TOPv2 benchmark, TreePiece shows $4.6$ times faster decoding speed than standard AR, and comparable speed but significantly higher accuracy compared to \emph{Non-Autoregressive} (NAR).
Submission Number: 831
Loading