Learning Metadata-Agnostic Representations for Text-to-SQL In-Context Example Selection

Published: 10 Oct 2024, Last Modified: 16 Oct 2024TRL @ NeurIPS 2024 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: text-to-sql, representation learning, phrase/sentence embedding, word/phrase alignment, in-context learning
TL;DR: We propose an approach that aligns fine-grained representations of natural language to those of SQL queries and adopt it to boost LLM performance through effective in-context demonstrations retrieval.
Abstract: In-context learning (ICL) is a powerful paradigm where large language models (LLMs) benefit from task demonstrations added to the prompt. Yet, selecting optimal demonstrations is not trivial, especially for complex or multi-modal tasks where input and output distributions differ. We hypothesize that forming task-specific representations of the input is key. In this paper, we propose a method to align representations of natural language questions and those of SQL queries in a shared embedding space. Our technique, dubbed MARLO—Metadata-Agnostic Representation Learning for Text-tO-SQL— uses query structure to model querying intent without over-indexing on underlying database metadata (i.e. tables, columns, or domain-specific entities of a database referenced in the question or query). This allows MARLO to select examples that are structurally and semantically relevant for the task rather than examples that are spuriously related to a certain domain or question phrasing. When used to retrieve examples based on question similarity, MARLO shows superior performance compared to generic embedding models (on average +2.9%pt. in execution accuracy) on the Spider benchmark. It also outperforms the next best method that masks metadata information by +0.8%pt. in execution accuracy on average, while imposing a significantly lower inference latency.
Submission Number: 17
Loading