Probing the topology of the space of tokens with structured prompts

Published: 01 Jan 2025, Last Modified: 09 May 2025CoRR 2025EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: This article presents a general and flexible method for prompting a large language model (LLM) to reveal its (hidden) token input embedding up to homeomorphism. Moreover, this article provides strong theoretical justification -- a mathematical proof for generic LLMs -- for why this method should be expected to work. With this method in hand, we demonstrate its effectiveness by recovering the token subspace of Llemma-7B. The results of this paper apply not only to LLMs but also to general nonlinear autoregressive processes.
Loading