Confirmation: I have read and agree with the workshop's policy on behalf of myself and my co-authors.
Track: long paper (4–8 pages excluding references)
Keywords: Spatial transcriptomics, zero-shot annotation, training-free inference, structured prediction, prototype-based methods, large language model verification, constrained reasoning, spatial graphs, neighborhood context, cross-platform generalization, interpretability
TL;DR: A training-free spatial transcriptomics annotator that uses biologically grounded prototypes and selective, ontology-constrained LLM verification to achieve robust, interpretable labeling across platforms.
Abstract: Spatial transcriptomics enables the analysis of cellular organization by measur-
ing gene expression in situ, but assigning coherent spatial region labels remains
challenging across platforms due to heterogeneous resolution, incomplete marker
panels, and ambiguous boundaries. Existing approaches typically rely on super-
vised training, dataset-specific tuning, or deep graph models, which can oversmooth
structure, generalize poorly across technologies, and offer limited interpretability.
We introduce NicheAgent, a training-free structured prediction framework that casts
spatial annotation as a constrained decision problem with selective language-based
verification. NicheAgent first performs deterministic prototype-based assignment
using curated region prototypes (“nichecards”) encoding canonical marker genes
and expression centroids. Only for low-confidence cases, a lightweight large
language model (LLM) is invoked as a closed-world verifier, arbitrating among
a fixed set of candidate labels using marker semantics and local neighborhood
context under a strict ontology. A single round of spatial smoothing enforces local
coherence without blurring anatomical boundaries.
Across Visium, MERFISH, and STARmap datasets, NicheAgent consistently out-
performs supervised, graph-based, and prior LLM-driven methods on standard
spatial annotation metrics, while remaining transparent and interpretable. More
broadly, our results highlight a general design pattern in which LLMs act as con-
strained adjudicators over symbolic hypotheses, improving structured prediction in
high-ambiguity settings without end-to-end learning or loss of interpretability.
Anonymization: This submission has been anonymized for double-blind review via the removal of identifying information such as names, affiliations, and identifying URLs.
Presenter: ~Sajib_Acharjee_Dip1
Format: Maybe: the presenting author will attend in person, contingent on other factors that still need to be determined (e.g., visa, funding).
Funding: Yes, the presenting author of this submission falls under ICLR’s funding aims, and funding would significantly impact their ability to attend the workshop in person.
Submission Number: 81
Loading