mRNA-GPT: End-to-end Generative Design and Optimization of Full-length mRNA

Published: 02 Mar 2026, Last Modified: 05 Mar 2026GEM 2026EveryoneRevisionsBibTeXCC BY 4.0
Keywords: Generative model; Full-length mRNA optimization; Langauge model
TL;DR: mRNA-GPT is a generative AI model that uses reinforcement learning to jointly optimize full-length mRNA sequences (5′ UTR, CDS, and 3′ UTR) for enhanced translation efficiency and stability, outperforming methods that optimize regions separately.
Abstract: We introduce mRNA-GPT, a generative model for end-to-end full-length mRNA sequence design and optimization. Unlike existing approaches that optimize isolated regions, mRNA-GPT jointly optimizes across all three regions (5′ UTR, CDS, and 3′ UTR) to capture cross-region regulatory interactions critical for therapeutic efficacy. Pre-trained on 10 million full-length natural mRNA sequences across diverse species and organisms, establishing a robust foundation for sequence generation. mRNA-GPT employs an iterative optimization framework with oracle-based rewards to progressively enhance target properties including translation efficiency and half-life. mRNA-GPT supports flexible generation modes: single regions, full-length sequences, or conditional generation given any region. Empirical results demonstrate superior performance over state-of-the-art methods: CDS optimization achieves higher predicted translation rates than LinearDesign and GEMORNA while maintaining structural stability, and full-length design captures critical cross-region interactions yielding enhanced translation efficiency. This unified approach establishes mRNA-GPT as a versatile platform for rational mRNA therapeutics design.
Presenter: ~Sizhen_Li1
Format: Maybe: the presenting author will attend in person, contingent on other factors that still need to be determined (e.g., visa, funding).
Funding: No, the presenting author of this submission does not fall under ICLR’s funding aims, or has sufficient alternate funding.
Submission Number: 26
Loading