SciMind: A Multimodal Mixture-of-Experts Model for Advancing Pharmaceutical Sciences

Published: 06 Jul 2024, Last Modified: 28 Jul 2024Language and Molecules ACL 2024 OralEveryoneRevisionsBibTeXCC BY 4.0
Keywords: LLM, Mixture-of-Experts, Multimodal, Pharmaceutical Sciences
Abstract: Large language models (LLMs) have made substantial strides, but their use in reliably tackling issues within specialized domains, particularly in interdisciplinary areas like pharmaceutical sciences, is hindered by data heterogeneity, knowledge complexity, unique objectives, and a spectrum of constraint conditions. In this area, diverse modalities such as nucleic acids, proteins, molecular structures, and natural language are often involved. We designed a specialized token set and introduced a new Mixture-of-Experts (MoEs) pre-training and fine-tuning strategy to unify these modalities in one model. With this strategy, we've created a multi-modal mixture-of-experts foundational model for pharmaceutical sciences, named SciMind. This model has undergone extensive pre-training on publicly accessible datasets including nucleic acid sequences, protein sequences, molecular structure strings, and biomedical texts, and delivers good performance on biomedical text comprehension, promoter prediction, protein function prediction, molecular description, and molecular generation.
Archival Option: The authors of this submission want it to appear in the archival proceedings.
Submission Number: 9
Loading