Xintao Fang
2024
SciMind: A Multimodal Mixture-of-Experts Model for Advancing Pharmaceutical Sciences
Zhaoping Xiong
|
Xintao Fang
|
Haotian Chu
|
Xiaozhe Wan
|
Liwei Liu
|
Yameng Li
|
Wenkai Xiang
|
Mingyue Zheng
Proceedings of the 1st Workshop on Language + Molecules (L+M 2024)
Large language models (LLMs) have made substantial strides, but their use in reliably tackling issues within specialized domains, particularly in interdisciplinary areas like pharmaceutical sciences, is hindered by data heterogeneity, knowledge complexity, unique objectives, and a spectrum of constraint conditions. In this area, diverse modalities such as nucleic acids, proteins, molecular structures, and natural language are often involved. We designed a specialized token set and introduced a new Mixture-of-Experts (MoEs) pre-training and fine-tuning strategy to unify these modalities in one model. With this strategy, we’ve created a multi-modal mixture-of-experts foundational model for pharmaceutical sciences, named SciMind. This model has undergone extensive pre-training on publicly accessible datasets including nucleic acid sequences, protein sequences, molecular structure strings, and biomedical texts, and delivers good performance on biomedical text comprehension, promoter prediction, protein function prediction, molecular description, and molecular generation.
Search
Co-authors
- Zhaoping Xiong 1
- Haotian Chu 1
- Xiaozhe Wan 1
- Liwei Liu 1
- Yameng Li 1
- show all...