Nature Language Model: Deciphering the Language of Nature for Scientific Discovery
- Yingce Xia ,
- Peiran Jin ,
- Shufang Xie ,
- Liang He ,
- Chuan Cao ,
- Renqian Luo ,
- Guoqing Liu ,
- Yue Wang ,
- Zequn Liu ,
- Yuan-Jyue Chen ,
- Yuan Chen ,
- Zekun Guo ,
- Yeqi Bai ,
- Pan Deng ,
- Yaosen Min ,
- Zi‐Ang Lu ,
- Hongxia Hao ,
- Han Yang ,
- Jielan Li ,
- Chang Liu ,
- Jia Zhang ,
- Jian-Bo Zhu ,
- Ke-Ming Wu ,
- Wei Zhang ,
- Kaiyuan Gao ,
- Qizhi Pei ,
- Qian Wang ,
- Xixian Liu ,
- Yanting Li ,
- Houtian Zhu ,
- Yeqing Lu ,
- Mingqian Ma ,
- Zun Wang ,
- Tian Xie ,
- Krzysztof Maziarz ,
- Marwin H. S. Segler ,
- Zhao Yang ,
- Zi-wei Chen ,
- Yu Shi ,
- Shuxin Zheng ,
- Lijun Wu ,
- Chen Hu ,
- Peggy Dai ,
- Tiemin Liu ,
- Hai-Long Liu ,
- Tao Qin
arXiv
Foundation models have revolutionized natural language processing and artificial intelligence, significantly enhancing how machines comprehend and generate human languages. Inspired by the success of these foundation models, researchers have developed foundation models for individual scientific domains, including small molecules, materials, proteins, DNA, RNA and even cells. However, these models are typically trained in isolation, lacking the ability to integrate across different scientific domains. Recognizing that entities within these domains can all be represented as sequences, which together form the”language of nature”, we introduce Nature Language Model (NatureLM), a sequence-based science foundation model designed for scientific discovery. Pre-trained with data from multiple scientific domains, NatureLM offers a unified, versatile model that enables various applications including: (i) generating and optimizing small molecules, proteins, RNA, and materials using text instructions; (ii) cross-domain generation/design, such as protein-to-molecule and protein-to-RNA generation; and (iii) top performance across different domains, matching or surpassing state-of-the-art specialist models. NatureLM offers a promising generalist approach for various scientific tasks, including drug discovery (hit generation/optimization, ADMET optimization, synthesis), novel material design, and the development of therapeutic proteins or nucleotides. We have developed NatureLM models in different sizes (1 billion, 8 billion, and 46.7 billion parameters) and observed a clear improvement in performance as the model size increases.