Issue 27, 2025

Pre-trained language models for protein and molecular design

Abstract

Pre-trained language models (PLMs) have recently emerged as a powerful tool, showcasing exceptional performance not just in natural language understanding but also in the realm of biological research. The advantage of PLMs lies in their ability to leverage the structural similarity between biological sequences and natural language. PLMs offer novel solutions for protein research and drug design applications. By pre-training on extensive unlabeled biological sequences and then fine-tuning for specific tasks, PLMs have delivered remarkable results. To summarize the growing landscape of PLMs in biological research, this paper integrates exemplary PLMs and common datasets, demonstrating the potential and application prospects of PLMs in prediction and generation tasks.

Graphical abstract: Pre-trained language models for protein and molecular design

Transparent peer review

To support increased transparency, we offer authors the option to publish the peer review history alongside their article.

View this article’s peer review history

Article information

Article type
Review Article
Submitted
27 Feb 2025
Accepted
12 Jun 2025
First published
16 Jun 2025

Phys. Chem. Chem. Phys., 2025,27, 14189-14216

Pre-trained language models for protein and molecular design

E. Zhang, Z. Pan, Z. Yao, T. Dong, G. Chen, T. Deng, S. Chen and C. Y. Chen, Phys. Chem. Chem. Phys., 2025, 27, 14189 DOI: 10.1039/D5CP00785B

To request permission to reproduce material from this article, please go to the Copyright Clearance Center request page.

If you are an author contributing to an RSC publication, you do not need to request permission provided correct acknowledgement is given.

If you are the author of this article, you do not need to request permission to reproduce figures and diagrams provided correct acknowledgement is given. If you want to reproduce the whole article in a third-party publication (excluding your thesis/dissertation for which permission is not required) please go to the Copyright Clearance Center request page.

Read more about how to correctly acknowledge RSC content.

Social activity

Spotlight

Advertisements