Pre-trained Language Models in Biomedical Domain: A Systematic Survey

Benyou Wang, Qianqian Xie, Jiahuan Pei, Zhihong Chen, Prayag Tiwari, Zhao Li, Jie Fu

Research output: Contribution to JournalArticleAcademicpeer-review

Abstract

Pre-trained language models (PLMs) have been the de facto paradigm for most natural language processing tasks. This also benefits the biomedical domain: researchers from informatics, medicine, and computer science communities propose various PLMs trained on biomedical datasets, e.g., biomedical text, electronic health records, protein, and DNA sequences for various biomedical tasks. However, the cross-discipline characteristics of biomedical PLMs hinder their spreading among communities; some existing works are isolated from each other without comprehensive comparison and discussions. It is nontrivial to make a survey that not only systematically reviews recent advances in biomedical PLMs and their applications but also standardizes terminology and benchmarks. This article summarizes the recent progress of pre-trained language models in the biomedical domain and their applications in downstream biomedical tasks. Particularly, we discuss the motivations of PLMs in the biomedical domain and introduce the key concepts of pre-trained language models. We then propose a taxonomy of existing biomedical PLMs that categorizes them from various perspectives systematically. Plus, their applications in biomedical downstream tasks are exhaustively discussed, respectively. Last, we illustrate various limitations and future trends, which aims to provide inspiration for the future research.
Original languageEnglish
Article number3611651
JournalACM Computing Surveys
Volume56
Issue number3
DOIs
Publication statusPublished - 5 Oct 2023
Externally publishedYes

Funding

This work is supported by Chinese Key-Area Research and Development Program of Guangdong Province (2020B0101350001), the Shenzhen Science and Technology Program (JCYJ20220818103001002), the Guangdong Provincial Key Laboratory of Big Data Computing, The Chinese University of Hong Kong, Shenzhen, Shenzhen Key Research Project (C10120230151) and Shenzhen Doctoral Startup Funding (RCBS20221008093330065).

FundersFunder number
Guangdong Provincial Key Laboratory of Big Data Computing
Shenzhen Doctoral Startup FundingRCBS20221008093330065
Shenzhen Key Research ProjectC10120230151
Shenzhen Graduate School, Peking University
Chinese University of Hong Kong
Science, Technology and Innovation Commission of Shenzhen MunicipalityJCYJ20220818103001002
Science, Technology and Innovation Commission of Shenzhen Municipality
Special Project for Research and Development in Key areas of Guangdong Province2020B0101350001
Special Project for Research and Development in Key areas of Guangdong Province

    Fingerprint

    Dive into the research topics of 'Pre-trained Language Models in Biomedical Domain: A Systematic Survey'. Together they form a unique fingerprint.

    Cite this