Innovative Application of Large Language Models in Biology and Medicine Under Artificial Intelligence for Science Paradigm

Xingli CUN, Chenjun DING, Fang CHEN, Xin ZHANG

Acta Academiae Medicinae Sinicae ›› 2025, Vol. 47 ›› Issue (6) : 969-981.

PDF(947 KB)
Home Journals Acta Academiae Medicinae Sinicae
Acta Academiae Medicinae Sinicae

Abbreviation (ISO4): Acta Academiae Medicinae Sinicae      Editor in chief: Xuetao CAO

About  /  Aim & scope  /  Editorial board  /  Indexed  /  Contact  / 
PDF(947 KB)
Acta Academiae Medicinae Sinicae ›› 2025, Vol. 47 ›› Issue (6) : 969-981. DOI: 10.3881/j.issn.1000-503X.16605
Medical Artificial Intelligence

Innovative Application of Large Language Models in Biology and Medicine Under Artificial Intelligence for Science Paradigm

Author information +
History +

Abstract

Currently,the artificial intelligence for science paradigm is evolving rapidly,with the new generation of artificial intelligence technologies represented by large language models(LLM)having injected transformative momentum into biological and medical research.Systematically examining the innovative applications of LLM in biomedical contexts under the artificial intelligence for science paradigm can provide critical references for methodological innovation and research paradigm transformation in deciphering complex disease mechanisms and advancing precision medicine.Through comparative analysis of high-impact peer-reviewed publications and preprints in the past two years,we elucidate cutting-edge research progress,developmental trajectories,and persistent challenges in LLM applications across biological and medical domains.Furthermore,we make an outlook on the future of this rapidly evolving field and proposes essential considerations for addressing emerging interdisciplinary challenges.

Key words

artificial intelligence / large language model / biology / medicine / artificial intelligence for science

Cite this article

Download Citations
Xingli CUN , Chenjun DING , Fang CHEN , et al. Innovative Application of Large Language Models in Biology and Medicine Under Artificial Intelligence for Science Paradigm[J]. Acta Academiae Medicinae Sinicae. 2025, 47(6): 969-981 https://doi.org/10.3881/j.issn.1000-503X.16605

References

[1]
Bengio Y, Ducharme R, Vincent P. A neural probabilistic language model[J]. J Mach Learn Res, 2003, 13:1137-1155.DOI:10.1145/3712001.
[2]
Kombrink S, Mikolov T, Karafiát M, et al. Recurrent neural network based language modeling in meeting recognition[J]. Interspeech, 2011, 11:2877-2880.DOI:10.21437/Interspeech.2011-720.
[3]
Kaplan J, McCandlish S, Henighan T, et al. Scaling laws for neural language models[EB/OL].[2020-01-23]. https://arxiv.org/pdf/2001.08361/1000.
[4]
Zhao WX, Zhou K, Li J, et al. A survey of large language models[EB/OL].[2023-11-24]. https://arxiv.org/abs/2303.18223.
[5]
Rosenfeld R. Two decades of statistical language modeling:where do we go from here[J]. Proc IEEE, 2000, 88(8):1270-1278.DOI:10.1109/5.880083.
[6]
Wei J, Tay Y, Bommasani R, et al. Emergent abilities of large language models[EB/OL].[2022-10-15].https://doi.org/10.48550/arXiv.2206.07682.
[7]
Brown T, Mann B, Ryder N, et al. Language models are few-shot learners[J]. NeurIPS, 2020, 33:1877-1901.DOI:10.5555/3495724.3495883.
[8]
Wei J, Bosma M, Zhao VY, et al. Finetuned language models are zero-shot learners[EB/OL].[2022-02-08].https://doi.org/10.48550/arXiv.2109.01652.
[9]
Sanh V, Webson A, Raffel C, et al. Multitask prompted training enables zero-shot task generalization[EB/OL].[2022-03-17].https://doi.org/10.48550/arXiv.2110.08207.
[10]
Wei J, Wang X, Schuurmans D, et al. Chain-of-thought prompting elicits reasoning in large language models[J]. NeurIPS, 2022, 35:24824-24837.DOI:10.5555/3600270.3602070.
[11]
Zhang S, Roller S, Goyal N, et al. Opt:open pre-trained transformer language models[EB/OL].[2022-06-21].https://doi.org/10.48550/arXiv.2205.01068.
[12]
Ouyang L, Wu J, Jiang X, et al. Training language models to follow instructions with human feedback[J]. NeurIPS, 2022, 35:27730-27744.DOI:10.5555/3600270.3602059.
[13]
Christiano PF, Leike J, Brown T, et al. Deep reinforcement learning from human preferences[J]. NeurIPS, 2017, 30:4302-4310.DOI:10.5555/3294996.3295184.
[14]
Schick T, Dwivedi-Yu J, Dessì R, et al. Toolformer:language models can teach themselves to use tools[EB/OL].[2023-02-09].https://doi.org/10.48550/arXiv.2302.04761.
[15]
Nakano R, Hilton J, Balaji S, et al. WebGPT:browser-assisted question-answering with human feedback[EB/OL].[2023-02-09].https://doi.org/10.48550/arXiv.2302.04761.
[16]
Gao L, Madaan A, Zhou S, et al. Pal:program-aided language models[J]. ICML, 2023, 202:10764-10799.DOI:10.5555/3618408.3618843.
[17]
Radford A, Jozefowicz R, Sutskever I. Learning to generate reviews and discovering sentiment[EB/OL].[2017-04-06].https://doi.org/10.48550/arXiv.1704.01444.
[18]
Radford A, Narasimhan K, Salimans T, et al. Improving language understanding by generative pre-training[EB/OL].[2018-06-11]. https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf.
[19]
Radford A, Wu J, Child R, et al. Language models are unsupervised multitask learners[EB/OL].[2019-02-14]. https://cdn.openai.com/better-language-models/language_models_are_unsupervised_multitask_learners.pdf.
[20]
Zhang Y, Sun S, Galley M, et al. DIALOGPT:large-scale generative pre-training for conversational response generation[EB/OL].[2020-05-02].https://doi.org/10.48550/arXiv.1911.00536.
[21]
Drori I, Tran S, Wang R, et al. A neural network solves,explains,and generates university math problems by program synthesis and few-shot learning at human level[J]. Proc Natl Acad Sci U S A, 2022, 119(32):e2123433119.DOI:10.1073/pnas.2123433119.
[22]
OpenAI. GPT-4 technical report[EB/OL].[2024-03-04].https://doi.org/10.48550/arXiv.2303.08774.
[23]
Bubeck S, Chandrasekaran V, Eldan R, et al. Sparks of artificial general intelligence:early experiments with GPT-4[EB/OL].[2024-03-04].https://doi.org/10.48550/arXiv.2303.08774.
[24]
Islam R, Moushi OM. Intelligent computing:proceedings of the 2025 computing conference[C].Online:Springer, Cham, 2025:47-60.DOI:10.1007/978-3-031-92611-2_4.
[25]
Touvron H, Lavril T, Izacard G, et al. Llama:open and efficient foundation language models[EB/OL].[2023-02-27].https://doi.org/10.48550/arXiv.2302.13971.
[26]
Chowdhery A, Narang S, Devlin J, et al. Palm:scaling language modeling with pathways[EB/OL].[2022-10-05].https://doi.org/10.48550/arXiv.2204.02311.
[27]
Cartwright O, Dunbar H, Radcliffe T. Evaluating privacy compliance in commercial large language models-ChatGPT,Claude,and Gemini[EB/OL].[2024-07-25].https://doi.org/10.21203/rs.3.rs-4792047/v1.
[28]
Li HL, Pang YH, Liu B. BioSeq-BLM:a platform for analyzing DNA,RNA and protein sequences based on biological language models[J]. Nucleic Acids Research, 2021, 49(22):e129.DOI:10.1093/nar/gkab829.
[29]
Wang C, Li M, He J, et al. A survey for large language models in biomedicine[EB/OL].[2024-08-29].https://doi.org/10.48550/arXiv.2409.00133.
[30]
Jumper J, Evans R, Pritzel A, et al. Highly accurate protein structure prediction with AlphaFold[J]. Nature, 2021, 596(7873):583-589.DOI:10.1038/s41586-021-03819-2.
[31]
Stahl K, Graziadei A, Dau T, et al. Protein structure prediction with in-cell photo-crosslinking mass spectrometry and deep learning[J]. Nat Biotechnol, 2023, 41:1810-1819.DOI:10.1038/s41587-023-01704-z.
[32]
Elnaggar A, Heinzinger M, Dallago C, et al. Prottrans:toward understanding the language of life through self-supervised learning[J]. IEEE Trans Pattern Anal Mach Intell, 2022, 44(10):7112-7127.DOI:10.1109/TPAMI.2021.3095381.
[33]
Heinzinger M, Weissenow K, Sanchez JG, et al. ProstT5:bilingual language model for protein sequence and structure[J]. NAR Genom Bioinform, 2024, 6(4):150.DOI:10.1093/nargab/lqae150.
[34]
Wu J, Wu F, Jiang B, et al. tFold-ab:fast and accurate antibody structure prediction without sequence homologs[EB/OL].[2022-11-13].https://doi.org/10.1101/2022.11.10.515918.
[35]
Kinch LN, Schaeffer RD, Kryshtafovych A, et al. Target classification in the 14th round of the critical assessment of protein structure prediction(CASP14)[J]. Proteins, 2021, 89(12):1618-1632.DOI:10.1002/prot.26202.
[36]
Lin Z, Akin H, Rao R, et al. Evolutionary-scale prediction of atomic-level protein structure with a language model[J]. Science, 2023, 379(6637):1123-1130.DOI:10.1126/science.ade2574.
[37]
Su J, Han C, Zhou Y, et al. Saprot:protein language modeling with structure-aware vocabulary[EB/OL].[2024-04-19].https://doi.org/10.1101/2023.10.01.560349.
[38]
Wang Z, Zhang Q, Shuang-Wei HU, et al. Multi-level protein structure pre-training via prompt learning[EB/OL].[2023-02-27]. https://openreview.net/forum?id=XGagtiJ8XC.
[39]
Nijkamp E, Ruffolo JA, Weinstein EN, et al. Progen2:exploring the boundaries of protein language models[J]. Cell Syst, 2023, 14(11):968-978.DOI:10.1016/j.cels.2023.10.002.
[40]
Hie B, Candido S, Lin Z, et al. A high-level programming language for generative protein design[EB/OL].[2022-12-22].https://doi.org/10.1101/2022.12.21.521526.
[41]
Verkuil R, Kabeli O, Du Y, et al. Language models generalize beyond natural proteins[EB/OL].[2022-12-22].https://doi.org/10.1101/2022.12.21.521521.
[42]
Zheng Z, Deng Y, Xue D, et al. Proceedings of the 40th international conference on machine learning[C]. Online:PMLR, 2023:42317-42338.DOI:10.5555/3618408.3620189.
[43]
Chen B, Cheng X, Geng Y, et al. xTrimoPGLM:unified 100B-scale pre-trained transformer for deciphering the language of protein[EB/OL].[2024-12-09].https://doi.org/10.48550/arXiv.2401.06199.
[44]
Shuai RW, Ruffolo JA, Gray JJ. IgLM:infilling language modeling for antibody sequence design[J]. Cell Syst, 2023, 14(11):979-989.DOI:10.1016/j.cels.2023.10.001.
[45]
Xu M, Yuan X, Miret S, et al. Protst:multi-modality learning of protein sequences and biomedical texts[EB/OL].[2023-01-28].https://doi.org/10.48550/arXiv.2301.12040.
[46]
Dalla-Torre H, Gonzalez L, Mendoza-Revilla J, et al. Nucleotide transformer:building and evaluating robust foundation models for human genomics[J]. Nat Methods, 2025, 22:287-297.DOI:10.1038/s41592-024-02523-z.
[47]
Nguyen E, Poli M, Faizi M, et al. Hyenadna:long-range genomic sequence modeling at single nucleotide resolution[EB/OL].[2023-11-14]. https://europepmc.org/article/med/37426456.
[48]
Mai DHA, Nguyen LT, Lee EY. TSSNote-CyaPromBERT:development of an integrated platform for highly accurate promoter prediction and visualization of Synechococcus sp.and Synechocystis sp.through a state-of-the-art natural language processing model BERT[J]. Front Genet, 2022, 13:1067562.DOI:10.3389/fgene.2022.1067562.
[49]
Zhang D, Zhang W, He B, et al. DNAGPT:a generalized pretrained tool for versatile DNA sequence analysis tasks[EB/OL].[2023-08-30].https://doi.org/10.48550/arXiv.2307.05628.
[50]
Luo H, Chen C, Shan W, et al. Proceedings of the 2022 computing conference(CompCom)[C]. Online:Springer, 2022:153-165.DOI:10.1007/978-3-031-13829-4_13.
[51]
Li J, Wu Z, Lin W, et al. iEnhancer-ELM:improve enhancer identification by extracting position-related multiscale contextual information based on enhancer language models[J]. Bioinform Adv, 2023, 3(1):vbad043.DOI:10.1093/bioadv/vbad043.
[52]
Zvyagin M, Brace A, Hippe K, et al. GenSLMs:genome-scale language models reveal SARS-CoV-2 evolutionary dynamics[J]. Int J High Perform Comput Appl, 2023, 37(6):683-705.DOI:10.1177/1094342023120115.
[53]
Chen K, Zhou Y, Ding M, et al. Self-supervised learning on millions of pre-mRNA sequences improves sequence-based RNA splicing prediction[EB/OL].[2023-05-09].https://doi.org/10.1101/2023.01.31.526427.
[54]
Wang X, Gao X, Wang G, et al. miProBERT:identification of microRNA promoters based on the pre-trained model BERT[J]. Briefings Bioinf, 2023, 24(3):bbad093.DOI:10.1093/bib/bbad093.
[55]
Zhang Y, Lang M, Jiang J, et al. Multiple sequence alignment-based RNA language model and its application to structural inference[J]. Nucleic Acids Res, 2024, 52(1):e3.DOI:10.1093/nar/gkad1031.
[56]
Wang X, Gu R, Chen Z, et al. UNI-RNA:universal pre-trained models revolutionize RNA research[EB/OL].[2023-07-12].https://doi.org/10.1101/2023.07.11.548588.
[57]
Babjac AN, Lu Z, Emrich SJ. CodonBERT:using BERT for sentiment analysis to better predict genes with low expression[J]. Proc ACM BCB, 2023, 29:1-6.DOI:10.1145/3584371.3613013.
[58]
Liu S, Li Y, Li Z, et al. A text-guided protein design framework[J]. Nat Mach Intell, 2025, 7:580-591.DOI:10.1038/s42256-025-01011-z.
[59]
Wang Z, Zhang Q, Ding K, et al. Instructprotein:aligning human and protein language via knowledge instruction[EB/OL].[2023-10-05].https://doi.org/10.48550/arXiv.2310.03269.
[60]
Li Y, Gao C, Song X, et al. DrugGPT:a GPT-based strategy for designing potential ligands targeting specific proteins[EB/OL].[2023-06-30].https://doi.org/10.1101/2023.06.29.543848.
[61]
Gao B, Qiang B, Tan H, et al. Drugclip:contrasive protein-molecule representation learning for virtual screening[EB/OL].[2023-10-10].https://doi.org/10.48550/arXiv.2310.06367.
[62]
Singhal K, Azizi S, Tu T, et al. Large language models encode clinical knowledge[J]. Nature, 2023, 620(7972):172-180.DOI:10.1038/s41586-023-06291-2.
[63]
Liu S, Wang J, Yang Y, et al. ChatGPT-powered conversational drug editing using retrieval and domain feedback[EB/OL].[2023-05-29].https://doi.org/10.48550/arXiv.2305.18090.
[64]
Luo Y, Zhang J, Fan S, et al. BiomMedGPT:open multimodal generative pre-trained transformer for biomedicine[EB/OL].[2023-08-18].https://doi.org/10.48550/arXiv.2308.09442.
[65]
Pei Q, Zhang W, Zhu J, et al. Biot5:enriching cross-modal integration in biology with chemical knowledge and natural language associations[EB/OL].[2024-01-29].https://doi.org/10.48550/arXiv.2310.07276.
[66]
Fang Y, Liang X, Zhang N, et al. Mol-instructions:a large-scale biomolecular instruction dataset for large language models[EB/OL].[2024-03-04].https://doi.org/10.48550/arXiv.2306.08018.
[67]
Yang F, Wang W, Wang F, et al. scBERT as a large-scale pretrained deep language model for cell type annotation of single-cell RNA-seq data[J]. Nat Mach Intell, 2022, 4:852-866.DOI:10.1038/s42256-022-00534-z.
[68]
Cui H, Wang C, Maan H, et al. scGPT:toward building a foundation model for single-cell multi-omics using generative AI[J]. Nat Methods, 2024, 21:1470-1480.DOI:10.1038/s41592-024-02201-0.
[69]
Theodoris CV, Xiao L, Chopra A, et al. Transfer learning enables predictions in network biology[J]. Nature, 2023, 618:616-624.DOI:10.1038/s41586-023-06139-9.
[70]
Gong J, Hao M, Cheng X, et al. xTrimoGene:an efficient and scalable representation learner for single-cell RNA-seq data[EB/OL].[2023-11-26].https://doi.org/10.48550/arXiv.2311.15156.
[71]
Hao M, Gong J, Zeng X, et al. Large-scale foundation model on single-cell transcriptomics[J]. Nat Methods, 2024, 21:1481-1491.DOI:10.1038/s41592-024-02305-7.
[72]
Levine D, Rizvi SA, Lévy S, et al. Cell2sentence:teaching large language models the language of biology[EB/OL].[2024-10-29].https://doi.org/10.1101/2023.09.11.557287.
[73]
Agrawal M, Hegselmann S, Lang H, et al. Large language models are few-shot clinical information extractors[EB/OL].[2022-11-30].https://doi.org/10.48550/arXiv.2205.12689.
[74]
Rajkomar A, Loreaux E, Liu Y, et al. Deciphering clinical abbreviations with a privacy protecting machine learning system[J]. Nat Commun, 2022, 13:7456.DOI:10.1038/s41467-022-35007-9.
[75]
Gu Y, Zhang S, Usuyama N, et al. Distilling large language models for biomedical knowledge extraction:a case study on adverse drug events[EB/OL].[2023-07-12].https://doi.org/10.48550/arXiv.2307.06439.
[76]
Swanson K, Chang H, Zou J. Predicting immune escape with pretrained protein language model embeddings[EB/OL].[2022-12-02].https://doi.org/10.1101/2022.11.30.518466.
[77]
Strubell E, Ganesh A, McCallum A. Energy and policy considerations for modern deep learning research[J]. Proc AAAI Conf Artif Intell, 2020, 34(9):13693-13696.DOI:10.1609/aaai.v34i09.7123.
[78]
Anil R, Dai AM, Firat O, et al. Palm 2 technical report[EB/OL].[2023-09-13].https://doi.org/10.48550/arXiv.2305.10403.
[79]
Thirunavukarasu AJ, Ting DSJ, Elangovan K, et al. Large language models in medicine[J]. Nat Med, 2023, 29:1930-1940.DOI:10.1038/s41591-023-02448-8.
[80]
Clusmann J, Kolbinger FR, Muti HS, et al. The future landscape of large language models in medicine[J]. Commun Med, 2023, 3:141.DOI:10.1038/s43856-023-00370-1.
[81]
Thirunavukarasu AJ, Hassan R, Mahmood S, et al. Trialling a large language model(ChatGPT)in general practice with the Applied Knowledge Test:observational study demonstrating opportunities and limitations in primary care[J]. JMIR Med Educ, 2023, 9(1):e46599.DOI:10.2196/46599.
[82]
Sarraju A, Bruemmer D, Van Iterson E, et al. Appropriateness of cardiovascular disease prevention recommendations obtained from a popular online chat-based artificial intelligence model[J]. JAMA, 2023, 329(10):842-844.DOI:10.1001/jama.2023.1044.
[83]
Van Veen D, Van Uden C, Blankemeier L, et al. Adapted large language models can outperform medical experts in clinical text summarization[J]. Nat Med, 2024, 30:1134-1142.DOI:10.1038/s41591-024-02855-5.
[84]
Li Y, Li Z, Zhang K, et al. ChatDoctor:a medical chat model fine-tuned on a large language model meta-AI(LLaMA)using medical domain knowledge[J]. Cureus, 2023, 15(6):e40895.DOI:10.7759/cureus.40895.
[85]
Liévin V, Hother CE, Winther O. Can large language models reason about medical questions[J]. Patterns, 2024, 5(3):100943.DOI:10.1016/j.patter.2024.100943.
[86]
Venigalla A, Frankle J, Carbin M. PubMed GPT:a domain-specific large language model for biomedical text[EB/OL].[2022-12-15]. https://crfm.stanford.edu/2022/12/15/pubmedgpt.html.
[87]
Peng C, Yang X, Chen A, et al. A study of generative large language model for medical research and healthcare[J]. Anpj Digit Med, 2023, 6:210.DOI:10.1038/s41746-023-00958-w.
[88]
Kung TH, Cheatham M, Medenilla A, et al. Performance of ChatGPT on USMLE:potential for AI-assisted medical education using large language models[J]. PLoS Digit Health, 2023, 2(2):e0000198.DOI:10.1371/journal.pdig.0000198.
[89]
Nori H, King N, McKinney SM, et al. Capabilities of GPT-4 on medical challenge problems[EB/OL].[2023-04-12].https://doi.org/10.48550/arXiv.2303.13375.
[90]
Singhal K, Tu T, Gottweis J, et al. Towards expert-level medical question answering with large language models[J]. Nat Med, 2025, 31:943-950.DOI:10.1038/s41591-024-03423-7.
[91]
Tu T, Azizi S, Driess D, et al. Towards generalist biomedical AI[J]. NEJM AI, 2024, 1(3):AIoa2300138.DOI:10.1056/AIoa2300138.
[92]
Perchik J. An AI revolution is brewing in medicine.What will it look like[J]. Nature, 2023, 622:686-688.DOI:10.1038/d41586-023-03302-0.
[93]
Sorin V, Klang E, Sklair-Levy M, et al. Large language model(ChatGPT)as a support tool for breast tumor board[J]. NPJ Breast Cancer, 2023, 9:44.DOI:10.1038/s41523-023-00557-8.
[94]
Moor M, Banerjee O, Abad ZSH, et al. Foundation models for generalist medical artificial intelligence[J]. Nature, 2023, 616:259-265.DOI:10.1038/s41586-023-05881-4.
[95]
Ahn S. The impending impacts of large language models on medical education[J]. Korean J Med Educ, 2023, 35(1):103-107.DOI:10.3946/kjme.2023.253.
[96]
Abd-Alrazaq A, AlSaad R, Alhuwail D, et al. Large language models in medical education:opportunities,challenges,and future directions[J]. JMIR Med Educ, 2023, 9:e48291.DOI:10.2196/48291.
[97]
Rosoł M, Głsior JS, Łaba J, et al. Evaluation of the performance of GPT-3.5 and GPT-4 on the medical final examination[J]. Sci Rep, 2023, 13:20512.DOI:10.1038/s41598-023-46995-z.
[98]
Esteva A, Robicquet A, Ramsundar B, et al. A guide to deep learning in healthcare[J]. Nat Med, 2019, 25:24-29.DOI:10.1038/s41591-018-0316-z.
[99]
Bommasani R, Hudson DA, Adeli E, et al. On the opportunities and risks of foundation models[EB/OL].[2022-07-12].https://doi.org/10.48550/arXiv.2108.07258.
[100]
Mahler M, Auza C, Albesa R, et al. Precision medicine and artificial intelligence[M]. Amsterdam:Elsevier, 2021:237-265.DOI:10.1016/B978-0-12-820239-5.00010-3.
[101]
李立果. 数据安全法下的医疗数据安全建设[J]. 信息安全, 2022, 4:128-133.
[102]
Soice EH, Rocha R, Cordova K, et al. Can large language models democratize access to dual-use biotechnology[EB/OL].[2023-06-06].https://doi.org/10.48550/arXiv.2306.03809.
[103]
Nazer LH, Zatarah R, Waldrip S, et al. Bias in artificial intelligence algorithms and recommendations for mitigation[J]. PLoS Digit Health, 2023, 2(6):e0000278.DOI:10.1371/journal.pdig.0000278.
[104]
Sau A, Sieliwonczyk E, Patlatzoglou K, et al. Artificial intelligence-enhanced electrocardiography for the identification of a sex-related cardiovascular risk continuum:a retrospective cohort study[J]. Lancet Digit Health, 2025, 7(3):e184 -e194.DOI:10.1016/S2589-7500(24)00267-8.
PDF(947 KB)

Accesses

Citation

Detail

Sections
Recommended

/