{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T21:54:12Z","timestamp":1770846852494,"version":"3.50.1"},"reference-count":32,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,12,15]],"date-time":"2025-12-15T00:00:00Z","timestamp":1765756800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,15]],"date-time":"2025-12-15T00:00:00Z","timestamp":1765756800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Nature Science Foundation of China","doi-asserted-by":"publisher","award":["62372326,62172300"],"award-info":[{"award-number":["62372326,62172300"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003395","name":"Shanghai Municipal Education Commission","doi-asserted-by":"publisher","award":["2024RGYB001"],"award-info":[{"award-number":["2024RGYB001"]}],"id":[{"id":"10.13039\/501100003395","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002858","name":"China Postdoctoral Science Foundation","doi-asserted-by":"publisher","award":["2025M771493"],"award-info":[{"award-number":["2025M771493"]}],"id":[{"id":"10.13039\/501100002858","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,12,15]]},"DOI":"10.1109\/bibm66473.2025.11356714","type":"proceedings-article","created":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T21:19:40Z","timestamp":1769721580000},"page":"70-75","source":"Crossref","is-referenced-by-count":0,"title":["BioBridge: Bridging Proteins and Language for Enhanced Biological Reasoning with LLMs"],"prefix":"10.1109","author":[{"given":"Yujia","family":"Wang","sequence":"first","affiliation":[{"name":"School of Computer Science and Technology, Tongji University,Shanghai,China"}]},{"given":"Jihong","family":"Guan","sequence":"additional","affiliation":[{"name":"School of Computer Science and Technology, Tongji University,Shanghai,China"}]},{"given":"Wengen","family":"Li","sequence":"additional","affiliation":[{"name":"School of Computer Science and Technology, Tongji University,Shanghai,China"}]},{"given":"Shuigeng","family":"Zhou","sequence":"additional","affiliation":[{"name":"College of Computer Science and Artificial Intelligence, Fudan University,Shanghai,China"}]},{"given":"Xuhong","family":"Wang","sequence":"additional","affiliation":[{"name":"Shanghai AI Laboratory,Shanghai,China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.52202\/079017-2810"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1146\/annurev-orgpsych-032414-111335"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.418"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.naacl-main.351"},{"key":"ref5","article-title":"Recyclable tuning for continual pretraining[J]","author":"Qin","year":"2023","journal-title":"arXiv preprint arXiv"},{"key":"ref6","article-title":"Conpet: Continual parameter-efficient tuning for large language models[J]","author":"Song","year":"2023","journal-title":"arXiv preprint arXiv"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2019.01.012"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i10.28948"},{"key":"ref9","article-title":"Multitask prompted training enables zero-shot task generalization[J]","author":"Sanh","year":"2021","journal-title":"arXiv preprint arXiv"},{"key":"ref10","article-title":"ProteinCLIP: enhancing protein language models with natural language[J]","author":"Wu","journal-title":"bioRxiv, 2024"},{"key":"ref11","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision[C]","volume-title":"International conference on machine learning","author":"Radford"},{"key":"ref12","article-title":"Protrek: Navigating the protein universe through tri-modal contrastive learning[J]","author":"Su","journal-title":"bioRxiv, 2024"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.acl-long.324"},{"key":"ref14","article-title":"Diffusion language models are versatile protein learners[J]","author":"Wang","year":"2024","journal-title":"arXiv preprint arXiv"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.5c00585"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19787-1_11"},{"key":"ref17","first-page":"38749","article-title":"Protst: Multi-modality learning of protein sequences and biomedical texts[C]","volume-title":"International Conference on Machine Learning","author":"Xu"},{"issue":"1","key":"ref18","article-title":"PubMed: the bibliographic database[J]","volume":"2","author":"Canese","year":"2013","journal-title":"The NCBI handbook, 2013"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btac598"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-59745-535-0_4"},{"key":"ref21","article-title":"Instruction tuning with gpt-4[J]","author":"Peng","year":"2023","journal-title":"arXiv preprint arXiv"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.410"},{"key":"ref23","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision[C]","volume-title":"International conference on machine learning","author":"Radford"},{"key":"ref24","article-title":"Ontoprotein: Protein pretraining with gene ontology embedding[J]","author":"Zhang","year":"2022","journal-title":"arXiv preprint arXiv"},{"key":"ref25","first-page":"19730","article-title":"Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models[C]","volume-title":"International conference on machine learning","author":"Li"},{"key":"ref26","article-title":"PFMBench: Protein Foundation Model Benchmark[J]","author":"Gao","year":"2025","journal-title":"arXiv preprint arXiv"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btaf396"},{"key":"ref28","article-title":"xTrimoPGLM: unified 100B-scale pretrained transformer for deciphering the language of protein[J]","author":"Chen","year":"2024","journal-title":"arXiv preprint arXiv"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3095381"},{"key":"ref30","article-title":"Saprot: Protein language modeling with structure-aware vocabulary[J]","author":"Su","journal-title":"BioRxiv, 2023"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1093\/nargab\/lqae150"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1126\/science.ads0018"}],"event":{"name":"2025 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)","location":"Wuhan, China","start":{"date-parts":[[2025,12,15]]},"end":{"date-parts":[[2025,12,18]]}},"container-title":["2025 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11355913\/11355975\/11356714.pdf?arnumber=11356714","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T20:54:34Z","timestamp":1770843274000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11356714\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,15]]},"references-count":32,"URL":"https:\/\/doi.org\/10.1109\/bibm66473.2025.11356714","relation":{},"subject":[],"published":{"date-parts":[[2025,12,15]]}}}