{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T03:01:28Z","timestamp":1775790088826,"version":"3.50.1"},"reference-count":44,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,9,30]],"date-time":"2025-09-30T00:00:00Z","timestamp":1759190400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,30]],"date-time":"2025-09-30T00:00:00Z","timestamp":1759190400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/100020595","name":"National Science and Technology Council","doi-asserted-by":"publisher","award":["NSTC112-2321-B-016-003"],"award-info":[{"award-number":["NSTC112-2321-B-016-003"]}],"id":[{"id":"10.13039\/100020595","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100019000","name":"Medical Affairs Bureau","doi-asserted-by":"publisher","award":["MND-MAB-C07-113021"],"award-info":[{"award-number":["MND-MAB-C07-113021"]}],"id":[{"id":"10.13039\/501100019000","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Med Syst"],"DOI":"10.1007\/s10916-025-02263-3","type":"journal-article","created":{"date-parts":[[2025,9,30]],"date-time":"2025-09-30T01:34:35Z","timestamp":1759196075000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["A Pretraining Approach for Small-sample Training Employing Radiographs (PASTER): a Multimodal Transformer Trained by Chest Radiography and Free-text Reports"],"prefix":"10.1007","volume":"49","author":[{"given":"Kai-Chieh","family":"Chen","sequence":"first","affiliation":[]},{"given":"Matthew","family":"Kuo","sequence":"additional","affiliation":[]},{"given":"Chun-Ho","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Hao-Chun","family":"Liao","sequence":"additional","affiliation":[]},{"given":"Dung-Jang","family":"Tsai","sequence":"additional","affiliation":[]},{"given":"Shing-An","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Chih-Wei","family":"Hsiang","sequence":"additional","affiliation":[]},{"given":"Cheng-Kuang","family":"Chang","sequence":"additional","affiliation":[]},{"given":"Kai-Hsiung","family":"Ko","sequence":"additional","affiliation":[]},{"given":"Yi-Chih","family":"Hsu","sequence":"additional","affiliation":[]},{"given":"Wei-Chou","family":"Chang","sequence":"additional","affiliation":[]},{"given":"Guo-Shu","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Wen-Hui","family":"Fang","sequence":"additional","affiliation":[]},{"given":"Chin-Sheng","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Shih-Hua","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Yuan-Hao","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Yi-Jen","family":"Hung","sequence":"additional","affiliation":[]},{"given":"Chien-Sung","family":"Tsai","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2337-2096","authenticated-orcid":false,"given":"Chin","family":"Lin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,30]]},"reference":[{"issue":"2","key":"2263_CR1","doi-asserted-by":"publisher","first-page":"545","DOI":"10.1378\/chest.10-1302","volume":"141","author":"S Raoof","year":"2012","unstructured":"Raoof, S., et al., Interpretation of plain chest roentgenogram. Chest, 2012. 141(2): p. 545\u2013558.","journal-title":"Chest"},{"key":"2263_CR2","doi-asserted-by":"crossref","unstructured":"\u00c7all\u0131, E., et al., Deep learning for chest X-ray analysis: A survey. 2021. 72: p. 102125.","DOI":"10.1016\/j.media.2021.102125"},{"key":"2263_CR3","doi-asserted-by":"crossref","unstructured":"Litjens, G., et al., A survey on deep learning in medical image analysis. 2017. 42: p. 60\u201388.","DOI":"10.1016\/j.media.2017.07.005"},{"issue":"6","key":"2263_CR4","first-page":"e271-e297","volume":"1","author":"X Liu","year":"2019","unstructured":"Liu, X., et al., A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis. Lancet Digit Health, 2019. 1(6): p. e271-e297.","journal-title":"Lancet Digit Health"},{"issue":"1","key":"2263_CR5","doi-asserted-by":"publisher","first-page":"65","DOI":"10.1038\/s41746-021-00438-z","volume":"4","author":"R Aggarwal","year":"2021","unstructured":"Aggarwal, R., et al., Diagnostic accuracy of deep learning in medical imaging: a systematic review and meta-analysis. NPJ Digit Med, 2021. 4(1): p. 65.","journal-title":"NPJ Digit Med"},{"issue":"1","key":"2263_CR6","doi-asserted-by":"publisher","first-page":"31","DOI":"10.1038\/s41591-021-01614-0","volume":"28","author":"P Rajpurkar","year":"2022","unstructured":"Rajpurkar, P., et al., AI in health and medicine. Nat Med, 2022. 28(1): p. 31\u201338.","journal-title":"Nat Med"},{"key":"2263_CR7","doi-asserted-by":"crossref","unstructured":"Willemink, M.J., et al., Preparing medical imaging data for machine learning. 2020. 295(1): p. 4\u201315.","DOI":"10.1148\/radiol.2020192224"},{"key":"2263_CR8","doi-asserted-by":"crossref","unstructured":"Joskowicz, L., et al., Inter-observer variability of manual contour delineation of structures in CT. 2019. 29: p. 1391\u20131399.","DOI":"10.1007\/s00330-018-5695-5"},{"key":"2263_CR9","doi-asserted-by":"crossref","unstructured":"Lutnick, B., et al., An integrated iterative annotation technique for easing neural network training in medical image analysis. 2019. 1(2): p. 112\u2013119.","DOI":"10.1038\/s42256-019-0018-3"},{"key":"2263_CR10","doi-asserted-by":"crossref","unstructured":"Esteva, A., et al., A guide to deep learning in healthcare. 2019. 25(1): p. 24\u201329.","DOI":"10.1038\/s41591-018-0316-z"},{"key":"2263_CR11","doi-asserted-by":"crossref","unstructured":"Kim, H.E., et al., Transfer learning for medical image classification: a literature review. 2022. 22(1): p. 69.","DOI":"10.1186\/s12880-022-00793-7"},{"key":"2263_CR12","unstructured":"Chen, T., et al. A simple framework for contrastive learning of visual representations. in International conference on machine learning. 2020. PMLR."},{"key":"2263_CR13","doi-asserted-by":"crossref","unstructured":"He, K., et al. Momentum contrast for unsupervised visual representation learning. in Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. 2020.","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"2263_CR14","doi-asserted-by":"crossref","unstructured":"He, K., et al. Masked autoencoders are scalable vision learners. in Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. 2022.","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"2263_CR15","unstructured":"Radford, A., et al. Learning transferable visual models from natural language supervision. in International conference on machine learning. 2021. PMLR."},{"key":"2263_CR16","unstructured":"Zhang, Y., et al. Contrastive learning of medical visual representations from paired images and text. in Machine Learning for Healthcare Conference. 2022. PMLR."},{"key":"2263_CR17","doi-asserted-by":"crossref","unstructured":"Zhang, S., et al., A Multimodal Biomedical Foundation Model Trained from Fifteen Million Image\u2013Text Pairs. Nejm Ai, 2025. 2(1).","DOI":"10.1056\/AIoa2400640"},{"key":"2263_CR18","doi-asserted-by":"crossref","unstructured":"Wang, Z., et al., MedCLIP: Contrastive Learning from Unpaired Medical Images and Text. Proc Conf Empir Methods Nat Lang Process, 2022. 2022: p. 3876\u20133887.","DOI":"10.18653\/v1\/2022.emnlp-main.256"},{"key":"2263_CR19","doi-asserted-by":"crossref","unstructured":"Benjamens, S., P. Dhunnoo, and B.J.N.d.m. Mesk\u00f3, The state of artificial intelligence-based FDA-approved medical devices and algorithms: an online database. 2020. 3(1): p. 118.","DOI":"10.1038\/s41746-020-00324-0"},{"key":"2263_CR20","doi-asserted-by":"crossref","unstructured":"Irvin, J., et al. Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison. in Proceedings of the AAAI conference on artificial intelligence. 2019.","DOI":"10.1609\/aaai.v33i01.3301590"},{"key":"2263_CR21","doi-asserted-by":"crossref","unstructured":"Wang, X., et al. Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. in Proceedings of the IEEE conference on computer vision and pattern recognition. 2017.","DOI":"10.1109\/CVPR.2017.369"},{"key":"2263_CR22","doi-asserted-by":"crossref","unstructured":"Huang, G., et al. Densely connected convolutional networks. in Proceedings of the IEEE conference on computer vision and pattern recognition. 2017.","DOI":"10.1109\/CVPR.2017.243"},{"key":"2263_CR23","doi-asserted-by":"crossref","unstructured":"Wang, H., et al., Scientific discovery in the age of artificial intelligence. 2023. 620(7972): p. 47\u201360.","DOI":"10.1038\/s41586-023-06221-2"},{"key":"2263_CR24","doi-asserted-by":"crossref","unstructured":"Manzari, O.N., et al., MedViT: a robust vision transformer for generalized medical image classification. 2023. 157: p. 106791.","DOI":"10.1016\/j.compbiomed.2023.106791"},{"key":"2263_CR25","unstructured":"Dosovitskiy, A., et al., An image is worth 16x16 words: Transformers for image recognition at scale. 2020."},{"key":"2263_CR26","doi-asserted-by":"crossref","unstructured":"Sellergren, A.B., et al., Simplified transfer learning for chest radiography models using less data. 2022. 305(2): p. 454\u2013465.","DOI":"10.1148\/radiol.212482"},{"key":"2263_CR27","doi-asserted-by":"crossref","unstructured":"Bannur, S., et al. Learning to exploit temporal structure for biomedical vision-language processing. in Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 2023.","DOI":"10.1109\/CVPR52729.2023.01442"},{"key":"2263_CR28","unstructured":"Sowrirajan, H., et al. Moco pretraining improves representation and transferability of chest x-ray models. in Medical Imaging with Deep Learning. 2021. PMLR."},{"issue":"6","key":"2263_CR29","first-page":"720","volume":"38","author":"PT Lauzier","year":"2022","unstructured":"Lauzier, P.T. and B.J.W. Chow, Artificial Intelligence Detection of Left Ventricular Systolic Dysfunction Using Chest X-Rays: Prospective Validation, Please. The Canadian journal of cardiology, 2022. 38(6): p. 720\u2013722.","journal-title":"The Canadian journal of cardiology"},{"key":"2263_CR30","doi-asserted-by":"crossref","unstructured":"Yu, A.C., B. Mohajer, and J.J.R.A.I. Eng, External validation of deep learning algorithms for radiologic diagnosis: a systematic review. 2022. 4(3): p. e210064.","DOI":"10.1148\/ryai.210064"},{"key":"2263_CR31","doi-asserted-by":"crossref","unstructured":"van Ginneken, B.J.R., Deep learning for triage of chest radiographs: should every institution train its own system?Radiology, 2019. Radiological Society of North America: Radiology. p. 545\u2013546.","DOI":"10.1148\/radiol.2018182318"},{"key":"2263_CR32","doi-asserted-by":"crossref","unstructured":"Dunnmon, J.A., et al., Assessment of convolutional neural networks for automated classification of chest radiographs. 2019. 290(2): p. 537\u2013544.","DOI":"10.1148\/radiol.2018181422"},{"key":"2263_CR33","doi-asserted-by":"crossref","unstructured":"Tiu, E., et al., Expert-level detection of pathologies from unannotated chest X-ray images via self-supervised learning. 2022. 6(12): p. 1399\u20131406.","DOI":"10.1038\/s41551-022-00936-9"},{"key":"2263_CR34","doi-asserted-by":"crossref","unstructured":"Vaid, A., et al., A foundational vision transformer improves diagnostic performance for electrocardiograms. 2023. 6(1): p. 108.","DOI":"10.1038\/s41746-023-00840-9"},{"key":"2263_CR35","unstructured":"Kumar, A., et al. Fine-Tuning can Distort Pretrained Features and Underperform Out-of-Distribution. in International Conference on Learning Representations. 2022."},{"issue":"2","key":"2263_CR36","doi-asserted-by":"publisher","first-page":"454","DOI":"10.1148\/radiol.212482","volume":"305","author":"AB Sellergren","year":"2022","unstructured":"Sellergren, A.B., et al., Simplified transfer learning for chest radiography models using less data. Radiology, 2022. 305(2): p. 454\u2013465.","journal-title":"Radiology"},{"key":"2263_CR37","doi-asserted-by":"crossref","unstructured":"Rieke, N., et al., The future of digital health with federated learning. 2020. 3(1): p. 1\u20137.","DOI":"10.1038\/s41746-020-00323-1"},{"key":"2263_CR38","doi-asserted-by":"crossref","unstructured":"Liu, W.-T., et al., A deep-learning algorithm-enhanced system integrating electrocardiograms and chest X-rays for diagnosing aortic dissection. 2022. 38(2): p. 160\u2013168.","DOI":"10.1016\/j.cjca.2021.09.028"},{"key":"2263_CR39","doi-asserted-by":"crossref","unstructured":"Rajpurkar, P., et al., Deep learning for chest radiograph diagnosis: A retrospective comparison of the CheXNeXt algorithm to practicing radiologists. 2018. 15(11): p. e1002686.","DOI":"10.1371\/journal.pmed.1002686"},{"key":"2263_CR40","doi-asserted-by":"crossref","unstructured":"Bluemke, D.A., et al., Assessing radiology research on artificial intelligence: a brief guide for authors, reviewers, and readers\u2014from the radiology editorial board.Radiology, 2020. Radiological Society of North America: Radiology. p. 487\u2013489.","DOI":"10.1148\/radiol.2019192515"},{"issue":"6","key":"2263_CR41","first-page":"763","volume":"38","author":"C Hsiang","year":"2022","unstructured":"Hsiang, C., et al., Detection of Left Ventricular Systolic Dysfunction Using an Artificial Intelligence-Enabled Chest X-Ray. The Canadian journal of cardiology, 2022. 38(6): p. 763\u2013773.","journal-title":"The Canadian journal of cardiology"},{"key":"2263_CR42","doi-asserted-by":"crossref","unstructured":"Seah, J.C., et al., Chest radiographs in congestive heart failure: visualizing neural network learning. 2019. 290(2): p. 514\u2013522.","DOI":"10.1148\/radiol.2018180887"},{"issue":"1","key":"2263_CR43","doi-asserted-by":"publisher","first-page":"4039","DOI":"10.1038\/s41467-023-39631-x","volume":"14","author":"A Pyrros","year":"2023","unstructured":"Pyrros, A., et al., Opportunistic detection of type 2 diabetes using deep learning from frontal chest radiographs. Nat Commun, 2023. 14(1): p. 4039.","journal-title":"Nat Commun"},{"key":"2263_CR44","doi-asserted-by":"crossref","unstructured":"Berland, L.L., et al., Managing incidental findings on abdominal CT: white paper of the ACR incidental findings committee. 2010. 7(10): p. 754\u2013773.","DOI":"10.1016\/j.jacr.2010.06.013"}],"container-title":["Journal of Medical Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10916-025-02263-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10916-025-02263-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10916-025-02263-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,30]],"date-time":"2025-09-30T14:19:47Z","timestamp":1759241987000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10916-025-02263-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,30]]},"references-count":44,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["2263"],"URL":"https:\/\/doi.org\/10.1007\/s10916-025-02263-3","relation":{},"ISSN":["1573-689X"],"issn-type":[{"value":"1573-689X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,9,30]]},"assertion":[{"value":"24 February 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 September 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 September 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}},{"value":"This study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Tri-Service General Hospital (IRB NO. C20230519). The IRB approved the study protocol and waived the requirement for individual informed consent due to the use of fully anonymized and retrospective data.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Institutional Review Board"}},{"value":"All the data were obtained from the hospital\u2019s quality control center, fully anonymized prior to analysis, and exempt from informed consent as approved by the Institutional Review Board.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed Consent"}},{"value":"The authors have no conflicts of interest to declare.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"120"}}