{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,18]],"date-time":"2026-04-18T07:47:04Z","timestamp":1776498424161,"version":"3.51.2"},"reference-count":47,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T00:00:00Z","timestamp":1770076800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T00:00:00Z","timestamp":1772582400000},"content-version":"vor","delay-in-days":29,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001691","name":"Japan Society for the Promotion of Science","doi-asserted-by":"publisher","award":["24KJ0830"],"award-info":[{"award-number":["24KJ0830"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["BMC Med Inform Decis Mak"],"DOI":"10.1186\/s12911-026-03370-y","type":"journal-article","created":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T08:07:48Z","timestamp":1770106068000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Textbook-level medical knowledge in large language models: comparative evaluation using Japanese National Medical Examination"],"prefix":"10.1186","volume":"26","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6320-544X","authenticated-orcid":false,"given":"Mingxin","family":"Liu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6251-3587","authenticated-orcid":false,"given":"Tsuyoshi","family":"Okuhara","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0363-7563","authenticated-orcid":false,"given":"Zhehao","family":"Dai","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0008-7910-1977","authenticated-orcid":false,"given":"Minghong","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Wenqiang","family":"Yin","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7877-9753","authenticated-orcid":false,"given":"Hiroko","family":"Okada","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1431-4786","authenticated-orcid":false,"given":"Emi","family":"Furukawa","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5934-0681","authenticated-orcid":false,"given":"Takahiro","family":"Kiuchi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,3]]},"reference":[{"key":"3370_CR1","unstructured":"OpenAI. ChatGPT. https:\/\/chat.openai.com\/chat. Accessed 8 Aug 2025."},{"key":"3370_CR2","doi-asserted-by":"publisher","unstructured":"Tsang R. Practical Applications of ChatGPT in undergraduate medical education. J Med Educ Curric Dev. 2023;10:23821205231178449. Published 24 May 2023. https:\/\/doi.org\/10.1177\/23821205231178449. PMID: 37255525.","DOI":"10.1177\/23821205231178449"},{"key":"3370_CR3","doi-asserted-by":"publisher","unstructured":"Hristidis V, Ruggiano N, Brown EL, Ganta SRR, Stewart S. ChatGPT vs Google for queries related to dementia and other cognitive decline: comparison of results. J Med Internet Res. 2023;25:e48966. Published 25 Jul 2023. https:\/\/doi.org\/10.2196\/48966. PMID: 37490317.","DOI":"10.2196\/48966"},{"key":"3370_CR4","doi-asserted-by":"publisher","first-page":"e60807","DOI":"10.2196\/60807","volume":"26","author":"M Liu","year":"2024","unstructured":"Liu M, Okuhara T, Chang X, Shirabe R, Nishiie Y, Okada H, Kiuchi T. Performance of ChatGPT across different versions in medical licensing examinations worldwide: systematic review and meta-analysis. J Med Internet Res. 2024;26:e60807. https:\/\/doi.org\/10.2196\/60807.","journal-title":"J Med Internet Res"},{"issue":"1","key":"3370_CR5","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1016\/j.identj.2024.10.014","volume":"75","author":"M Liu","year":"2025","unstructured":"Liu M, Okuhara T, Huang W, Ogihara A, Nagao HS, Okada H, Kiuchi T. Large Language models in dental licensing examinations: systematic review and Meta-Analysis. Int Dent J. 2025;75(1):213\u201322. https:\/\/doi.org\/10.1016\/j.identj.2024.10.014.","journal-title":"Int Dent J"},{"issue":"10","key":"3370_CR6","doi-asserted-by":"publisher","first-page":"e0312771","DOI":"10.1371\/journal.pone.0312771","volume":"19","author":"M Liu","year":"2024","unstructured":"Liu M, Okuhara T, Chang X, Okada H, Kiuchi T. Performance of ChatGPT in medical licensing examinations in countries worldwide: A systematic review and meta-analysis protocol. PLoS ONE. 2024;19(10):e0312771. https:\/\/doi.org\/10.1371\/journal.pone.0312771.","journal-title":"PLoS ONE"},{"key":"3370_CR7","doi-asserted-by":"publisher","first-page":"e67244","DOI":"10.2196\/67244","volume":"11","author":"O Bolgova","year":"2025","unstructured":"Bolgova O, Shypilova I, Mavrych V. Large Language models in biochemistry education: comparative evaluation of performance. JMIR Med Educ. 2025;11:e67244. https:\/\/doi.org\/10.2196\/67244. PMID: 40209205; PMCID: PMC12005600.","journal-title":"JMIR Med Educ"},{"issue":"3","key":"3370_CR8","doi-asserted-by":"publisher","first-page":"378","DOI":"10.1111\/1471-0528.17641","volume":"131","author":"G Levin","year":"2024","unstructured":"Levin G, Horesh N, Brezinov Y, Meyer R. Performance of ChatGPT in medical examinations: a systematic review and a meta-analysis. BJOG. 2024;131(3):378\u201380. https:\/\/doi.org\/10.1111\/1471-0528.17641.","journal-title":"BJOG"},{"issue":"4","key":"3370_CR9","doi-asserted-by":"publisher","first-page":"745","DOI":"10.1007\/s10439-023-03318-7","volume":"52","author":"M Alessandri Bonetti","year":"2024","unstructured":"Alessandri Bonetti M, Giorgino R, Gallo Afflitto G, De Lorenzi F, Egro FM. How does ChatGPT perform on the Italian residency admission National exam compared to 15,869 medical graduates? Ann Biomed Eng. 2024;52(4):745\u20139. https:\/\/doi.org\/10.1007\/s10439-023-03318-7.","journal-title":"Ann Biomed Eng"},{"key":"3370_CR10","doi-asserted-by":"publisher","unstructured":"Aljindan FK, Al Qurashi AA, Albalawi IAS et al. ChatGPT Conquers the saudi medical licensing exam: exploring the accuracy of artificial intelligence in medical knowledge assessment and implications for modern medical education. Cureus. 2023;15(9):e45043. Published 11 Sep 2023. https:\/\/doi.org\/10.7759\/cureus.45043. PMID: 37829968.","DOI":"10.7759\/cureus.45043"},{"key":"3370_CR11","doi-asserted-by":"publisher","unstructured":"Armitage RC. Performance of generative pre-trained transformer-4 (GPT-4) in membership of the royal college of general practitioners (MRCGP)-style examination questions. Postgrad Med J., 2024;100(1182): 274\u2013275. https:\/\/doi.org\/10.1093\/postmj\/qgad128. PMID: 38142282.","DOI":"10.1093\/postmj\/qgad128"},{"key":"3370_CR12","doi-asserted-by":"publisher","unstructured":"Ebrahimian M, Behnam B, Ghayebi N, Sobhrakhshankhah E. ChatGPT in Iranian medical licensing examination: evaluating the diagnostic accuracy and decision-making capabilities of an AI-based model. BMJ Health Care Inform. 2023;30(1):e100815. Published 11 Dec 2023. https:\/\/doi.org\/10.1136\/bmjhci-2023-100815. PMID: 38081765.","DOI":"10.1136\/bmjhci-2023-100815"},{"key":"3370_CR13","doi-asserted-by":"publisher","unstructured":"Fang C, Wu Y, Fu W et al. How does GPT-4 preform on non-English national medical licensing examination? an evaluation in Chinese language. PLOS Digit Health. 2023;2(12):e0000397. Published 1 Dec 2023. https:\/\/doi.org\/10.1371\/journal.pdig.0000397. PMID: 38039286.","DOI":"10.1371\/journal.pdig.0000397"},{"key":"3370_CR14","doi-asserted-by":"publisher","unstructured":"Flores-Cohaila JA, Garc\u00eda-Vicente A, Vizcarra-Jim\u00e9nez SF, et al. Performance of ChatGPT on the peruvian national licensing medical examination: cross-sectional study. JMIR Med Educ. 2023;9:e48039. Published 28 Sep 2023. https:\/\/doi.org\/10.2196\/48039. PMID: 37768724.","DOI":"10.2196\/48039"},{"key":"3370_CR15","doi-asserted-by":"publisher","unstructured":"Garabet R, Mackey BP, Cross J, Weingarten M. GPT-4 Performance on USMLE Step 1 Style Questions and Its Implications for Medical Education: A Comparative Study Across Systems and Disciplines. Med Sci Educ. 2023;34(1):145\u2013152. Published 27 Dec 2023. https:\/\/doi.org\/10.1007\/s40670-023-01956-z. PMID: 38510401.","DOI":"10.1007\/s40670-023-01956-z"},{"key":"3370_CR16","doi-asserted-by":"publisher","unstructured":"Gilson A, Safranek CW, Huang T et al. How Does ChatGPT perform on the united states medical licensing examination (USMLE)? The implications of large language models for medical education and knowledge assessment [published correction appears in JMIR Med Educ. 2024;10:e57594]. JMIR Med Educ. 2023;9:e45312. Published 8 Feb 2023. https:\/\/doi.org\/10.2196\/45312. PMID: 36753318.","DOI":"10.2196\/45312"},{"issue":"10","key":"3370_CR17","doi-asserted-by":"publisher","first-page":"e20230848","DOI":"10.1590\/1806-9282.20230848","volume":"69","author":"M Gobira","year":"2023","unstructured":"Gobira M, Nakayama LF, Moreira R, Andrade E, Regatieri CVS, R Belfort Jr. Performance of GPT-4 in answering questions from the Brazilian National examination for medical degree revalidation. Rev Assoc Med Bras (1992). 2023;69(10):e20230848. https:\/\/doi.org\/10.1590\/1806-9282.20230848. Published 2023 Sep 25. PMID: 37792871.","journal-title":"Rev Assoc Med Bras (1992)"},{"issue":"2","key":"3370_CR18","doi-asserted-by":"publisher","first-page":"192","DOI":"10.1097\/ACM.0000000000005549","volume":"99","author":"V Yaneva","year":"2024","unstructured":"Yaneva V, Baldwin P, Jurich DP, Swygert K, Clauser BE. Examining ChatGPT performance on USMLE sample items and implications for assessment. Acad Med. 2024;99(2):192\u20137. https:\/\/doi.org\/10.1097\/ACM.0000000000005549.","journal-title":"Acad Med"},{"key":"3370_CR19","doi-asserted-by":"publisher","unstructured":"Liu, M., Okuhara, T., Dai, Z., Huang, W., Gu, L., Okada, H., \u2026 Kiuchi, T. (2025).Evaluating the Effectiveness of advanced large language models in medical Knowledge:A Comparative study using Japanese national medical examination. Int J Med Inform. 193;105673. https:\/\/doi.org\/10.1016\/j.ijmedinf.2024.105673","DOI":"10.1016\/j.ijmedinf.2024.105673"},{"issue":"1","key":"3370_CR20","doi-asserted-by":"publisher","first-page":"2534065","DOI":"10.1080\/10872981.2025.2534065","volume":"30","author":"V Mavrych","year":"2025","unstructured":"Mavrych V, Yousef EM, Yaqinuddin A, Bolgova O. Large Language models in medical education: a comparative cross-platform evaluation in answering histological questions. Med Educ Online. 2025;30(1):2534065. https:\/\/doi.org\/10.1080\/10872981.2025.2534065. Epub 2025 Jul 12. PMID: 40651009; PMCID: PMC12258195.","journal-title":"Med Educ Online"},{"key":"3370_CR21","doi-asserted-by":"publisher","unstructured":"Bolgova O, Ganguly P, Mavrych V. Comparative analysis of LLMs performance in medical embryology: A cross-platform study of ChatGPT, Claude, Gemini, and Copilot. Anat Sci Educ. 2025;18(7):718\u2013726. https:\/\/doi.org\/10.1002\/ase.70044. Epub 2025 May 11. PMID: 40350555.","DOI":"10.1002\/ase.70044"},{"issue":"1","key":"3370_CR22","doi-asserted-by":"publisher","first-page":"74","DOI":"10.1007\/s10916-025-02213-z","volume":"49","author":"J Wu","year":"2025","unstructured":"Wu J, Wang Z, Qin Y. Performance of DeepSeek-R1 and ChatGPT-4o on the Chinese National medical licensing examination: a comparative study. J Med Syst. 2025;49(1):74. https:\/\/doi.org\/10.1007\/s10916-025-02213-z.","journal-title":"J Med Syst"},{"key":"3370_CR23","unstructured":"Gemini 2.5 pro. Google Cloud. https:\/\/cloud.google.com\/vertex-ai\/generative-ai\/docs\/models\/gemini\/2-5-pro. Accessed 24 Aug\u00a0 2025."},{"key":"3370_CR24","unstructured":"Grok 4. xAI. https:\/\/x.ai\/news\/grok-4. Accessed August 24, 2025."},{"key":"3370_CR25","unstructured":"Claude Opus 4. 1. Anthropic. https:\/\/www.anthropic.com\/news\/claude-opus-4-1. Accessed 24 Aug 2025."},{"key":"3370_CR26","unstructured":"Introducing, GPT-5. OpenAI. https:\/\/openai.com\/ja-JP\/index\/introducing-gpt-5\/. Accessed 24 Aug 2025."},{"key":"3370_CR27","unstructured":"Ministry of Health, Labour and Welfare. Japanese National Medical Examination. https:\/\/www.mhlw.go.jp\/kouseiroudoushou\/shikaku_shiken\/ishi\/. Accessed 8 Aug 2025."},{"key":"3370_CR28","unstructured":"Information Available in English. Ministry of Health, Labour and Welfare. https:\/\/www.mhlw.go.jp\/stf\/seisakunitsuite\/bunya\/topics_150873_139_140.html. Accessed 8 Aug 2025."},{"key":"3370_CR29","unstructured":"GPT-5 Chat. OpenAI Platform. https:\/\/platform.openai.com\/docs\/models\/gpt-5-chat-latest. Accessed 24 Aug 2025."},{"key":"3370_CR30","unstructured":"Models and Pricing. xAI. https:\/\/docs.x.ai\/docs\/models. Accessed August 24, 2025."},{"key":"3370_CR31","unstructured":"Gemini. GoogleDeepMind. https:\/\/deepmind.google\/models\/gemini\/pro\/. Accessed 24 Aug 2025."},{"key":"3370_CR32","unstructured":"Models overview. Anthropic. https:\/\/docs.anthropic.com\/en\/docs\/about-claude\/models\/overview. Accessed 24 Aug 2025."},{"key":"3370_CR33","unstructured":"Trends in Average Scores for the national medical licensing examination. (Past 5 Years) explanation of passing standards and regional trends. Ishin-Kai. https:\/\/ishin-kai.info\/column\/exam\/4706?utm_source=chatgpt.com. Accessed 21 Dec 2025."},{"key":"3370_CR34","doi-asserted-by":"crossref","unstructured":"Wei J, Wang X, Schuurmans D, Bosma M, Xia F, Chi E, Zhou D. Chain-of-thought prompting elicits reasoning in large language models. Adv Neural Inform Proc Syst. 2022;35:24824\u201324837.","DOI":"10.52202\/068431-1800"},{"key":"3370_CR35","unstructured":"Lightman H, Kosaraju V, Burda Y, Edwards H, Baker B, Lee T, Cobbe K. Let\u2019s verify step by step. In the twelfth international conference on learning representations. 2023."},{"key":"3370_CR36","unstructured":"Jiang J, Huang J, Aizawa A. JMedBench: a benchmark for evaluating Japanese biomedical large language models. 2024."},{"key":"3370_CR37","unstructured":"Lewis P, Perez E, Piktus A, Petroni F, Karpukhin V, Goyal N, Kiela D. Retrieval-augmented generation for knowledge-intensive nlp tasks. Adv Neural Inform Proc Syst. 2020;33:9459\u20139474."},{"key":"3370_CR38","doi-asserted-by":"publisher","unstructured":"Guillen-Grima F, Guillen-Aguinaga S, Guillen-Aguinaga L et al. Evaluating the efficacy of ChatGPT in navigating the Spanish medical residency entrance examination (MIR): Promising horizons for ai in clinical medicine. Clin Pract. 2023;13(6):1460\u20131487. Published 20 Nov 2023. https:\/\/doi.org\/10.3390\/clinpract13060130. PMID: 37987431.","DOI":"10.3390\/clinpract13060130"},{"key":"3370_CR39","doi-asserted-by":"publisher","unstructured":"Nakao T, Miki S, Nakamura Y et al. Capability of GPT-4V(ision) in the Japanese national medical licensing examination: evaluation study. JMIR Med Educ. 2024;10:e54393. Published 12 Mar 2024. https:\/\/doi.org\/10.2196\/54393. PMID: 38470459.","DOI":"10.2196\/54393"},{"key":"3370_CR40","doi-asserted-by":"publisher","unstructured":"Khorshidi H, Mohammadi A, Yousem DM, Abolghasemi J, Ansari G, Mirza-Aghazadeh-Attari M, Ardakani AA. Application of ChatGPT in multilingual medical education: How does ChatGPT fare in 2023\u2019s Iranian residency entrance examination. Inform Med Unl. 2023;41:101314. https:\/\/doi.org\/10.1016\/j.imu.2023.101314","DOI":"10.1016\/j.imu.2023.101314"},{"key":"3370_CR41","doi-asserted-by":"publisher","unstructured":"Knoedler L, Alfertshofer M, Knoedler S, Hoch CC, Funk PF, Cotofana S, Lamby P. Pure wisdom or Potemkin villages? a comparison of ChatGPT 3.5 and ChatGPT 4 on USMLE step 3 style questions: quantitative analysis. JMIR Med Edu. 2024:10(1):e51148. https:\/\/doi.org\/10.2196\/51148","DOI":"10.2196\/51148"},{"issue":"1","key":"3370_CR42","doi-asserted-by":"publisher","first-page":"20512","DOI":"10.1038\/s41598-023-46995-z","volume":"13","author":"M Roso\u0142","year":"2023","unstructured":"Roso\u0142 M, G\u0105sior JS, \u0141aba J, Korzeniewski K, M\u0142y\u0144czak M. Evaluation of the performance of GPT-3.5 and GPT-4 on the Polish medical final examination. Sci Rep. 2023;13(1):20512. https:\/\/doi.org\/10.1038\/s41598-023-46995-z.","journal-title":"Sci Rep"},{"issue":"1","key":"3370_CR43","doi-asserted-by":"publisher","first-page":"e48002","DOI":"10.2196\/48002","volume":"9","author":"S Takagi","year":"2023","unstructured":"Takagi S, Watari T, Erabi A, Sakaguchi K. Performance of GPT-3.5 and GPT-4 on the Japanese medical licensing examination: comparison study. JMIR Med Educ. 2023;9(1):e48002. https:\/\/doi.org\/10.2196\/48002.","journal-title":"JMIR Med Educ"},{"key":"3370_CR44","doi-asserted-by":"publisher","first-page":"e52202","DOI":"10.2196\/52202","volume":"9","author":"T Watari","year":"2023","unstructured":"Watari T, Takagi S, Sakaguchi K, Nishizaki Y, Shimizu T, Yamamoto Y, Tokuda Y. Performance comparison of ChatGPT-4 and Japanese medical residents in the general medicine in-training examination: comparison study. JMIR Med Educ. 2023;9:e52202. https:\/\/doi.org\/10.2196\/52202.","journal-title":"JMIR Med Educ"},{"key":"3370_CR45","doi-asserted-by":"publisher","unstructured":"Du X, Zhou Z, Wang Y, Chuang YW, Li Y, Yang R, Zhang W, Wang X, Chen X, Guan H, Lian J, Hong P, Bates DW, Zhou L. Testing and evaluation of generative large language models in electronic health record applications: a systematic review. medRxiv [Preprint]. Int J Med Inform. 2026;205:106091. https:\/\/doi.org\/10.1016\/j.ijmedinf.2025.106091. PMID: 39228726; PMCID: PMC11370524.","DOI":"10.1016\/j.ijmedinf.2025.106091"},{"issue":"2","key":"3370_CR46","doi-asserted-by":"publisher","first-page":"ooae028","DOI":"10.1093\/jamiaopen\/ooae028","volume":"7","author":"SL Baxter","year":"2024","unstructured":"Baxter SL, Longhurst CA, Millen M, Sitapati AM, Tai-Seale M. Generative artificial intelligence responses to patient messages in the electronic health record: early lessons learned. JAMIA Open. 2024;7(2):ooae028. https:\/\/doi.org\/10.1093\/jamiaopen\/ooae028. PMID: 38601475; PMCID: PMC11006101.","journal-title":"JAMIA Open"},{"issue":"12","key":"3370_CR47","doi-asserted-by":"publisher","first-page":"e0000416","DOI":"10.1371\/journal.pdig.0000416","volume":"2","author":"D Jang","year":"2023","unstructured":"Jang D, Yun TR, Lee CY, Kwon YK, Kim CE. GPT-4 can pass the Korean National licensing examination for Korean medicine Doctors. PLOS Digit Health. 2023;2(12):e0000416. https:\/\/doi.org\/10.1371\/journal.pdig.0000416.","journal-title":"PLOS Digit Health"}],"container-title":["BMC Medical Informatics and Decision Making"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/article\/10.1186\/s12911-026-03370-y","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s12911-026-03370-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s12911-026-03370-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T11:23:18Z","timestamp":1772623398000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1186\/s12911-026-03370-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,3]]},"references-count":47,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2026,12]]}},"alternative-id":["3370"],"URL":"https:\/\/doi.org\/10.1186\/s12911-026-03370-y","relation":{},"ISSN":["1472-6947"],"issn-type":[{"value":"1472-6947","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,3]]},"assertion":[{"value":"23 October 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 January 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 February 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The ethics approval of this study was waived by the Research Ethics Committee of the Graduate School of Medicine, The University of Tokyo, in accordance with the Ethical Guidelines for Medical and Biological Research Involving Human Subjects (MEXT\/MHLW\/METI, 2021). The study did not involve human participants, identifiable personal data, or human tissue. The study adhered to the principles of the Declaration of Helsinki. Additionally, informed consent was obtained from both the licensed physician and medical student prior to their participation in the evaluation process.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval and consent to participate"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}},{"value":"The authors declare no competing interests.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"65"}}