{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T22:21:28Z","timestamp":1776118888698,"version":"3.50.1"},"reference-count":36,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,11,6]],"date-time":"2025-11-06T00:00:00Z","timestamp":1762387200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,11,6]],"date-time":"2025-11-06T00:00:00Z","timestamp":1762387200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["npj Digit. Med."],"DOI":"10.1038\/s41746-025-02135-7","type":"journal-article","created":{"date-parts":[[2025,11,6]],"date-time":"2025-11-06T14:47:25Z","timestamp":1762440445000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["The perils of politeness: how large language models may amplify medical misinformation"],"prefix":"10.1038","volume":"8","author":[{"given":"Kyra L.","family":"Rosen","sequence":"first","affiliation":[]},{"given":"Margaret","family":"Sui","sequence":"additional","affiliation":[]},{"given":"Kimia","family":"Heydari","sequence":"additional","affiliation":[]},{"given":"Elizabeth J.","family":"Enichen","sequence":"additional","affiliation":[]},{"given":"Joseph C.","family":"Kvedar","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,6]]},"reference":[{"key":"2135_CR1","unstructured":"Henry, T. A. 2 in 3 physicians are using health AI\u2014up 78% from 2023. 2025. AMA News Wire. February 26, 2025. https:\/\/www.ama-assn.org\/practice-management\/digital-health\/2-3-physicians-are-using-health-ai-78-2023."},{"key":"2135_CR2","unstructured":"Presiado, M., Montero, A., Lopes, L. & Hamel, L. KFF Health Misinformation Tracking Poll: Artificial Intelligence and Health Information https:\/\/www.kff.org\/public-opinion\/kff-health-misinformation-tracking-poll-artificial-intelligence-and-health-information\/ (2024)."},{"key":"2135_CR3","first-page":"e76941","volume":"27","author":"L Qiu","year":"2025","unstructured":"Qiu, L. et al. Physician use of large language models: a quantitative study based on large-scale query-level data. J. Med. Int. Res. 27, e76941 (2025).","journal-title":"J. Med. Int. Res."},{"key":"2135_CR4","first-page":"e68560","volume":"27","author":"HS Yun","year":"2025","unstructured":"Yun, H. S. & Bickmore, T. Online health information\u2013seeking in the era of large language models: cross-sectional web-based survey study. J. Med. Int. Res. 27, e68560 (2025).","journal-title":"J. Med. Int. Res."},{"key":"2135_CR5","doi-asserted-by":"publisher","first-page":"625","DOI":"10.1038\/s41586-024-07421-0","volume":"630","author":"S Farquhar","year":"2024","unstructured":"Farquhar, S., Kossen, J., Kuhn, L. & Gal, Y. Detecting hallucinations in large language models using semantic entropy. Nature 630, 625\u2013630 (2024).","journal-title":"Nature"},{"key":"2135_CR6","first-page":"e46924","volume":"25","author":"M M\u00e1jovsk\u00fd","year":"2023","unstructured":"M\u00e1jovsk\u00fd, M., \u010cern\u00fd, M., Kasal, M., Komarc, M. & Netuka, D. Artificial intelligence can generate fraudulent but authentic-looking scientific medical articles: Pandora\u2019s box has been opened. J. Med. Int. Res. 25, e46924 (2023).","journal-title":"J. Med. Int. Res."},{"key":"2135_CR7","doi-asserted-by":"publisher","first-page":"92","DOI":"10.1001\/jamainternmed.2023.5947","volume":"184","author":"BD Menz","year":"2024","unstructured":"Menz, B. D., Modi, N. D., Sorich, M. J. & Hopkins, A. M. Health disinformation use case highlighting the urgent need for artificial intelligence vigilance: weapons of mass disinformation. JAMA Intern. Med. 184, 92\u201396 (2024).","journal-title":"JAMA Intern. Med."},{"key":"2135_CR8","doi-asserted-by":"publisher","first-page":"e078538","DOI":"10.1136\/bmj-2023-078538","volume":"384","author":"BD Menz","year":"2024","unstructured":"Menz, B. D. et al. Current safeguards, risk mitigation, and transparency measures of large language models against the generation of health disinformation: repeated cross sectional analysis. Bmj 384, e078538 (2024).","journal-title":"Bmj"},{"key":"2135_CR9","unstructured":"Sharma M. et al. Towards Understanding Sycophancy in Language Models. arXiv preprint arXiv:231013548. 2025."},{"key":"2135_CR10","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1038\/d41586-025-03390-0","volume":"647","author":"M Naddaf","year":"2025","unstructured":"Naddaf, M. AI chatbots are sycophants \u2014 researchers say it\u2019s harming science. Nature 647, 13\u201314 (2025).","journal-title":"Nature"},{"key":"2135_CR11","doi-asserted-by":"crossref","unstructured":"Malmqvist L. Sycophancy in Large Language Models: Causes and Mitigations. arXiv preprint arXiv:241115287. 2024.","DOI":"10.1007\/978-3-031-92611-2_5"},{"key":"2135_CR12","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-025-02008-z","volume":"8","author":"S Chen","year":"2025","unstructured":"Chen, S. et al. When helpfulness backfires: LLMs and the risk of false medical information due to sycophantic behavior. npj Digital Med. 8, 605 (2025).","journal-title":"npj Digital Med."},{"key":"2135_CR13","doi-asserted-by":"publisher","first-page":"39","DOI":"10.3122\/jabfm.19.1.39","volume":"19","author":"KL Schwartz","year":"2006","unstructured":"Schwartz, K. L. et al. Family medicine patients\u2019 use of the Internet for health information: a MetroNet study. J. Am. Board Fam. Med. 19, 39\u201345 (2006).","journal-title":"J. Am. Board Fam. Med."},{"key":"2135_CR14","first-page":"e24","volume":"5","author":"KS Shuyler","year":"2003","unstructured":"Shuyler, K. S. & Knight, K. M. What are patients seeking when they turn to the Internet? Qualitative content analysis of questions asked by visitors to an orthopaedics Web site. J. Med. Int. Res. 5, e24 (2003).","journal-title":"J. Med. Int. Res."},{"key":"2135_CR15","doi-asserted-by":"publisher","first-page":"609","DOI":"10.1007\/s11606-016-3612-7","volume":"31","author":"AS Kesselheim","year":"2016","unstructured":"Kesselheim, A. S. et al. Variations in Patients\u2019 Perceptions and Use of Generic Drugs: Results of a National Survey. J. Gen. Intern. Med. 31, 609\u2013614 (2016).","journal-title":"J. Gen. Intern. Med."},{"key":"2135_CR16","doi-asserted-by":"publisher","first-page":"e008915","DOI":"10.1136\/bmjopen-2015-008915","volume":"5","author":"S Colgan","year":"2015","unstructured":"Colgan, S. et al. Perceptions of generic medication in the general population, doctors and pharmacists: a systematic review. BMJ Open 5, e008915 (2015).","journal-title":"BMJ Open"},{"key":"2135_CR17","doi-asserted-by":"publisher","first-page":"e2461940","DOI":"10.1001\/jamanetworkopen.2024.61940","volume":"8","author":"B Nickel","year":"2025","unstructured":"Nickel, B. et al. Social Media Posts About Medical Tests With Potential for Overdiagnosis. JAMA Network Open 8, e2461940\u2013e2461940 (2025).","journal-title":"JAMA Network Open"},{"key":"2135_CR18","unstructured":"Jamieson, K. H., Winneg, K., Jr. S. P., Gibson, L. A., Jamieson, P. E. Annenberg Science and Public Health Knowledge Monitor. 2024. https:\/\/www.annenbergpublicpolicycenter.org\/wp-content\/uploads\/asaph-report-summer-2024-v3-1.pdf."},{"key":"2135_CR19","doi-asserted-by":"publisher","DOI":"10.1126\/sciadv.adh1850","volume":"9","author":"G Spitale","year":"2023","unstructured":"Spitale, G., Biller-Andorno, N. & Germani, F. AI model GPT-3 (dis)informs us better than humans. Sci. Adv. 9, eadh1850 (2023).","journal-title":"Sci. Adv."},{"key":"2135_CR20","doi-asserted-by":"publisher","first-page":"1645","DOI":"10.1038\/s41562-025-02194-6","volume":"9","author":"F Salvi","year":"2025","unstructured":"Salvi, F., Horta Ribeiro, M., Gallotti, R. & West, R. On the conversational persuasiveness of GPT-4. Nat Human Behav 9, 1645\u20131653 (2025).","journal-title":"Nat Human Behav"},{"key":"2135_CR21","doi-asserted-by":"publisher","first-page":"e0263381","DOI":"10.1371\/journal.pone.0263381","volume":"17","author":"K Singh","year":"2022","unstructured":"Singh, K. et al. Misinformation, believability, and vaccine acceptance over 40 countries: Takeaways from the initial phase of the COVID-19 infodemic. PLoS One 17, e0263381 (2022).","journal-title":"PLoS One"},{"key":"2135_CR22","doi-asserted-by":"publisher","first-page":"544","DOI":"10.2471\/BLT.21.287654","volume":"100","author":"IJ Borges do Nascimento","year":"2022","unstructured":"Borges do Nascimento, I. J. et al. Infodemics and health misinformation: a systematic review of reviews. Bull World Health Organ 100, 544\u2013561 (2022).","journal-title":"Bull World Health Organ"},{"key":"2135_CR23","doi-asserted-by":"publisher","first-page":"1459","DOI":"10.1001\/jamaoncol.2023.2954","volume":"9","author":"S Chen","year":"2023","unstructured":"Chen, S. et al. Use of artificial intelligence chatbots for cancer treatment information. JAMA Oncology 9, 1459\u20131462 (2023).","journal-title":"JAMA Oncology"},{"key":"2135_CR24","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-024-01283-6","volume":"7","author":"S Schmidgall","year":"2024","unstructured":"Schmidgall, S. et al. Evaluation and mitigation of cognitive biases in medical language models. npj Digital Medicine 7, 295 (2024).","journal-title":"npj Digital Medicine"},{"key":"2135_CR25","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2024.109553","volume":"139","author":"Z Wang","year":"2025","unstructured":"Wang, Z. et al. Word-Sequence Entropy: Towards uncertainty estimation in free-form medical question answering applications and beyond. Eng. Appl. Artif. Intell. 139, 109553 (2025).","journal-title":"Eng. Appl. Artif. Intell."},{"key":"2135_CR26","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1093\/jamia\/ocae254","volume":"32","author":"T Savage","year":"2025","unstructured":"Savage, T. et al. Large language model uncertainty proxies: discrimination and calibration for medical diagnosis and treatment. J. Am. Med. Inform. Assoc. 32, 139\u2013149 (2025).","journal-title":"J. Am. Med. Inform. Assoc."},{"key":"2135_CR27","unstructured":"Lee, H. et al. RLAIF vs. RLHF: Scaling Reinforcement Learning from Human Feedback with AI Feedback. arXiv preprint arXiv:230900267. 2024."},{"key":"2135_CR28","unstructured":"Bai, Y. et al. Constitutional AI: Harmlessness from AI Feedback. arXiv preprint arXiv:221208073. 2022."},{"key":"2135_CR29","doi-asserted-by":"publisher","first-page":"605","DOI":"10.1093\/jamia\/ocaf008","volume":"32","author":"S Liu","year":"2025","unstructured":"Liu, S., McCoy, A. B. & Wright, A. Improving large language model applications in biomedicine with retrieval-augmented generation: a systematic review, meta-analysis, and clinical development guidelines. J. Am. Med. Inform. Assoc. 32, 605\u2013615 (2025).","journal-title":"J. Am. Med. Inform. Assoc."},{"key":"2135_CR30","doi-asserted-by":"publisher","first-page":"2396","DOI":"10.1038\/s41591-023-02412-6","volume":"29","author":"S Gilbert","year":"2023","unstructured":"Gilbert, S., Harvey, H., Melvin, T., Vollebregt, E. & Wicks, P. Large language model AI chatbots require approval as medical devices. Nat. Med. 29, 2396\u20132398 (2023).","journal-title":"Nat. Med."},{"key":"2135_CR31","doi-asserted-by":"publisher","first-page":"148","DOI":"10.1038\/s41746-025-01544-y","volume":"8","author":"GE Weissman","year":"2025","unstructured":"Weissman, G. E., Mankowitz, T. & Kanter, G. P. Unregulated large language models produce medical device-like output. NPJ Digit. Med. 8, 148 (2025).","journal-title":"NPJ Digit. Med."},{"key":"2135_CR32","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-025-01443-2","volume":"8","author":"DS Comeau","year":"2025","unstructured":"Comeau, D. S., Bitterman, D. S. & Celi, L. A. Preventing unrestricted and unmonitored AI experimentation in healthcare through transparency and accountability. npj Digital Med. 8, 42 (2025).","journal-title":"npj Digital Med."},{"key":"2135_CR33","first-page":"23","volume":"12","author":"JY Bo","year":"2024","unstructured":"Bo, J. Y., Kumar, H., Liut, M. & Anderson, A. Disclosures & disclaimers: investigating the impact of transparency disclosures and reliability disclaimers on learner-LLM interactions. Proc. AAAI Conf. Human Comput. Crowdsourcing 12, 23\u201332 (2024).","journal-title":"Proc. AAAI Conf. Human Comput. Crowdsourcing"},{"key":"2135_CR34","doi-asserted-by":"publisher","DOI":"10.3389\/fmed.2025.1527864","volume":"12","author":"S Aydin","year":"2025","unstructured":"Aydin, S., Karabacak, M., Vlachos, V. & Margetis, K. Navigating the potential and pitfalls of large language models in patient-centered medication guidance and self-decision support. Front Med. (Lausanne) 12, 1527864 (2025).","journal-title":"Front Med. (Lausanne)"},{"key":"2135_CR35","doi-asserted-by":"publisher","first-page":"26","DOI":"10.1038\/s43856-024-00717-2","volume":"5","author":"F Busch","year":"2025","unstructured":"Busch, F. et al. Current applications and challenges in large language models for patient care: a systematic review. Commun. Med. (Lond) 5, 26 (2025).","journal-title":"Commun. Med. (Lond)"},{"key":"2135_CR36","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-025-01790-0","volume":"8","author":"A Mahajan","year":"2025","unstructured":"Mahajan, A., Obermeyer, Z., Daneshjou, R., Lester, J. & Powell, D. Cognitive bias in clinical large language models. npj Digit. Med. 8, 428 (2025).","journal-title":"npj Digit. Med."}],"container-title":["npj Digital Medicine"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-02135-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-02135-7","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-02135-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,6]],"date-time":"2025-11-06T14:47:28Z","timestamp":1762440448000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-02135-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,6]]},"references-count":36,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["2135"],"URL":"https:\/\/doi.org\/10.1038\/s41746-025-02135-7","relation":{},"ISSN":["2398-6352"],"issn-type":[{"value":"2398-6352","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,6]]},"assertion":[{"value":"21 October 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 October 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 November 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Authors K.R., M.S., K.H., and E.E. declare no financial or non-financial competing interests. Author J.K. serves as Editor-in-Chief of this journal and had no role in the peer-review or decision to publish this manuscript. Author J.K. declares no financial competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"644"}}