{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T23:58:51Z","timestamp":1774396731567,"version":"3.50.1"},"reference-count":71,"publisher":"MDPI AG","issue":"2","license":[{"start":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T00:00:00Z","timestamp":1770249600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"European Union\u2014Next Generation EU","award":["AGE-IT-PE00000015 CUP: H43C22000840006"],"award-info":[{"award-number":["AGE-IT-PE00000015 CUP: H43C22000840006"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["MAKE"],"abstract":"<jats:p>The increasing integration of artificial intelligence (AI) in decision-making processes has amplified discussions surrounding algorithmic authority\u2014the perceived epistemic legitimacy of AI systems over human judgment. This study investigates how individuals attribute epistemic authority to AI, focusing on psychological, contextual, and sociotechnical factors. Existing research highlights the importance of trust in automation, perceived performance, and moral frameworks in shaping such attributions. Unlike prior conceptual or philosophical accounts of algorithmic authority, our study adopts a relational and empirically grounded perspective by operationalizing algority through psychometric measures and contextual assessments. To address knowledge gaps in the micro-level dynamics of this phenomenon, we conducted an empirical study using psychometric tools and scenario-based assessments. Here, we report key findings from a survey of 610 participants, revealing significant correlations between trust in automation (TiA), perceptions of automated performance (PAS), and the propensity to defer to AI, particularly in high-stakes scenarios like criminal justice and job-matching. Trust in automation emerged as a primary factor, while moral attitudes moderated deference in ethically sensitive contexts. Our findings highlight the practical relevance of transparency and explainability for supporting critical engagement with AI outputs and for informing the design of contextually appropriate decision support. This study contributes to understanding algorithmic authority as a multidimensional construct, offering empirically grounded insights for designing AI systems that are trustworthy and context-sensitive.<\/jats:p>","DOI":"10.3390\/make8020036","type":"journal-article","created":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T10:35:37Z","timestamp":1770287737000},"page":"36","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Perceiving AI as an Epistemic Authority or Algority: A User Study on the Human Attribution of Authority to AI"],"prefix":"10.3390","volume":"8","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0522-2804","authenticated-orcid":false,"given":"Frida","family":"Milella","sequence":"first","affiliation":[{"name":"Department of Informatics, Systems and Communication (DISCo), University of Milano-Bicocca, Viale Sarca 336, 20126 Milan, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4065-3415","authenticated-orcid":false,"given":"Federico","family":"Cabitza","sequence":"additional","affiliation":[{"name":"Department of Informatics, Systems and Communication (DISCo), University of Milano-Bicocca, Viale Sarca 336, 20126 Milan, Italy"},{"name":"Digital Health and Wellbeing Center, Fondazione Bruno Kessler (FBK), Via Sommarive 18, 38123 Trento, Italy"}]}],"member":"1968","published-online":{"date-parts":[[2026,2,5]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"241","DOI":"10.1111\/puar.13483","article-title":"Explaining why the computer says no: Algorithmic transparency affects the perceived trustworthiness of automated decision-making","volume":"83","author":"Grimmelikhuijsen","year":"2023","journal-title":"Public Adm. Rev."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"140","DOI":"10.1093\/idpl\/ipy005","article-title":"Why the \u2018Computer says no\u2019: Illustrating big data\u2019s discrimination risk through complex systems science","volume":"8","author":"Rhoen","year":"2018","journal-title":"Int. Data Priv. Law"},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"870","DOI":"10.12968\/bjon.2021.30.14.870","article-title":"The computer says no: AI, health law, ethics and patient safety","volume":"30","author":"Tingle","year":"2021","journal-title":"Br. J. Nurs."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"231","DOI":"10.1080\/00076791.2016.1177024","article-title":"\u2018The computer says no\u2019: The demise of the traditional bank manager and the depersonalisation of British banking, 1960\u20132010","volume":"59","author":"Vik","year":"2017","journal-title":"Bus. Hist."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Wihlborg, E., Larsson, H., and Hedstr\u00f6m, K. (2016, January 5\u20138). \u201cThe Computer Says No!\u201d\u2013A Case Study on Automated Decision-Making in Public Authorities. Proceedings of the 2016 49th Hawaii International Conference on System Sciences (HICSS), Koloa, HI, USA.","DOI":"10.1109\/HICSS.2016.364"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"224","DOI":"10.1108\/JD-06-2016-0081","article-title":"The search-ification of everyday life and the mundane-ification of search","volume":"73","author":"Sundin","year":"2017","journal-title":"J. Doc."},{"key":"ref_7","unstructured":"Shirky, C. (2026, January 24). A Speculative Post on the Idea of Algorithmic Authority. Available online: https:\/\/www.bibsonomy.org\/url\/a4f71c8404afbb43b64a2c03196fe5e5."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"35","DOI":"10.1007\/s11412-015-9210-6","article-title":"The new information literate: Open collaboration and information production in schools","volume":"10","author":"Forte","year":"2015","journal-title":"Int. J. Comput.-Support. Collab. Learn."},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Lustig, C., and Nardi, B. (2015, January 5\u20138). Algorithmic authority: The case of Bitcoin. Proceedings of the 2015 48th Hawaii International Conference on System Sciences, Kauai, HI, USA.","DOI":"10.1109\/HICSS.2015.95"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Lustig, C., Pine, K., Nardi, B., Irani, L., Lee, M.K., Nafus, D., and Sandvig, C. (2016, January 7\u201312). Algorithmic authority: The ethics, politics, and economics of algorithms that interpret, decide, and manage. Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems, San Jose, CA, USA.","DOI":"10.1145\/2851581.2886426"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"245","DOI":"10.1007\/s13347-015-0211-1","article-title":"The threat of algocracy: Reality, resistance and accommodation","volume":"29","author":"Danaher","year":"2016","journal-title":"Philos. Technol."},{"key":"ref_12","first-page":"726","article-title":"Epistemic beliefs and internet reliance\u2014Is algorithmic authority part of the picture?","volume":"122","author":"Sormunen","year":"2021","journal-title":"Inf. Learn. Sci."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"102696","DOI":"10.1016\/j.ijhcs.2021.102696","article-title":"The need to move away from agential-AI: Empirical investigations, useful concepts and open issues","volume":"155","author":"Cabitza","year":"2021","journal-title":"Int. J. Hum.-Comput. Stud."},{"key":"ref_14","unstructured":"Schwarz, O. (2021). Sociological Theory for Digital Society: The Codes That Bind Us Together, John Wiley & Sons."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"213","DOI":"10.1146\/annurev-soc-090820-020800","article-title":"The Society of Algorithms","volume":"47","author":"Burrell","year":"2021","journal-title":"Annu. Rev. Sociol."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"985","DOI":"10.1177\/1461444809336551","article-title":"Power Through the Algorithm? Participatory Web Cultures and the Technological Unconscious","volume":"11","author":"Beer","year":"2009","journal-title":"New Media Soc."},{"key":"ref_17","unstructured":"Noble, S.U. (2018). Algorithms of Oppression, New York University Press."},{"key":"ref_18","unstructured":"Cheney-Lippold, J. (2017). We Are Data, New York University Press."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"611","DOI":"10.1007\/s00146-019-00931-w","article-title":"In AI we trust? Perceptions about automated decision-making by artificial intelligence","volume":"35","author":"Araujo","year":"2020","journal-title":"AI Soc."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"627","DOI":"10.5465\/annals.2018.0057","article-title":"Human trust in artificial intelligence: Review of empirical research","volume":"14","author":"Glikson","year":"2020","journal-title":"Acad. Manag. Ann."},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Kizilcec, R.F. (2016, January 7\u201312). How much information? Effects of transparency on trust in an algorithmic interface. Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, San Jose, CA, USA.","DOI":"10.1145\/2858036.2858402"},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"2053951718756684","DOI":"10.1177\/2053951718756684","article-title":"Understanding perception of algorithmic decisions: Fairness, trust, and emotion in response to algorithmic management","volume":"5","author":"Lee","year":"2018","journal-title":"Big Data Soc."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"352","DOI":"10.1080\/14639220110110306","article-title":"Automated diagnostic aids: The effects of aid reliability on users\u2019 trust and reliance","volume":"2","author":"Wiegmann","year":"2001","journal-title":"Theor. Issues Ergon. Sci."},{"key":"ref_24","unstructured":"Zeigler-Hill, V., and Shackelford, T.K. (2016). Moral Foundations Questionnaire. Encyclopedia of Personality and Individual Differences, Springer International Publishing."},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Kohn, S.C., De Visser, E.J., Wiese, E., Lee, Y.C., and Shaw, T.H. (2021). Measurement of trust in automation: A narrative review and reference guide. Front. Psychol., 12.","DOI":"10.3389\/fpsyg.2021.604977"},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"740","DOI":"10.1177\/0018720815581247","article-title":"Measuring individual differences in the perfect automation schema","volume":"57","author":"Merritt","year":"2015","journal-title":"Hum. Factors"},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"12041","DOI":"10.1007\/s11229-020-02849-z","article-title":"A deference model of epistemic authority","volume":"198","author":"Bokros","year":"2021","journal-title":"Synthese"},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"716","DOI":"10.1080\/02691728.2025.2449602","article-title":"Artificial epistemic authorities","volume":"39","author":"Hauswald","year":"2025","journal-title":"Soc. Epistemol."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Lange, B. (2024). Epistemic Deference to AI. Proceedings of the International Conference on Bridging the Gap Between AI and Reality, Springer Nature.","DOI":"10.1007\/978-3-032-01377-4_9"},{"key":"ref_30","unstructured":"Yang, S., and Ma, R. (2025). Classifying Epistemic Relationships in Human-AI Interaction: An Exploratory Approach. arXiv."},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"106879","DOI":"10.1016\/j.chb.2021.106879","article-title":"To err is human, not algorithmic\u2014Robust reactions to erring algorithms","volume":"124","author":"Renier","year":"2021","journal-title":"Comput. Hum. Behav."},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"1291","DOI":"10.3390\/nursrep13030109","article-title":"The social and professional image of the nurse: Results of an online snowball sampling survey among the general population in the post-pandemic period","volume":"13","author":"Rubbi","year":"2023","journal-title":"Nurs. Rep."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"1918","DOI":"10.1093\/aje\/kwab098","article-title":"Snowball Sampling Study Design for Serosurveys Early in Disease Outbreaks","volume":"190","author":"Qiu","year":"2021","journal-title":"Am. J. Epidemiol."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"315","DOI":"10.1146\/annurev-orgpsych-120920-052946","article-title":"Innovations in Sampling: Improving the Appropriateness and Quality of Samples in Organizational Research","volume":"10","author":"Zickar","year":"2022","journal-title":"Annu. Rev. Organ. Psychol. Organ. Behav."},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"1029","DOI":"10.1037\/a0015141","article-title":"Liberals and conservatives rely on different sets of moral foundations","volume":"96","author":"Graham","year":"2009","journal-title":"J. Personal. Soc. Psychol."},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"1303","DOI":"10.1111\/bjso.12452","article-title":"Reanalysing the factor structure of the moral foundations questionnaire","volume":"60","author":"Harper","year":"2021","journal-title":"Br. J. Soc. Psychol."},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"K\u00f6rber, M. (2019). Theoretical considerations and development of a questionnaire to measure trust in automation. Proceedings of the 20th Congress of the International Ergonomics Association (IEA 2018) Volume VI: Transport Ergonomics and Human Factors (TEHF); Aerospace Human Factors and Ergonomics 20, Springer.","DOI":"10.1007\/978-3-319-96074-6_2"},{"key":"ref_38","first-page":"28","article-title":"Validity and reliability of the research instrument; how to test the validation of a questionnaire\/survey in a research","volume":"5","author":"Taherdoost","year":"2016","journal-title":"Int. J. Acad. Res. Manag. (IJARM)"},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"679","DOI":"10.1016\/S2212-5671(15)00123-9","article-title":"How reliable are measurement scales? External factors with indirect influence on reliability estimators","volume":"20","author":"Ursachi","year":"2015","journal-title":"Procedia Econ. Financ."},{"key":"ref_40","doi-asserted-by":"crossref","first-page":"115","DOI":"10.1007\/s40685-018-0072-4","article-title":"Partial least squares structural equation modeling-based discrete choice modeling: An illustration in modeling retailer choice","volume":"12","author":"Hair","year":"2019","journal-title":"Bus. Res."},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Okoye, K., and Hosseini, S. (2024). Correlation tests in R: Pearson cor, kendall\u2019s tau, and spearman\u2019s rho. R Programming: Statistical Data Analysis in Research, Springer.","DOI":"10.1007\/978-981-97-3385-9_12"},{"key":"ref_42","doi-asserted-by":"crossref","unstructured":"McKight, P.E., and Najab, J. (2010). Kruskal-wallis test. The Corsini Encyclopedia of Psychology, John Wiley & Sons, Inc.","DOI":"10.1002\/9780470479216.corpsy0491"},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Fang, Z., Du, R., and Cui, X. (2012). Uniform approximation is more appropriate for wilcoxon rank-sum test in gene set analysis. PLoS ONE, 7.","DOI":"10.1371\/journal.pone.0031505"},{"key":"ref_44","first-page":"19","article-title":"The need to report effect size estimates revisited. An overview of some recommended measures of effect size","volume":"1","author":"Tomczak","year":"2014","journal-title":"Trends Sport Sci."},{"key":"ref_45","doi-asserted-by":"crossref","first-page":"612","DOI":"10.1080\/02691728.2025.2472780","article-title":"Expert Authority and Its Assessment","volume":"39","author":"Keren","year":"2025","journal-title":"Soc. Epistemol."},{"key":"ref_46","doi-asserted-by":"crossref","first-page":"e70005","DOI":"10.1111\/phc3.70005","article-title":"Experts\u2014Part II: The Sources of Epistemic Authority","volume":"19","author":"Croce","year":"2024","journal-title":"Philos. Compass"},{"key":"ref_47","doi-asserted-by":"crossref","first-page":"225","DOI":"10.1093\/ijpor\/edr032","article-title":"The role of media and deference to scientific authority in cultivating trust in sources of information about emerging technologies","volume":"24","author":"Anderson","year":"2012","journal-title":"Int. J. Public Opin. Res."},{"key":"ref_48","first-page":"330","article-title":"Developing an assessment of epistemic trust: A research protocol","volume":"21","author":"Talia","year":"2018","journal-title":"Res. Psychother. Psychopathol. Process Outcome"},{"key":"ref_49","doi-asserted-by":"crossref","first-page":"800","DOI":"10.1177\/0963662520962741","article-title":"Deference and decision-making in science and society: How deference to scientific authority goes beyond confidence in science and scientists to become authoritarianism","volume":"29","author":"Howell","year":"2020","journal-title":"Public Underst. Sci."},{"key":"ref_50","doi-asserted-by":"crossref","unstructured":"Kapania, S., Siy, O., Clapper, G., Sp, A.M., and Sambasivan, N. (May, January 29). \u201cBecause AI is 100% right and safe\u201d: User attitudes and sources of AI authority in India. Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems, New Orleans, LA, USA.","DOI":"10.1145\/3491102.3517533"},{"key":"ref_51","doi-asserted-by":"crossref","first-page":"30","DOI":"10.1007\/s11023-024-09681-1","article-title":"Experts or Authorities? The Strange Case of the Presumed Epistemic Superiority of Artificial Intelligence Systems","volume":"34","author":"Ferrario","year":"2024","journal-title":"Minds Mach."},{"key":"ref_52","doi-asserted-by":"crossref","first-page":"205","DOI":"10.1136\/medethics-2019-105586","article-title":"On the ethics of algorithmic decision-making in healthcare","volume":"46","author":"Grote","year":"2020","journal-title":"J. Med. Ethics"},{"key":"ref_53","doi-asserted-by":"crossref","first-page":"230","DOI":"10.1518\/001872097778543886","article-title":"Humans and automation: Use, misuse, disuse, abuse","volume":"39","author":"Parasuraman","year":"1997","journal-title":"Hum. Factors"},{"key":"ref_54","unstructured":"Moes, M., Knox, K., Pierce, L., and Beck, H. (1999, January 18\u201321). Should I decide or let the machine decide for me. Proceedings of the Poster Presented at the Meeting of the Southeastern Psychological Association, Savannah, GA, USA."},{"key":"ref_55","doi-asserted-by":"crossref","first-page":"61","DOI":"10.1109\/TTS.2024.3370095","article-title":"When AI fails, who do we blame? Attributing responsibility in human\u2013AI interactions","volume":"5","author":"Schoenherr","year":"2024","journal-title":"IEEE Trans. Technol. Soc."},{"key":"ref_56","doi-asserted-by":"crossref","first-page":"22","DOI":"10.1145\/3241365","article-title":"Overtrust in the robotic age","volume":"61","author":"Wagner","year":"2018","journal-title":"Commun. ACM"},{"key":"ref_57","first-page":"47","article-title":"Building trust in artificial intelligence, machine learning, and robotics","volume":"31","author":"Siau","year":"2018","journal-title":"Cut. Bus. Technol. J."},{"key":"ref_58","doi-asserted-by":"crossref","first-page":"403","DOI":"10.1002\/bdm.2118","article-title":"Making sense of recommendations","volume":"32","author":"Yeomans","year":"2019","journal-title":"J. Behav. Decis. Mak."},{"key":"ref_59","doi-asserted-by":"crossref","first-page":"90","DOI":"10.1016\/j.obhdp.2018.12.005","article-title":"Algorithm appreciation: People prefer algorithmic to human judgment","volume":"151","author":"Logg","year":"2019","journal-title":"Organ. Behav. Hum. Decis. Processes"},{"key":"ref_60","doi-asserted-by":"crossref","first-page":"809","DOI":"10.1177\/0022243719851788","article-title":"Task-dependent algorithm aversion","volume":"56","author":"Castelo","year":"2019","journal-title":"J. Mark. Res."},{"key":"ref_61","doi-asserted-by":"crossref","unstructured":"Brailsford, J., Vetere, F., and Velloso, E. (2024, January 11\u201316). Exploring the Association between Moral Foundations and Judgements of AI Behaviour. Proceedings of the CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA.","DOI":"10.1145\/3613904.3642712"},{"key":"ref_62","doi-asserted-by":"crossref","first-page":"100154","DOI":"10.1016\/j.chbr.2021.100154","article-title":"Perceptions of violations by artificial and human actors across moral foundations","volume":"5","author":"Maninger","year":"2022","journal-title":"Comput. Hum. Behav. Rep."},{"key":"ref_63","doi-asserted-by":"crossref","first-page":"407","DOI":"10.1177\/0018720814547570","article-title":"Trust in Automation: Integrating Empirical Evidence on Factors That Influence Trust: Integrating Empirical Evidence on Factors That Influence Trust","volume":"57","author":"Hoff","year":"2014","journal-title":"Hum. Factors J. Hum. Factors Ergon. Soc."},{"key":"ref_64","first-page":"100040","article-title":"Task-specific algorithm advice acceptance: A review and directions for future research","volume":"7","author":"Kaufmann","year":"2023","journal-title":"Data Inf. Manag."},{"key":"ref_65","unstructured":"Milella, F., Natali, C., Scantamburlo, T., Campagner, A., and Cabitza, F. (September, January 28). The impact of gender and personality in human-AI teaming: The case of collaborative question answering. Proceedings of the IFIP Conference on Human-Computer Interaction, York, UK."},{"key":"ref_66","doi-asserted-by":"crossref","unstructured":"Pasquale, F. (2015). The Black Box Society: The Secret Algorithms That Control Money and Information, Harvard University Press.","DOI":"10.4159\/harvard.9780674736061"},{"key":"ref_67","doi-asserted-by":"crossref","first-page":"277","DOI":"10.1016\/j.chb.2019.04.019","article-title":"Role of fairness, accountability, and transparency in algorithmic affordance","volume":"98","author":"Shin","year":"2019","journal-title":"Comput. Hum. Behav."},{"key":"ref_68","doi-asserted-by":"crossref","unstructured":"Kim, T., Molina, M.D., Rheu, M., Zhan, E.S., and Peng, W. (2023, January 23\u201328). One AI does not fit all: A cluster analysis of the laypeople\u2019s perception of AI roles. Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, Hamburg, Germany.","DOI":"10.1145\/3544548.3581340"},{"key":"ref_69","doi-asserted-by":"crossref","first-page":"105945","DOI":"10.1016\/j.actpsy.2025.105945","article-title":"Trust in AI vs. human doctors: The roles of subjective understanding, perceived epistemic authority and social proof","volume":"261","author":"Ding","year":"2025","journal-title":"Acta Psychol."},{"key":"ref_70","doi-asserted-by":"crossref","first-page":"32","DOI":"10.1007\/s11948-023-00451-3","article-title":"AI as an epistemic technology","volume":"29","author":"Alvarado","year":"2023","journal-title":"Sci. Eng. Ethics"},{"key":"ref_71","doi-asserted-by":"crossref","unstructured":"Facchini, A., Fregosi, C., Natali, C., Termine, A., and Wilson, B. (2024, January 24\u201327). Algorithmic Authority & AI Influence in Decision Settings: Theories and Implications for Design. Proceedings of the 12th International Conference on Human-Agent Interaction, Swansea, UK.","DOI":"10.1145\/3687272.3691363"}],"container-title":["Machine Learning and Knowledge Extraction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2504-4990\/8\/2\/36\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T11:08:58Z","timestamp":1770289738000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2504-4990\/8\/2\/36"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,5]]},"references-count":71,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2026,2]]}},"alternative-id":["make8020036"],"URL":"https:\/\/doi.org\/10.3390\/make8020036","relation":{},"ISSN":["2504-4990"],"issn-type":[{"value":"2504-4990","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,5]]}}}