{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T05:25:07Z","timestamp":1761110707463,"version":"build-2065373602"},"reference-count":73,"publisher":"MDPI AG","issue":"3","license":[{"start":{"date-parts":[[2025,9,5]],"date-time":"2025-09-05T00:00:00Z","timestamp":1757030400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["MAKE"],"abstract":"<jats:p>Deep neural networks (DNNs) are highly effective across many domains but are sensitive to noisy or corrupted training data. Existing noise mitigation strategies often rely on strong assumptions about noise distributions or require costly retraining, limiting their scalability. Inspired by machine unlearning, we propose a novel framework that integrates attribution-guided data partitioning, neuron pruning, and targeted fine-tuning to enhance robustness. Our method uses gradient-based attribution to probabilistically identify clean samples without assuming specific noise characteristics. It then applies sensitivity-based neuron pruning to remove components most susceptible to noise, followed by fine-tuning on the retained high-quality subset. This approach jointly addresses data and model-level noise, offering a practical alternative to full retraining or explicit noise modeling. We evaluate our method on CIFAR-10 image classification and keyword spotting tasks under varying levels of label corruption. On CIFAR-10, our framework improves accuracy by up to 10% (F-FT vs. retrain) and reduces retraining time by 47% (L-FT vs. retrain), highlighting both accuracy and efficiency gains. These results highlight its effectiveness and efficiency in noisy settings, making it a scalable solution for robust generalization.<\/jats:p>","DOI":"10.3390\/make7030095","type":"journal-article","created":{"date-parts":[[2025,9,5]],"date-time":"2025-09-05T14:53:01Z","timestamp":1757083981000},"page":"95","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Machine Unlearning for Robust DNNs: Attribution-Guided Partitioning and Neuron Pruning in Noisy Environments"],"prefix":"10.3390","volume":"7","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-3863-1478","authenticated-orcid":false,"given":"Deliang","family":"Jin","sequence":"first","affiliation":[{"name":"School of Computer Science and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6290-7041","authenticated-orcid":false,"given":"Gang","family":"Chen","sequence":"additional","affiliation":[{"name":"School of Computer Science and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China"},{"name":"School of Cyber Science and Engineering, Zhengzhou University, Zhengzhou 450001, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-5456-8782","authenticated-orcid":false,"given":"Shuo","family":"Feng","sequence":"additional","affiliation":[{"name":"School of Computer Science and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China"}]},{"given":"Yufeng","family":"Ling","sequence":"additional","affiliation":[{"name":"School of Computer Science and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China"}]},{"given":"Haoran","family":"Zhu","sequence":"additional","affiliation":[{"name":"School of Computer Science and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China"}]}],"member":"1968","published-online":{"date-parts":[[2025,9,5]]},"reference":[{"key":"ref_1","unstructured":"Krizhevsky, A., Sutskever, I., and Hinton, G.E. (2012, January 3\u20136). ImageNet Classification with Deep Convolutional Neural Networks. Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Lake Tahoe, NV, USA."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep Residual Learning for Image Recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_3","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., and Gelly, S. (2020). An image is worth 16x16 words: Transformers for image recognition at scale. arXiv."},{"key":"ref_4","unstructured":"Devlin, J., Chang, M.W., Lee, K., and Toutanova, K. (August, January 28). BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics (NAACL), Minneapolis, MN, USA."},{"key":"ref_5","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., and Polosukhin, I. (2017, January 4\u20139). Attention is All You Need. Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA."},{"key":"ref_6","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Graves, A., Mohamed, A.r., and Hinton, G. (2013, January 26\u201331). Speech recognition with deep recurrent neural networks. Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, BC, Canada.","DOI":"10.1109\/ICASSP.2013.6638947"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"82","DOI":"10.1109\/MSP.2012.2205597","article-title":"Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups","volume":"29","author":"Hinton","year":"2012","journal-title":"IEEE Signal Process. Mag."},{"key":"ref_9","unstructured":"Hannun, A., Case, C., Casper, J., Catanzaro, B., Diamos, G., Elsen, E., Prenger, R., Satheesh, S., Sengupta, S., and Coates, A. (2014). Deep speech: Scaling up end-to-end speech recognition. arXiv."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"115","DOI":"10.1038\/nature21056","article-title":"Dermatologist-level classification of skin cancer with deep neural networks","volume":"542","author":"Esteva","year":"2017","journal-title":"Nature"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"60","DOI":"10.1016\/j.media.2017.07.005","article-title":"A survey on deep learning in medical image analysis","volume":"42","author":"Litjens","year":"2017","journal-title":"Med. Image Anal."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"529","DOI":"10.1038\/nature14236","article-title":"Human-level control through deep reinforcement learning","volume":"518","author":"Mnih","year":"2015","journal-title":"Nature"},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"484","DOI":"10.1038\/nature16961","article-title":"Mastering the game of Go with deep neural networks and tree search","volume":"529","author":"Silver","year":"2016","journal-title":"Nature"},{"key":"ref_14","unstructured":"Schulman, J., Wolski, F., Dhariwal, P., Radford, A., and Klimov, O. (2017). Proximal policy optimization algorithms. arXiv."},{"key":"ref_15","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., and Bengio, Y. (2014, January 8\u201313). Generative Adversarial Nets. Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Montreal, QC, Canada."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Biggio, B., Corona, I., Maiorca, D., Nelson, B., \u0160rndi\u0107, N., Laskov, P., Giacinto, G., and Roli, F. (2013, January 23\u201327). Evasion attacks against machine learning at test time. Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases (ECML-PKDD), Prague, Czech Republic.","DOI":"10.1007\/978-3-642-40994-3_25"},{"key":"ref_17","first-page":"845","article-title":"Classification in the presence of label noise: A survey","volume":"25","author":"Verleysen","year":"2013","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"ref_18","unstructured":"Northcutt, C.G., Athalye, A., and Mueller, J. (2021). Pervasive label errors in test sets destabilize machine learning benchmarks. arXiv."},{"key":"ref_19","unstructured":"Song, H., Han, B., Liu, Y., and Sugiyama, M. (2019, January 10\u201315). SELFIE: Refurbishing Unclean Samples for Robust Deep Learning. Proceedings of the International Conference on Machine Learning (ICML), Long Beach, CA, USA."},{"key":"ref_20","unstructured":"Rolnick, D., Veit, A., Belongie, S., and Shavit, N. (2017). Deep learning is robust to massive label noise. arXiv."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1038\/sdata.2016.35","article-title":"MIMIC-III, a freely accessible critical care database","volume":"3","author":"Johnson","year":"2016","journal-title":"Sci. Data"},{"key":"ref_22","unstructured":"Arpit, D., Jastrz\u0119bski, S., Ballas, N., Krueger, D., Bengio, E., Kanwal, M.S., Maharaj, T., Fischer, A., Courville, A., and Bengio, Y. (2017, January 6\u201311). A closer look at memorization in deep networks. Proceedings of the International Conference on Machine Learning (ICML), Sydney, NSW, Australia."},{"key":"ref_23","unstructured":"Zhang, C., Bengio, S., Hardt, M., Recht, B., and Vinyals, O. (2016). Understanding deep learning requires rethinking generalization. arXiv."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"15849","DOI":"10.1073\/pnas.1903070116","article-title":"Reconciling modern machine-learning practice and the classical bias\u2013variance trade-off","volume":"116","author":"Belkin","year":"2019","journal-title":"Proc. Natl. Acad. Sci. USA"},{"key":"ref_25","unstructured":"Amodei, D., Olah, C., Steinhardt, J., Christiano, P., Schulman, J., and Man\u00e9, D. (2016). Concrete problems in AI safety. arXiv."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Lee, K.H., He, X., Zhang, L., and Yang, L. (2018, January 18\u201323). Cleannet: Transfer learning for scalable image classifier training with label noise. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00571"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Lai, K.H., Zha, D., Wang, G., Xu, J., Zhao, Y., Kumar, D., Chen, Y., Zumkhawaka, P., Wan, M., and Martinez, D. (2021, January 2\u20139). Tods: An automated time series outlier detection system. Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Online.","DOI":"10.1609\/aaai.v35i18.18012"},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"131","DOI":"10.1613\/jair.606","article-title":"Identifying mislabeled training data","volume":"11","author":"Brodley","year":"1999","journal-title":"J. Artif. Intell. Res."},{"key":"ref_29","unstructured":"Wang, Y., Ma, X., Chen, Z., Luo, Y., Yi, J., and Bailey, J. (November, January 27). Symmetric cross entropy for robust learning with noisy labels. Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea."},{"key":"ref_30","first-page":"8792","article-title":"Generalized cross entropy loss for training deep neural networks with noisy labels","volume":"31","author":"Zhang","year":"2018","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_31","unstructured":"Ma, X., Wang, Y., Houle, M.E., Zhou, S., Erfani, S., Xia, S., Wijewickrema, S., and Bailey, J. (2018, January 10\u201315). Dimensionality-driven learning with noisy labels. Proceedings of the International Conference on Machine Learning, Stockholm, Sweden."},{"key":"ref_32","first-page":"8536","article-title":"Co-teaching: Robust training of deep neural networks with extremely noisy labels","volume":"31","author":"Han","year":"2018","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_33","unstructured":"Jiang, L., Zhou, Z., Leung, T., Li, L.J., and Fei-Fei, L. (2018, January 10\u201315). Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels. Proceedings of the International Conference on Machine Learning (ICML), Stockholm, Sweden."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Patrini, G., Rozza, A., Krishna Menon, A., Nock, R., and Qu, L. (2017, January 21\u201326). Making deep neural networks robust to label noise: A loss correction approach. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.240"},{"key":"ref_35","unstructured":"Ren, M., Zeng, W., Yang, B., and Urtasun, R. (2018, January 10\u201315). Learning to reweight examples for robust deep learning. Proceedings of the International Conference on Machine Learning (ICML), Stockholm, Sweden."},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"447","DOI":"10.1109\/TPAMI.2015.2456899","article-title":"Classification with noisy labels by importance reweighting","volume":"38","author":"Liu","year":"2015","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_37","first-page":"6838","article-title":"Are anchor points really indispensable in label-noise learning?","volume":"32","author":"Xia","year":"2019","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Li, J., Wong, Y., Zhao, Q., and Kankanhalli, M.S. (2019, January 16\u201320). Learning to learn from noisy labeled data. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00519"},{"key":"ref_39","first-page":"1919","article-title":"Meta-weight-net: Learning an explicit mapping for sample weighting","volume":"32","author":"Shu","year":"2019","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Cao, Y., and Yang, J. (2015, January 17\u201321). Towards making systems forget with machine unlearning. Proceedings of the IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA.","DOI":"10.1109\/SP.2015.35"},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Bourtoule, L., Chandrasekaran, V., Choquette-Choo, C.A., Jia, H., Travers, A., Zhang, B., Lie, D., and Papernot, N. (2021, January 24\u201327). Machine unlearning. Proceedings of the IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA.","DOI":"10.1109\/SP40001.2021.00019"},{"key":"ref_42","doi-asserted-by":"crossref","unstructured":"Graves, L., Nagisetty, V., and Ganesh, V. (2021, January 2\u20139). Amnesiac machine learning. Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Online.","DOI":"10.1609\/aaai.v35i13.17371"},{"key":"ref_43","unstructured":"Thudi, A., Jia, H., Shumailov, I., and Papernot, N. (2022, January 10\u201312). On the necessity of auditable algorithmic definitions for machine unlearning. Proceedings of the USENIX Security 22, Boston, MA, USA."},{"key":"ref_44","doi-asserted-by":"crossref","unstructured":"Golatkar, A., Achille, A., and Soatto, S. (2020, January 13\u201319). Eternal sunshine of the spotless net: Selective forgetting in deep networks. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00932"},{"key":"ref_45","unstructured":"Sundararajan, M., Taly, A., and Yan, Q. (2017, January 6\u201311). Axiomatic attribution for deep networks. Proceedings of the International Conference on Machine Learning (ICML), Sydney, Australia."},{"key":"ref_46","unstructured":"Simonyan, K., Vedaldi, A., and Zisserman, A. (2013). Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv."},{"key":"ref_47","first-page":"1135","article-title":"Learning both weights and connections for efficient neural networks","volume":"28","author":"Han","year":"2015","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_48","unstructured":"Louizos, C., Welling, M., and Kingma, D.P. (2017). Learning sparse neural networks through L0 regularization. arXiv."},{"key":"ref_49","doi-asserted-by":"crossref","unstructured":"Howard, J., and Ruder, S. (2018). Universal language model fine-tuning for text classification. arXiv.","DOI":"10.18653\/v1\/P18-1031"},{"key":"ref_50","doi-asserted-by":"crossref","unstructured":"Ruder, S. (2019). Neural Transfer Learning for Natural Language Processing. [Ph.D. Thesis, NUI Galway].","DOI":"10.18653\/v1\/N19-5004"},{"key":"ref_51","doi-asserted-by":"crossref","unstructured":"Zeiler, M.D., and Fergus, R. (2014, January 6\u201312). Visualizing and Understanding Convolutional Networks. Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland.","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"ref_52","first-page":"4765","article-title":"A Unified Approach to Interpreting Model Predictions","volume":"30","author":"Lundberg","year":"2017","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_53","unstructured":"Ghorbani, A., Abid, A., and Zou, J. (February, January 27). Interpretation of neural networks is fragile. Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Honolulu, HI, USA."},{"key":"ref_54","first-page":"841","article-title":"Counterfactual explanations without opening the black box: Automated decisions and the GDPR","volume":"31","author":"Wachter","year":"2017","journal-title":"Harv. JL Tech."},{"key":"ref_55","doi-asserted-by":"crossref","unstructured":"Kindermans, P.J., Hooker, S., Adebayo, J., Alber, M., Sch\u00fctt, K.T., D\u00e4hne, S., Erhan, D., and Kim, B. (2019). The (un) reliability of saliency methods. Explainable AI: Interpreting, Explaining and Visualizing Deep Learning, Springer.","DOI":"10.1007\/978-3-030-28954-6_14"},{"key":"ref_56","unstructured":"Sixt, L., Granz, M., and Landgraf, T. (2020, January 13\u201318). When explanations lie: Why many modified bp attributions fail. Proceedings of the International Conference on Machine Learning (ICML), Online."},{"key":"ref_57","doi-asserted-by":"crossref","unstructured":"Geva, M., Schuster, R., Berant, J., and Levy, O. (2021, January 7\u201311). Transformer Feed-Forward Layers Are Key-Value Memories. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP), Punta Cana, Dominican Republic.","DOI":"10.18653\/v1\/2021.emnlp-main.446"},{"key":"ref_58","unstructured":"Alain, G., and Bengio, Y. (2016, January 2\u20134). Understanding intermediate layers using linear classifier probes. Proceedings of the International Conference on Learning Representations (ICLR) Workshop, San Juan, Puerto Rico."},{"key":"ref_59","unstructured":"Morcos, A.S., Barrett, D.G., Rabinowitz, N.C., and Botvinick, M. (May, January 30). On the importance of single directions for generalization. Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada."},{"key":"ref_60","first-page":"9737","article-title":"A benchmark for interpretability methods in deep neural networks","volume":"32","author":"Hooker","year":"2019","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_61","doi-asserted-by":"crossref","unstructured":"Agarwal, C., D\u2019souza, D., and Hooker, S. (2022, January 19\u201324). Estimating example difficulty using variance of gradients. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA.","DOI":"10.1109\/CVPR52688.2022.01012"},{"key":"ref_62","first-page":"20378","article-title":"Movement pruning: Adaptive sparsity by fine-tuning","volume":"33","author":"Sanh","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_63","unstructured":"Frankle, J., and Carbin, M. (2019, January 6\u20139). The lottery ticket hypothesis: Finding sparse, trainable neural networks. Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA."},{"key":"ref_64","doi-asserted-by":"crossref","unstructured":"Molchanov, P., Mallya, A., Tyree, S., Frosio, I., and Kautz, J. (2019, January 15\u201320). Importance estimation for neural network pruning. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.01152"},{"key":"ref_65","unstructured":"Theis, L., Korshunova, I., Tejani, A., and Husz\u00e1r, F. (2018). Faster gaze prediction with dense networks and fisher pruning. arXiv."},{"key":"ref_66","doi-asserted-by":"crossref","first-page":"71","DOI":"10.1098\/rsta.1894.0003","article-title":"Contributions to the mathematical theory of evolution","volume":"185","author":"Pearson","year":"1894","journal-title":"Philos. Trans. R. Soc. Lond. A"},{"key":"ref_67","unstructured":"Arthur, D., and Vassilvitskii, S. (2006). k-Means++: The Advantages of Careful Seeding, Technical Report; Stanford."},{"key":"ref_68","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1111\/j.2517-6161.1977.tb01600.x","article-title":"Maximum likelihood from incomplete data via the EM algorithm","volume":"39","author":"Dempster","year":"1977","journal-title":"J. R. Stat. Soc. Ser. B"},{"key":"ref_69","first-page":"100","article-title":"Algorithm AS 136: A k-means clustering algorithm","volume":"28","author":"Hartigan","year":"1979","journal-title":"J. R. Stat. Soc. Ser. C"},{"key":"ref_70","unstructured":"Bishop, C.M., and Nasrabadi, N.M. (2006). Pattern Recognition and Machine Learning, Springer."},{"key":"ref_71","doi-asserted-by":"crossref","first-page":"611","DOI":"10.1198\/016214502760047131","article-title":"Model-based clustering, discriminant analysis, and density estimation","volume":"97","author":"Fraley","year":"2002","journal-title":"J. Am. Stat. Assoc."},{"key":"ref_72","unstructured":"Han, S., Mao, H., and Dally, W.J. (2015). Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv."},{"key":"ref_73","unstructured":"Pochinkov, N., and Schoots, N. (2024). Dissecting language models: Machine unlearning via selective pruning. arXiv."}],"container-title":["Machine Learning and Knowledge Extraction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2504-4990\/7\/3\/95\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T18:40:19Z","timestamp":1760035219000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2504-4990\/7\/3\/95"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,5]]},"references-count":73,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2025,9]]}},"alternative-id":["make7030095"],"URL":"https:\/\/doi.org\/10.3390\/make7030095","relation":{},"ISSN":["2504-4990"],"issn-type":[{"type":"electronic","value":"2504-4990"}],"subject":[],"published":{"date-parts":[[2025,9,5]]}}}