{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T07:31:52Z","timestamp":1774423912288,"version":"3.50.1"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,12,31]],"date-time":"2024-12-31T00:00:00Z","timestamp":1735603200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,12,31]],"date-time":"2024-12-31T00:00:00Z","timestamp":1735603200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62402525"],"award-info":[{"award-number":["62402525"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["2462023YJRC023"],"award-info":[{"award-number":["2462023YJRC023"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["CCF Trans. HPC"],"published-print":{"date-parts":[[2025,2]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>With the support of more precision formats in emerging hardware architectures, mixed-precision has become a popular approach to accelerate deep learning (DL) training. Applying low-precision formats such as FP16 and BF16 to neural operators can save GPU memory while improving bandwidth. However, DL frameworks use black and white lists as default mixed-precision selections and cannot flexibly adapt to a variety of neural networks. In addition, existing work on automatic precision adjustment does not consider model convergence, and the decision cost of precision selection is high. To address the above problems, this paper proposes <jats:italic>CoMP<\/jats:italic>, a non-intrusive framework for Convergence-aware operator-wise Mixed-precision training. <jats:italic>CoMP<\/jats:italic> uses two-stage precision adjustment based on epochs and batches to ensure convergence and performance respectively. After that, <jats:italic>CoMP<\/jats:italic> performs subsequent training according to the searched optimal operator-wise mixed-precision plan. The experimental results on A100 GPU show that <jats:italic>CoMP<\/jats:italic> achieves a maximum performance speedup of 1.15<jats:inline-formula>\n              <jats:alternatives>\n                <jats:tex-math>$$\\times$$<\/jats:tex-math>\n                <mml:math xmlns:mml=\"http:\/\/www.w3.org\/1998\/Math\/MathML\">\n                  <mml:mo>\u00d7<\/mml:mo>\n                <\/mml:math>\n              <\/jats:alternatives>\n            <\/jats:inline-formula> compared with PyTorch AMP implementation, while also saving up to 29.81% of GPU memory.<\/jats:p>","DOI":"10.1007\/s42514-024-00208-9","type":"journal-article","created":{"date-parts":[[2024,12,31]],"date-time":"2024-12-31T10:51:05Z","timestamp":1735642265000},"page":"43-57","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Convergence-aware operator-wise mixed-precision training"],"prefix":"10.1007","volume":"7","author":[{"given":"Wenhao","family":"Dai","sequence":"first","affiliation":[]},{"given":"Ziyi","family":"Jia","sequence":"additional","affiliation":[]},{"given":"Yuesi","family":"Bai","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2927-362X","authenticated-orcid":false,"given":"Qingxiao","family":"Sun","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,31]]},"reference":[{"key":"208_CR1","unstructured":"Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., Devin, M., Ghemawat, S., Irving, G., Isard, M., et al.: Tensorflow: A system for large-scale machine learning. In: USENIX Symposium on Operating Systems Design and Implementation (2016)"},{"issue":"12","key":"208_CR2","doi-asserted-by":"publisher","first-page":"2526","DOI":"10.1016\/j.cpc.2008.11.005","volume":"180","author":"M Baboulin","year":"2009","unstructured":"Baboulin, M., Buttari, A., Dongarra, J., Kurzak, J., Langou, J., Langou, J., Luszczek, P., Tomov, S.: Accelerating scientific computations with mixed precision algorithms. Comput. Phys. Commun. 180(12), 2526\u20132533 (2009)","journal-title":"Comput. Phys. Commun."},{"issue":"4","key":"208_CR3","doi-asserted-by":"publisher","first-page":"457","DOI":"10.1177\/1094342007084026","volume":"21","author":"A Buttari","year":"2007","unstructured":"Buttari, A., Dongarra, J., Langou, J., Langou, J., Luszczek, P., Kurzak, J.: Mixed precision iterative refinement techniques for the solution of dense linear systems. Int. J. High Perform. Comput. Appl. 21(4), 457\u2013466 (2007)","journal-title":"Int. J. High Perform. Comput. Appl."},{"key":"208_CR4","unstructured":"Bylina, B., Bylina, J.: Mixed precision iterative refinement techniques for the wz factorization. In: Federated Conference on Computer Science and Information Systems, pp. 425\u2013431 (2013)"},{"key":"208_CR5","doi-asserted-by":"crossref","unstructured":"Cai, Z., Vasconcelos, N.: Rethinking differentiable search for mixed-precision neural networks. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2349\u20132358 (2020)","DOI":"10.1109\/CVPR42600.2020.00242"},{"key":"208_CR6","doi-asserted-by":"crossref","unstructured":"Chitty-Venkata, K.T., Emani, M., Vishwanath, V., Somani, A.K.: Efficient design space exploration for sparse mixed precision neural architectures. In: International Symposium on High-Performance Parallel and Distributed Computing, pp. 265\u2013276 (2022)","DOI":"10.1145\/3502181.3531463"},{"key":"208_CR7","doi-asserted-by":"crossref","unstructured":"Cho, Y., Demmel, J.W., King, J., Li, X.S., Liu, Y., Luo, H.: Harnessing the crowd for autotuning high-performance computing applications. In: IEEE International Parallel and Distributed Processing Symposium, pp. 635\u2013645 (2023)","DOI":"10.1109\/IPDPS54959.2023.00069"},{"issue":"11","key":"208_CR8","doi-asserted-by":"publisher","first-page":"2040","DOI":"10.1109\/JPROC.2018.2868961","volume":"106","author":"J Dongarra","year":"2018","unstructured":"Dongarra, J., Gates, M., Kurzak, J., Luszczek, P., Tsai, Y.M.: Autotuning numerical dense linear algebra for batched computation with gpu hardware accelerators. Proc. IEEE 106(11), 2040\u20132055 (2018)","journal-title":"Proc. IEEE"},{"key":"208_CR9","doi-asserted-by":"crossref","unstructured":"Gao, W., Zhuang, W., Li, M., Sun, P., Wen, Y., Zhang, T.: Ymir: A scheduler for foundation model fine-tuning workloads in datacenters. In: ACM International Conference on Supercomputing, pp. 259\u2013271 (2024)","DOI":"10.1145\/3650200.3656599"},{"key":"208_CR10","doi-asserted-by":"crossref","unstructured":"Gholami, A., Kim, S., Dong, Z., Yao, Z., Mahoney, M.W., Keutzer, K.: A survey of quantization methods for efficient neural network inference. In: Low-Power Computer Vision, pp. 291\u2013326 (2022)","DOI":"10.1201\/9781003162810-13"},{"key":"208_CR11","doi-asserted-by":"crossref","unstructured":"Gong, C., Jiang, Z., Wang, D., Lin, Y., Liu, Q., Pan, D.Z.: Mixed precision neural architecture search for energy efficient deep learning. In: IEEE\/ACM International Conference on Computer-Aided Design, pp. 1\u20137 (2019). IEEE","DOI":"10.1109\/ICCAD45719.2019.8942147"},{"issue":"11","key":"208_CR12","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial networks. Commun. ACM 63(11), 139\u2013144 (2020)","journal-title":"Commun. ACM"},{"key":"208_CR13","doi-asserted-by":"crossref","unstructured":"Haidar, A., Tomov, S., Dongarra, J., Higham, N.J.: Harnessing gpu tensor cores for fast fp16 arithmetic to speed up mixed-precision iterative refinement solvers. In: International Conference for High Performance Computing, Networking, Storage and Analysis, pp. 603\u2013613 (2018)","DOI":"10.1109\/SC.2018.00050"},{"key":"208_CR14","doi-asserted-by":"crossref","unstructured":"Haidar, A., Wu, P., Tomov, S., Dongarra, J.: Investigating half precision arithmetic to accelerate dense linear system solvers. In: Workshop on Latest Advances in Scalable Algorithms for Large-scale Systems, pp. 1\u20138 (2017)","DOI":"10.1145\/3148226.3148237"},{"key":"208_CR15","unstructured":"He, X., Sun, J., Chen, H., Li, D.: Campo: Cost-aware performance optimization for mixed-precision neural network training. In: USENIX Annual Technical Conference, pp. 505\u2013518 (2022)"},{"issue":"2","key":"208_CR16","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3441830","volume":"18","author":"N-M Ho","year":"2021","unstructured":"Ho, N.-M., Silva, H.D., Wong, W.-F.: Gram: A framework for dynamically mixing precisions in gpu applications. ACM Trans. Architect. Code Optim. 18(2), 1\u201324 (2021)","journal-title":"ACM Trans. Architect. Code Optim."},{"key":"208_CR17","doi-asserted-by":"crossref","unstructured":"Huang, K., Zhai, J., Zheng, Z., Yi, Y., Shen, X.: Understanding and bridging the gaps in current gnn performance optimizations. In: ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming, pp. 119\u2013132 (2021)","DOI":"10.1145\/3437801.3441585"},{"key":"208_CR18","unstructured":"Jia, X., Song, S., He, W., Wang, Y., Rong, H., Zhou, F., Xie, L., Guo, Z., Yang, Y., Yu, L., et al.: Highly scalable deep learning training system with mixed-precision: Training imagenet in four minutes. arXiv preprint arXiv:1807.11205 (2018)"},{"key":"208_CR19","unstructured":"Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016)"},{"key":"208_CR20","doi-asserted-by":"crossref","unstructured":"Kotipalli, P.V., Singh, R., Wood, P., Laguna, I., Bagchi, S.: Ampt-ga: automatic mixed precision floating point tuning for gpu applications. In: ACM International Conference on Supercomputing, pp. 160\u2013170 (2019)","DOI":"10.1145\/3330345.3330360"},{"key":"208_CR21","unstructured":"Krizhevsky, A., Hinton, G., et al.: Learning multiple layers of features from tiny images (2009)"},{"key":"208_CR22","doi-asserted-by":"publisher","DOI":"10.1145\/3065386","author":"A Krizhevsky","year":"2012","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Adv. Neural Inform. Process. Syst. (2012). https:\/\/doi.org\/10.1145\/3065386","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"208_CR23","doi-asserted-by":"crossref","unstructured":"Lam, M.O., Hollingsworth, J.K., Supinski, B.R., Legendre, M.P.: Automatically adapting programs for mixed-precision floating-point computation. In: ACM International Conference on Supercomputing, pp. 369\u2013378 (2013)","DOI":"10.1145\/2464996.2465018"},{"key":"208_CR24","unstructured":"Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., Soricut, R.: Albert: A lite bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942 (2019)"},{"issue":"9","key":"208_CR25","doi-asserted-by":"publisher","first-page":"6690","DOI":"10.1109\/TGRS.2019.2907932","volume":"57","author":"S Li","year":"2019","unstructured":"Li, S., Song, W., Fang, L., Chen, Y., Ghamisi, P., Benediktsson, J.A.: Deep learning for hyperspectral image classification: an overview. IEEE Trans. Geosci. Remote Sens. 57(9), 6690\u20136709 (2019)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"208_CR26","unstructured":"Micikevicius, P., Narang, S., Alben, J., Diamos, G., Elsen, E., Garcia, D., Ginsburg, B., Houston, M., Kuchaiev, O., Venkatesh, G., et al.: Mixed precision training. arXiv preprint arXiv:1710.03740 (2017)"},{"key":"208_CR27","doi-asserted-by":"crossref","unstructured":"Nandakumar, S., Le\u00a0Gallo, M., Boybat, I., Rajendran, B., Sebastian, A., Eleftheriou, E.: Mixed-precision architecture based on computational memory for training deep neural networks. In: IEEE International Symposium on Circuits and Systems, pp. 1\u20135 (2018)","DOI":"10.1109\/ISCAS.2018.8351656"},{"key":"208_CR28","doi-asserted-by":"crossref","unstructured":"Parasyris, K., Georgakoudis, G., Rangel, E., Laguna, I., Doerfert, J.: Scalable tuning of (openmp) gpu applications via kernel record and replay. In: International Conference for High Performance Computing, Networking, Storage and Analysis, pp. 1\u201314 (2023)","DOI":"10.1145\/3581784.3607098"},{"key":"208_CR29","unstructured":"Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., et al.: Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inform. Process. Syst. 32 (2019)"},{"key":"208_CR30","doi-asserted-by":"crossref","unstructured":"Pfaffe, P., Grosser, T., Tillmann, M.: Efficient hierarchical online-autotuning: a case study on polyhedral accelerator mapping. In: ACM International Conference on Supercomputing, pp. 354\u2013366 (2019)","DOI":"10.1145\/3330345.3330377"},{"key":"208_CR31","doi-asserted-by":"crossref","unstructured":"Rajpurkar, P., Jia, R., Liang, P.: Know what you don\u2019t know: Unanswerable questions for squad. arXiv preprint arXiv:1806.03822 (2018)","DOI":"10.18653\/v1\/P18-2124"},{"key":"208_CR32","doi-asserted-by":"crossref","unstructured":"Randall, T., Koo, J., Videau, B., Kruse, M., Wu, X., Hovland, P., Hall, M., Ge, R., Balaprakash, P.: Transfer-learning-based autotuning using gaussian copula. In: ACM International Conference on Supercomputing, pp. 37\u201349 (2023)","DOI":"10.1145\/3577193.3593712"},{"key":"208_CR33","doi-asserted-by":"crossref","unstructured":"Reggiani, E., Pappalardo, A., Doblas, M., Moreto, M., Olivieri, M., Unsal, O.S., Cristal, A.: Mix-gemm: An efficient hw-sw architecture for mixed-precision quantized deep neural networks inference on edge devices. In: IEEE International Symposium on High-Performance Computer Architecture, pp. 1085\u20131098 (2023)","DOI":"10.1109\/HPCA56546.2023.10071076"},{"issue":"3","key":"208_CR34","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., Berg, A.C., Fei-Fei, L.: Imagenet large scale visual recognition challenge. Int. J. Comput. Vision 115(3), 211\u2013252 (2015)","journal-title":"Int. J. Comput. Vision"},{"key":"208_CR35","doi-asserted-by":"crossref","unstructured":"Sevilla, J., Heim, L., Ho, A., Besiroglu, T., Hobbhahn, M., Villalobos, P.: Compute trends across three eras of machine learning. In: International Joint Conference on Neural Networks, pp. 1\u20138 (2022)","DOI":"10.1109\/IJCNN55064.2022.9891914"},{"key":"208_CR36","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"208_CR37","doi-asserted-by":"crossref","unstructured":"Sun, Q., Liu, Y., Yang, H., Jiang, Z., Liu, X., Dun, M., Luan, Z., Qian, D.: cstuner: Scalable auto-tuning framework for complex stencil computation on gpus. In: IEEE International Conference on Cluster Computing, pp. 192\u2013203 (2021)","DOI":"10.1109\/Cluster48925.2021.00037"},{"key":"208_CR38","doi-asserted-by":"crossref","unstructured":"Sun, Q., Zhang, X., Geng, H., Zhao, Y., Bai, Y., Zheng, H., Yu, B.: Gtuner: Tuning dnn computations on gpu via graph attention network. In: ACM\/IEEE Design Automation Conference, pp. 1045\u20131050 (2022)","DOI":"10.1145\/3489517.3530584"},{"issue":"12","key":"208_CR39","doi-asserted-by":"publisher","first-page":"1614","DOI":"10.1109\/TC.2008.89","volume":"57","author":"J Sun","year":"2008","unstructured":"Sun, J., Peterson, G.D., Storaasli, O.O.: High-performance mixed-precision linear solver for fpgas. IEEE Trans. Comput. 57(12), 1614\u20131623 (2008)","journal-title":"IEEE Trans. Comput."},{"key":"208_CR40","first-page":"21508","volume":"35","author":"Z Sun","year":"2022","unstructured":"Sun, Z., Ge, C., Wang, J., Lin, M., Chen, H., Li, H., Sun, X.: Entropy-driven mixed-precision quantization for deep network design. Adv. Neural. Inf. Process. Syst. 35, 21508\u201321520 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"1","key":"208_CR41","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1109\/TPDS.2023.3325630","volume":"35","author":"Q Sun","year":"2024","unstructured":"Sun, Q., Liu, Y., Yang, H., Jiang, Z., Luan, Z., Qian, D.: Adaptive auto-tuning framework for global exploration of stencil optimization on gpus. IEEE Trans. Parallel Distrib. Syst. 35(1), 20\u201333 (2024)","journal-title":"IEEE Trans. Parallel Distrib. Syst."},{"key":"208_CR42","unstructured":"Tian, R., Zhao, Z., Liu, W., Liu, H., Mao, W., Zhao, Z., Yan, K.: Samp: A toolkit for model inference with self-adaptive mixed-precision. arXiv preprint arXiv:2209.09130 (2022)"},{"key":"208_CR43","unstructured":"Veli\u010dkovi\u0107, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., Bengio, Y.: Graph attention networks. arXiv preprint arXiv:1710.10903 (2017)"},{"key":"208_CR44","doi-asserted-by":"crossref","unstructured":"Wang, K., Liu, Z., Lin, Y., Lin, J., Han, S.: Haq: Hardware-aware automated quantization with mixed precision. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8612\u20138620 (2019)","DOI":"10.1109\/CVPR.2019.00881"},{"key":"208_CR45","doi-asserted-by":"crossref","unstructured":"Wang, Y., Zhang, C., Xie, Z., Guo, C., Liu, Y., Leng, J.: Dual-side sparse tensor core. In: ACM\/IEEE Annual International Symposium on Computer Architecture, pp. 1083\u20131095 (2021)","DOI":"10.1109\/ISCA52012.2021.00088"},{"key":"208_CR46","unstructured":"Wang, M., Zheng, D., Ye, Z., Gan, Q., Li, M., Song, X., Zhou, J., Ma, C., Yu, L., Gai, Y., Xiao, T., He, T., Karypis, G., Li, J., Zhang, Z.: Deep graph library: A graph-centric, highly-performant package for graph neural networks. arXiv preprint arXiv:1909.01315 (2019)"},{"key":"208_CR47","unstructured":"Wu, B., Wang, Y., Zhang, P., Tian, Y., Vajda, P., Keutzer, K.: Mixed precision quantization of convnets via differentiable neural architecture search. arXiv preprint arXiv:1812.00090 (2018)"},{"key":"208_CR48","doi-asserted-by":"crossref","unstructured":"Xu, J., Song, G., Zhou, B., Li, F., Hao, J., Zhao, J.: A holistic approach to automatic mixed-precision code generation and tuning for affine programs. In: ACM SIGPLAN Annual Symposium on Principles and Practice of Parallel Programming, pp. 55\u201367 (2024)","DOI":"10.1145\/3627535.3638484"},{"key":"208_CR49","doi-asserted-by":"crossref","unstructured":"Zhai, Y., Zhang, Y., Liu, S., Chu, X., Peng, J., Ji, J., Zhang, Y.: Tlp: A deep learning-based cost model for tensor program tuning. In: ACM International Conference on Architectural Support for Programming Languages and Operating Systems, pp. 833\u2013845 (2023)","DOI":"10.1145\/3575693.3575737"},{"issue":"7","key":"208_CR50","doi-asserted-by":"publisher","first-page":"1035","DOI":"10.1109\/TC.2019.2895031","volume":"68","author":"H Zhang","year":"2019","unstructured":"Zhang, H., Chen, D., Ko, S.-B.: Efficient multiple-precision floating-point fused multiply-add with mixed-precision support. IEEE Trans. Comput. 68(7), 1035\u20131048 (2019)","journal-title":"IEEE Trans. Comput."},{"key":"208_CR51","doi-asserted-by":"crossref","unstructured":"Zhou, X., Zhang, L., Guo, C., Yin, X., Zhuo, C.: A convolutional neural network accelerator architecture with fine-granular mixed precision configurability. In: IEEE International Symposium on Circuits and Systems, pp. 1\u20135 (2020)","DOI":"10.1109\/ISCAS45731.2020.9180844"},{"key":"208_CR52","doi-asserted-by":"crossref","unstructured":"Zhu, Z., Li, F., Li, G., Liu, Z., Mo, Z., Hu, Q., Liang, X., Cheng, J.: Mega: A memory-efficient gnn accelerator exploiting degree-aware mixed-precision quantization. In: IEEE International Symposium on High-Performance Computer Architecture, pp. 124\u2013138 (2024)","DOI":"10.1109\/HPCA57654.2024.00020"}],"container-title":["CCF Transactions on High Performance Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42514-024-00208-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s42514-024-00208-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42514-024-00208-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,24]],"date-time":"2025-02-24T10:50:12Z","timestamp":1740394212000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s42514-024-00208-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,31]]},"references-count":52,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,2]]}},"alternative-id":["208"],"URL":"https:\/\/doi.org\/10.1007\/s42514-024-00208-9","relation":{},"ISSN":["2524-4922","2524-4930"],"issn-type":[{"value":"2524-4922","type":"print"},{"value":"2524-4930","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,31]]},"assertion":[{"value":"29 July 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 November 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 December 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"On behalf of all authors, the corresponding author states that there is no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}