{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T14:38:34Z","timestamp":1769006314968,"version":"3.49.0"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,12,16]],"date-time":"2024-12-16T00:00:00Z","timestamp":1734307200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,16]],"date-time":"2024-12-16T00:00:00Z","timestamp":1734307200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62302086"],"award-info":[{"award-number":["62302086"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005047","name":"Natural Science Foundation of Liaoning Province","doi-asserted-by":"publisher","award":["2023-MSBA-070"],"award-info":[{"award-number":["2023-MSBA-070"]}],"id":[{"id":"10.13039\/501100005047","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["N2317005"],"award-info":[{"award-number":["N2317005"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Real-Time Image Proc"],"published-print":{"date-parts":[[2025,2]]},"DOI":"10.1007\/s11554-024-01598-3","type":"journal-article","created":{"date-parts":[[2024,12,16]],"date-time":"2024-12-16T07:04:59Z","timestamp":1734332699000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["EAPoseNet: efficient animal pose network in low computing power scenarios"],"prefix":"10.1007","volume":"22","author":[{"given":"Yuming","family":"Chen","sequence":"first","affiliation":[]},{"given":"Chaopeng","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Tianzhe","family":"Jiao","sequence":"additional","affiliation":[]},{"given":"Zhe","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Jie","family":"Song","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,16]]},"reference":[{"key":"1598_CR1","unstructured":"Li, Shuyuan, Li, Jianguo, Tang, Hanlin, Qian, Rui, Lin, Weiyao: Atrw: a benchmark for amur tiger re-identification in the wild. arXiv preprint arXiv:1906.05586, (2019)"},{"key":"1598_CR2","unstructured":"Perez, Michael, Toler-Franklin, Corey: Cnn-based action recognition and pose estimation for classifying animal behavior from videos: A survey. arXiv preprint arXiv:2301.06187, (2023)"},{"key":"1598_CR3","doi-asserted-by":"crossref","unstructured":"Bambori, Viral, Ghatiya, Yash, Sethi, Tathya, Sharma, Abhishek, Sharma, Sanjeev: Efficient pose estimation in social insects residing in colonies using deepposekit. In 2023 IEEE International Conference on Advanced Networks and Telecommunications Systems (ANTS), pages 144\u2013147. IEEE, (2023)","DOI":"10.1109\/ANTS59832.2023.10468853"},{"key":"1598_CR4","doi-asserted-by":"crossref","unstructured":"Pascual-Hern\u00e1ndez, David, de Frutos, Nuria Oyaga, Mora-Jim\u00e9nez, Inmaculada, Ca\u00f1as-Plaza, Jos\u00e9\u00a0Mar\u00eda: Efficient 3d human pose estimation from rgbd sensors. Displays, 74:102225, (2022)","DOI":"10.1016\/j.displa.2022.102225"},{"key":"1598_CR5","doi-asserted-by":"crossref","unstructured":"Cao, Jinkun, Tang, Hongyang, Fang, Hao-Shu, Shen, Xiaoyong, Lu, Cewu, Tai, Yu-Wing: Cross-domain adaptation for animal pose estimation. In Proceedings of the IEEE\/CVF international conference on computer vision, pages 9498\u20139507, (2019)","DOI":"10.1109\/ICCV.2019.00959"},{"key":"1598_CR6","doi-asserted-by":"crossref","unstructured":"Sun, Ke, Xiao, Bin, Liu, Dong, Wang, Jingdong: Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pages 5693\u20135703, (2019)","DOI":"10.1109\/CVPR.2019.00584"},{"key":"1598_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.displa.2024.102723","volume":"83","author":"R Wang","year":"2024","unstructured":"Wang, R., Liu, Z., Zhang, H., Wang, X., Cao, W.: Cross dimensional subspace attention for human pose estimation. Displays 83, 102723 (2024)","journal-title":"Displays"},{"key":"1598_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2024.104093","volume":"99","author":"Q Zheng","year":"2024","unstructured":"Zheng, Q., Guo, H., Yin, Y., Zheng, B., Jiang, H.: Lfsimcc: Spatial fusion lightweight network for human pose estimation. J. Vis. Commun. Image Represent. 99, 104093 (2024)","journal-title":"J. Vis. Commun. Image Represent."},{"key":"1598_CR9","doi-asserted-by":"crossref","unstructured":"Xiao, Bin, Wu, Haiping, Wei, Yichen: Simple baselines for human pose estimation and tracking. In Proceedings of the European conference on computer vision (ECCV), pages 466\u2013481, (2018)","DOI":"10.1007\/978-3-030-01231-1_29"},{"key":"1598_CR10","unstructured":"Yang, Lingxiao, Zhang, Ru-Yuan, Li, Lida, Xie, Xiaohua: Simam: A simple, parameter-free attention module for convolutional neural networks. In International conference on machine learning, pages 11863\u201311874. PMLR, (2021)"},{"key":"1598_CR11","doi-asserted-by":"crossref","unstructured":"Wang, Chien-Yao, Mark Liao, Hong-Yuan, Wu, Yueh-Hua, Chen, Ping-Yang, Hsieh, Jun-Wei, Yeh, I-Hau: Cspnet: A new backbone that can enhance learning capability of cnn. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition workshops, pages 390\u2013391, (2020)","DOI":"10.1109\/CVPRW50498.2020.00203"},{"key":"1598_CR12","doi-asserted-by":"crossref","unstructured":"Kresovic, Milan, Nguyen, Thong, Ullah, Mohib, Afridi, Hina, Cheikh, Faouzi\u00a0Alaya: Pigpose: A realtime framework for farm animal pose estimation and tracking. In IFIP International Conference on Artificial Intelligence Applications and Innovations, pages 204\u2013215. Springer, (2022)","DOI":"10.1007\/978-3-031-08333-4_17"},{"key":"1598_CR13","doi-asserted-by":"crossref","unstructured":"Kitamura, Takumi, Teshima, Hitoshi, Thomas, Diego, Kawasaki, Hiroshi: Refining openpose with a new sports dataset for robust 2d pose estimation. In Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pages 672\u2013681, (2022)","DOI":"10.1109\/WACVW54805.2022.00074"},{"issue":"1","key":"1598_CR14","doi-asserted-by":"publisher","first-page":"117","DOI":"10.1038\/s41592-018-0234-5","volume":"16","author":"TD Pereira","year":"2019","unstructured":"Pereira, T.D., Aldarondo, D.E., Willmore, L., Kislin, M., Wang, S.S.-H., Murthy, M., Shaevitz, J.W.: Fast animal pose estimation using deep neural networks. Nat. Methods 16(1), 117\u2013125 (2019)","journal-title":"Nat. Methods"},{"key":"1598_CR15","doi-asserted-by":"crossref","unstructured":"Cheng, Bowen, Xiao, Bin, Wang, Jingdong, Shi, Honghui, Huang, Thomas\u00a0S., Zhang, Lei: Higherhrnet: Scale-aware representation learning for bottom-up human pose estimation. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pages 5386\u20135395, (2020)","DOI":"10.1109\/CVPR42600.2020.00543"},{"key":"1598_CR16","doi-asserted-by":"crossref","unstructured":"Bras\u00f3, Guillem, Kister, Nikita, Leal-Taix\u00e9, Laura: The center of attention: Center-keypoint grouping via attention for multi-person pose estimation. In Proceedings of the IEEE\/CVF International Conference on Computer Vision, pages 11853\u201311863, (2021)","DOI":"10.1109\/ICCV48922.2021.01164"},{"key":"1598_CR17","doi-asserted-by":"crossref","unstructured":"Geng, Zigang, Sun, Ke, Xiao, Bin, Zhang, Zhaoxiang, Wang, Jingdong: Bottom-up human pose estimation via disentangled keypoint regression. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pages 14676\u201314686, (2021)","DOI":"10.1109\/CVPR46437.2021.01444"},{"key":"1598_CR18","doi-asserted-by":"crossref","unstructured":"Huang, Junjie, Zhu, Zheng, Guo, Feng, Huang, Guan: The devil is in the details: Delving into unbiased data processing for human pose estimation. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pages 5700\u20135709, (2020)","DOI":"10.1109\/CVPR42600.2020.00574"},{"key":"1598_CR19","doi-asserted-by":"crossref","unstructured":"Lauer, Jessy, Zhou, Ye., Shaokai, Menegas, William, Schneider, Steffen, Nath, Tanmay, Rahman, Mostafizur, Mohammed, Santo, Di., Valentina, Soberanes, Daniel, Feng, Guoping, et al.: Multi-animal pose estimation, identification and tracking with deeplabcut. Nature Methods 19(4), 496\u2013504 (2022)","DOI":"10.1038\/s41592-022-01443-0"},{"issue":"10","key":"1598_CR20","doi-asserted-by":"publisher","first-page":"1938","DOI":"10.3390\/agriculture13101938","volume":"13","author":"X Li","year":"2023","unstructured":"Li, X., Sun, K., Fan, H., He, Z.: Real-time cattle pose estimation based on improved rtmpose. Agriculture 13(10), 1938 (2023)","journal-title":"Agriculture"},{"key":"1598_CR21","first-page":"7281","volume":"34","author":"Y Yuan","year":"2021","unstructured":"Yuan, Y., Rao, F., Huang, L., Lin, W., Zhang, C., Chen, X., Wang, J.: Hrformer: High-resolution vision transformer for dense predict. Adv. Neural. Inf. Process. Syst. 34, 7281\u20137293 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1598_CR22","doi-asserted-by":"crossref","unstructured":"Li, Chen, Lee, Gim\u00a0Hee: From synthetic to real: Unsupervised domain adaptation for animal pose estimation. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 1482\u20131491, June (2021)","DOI":"10.1109\/CVPR46437.2021.00153"},{"key":"1598_CR23","first-page":"38571","volume":"35","author":"X Yufei","year":"2022","unstructured":"Yufei, X., Zhang, J., Zhang, Q., Tao, D.: Vitpose: Simple vision transformer baselines for human pose estimation. Adv. Neural. Inf. Process. Syst. 35, 38571\u201338584 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1598_CR24","unstructured":"Jiang, Tao, Lu, Peng, Zhang, Li, Ma, Ningsheng, Han, Rui, Lyu, Chengqi, Li, Yining, Chen, Kai: Rtmpose: Real-time multi-person pose estimation based on mmpose. arXiv preprint arXiv:2303.07399, (2023)"},{"key":"1598_CR25","doi-asserted-by":"crossref","unstructured":"Maji, Debapriya, Nagori, Soyeb, Mathew, Manu, Poddar, Deepak: Yolo-pose: Enhancing yolo for multi person pose estimation using object keypoint similarity loss. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pages 2637\u20132646, (2022)","DOI":"10.1109\/CVPRW56347.2022.00297"},{"issue":"8","key":"1598_CR26","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0306530","volume":"19","author":"E Zhao","year":"2024","unstructured":"Zhao, E., Chen, B., Zhao, H., Liu, G., Jiang, J., Li, Y., Zhang, J., Luo, C.: Lightweight cattle pose estimation with fusion of reparameterization and an attention mechanism. PLoS ONE 19(8), e0306530 (2024)","journal-title":"PLoS ONE"},{"key":"1598_CR27","doi-asserted-by":"crossref","unstructured":"Yu, Dongyang, Zhang, Haoyue, Zhao, Ruisheng, Chen, Guoqi, An, Yang, Yanhong: Movepose: A high-performance human pose estimation algorithm on mobile and edge devices. In International Conference on Artificial Neural Networks, pages 144\u2013158. Springer, (2024)","DOI":"10.1007\/978-3-031-72338-4_11"},{"key":"1598_CR28","doi-asserted-by":"publisher","first-page":"5314","DOI":"10.1109\/ACCESS.2022.3231750","volume":"11","author":"W Zhang","year":"2022","unstructured":"Zhang, W., Yang, X., Bai, R., Li, L.: Animal pose estimation algorithm based on the lightweight stacked hourglass network. IEEE Access 11, 5314\u20135327 (2022)","journal-title":"IEEE Access"},{"key":"1598_CR29","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2024.108662","volume":"218","author":"L Chen","year":"2024","unstructured":"Chen, L., Zhang, L., Tang, J., Tang, C., An, R., Han, R., Zhang, Y.: Grmpose: Gcn-based real-time dairy goat pose estimation. Comput. Electron. Agric. 218, 108662 (2024)","journal-title":"Comput. Electron. Agric."},{"key":"1598_CR30","doi-asserted-by":"crossref","unstructured":"Yu, Weihao, Zhou, Pan, Yan, Shuicheng, Wang, Xinchao: Inceptionnext: When inception meets convnext. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pages 5672\u20135683, (2024)","DOI":"10.1109\/CVPR52733.2024.00542"},{"key":"1598_CR31","unstructured":"Yang, Songlin, Wang, Bailin, Shen, Yikang, Panda, Rameswar, Kim, Yoon: Gated linear attention transformers with hardware-efficient training. arXiv preprint arXiv:2312.06635, (2023)"},{"key":"1598_CR32","doi-asserted-by":"crossref","unstructured":"He, Kaiming, Zhang, Xiangyu, Ren, Shaoqing, Sun, Jian: Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770\u2013778, (2016)","DOI":"10.1109\/CVPR.2016.90"},{"issue":"6","key":"1598_CR33","doi-asserted-by":"publisher","first-page":"415","DOI":"10.1016\/j.irbm.2021.03.002","volume":"42","author":"R El Jurdi","year":"2021","unstructured":"El Jurdi, R., Petitjean, C., Honeine, P., Abdallah, F.: Coordconv-unet: investigating coordconv for organ segmentation. IRBM 42(6), 415\u2013423 (2021)","journal-title":"IRBM"},{"key":"1598_CR34","doi-asserted-by":"crossref","unstructured":"Hu, Jie, Shen, Li, Sun, Gang: Squeeze-and-excitation networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7132\u20137141, (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"1598_CR35","doi-asserted-by":"crossref","unstructured":"Woo, Sanghyun, Park, Jongchan, Lee, Joon-Young, Kweon, In\u00a0So: Cbam: Convolutional block attention module. In Proceedings of the European conference on computer vision (ECCV), pages 3\u201319, (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"issue":"9","key":"1598_CR36","doi-asserted-by":"publisher","first-page":"1904","DOI":"10.1109\/TPAMI.2015.2389824","volume":"37","author":"K He","year":"2015","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 37(9), 1904\u20131916 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1598_CR37","doi-asserted-by":"crossref","unstructured":"Wang, Chien-Yao, Yeh, I-Hau, Mark Liao, Hong-Yuan: Yolov9: Learning what you want to learn using programmable gradient information. arXiv preprint arXiv:2402.13616, (2024)","DOI":"10.1007\/978-3-031-72751-1_1"},{"key":"1598_CR38","doi-asserted-by":"crossref","unstructured":"Chollet, Fran\u00e7ois: Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1251\u20131258, (2017)","DOI":"10.1109\/CVPR.2017.195"},{"key":"1598_CR39","doi-asserted-by":"crossref","unstructured":"Yang, Zhendong, Zeng, Ailing, Yuan, Chun, Li, Yu: Effective whole-body pose estimation with two-stages distillation. In Proceedings of the IEEE\/CVF International Conference on Computer Vision, pages 4210\u20134220, (2023)","DOI":"10.1109\/ICCVW60793.2023.00455"},{"key":"1598_CR40","unstructured":"Yu, Hang, Xu, Yufei, Zhang, Jing, Zhao, Wei, Guan, Ziyu, Tao, Dacheng: Ap-10k: A benchmark for animal pose estimation in the wild. arXiv preprint arXiv:2108.12617, (2021)"},{"key":"1598_CR41","unstructured":"Graving, Jacob\u00a0M, Chae, Daniel, Naik, Hemal, Li, Liang, Koger, Benjamin, Costelloe, Blair\u00a0R., Couzin,Iain\u00a0D: Fast and robust animal pose estimation. bioRxiv, page 620245, (2019)"},{"key":"1598_CR42","unstructured":"Yang, Jie, Zeng, Ailing, Zhang, Ruimao, Zhang, Lei: Unipose: Detecting any keypoints. arXiv preprint arXiv:2310.08530, (2023)"},{"key":"1598_CR43","doi-asserted-by":"crossref","unstructured":"Imambi, Sagar, Prakash, Kolla\u00a0Bhanu, Kanagachidambaresan, G.R.: Pytorch. Programming with TensorFlow: solution for edge computing applications, pages 87\u2013104, (2021)","DOI":"10.1007\/978-3-030-57077-4_10"},{"key":"1598_CR44","unstructured":"MMPose Contributors. Openmmlab pose estimation toolbox and benchmark. https:\/\/github.com\/open-mmlab\/mmpose, (2020)"},{"key":"1598_CR45","unstructured":"MMCV Contributors. MMCV: OpenMMLab computer vision foundation. https:\/\/github.com\/open-mmlab\/mmcv, (2018)"},{"key":"1598_CR46","unstructured":"Chen, Kai, Wang, Jiaqi, Pang, Jiangmiao, Cao, Yuhang, Xiong, Yu, Li, Xiaoxiao, Sun, Shuyang, Feng, Wansen, Liu, Ziwei, Xu, Jiarui, Zhang, Zheng, Cheng, Dazhi, Zhu, Chenchen, Cheng, Tianheng, Zhao, Qijie, Li, Buyu, Lu, Xin, Zhu, Rui, Wu, Yue, Dai, Jifeng, Wang, Jingdong, Shi, Jianping, Ouyang, Wanli, Loy, Chen\u00a0Change, Lin, Dahua: MMDetection: Open mmlab detection toolbox and benchmark. arXiv preprint arXiv:1906.07155, (2019)"},{"issue":"2","key":"1598_CR47","doi-asserted-by":"publisher","first-page":"496","DOI":"10.1007\/s11263-022-01711-8","volume":"131","author":"L Zhang","year":"2023","unstructured":"Zhang, L., Gao, J., Xiao, Z., Fan, H.: Animaltrack: A benchmark for multi-animal tracking in the wild. Int. J. Comput. Vision 131(2), 496\u2013513 (2023)","journal-title":"Int. J. Comput. Vision"},{"key":"1598_CR48","first-page":"17301","volume":"35","author":"Y Yang","year":"2022","unstructured":"Yang, Y., Yang, J., Yufei, X., Zhang, J., Lan, L., Tao, D.: Apt-36k: A large-scale benchmark for animal pose estimation and tracking. Adv. Neural. Inf. Process. Syst. 35, 17301\u201317313 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1598_CR49","doi-asserted-by":"crossref","unstructured":"Wang, Xinlong, Kong, Tao, Shen, Chunhua, Jiang, Yuning, Li, Lei: Solo: Segmenting objects by locations. In Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XVIII 16, pages 649\u2013665. Springer, (2020)","DOI":"10.1007\/978-3-030-58523-5_38"},{"key":"1598_CR50","doi-asserted-by":"crossref","unstructured":"Hou, Qibin, Zhou, Daquan, Feng, Jiashi: Coordinate attention for efficient mobile network design. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pages 13713\u201313722, (2021)","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"1598_CR51","doi-asserted-by":"crossref","unstructured":"Wang, Qilong, Wu, Banggu, Zhu, Pengfei, Li, Peihua, Zuo, Wangmeng, Hu, Qinghua: Eca-net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pages 11534\u201311542, (2020)","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"1598_CR52","unstructured":"MMDeploy Contributors. Openmmlab\u2019s model deployment toolbox. https:\/\/github.com\/open-mmlab\/mmdeploy, (2021)"}],"container-title":["Journal of Real-Time Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01598-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11554-024-01598-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01598-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,3]],"date-time":"2025-02-03T17:18:33Z","timestamp":1738603113000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11554-024-01598-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,16]]},"references-count":52,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,2]]}},"alternative-id":["1598"],"URL":"https:\/\/doi.org\/10.1007\/s11554-024-01598-3","relation":{},"ISSN":["1861-8200","1861-8219"],"issn-type":[{"value":"1861-8200","type":"print"},{"value":"1861-8219","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,16]]},"assertion":[{"value":"7 August 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 November 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 December 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"21"}}