{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T03:17:48Z","timestamp":1740107868716,"version":"3.37.3"},"reference-count":49,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2024,7,26]],"date-time":"2024-07-26T00:00:00Z","timestamp":1721952000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,7,26]],"date-time":"2024-07-26T00:00:00Z","timestamp":1721952000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62202207"],"award-info":[{"award-number":["62202207"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Natural Science Foundation of Jiangsu Higher Education Institutions","award":["22KJB520015"],"award-info":[{"award-number":["22KJB520015"]}]},{"name":"Scientific Research Foundation of Jiangsu University","award":["21JDG051"],"award-info":[{"award-number":["21JDG051"]}]},{"name":"National Key Research and Development Program for Young Scientists","award":["2022YFD2000200"],"award-info":[{"award-number":["2022YFD2000200"]}]},{"name":"Jiangsu Provincial Key Laboratory of Smart Grid Technology and Equipment, Southeast University"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1007\/s00530-024-01426-5","type":"journal-article","created":{"date-parts":[[2024,7,26]],"date-time":"2024-07-26T21:05:15Z","timestamp":1722027915000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["SAM-guided contrast based self-training for source-free cross-domain semantic segmentation"],"prefix":"10.1007","volume":"30","author":[{"given":"Qinghua","family":"Ren","sequence":"first","affiliation":[]},{"given":"Ke","family":"Hou","sequence":"additional","affiliation":[]},{"given":"Yongzhao","family":"Zhan","sequence":"additional","affiliation":[]},{"given":"Chen","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,26]]},"reference":[{"key":"1426_CR1","doi-asserted-by":"publisher","unstructured":"Kang, G., Jiang, L., Yang, Y., et al.: Contrastive adaptation network for unsupervised domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4893\u20134902 (2019). https:\/\/doi.org\/10.1109\/cvpr.2019.00503","DOI":"10.1109\/cvpr.2019.00503"},{"key":"1426_CR2","doi-asserted-by":"publisher","unstructured":"Peng, X., Bai, Q., Xia, X., et al.: Moment matching for multi-source domain adaptation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1406\u20131415 (2019). https:\/\/doi.org\/10.1109\/iccv.2019.00149","DOI":"10.1109\/iccv.2019.00149"},{"key":"1426_CR3","doi-asserted-by":"publisher","unstructured":"Gu, X., Sun, J., Xu, Z.: Spherical space domain adaptation with robust pseudo-label loss. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9101\u20139110 (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.00912","DOI":"10.1109\/CVPR42600.2020.00912"},{"key":"1426_CR4","doi-asserted-by":"publisher","unstructured":"Fleuret, F., et al.: Uncertainty reduction for model adaptation in semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9613\u20139623 (2021). https:\/\/doi.org\/10.1109\/cvpr46437.2021.00949","DOI":"10.1109\/cvpr46437.2021.00949"},{"key":"1426_CR5","doi-asserted-by":"publisher","unstructured":"Klingner, M., Term\u00f6hlen, J.A., Ritterbach, J., et al.: Unsupervised batchnorm adaptation (UBNA): a domain adaptation method for semantic segmentation without using source domain representations. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 210\u2013220 (2022). https:\/\/doi.org\/10.1109\/wacvw54805.2022.00027","DOI":"10.1109\/wacvw54805.2022.00027"},{"key":"1426_CR6","doi-asserted-by":"publisher","unstructured":"Wang, H., Shen, T., Zhang, W., et al.: Classes matter: a fine-grained adversarial approach to cross-domain semantic segmentation. In: European Conference on Computer Vision, pp. 642\u2013659. Springer (2020). https:\/\/doi.org\/10.1007\/978-3-030-58568-6_38","DOI":"10.1007\/978-3-030-58568-6_38"},{"key":"1426_CR7","doi-asserted-by":"publisher","unstructured":"Yu, F., Zhang, M., Dong, H., et al.: Dast: unsupervised domain adaptation in semantic segmentation based on discriminator attention and self-training. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 10754\u201310762 (2021). https:\/\/doi.org\/10.1609\/aaai.v35i12.17285","DOI":"10.1609\/aaai.v35i12.17285"},{"issue":"7","key":"1426_CR8","doi-asserted-by":"publisher","first-page":"9529","DOI":"10.1109\/tits.2022.3140481","volume":"23","author":"X Zhang","year":"2022","unstructured":"Zhang, X., Chen, Y., Shen, Z., et al.: Confidence-and-refinement adaptation model for cross-domain semantic segmentation. IEEE Trans. Intell. Transp. Syst. 23(7), 9529\u20139542 (2022). https:\/\/doi.org\/10.1109\/tits.2022.3140481","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"1426_CR9","doi-asserted-by":"publisher","DOI":"10.1109\/tcsvt.2023.3260246","author":"Z Mei","year":"2023","unstructured":"Mei, Z., Ye, P., Ye, H., et al.: Automatic loss function search for adversarial unsupervised domain adaptation. IEEE Trans. Circuits Syst. Video Technol. (2023). https:\/\/doi.org\/10.1109\/tcsvt.2023.3260246","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1426_CR10","doi-asserted-by":"publisher","unstructured":"Zou, Y., Yu, Z., Liu, X., et al.: Confidence regularized self-training. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5982\u20135991 (2019). https:\/\/doi.org\/10.1109\/iccv.2019.00608","DOI":"10.1109\/iccv.2019.00608"},{"key":"1426_CR11","doi-asserted-by":"publisher","unstructured":"Zhang, P., Zhang, B., Zhang, T., et al.: Prototypical pseudo label denoising and target structure learning for domain adaptive semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12414\u201312424 (2021). https:\/\/doi.org\/10.1109\/cvpr46437.2021.01223","DOI":"10.1109\/cvpr46437.2021.01223"},{"key":"1426_CR12","doi-asserted-by":"publisher","DOI":"10.1109\/tmm.2023.3266892","author":"Q Ren","year":"2023","unstructured":"Ren, Q., Mao, Q., Lu, S.: Prototypical bidirectional adaptation and learning for cross-domain semantic segmentation. IEEE Trans. Multimedia (2023). https:\/\/doi.org\/10.1109\/tmm.2023.3266892","journal-title":"IEEE Trans. Multimedia"},{"key":"1426_CR13","doi-asserted-by":"publisher","unstructured":"Yang, C.Y., Kuo, Y.J., Hsu, C.T.: Source free domain adaptation for semantic segmentation via distribution transfer and adaptive class-balanced self-training. In: 2022 IEEE International Conference on Multimedia and Expo (ICME), pp. 1\u20136. IEEE (2022). https:\/\/doi.org\/10.1109\/icme52920.2022.9859581","DOI":"10.1109\/icme52920.2022.9859581"},{"key":"1426_CR14","doi-asserted-by":"publisher","unstructured":"Karim, N., Mithun, N.C., Rajvanshi, A., et al.: C-sfda: a curriculum learning aided self-training framework for efficient source free domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 24120\u201324131 (2023). https:\/\/doi.org\/10.1109\/cvpr52729.2023.02310","DOI":"10.1109\/cvpr52729.2023.02310"},{"key":"1426_CR15","doi-asserted-by":"publisher","unstructured":"Wang, Z., Luo, Y., Chen, Z., et al.: Cal-sfda: source-free domain-adaptive semantic segmentation with differentiable expected calibration error. In: Proceedings of the 31st ACM International Conference on Multimedia, pp. 1167\u20131178 (2023). https:\/\/doi.org\/10.1145\/3581783.3611808","DOI":"10.1145\/3581783.3611808"},{"key":"1426_CR16","doi-asserted-by":"publisher","first-page":"3635","DOI":"10.1109\/cvpr42600.2020.00966","volume":"34","author":"J Huang","year":"2021","unstructured":"Huang, J., Guan, D., Xiao, A., et al.: Model adaptation: historical contrastive learning for unsupervised domain adaptation without source data. Adv. Neural Inf. Process. Syst. 34, 3635\u20133649 (2021). https:\/\/doi.org\/10.1109\/cvpr42600.2020.00966","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"1426_CR17","doi-asserted-by":"crossref","unstructured":"Cao, Y., Zhang, H., Lu, X., et al.: (2024) Towards source-free domain adaptive semantic segmentation via importance-aware and prototype-contrast learning. IEEE Trans. Intell. Veh","DOI":"10.1109\/TIV.2024.3383157"},{"key":"1426_CR18","doi-asserted-by":"publisher","unstructured":"Kirillov, A., Mintun, E., Ravi, N., et al.: Segment anything. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4015\u20134026 (2023). https:\/\/doi.org\/10.1109\/iccv51070.2023.00371","DOI":"10.1109\/iccv51070.2023.00371"},{"key":"1426_CR19","doi-asserted-by":"publisher","unstructured":"Csurka, G.: A comprehensive survey on domain adaptation for visual applications. In: Domain adaptation in computer vision applications, pp. 1\u201335 (2017). https:\/\/doi.org\/10.1007\/978-3-319-58347-1_1","DOI":"10.1007\/978-3-319-58347-1_1"},{"key":"1426_CR20","doi-asserted-by":"publisher","first-page":"135","DOI":"10.1016\/j.neucom.2018.05.083","volume":"312","author":"M Wang","year":"2018","unstructured":"Wang, M., Deng, W.: Deep visual domain adaptation: a survey. Neurocomputing 312, 135\u2013153 (2018)","journal-title":"Neurocomputing"},{"key":"1426_CR21","doi-asserted-by":"publisher","DOI":"10.1109\/tcsvt.2024.3370570","author":"Q Ren","year":"2024","unstructured":"Ren, Q., Lu, S., Mao, Q., et al.: Exploring prototype-anchor contrast for semantic segmentation. IEEE Trans. Circuits Syst. Video Technol. (2024). https:\/\/doi.org\/10.1109\/tcsvt.2024.3370570","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1426_CR22","unstructured":"Long, M., Cao, Y., Wang, J., et al.: Learning transferable features with deep adaptation networks. In: International Conference on Machine Learning, PMLR, pp. 97\u2013105 (2015)"},{"key":"1426_CR23","unstructured":"Long, M., Cao, Z., Wang, J., et al.: Conditional adversarial domain adaptation. In: Advances in Neural Information Processing Systems, vol. 31. Curran Associates, Inc. (2018)"},{"key":"1426_CR24","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. In: International Conference on Learning Representations (2021)"},{"key":"1426_CR25","unstructured":"Hoffman, J., Wang, D., Yu, F., et al.: Fcns in the wild: pixel-level adversarial and constraint-based adaptation (2016). arXiv preprint arXiv:1612.02649"},{"key":"1426_CR26","doi-asserted-by":"publisher","unstructured":"Tsai, Y.H., Hung, W.C., Schulter, S., et al.: Learning to adapt structured output space for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7472\u20137481 (2018). https:\/\/doi.org\/10.1109\/cvpr.2018.00780","DOI":"10.1109\/cvpr.2018.00780"},{"key":"1426_CR27","doi-asserted-by":"publisher","unstructured":"Hoyer, L., Dai, D., Van Gool, L.: Daformer: improving network architectures and training strategies for domain-adaptive semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9924\u20139935 (2022). https:\/\/doi.org\/10.1109\/cvpr52688.2022.00969","DOI":"10.1109\/cvpr52688.2022.00969"},{"key":"1426_CR28","doi-asserted-by":"publisher","unstructured":"Tzeng, E., Hoffman, J., Saenko, K., et al.: Adversarial discriminative domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7167\u20137176 (2017). https:\/\/doi.org\/10.1109\/cvpr.2017.316","DOI":"10.1109\/cvpr.2017.316"},{"key":"1426_CR29","unstructured":"Hoffman, J., Tzeng, E., Park, T., et al.: Cycada: cycle-consistent adversarial domain adaptation. In: International Conference on Machine Learning, PMLR, pp. 1989\u20131998 (2018)"},{"key":"1426_CR30","doi-asserted-by":"publisher","unstructured":"Vu, T.H., Jain, H., Bucher, M., et al.: Advent: adversarial entropy minimization for domain adaptation in semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2517\u20132526 (2019). https:\/\/doi.org\/10.1109\/cvpr.2019.00262","DOI":"10.1109\/cvpr.2019.00262"},{"key":"1426_CR31","doi-asserted-by":"publisher","unstructured":"Zou, Y., Yu, Z., Kumar, B., et al.: Unsupervised domain adaptation for semantic segmentation via class-balanced self-training. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 289\u2013305 (2018). https:\/\/doi.org\/10.1007\/978-3-030-01219-9_18","DOI":"10.1007\/978-3-030-01219-9_18"},{"key":"1426_CR32","doi-asserted-by":"publisher","unstructured":"Mei, K., Zhu, C., Zou, J., et al.: Instance adaptive self-training for unsupervised domain adaptation. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, 23\u201328 August 2020, Proceedings, Part XXVI 16, pp. 415\u2013430. Springer (2020). https:\/\/doi.org\/10.1007\/978-3-030-58574-7_25","DOI":"10.1007\/978-3-030-58574-7_25"},{"issue":"1","key":"1426_CR33","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1109\/MSP.2017.2765202","volume":"35","author":"A Creswell","year":"2018","unstructured":"Creswell, A., White, T., Dumoulin, V., et al.: Generative adversarial networks: an overview. IEEE Signal Process. Mag. 35(1), 53\u201365 (2018)","journal-title":"IEEE Signal Process. Mag."},{"key":"1426_CR34","doi-asserted-by":"publisher","unstructured":"Luo, Y., Zheng, L., Guan, T., et\u00a0al.: Taking a closer look at domain shift: category-level adversaries for semantics consistent domain adaptation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2507\u20132516 (2019). https:\/\/doi.org\/10.1109\/cvpr.2019.00261","DOI":"10.1109\/cvpr.2019.00261"},{"key":"1426_CR35","unstructured":"Zhang,. Q., Zhang, J., Liu, W., et al.: Category anchor-guided unsupervised domain adaptation for semantic segmentation. In: Proceedings of the 33rd International Conference on Neural Information Processing Systems, vol. 32. Curran Associates Inc., Red Hook, NY (2019)"},{"key":"1426_CR36","doi-asserted-by":"publisher","unstructured":"Tranheden, W., Olsson, V., Pinto, J., et\u00a0al.: Dacs: domain adaptation via cross-domain mixed sampling. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1379\u20131389 (2021). https:\/\/doi.org\/10.1109\/wacv48630.2021.00142","DOI":"10.1109\/wacv48630.2021.00142"},{"key":"1426_CR37","unstructured":"Liang, J., Hu, D., Feng, J.: Do we really need to access the source data? Source hypothesis transfer for unsupervised domain adaptation. In: International Conference on Machine Learning, PMLR, pp. 6028\u20136039 (2020)"},{"key":"1426_CR38","doi-asserted-by":"publisher","unstructured":"Wu, Z., Xiong, Y., Yu, S.X., et al.: Unsupervised feature learning via non-parametric instance discrimination. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3733\u20133742 (2018). https:\/\/doi.org\/10.1109\/cvpr.2018.00393","DOI":"10.1109\/cvpr.2018.00393"},{"key":"1426_CR39","unstructured":"Oord, A., Li, Y., Vinyals, O.: Representation learning with contrastive predictive coding (2018). arXiv preprint arXiv:1807.03748"},{"key":"1426_CR40","unstructured":"Chen, T., Kornblith, S., Norouzi, M., et al.: A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning, PMLR, pp. 1597\u20131607 (2020)"},{"key":"1426_CR41","doi-asserted-by":"publisher","unstructured":"Ye, M., Zhang, X., Yuen, P.C., et al.: Unsupervised embedding learning via invariant and spreading instance feature. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6210\u20136219 (2019). https:\/\/doi.org\/10.1109\/cvpr.2019.00637","DOI":"10.1109\/cvpr.2019.00637"},{"key":"1426_CR42","doi-asserted-by":"publisher","unstructured":"He, K., Fan, H., Wu, Y., et al.: Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9729\u20139738 (2020). https:\/\/doi.org\/10.1109\/cvpr42600.2020.00975","DOI":"10.1109\/cvpr42600.2020.00975"},{"key":"1426_CR43","unstructured":"Tarvainen, A., Valpola, H.: Mean teachers are better role models: weight-averaged consistency targets improve semi-supervised deep learning results. In: Advances in Neural Information Processing Systems, vol. 30 (2017)"},{"key":"1426_CR44","doi-asserted-by":"publisher","unstructured":"Wang, Y., Ma, X., Chen, Z., et al.: Symmetric cross entropy for robust learning with noisy labels. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 322\u2013330 (2019). https:\/\/doi.org\/10.1109\/iccv.2019.00041","DOI":"10.1109\/iccv.2019.00041"},{"key":"1426_CR45","doi-asserted-by":"publisher","unstructured":"Richter,. S.R., Vineet, V., Roth, S., et al.: Playing for data: ground truth from computer games. In: Computer Vision\u2014ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11\u201314 October 2016, Proceedings, Part II 14, pp. 102\u2013118. Springer (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_7","DOI":"10.1007\/978-3-319-46475-6_7"},{"key":"1426_CR46","doi-asserted-by":"publisher","unstructured":"Ros, G., Sellart, L., Materzynska, J., et al.: The synthia dataset: a large collection of synthetic images for semantic segmentation of urban scenes. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3234\u20133243 (2016). https:\/\/doi.org\/10.1109\/cvpr.2016.352","DOI":"10.1109\/cvpr.2016.352"},{"key":"1426_CR47","doi-asserted-by":"publisher","unstructured":"Cordts, M., Omran, M., Ramos, S., et al.: The cityscapes dataset for semantic urban scene understanding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3213\u20133223 (2016). https:\/\/doi.org\/10.1109\/cvpr.2016.350","DOI":"10.1109\/cvpr.2016.350"},{"issue":"4","key":"1426_CR48","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1109\/tpami.2017.2699184","volume":"40","author":"LC Chen","year":"2017","unstructured":"Chen, L.C., Papandreou, G., Kokkinos, I., et al.: Deeplab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 40(4), 834\u2013848 (2017). https:\/\/doi.org\/10.1109\/tpami.2017.2699184","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1426_CR49","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., et al.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016). https:\/\/doi.org\/10.1109\/cvpr.2016.90","DOI":"10.1109\/cvpr.2016.90"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01426-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01426-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01426-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,22]],"date-time":"2024-08-22T08:33:30Z","timestamp":1724315610000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01426-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,26]]},"references-count":49,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["1426"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01426-5","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"type":"print","value":"0942-4962"},{"type":"electronic","value":"1432-1882"}],"subject":[],"published":{"date-parts":[[2024,7,26]]},"assertion":[{"value":"1 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 July 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 July 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"217"}}