{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T07:04:31Z","timestamp":1766127871355,"version":"3.48.0"},"reference-count":45,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T00:00:00Z","timestamp":1759968000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T00:00:00Z","timestamp":1759968000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"University Synergy Innovation Program of Anhui Province, China","award":["GXXT-2021-030, GXXT-2022-031"],"award-info":[{"award-number":["GXXT-2021-030, GXXT-2022-031"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62076005, U20A20398"],"award-info":[{"award-number":["62076005, U20A20398"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Dreams Foundation of Jianghuai Advance Technology Center","award":["2023-ZM01Z015"],"award-info":[{"award-number":["2023-ZM01Z015"]}]},{"DOI":"10.13039\/501100003995","name":"Natural Science Foundation of Anhui Province","doi-asserted-by":"publisher","award":["2008085MF191"],"award-info":[{"award-number":["2008085MF191"]}],"id":[{"id":"10.13039\/501100003995","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s00530-025-01992-2","type":"journal-article","created":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T09:34:18Z","timestamp":1760002458000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Dual-level semantic alignment for video moment retrieval and highlight detection"],"prefix":"10.1007","volume":"31","author":[{"given":"Haifeng","family":"Zhao","sequence":"first","affiliation":[]},{"given":"Qinghua","family":"Ling","sequence":"additional","affiliation":[]},{"given":"Wenhai","family":"Qin","sequence":"additional","affiliation":[]},{"given":"Leilei","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Dengdi","family":"Sun","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,9]]},"reference":[{"issue":"11","key":"1992_CR1","doi-asserted-by":"publisher","first-page":"1838","DOI":"10.1109\/JPROC.2021.3117472","volume":"109","author":"E Apostolidis","year":"2021","unstructured":"Apostolidis, E., Adamantidou, E., Metsai, A.I., et al.: Video summarization using deep neural networks: a survey. Proc. IEEE 109(11), 1838\u20131863 (2021)","journal-title":"Proc. IEEE"},{"key":"1992_CR2","doi-asserted-by":"publisher","unstructured":"Badamdorj, T., Rochan, M., Wang, Y., et\u00a0al.: Joint visual and audio learning for video highlight detection. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV). IEEE, Montreal, Canada, pp. 8107\u20138117. https:\/\/doi.org\/10.1109\/ICCV48922.2021.00802(2021)","DOI":"10.1109\/ICCV48922.2021.00802"},{"key":"1992_CR3","doi-asserted-by":"crossref","unstructured":"Badamdorj, T., Rochan, M., Wang, Y., et\u00a0al.: Contrastive learning for unsupervised video highlight detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. IEEE, New Orleans, USA, pp. 14042\u201314052 (2022)","DOI":"10.1109\/CVPR52688.2022.01365"},{"key":"1992_CR4","doi-asserted-by":"publisher","unstructured":"Cao, M., Chen, L., Shou, M. Z., et\u00a0al.: On pursuit of designing multi-modal transformer for video grounding. In: EI Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing:9810\u20139823. https:\/\/doi.org\/10.18653\/v1\/2021.emnlp-main.773. https:\/\/aclanthology.org\/2021.emnlp-main.773 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.773"},{"key":"1992_CR5","doi-asserted-by":"crossref","unstructured":"Chen, S., Jiang, Y.G.: Hierarchical visual-textual graph for temporal activity localization via language. In: Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XX 16, pp. 601\u2013618. Springer, Cham (2020)","DOI":"10.1007\/978-3-030-58565-5_36"},{"key":"1992_CR6","unstructured":"Chen, T., Kornblith, S., Norouzi, M., et al.: A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning, pp. 1597\u20131607. PMLR. JMLR.org, Vienna, Austria (2020)"},{"key":"1992_CR7","unstructured":"Escorcia, V., Soldan, M., Sivic, J., et\u00a0al.: Temporal localization of moments in video collections with natural language. arXiv: abs\/1907.12763 (2019)"},{"key":"1992_CR8","doi-asserted-by":"crossref","unstructured":"Gao, J., Xu, C.: Fast video moment retrieval. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. IEEE, Montreal, Canada, pp. 1523\u20131532 (2021)","DOI":"10.1109\/ICCV48922.2021.00155"},{"key":"1992_CR9","doi-asserted-by":"crossref","unstructured":"Gao, J., Sun, C., Yang, Z., et\u00a0al.: Tall: Temporal activity localization via language query. In: Proceedings of the IEEE International Conference on Computer Vision. IEEE, Venice, Italy, pp. 5267\u20135275 (2017)","DOI":"10.1109\/ICCV.2017.563"},{"key":"1992_CR10","doi-asserted-by":"publisher","unstructured":"Hendricks, L. A., Wang, O., Shechtman, E., et\u00a0al.: Localizing moments in video with natural language. In: 2017 IEEE International Conference on Computer Vision (ICCV). IEEE, Venice, Italy, pp. 5804\u20135813. https:\/\/doi.org\/10.1109\/ICCV.2017.618 (2017)","DOI":"10.1109\/ICCV.2017.618"},{"key":"1992_CR11","doi-asserted-by":"crossref","unstructured":"Jang, J., Park, J., Kim, J., et\u00a0al.: Knowing where to focus: Event-aware transformer for video grounding. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. IEEE, Paris, France, pp. 13846\u201313856 (2023)","DOI":"10.1109\/ICCV51070.2023.01273"},{"key":"1992_CR12","doi-asserted-by":"crossref","unstructured":"Lei, J., Yu, L., Berg, T.L., et al.: Tvr: A large-scale dataset for video-subtitle moment retrieval. In: Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXI 16, pp. 447\u2013463. Springer International Publishing, Cham (2020)","DOI":"10.1007\/978-3-030-58589-1_27"},{"key":"1992_CR13","first-page":"11846","volume":"34","author":"J Lei","year":"2021","unstructured":"Lei, J., Berg, T.L., Bansal, M.: Detecting moments and highlights in videos via natural language queries. Adv. Neural. Inf. Process. Syst. 34, 11846\u201311858 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1992_CR14","doi-asserted-by":"publisher","unstructured":"Li, J., Zhang, F., Lin, S., et\u00a0al.: Mim: Lightweight multi-modal interaction model for joint video moment retrieval and highlight detection. In: 2023 IEEE International Conference on Multimedia and Expo (ICME), pp. 1961\u20131966. https:\/\/doi.org\/10.1109\/ICME55011.2023.00336 (2023a)","DOI":"10.1109\/ICME55011.2023.00336"},{"key":"1992_CR15","doi-asserted-by":"crossref","unstructured":"Li, K., Guo, D., Wang, M.: Proposal-free video grounding with contextual pyramid network. In: Proceedings of the AAAI Conference on Artificial Intelligence. AAAI, Vancouver, Canada, pp. 1902\u20131910 (2021)","DOI":"10.1609\/aaai.v35i3.16285"},{"key":"1992_CR16","unstructured":"Li, P., Xie, C. W., Xie, H., et\u00a0al.: Momentdiff: Generative video moment retrieval from random to real. In: Oh, A., Naumann, T., Globerson, A., et\u00a0al. (eds) Advances in Neural Information Processing Systems, vol\u00a036. Curran Associates, Inc., New Orleans, LA, USA, pp. 65948\u201365966. https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2023\/file\/d01bda31bbcd780774ff15b534e03c40-Paper-Conference.pdf (2023b)"},{"key":"1992_CR17","doi-asserted-by":"crossref","unstructured":"Lin, K. Q., Zhang, P., Chen, J., et\u00a0al. Univtg: Towards unified video-language temporal grounding. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. IEEE, Paris, France, pp. 2794\u20132804 (2023)","DOI":"10.1109\/ICCV51070.2023.00262"},{"key":"1992_CR18","doi-asserted-by":"crossref","unstructured":"Liu, W., Mei, T., Zhang, Y., et\u00a0al. Multi-task deep visual-semantic embedding for video thumbnail selection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE, Boston, USA, pp. 3707\u20133715 (2015)","DOI":"10.1109\/CVPR.2015.7298994"},{"key":"1992_CR19","doi-asserted-by":"crossref","unstructured":"Liu, Y., Li, S., Wu, Y., et\u00a0al.: Umt: Unified multi-modal transformers for joint video moment retrieval and highlight detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. IEEE, New Orleans, USA, pp. 3042\u20133051 (2022)","DOI":"10.1109\/CVPR52688.2022.00305"},{"key":"1992_CR20","unstructured":"Ma, Y., Yang, T., Shan, Y., et\u00a0al.: Simvtp: Simple video text pre-training with masked autoencoders. arXiv preprint arXiv:2212.03490 abs\/2212.03490 (2022)"},{"key":"1992_CR21","doi-asserted-by":"crossref","unstructured":"Mahasseni, B., Lam, M., Todorovic, S.: Unsupervised video summarization with adversarial lstm networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE, Hawaii, USA, pp. 202\u2013211 (2017)","DOI":"10.1109\/CVPR.2017.318"},{"key":"1992_CR22","doi-asserted-by":"crossref","unstructured":"Moon, W., Hyun, S., Park, S., et\u00a0al.: Query-dependent video representation for moment retrieval and highlight detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. IEEE, Vancouver, Canada, pp. 23023\u201323033 (2023)","DOI":"10.1109\/CVPR52729.2023.02205"},{"key":"1992_CR23","doi-asserted-by":"crossref","unstructured":"Mun, J., Cho, M., Han, B.: Local-global video-text interactions for temporal grounding. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. IEEE, Seattle, WA, USA, pp. 10810\u201310819 (2020)","DOI":"10.1109\/CVPR42600.2020.01082"},{"key":"1992_CR24","doi-asserted-by":"publisher","unstructured":"Otsuka, I., Nakane, K., Divakaran, A., et\u00a0al.: A highlight scene detection and video summarization system using \u2019audio\u2019 feature for a personal video recorder. In: 2005 Digest of Technical Papers. International Conference on Consumer Electronics, 2005. ICCE. IEEE, Las Vegas, USA, pp. 223\u2013224. https:\/\/doi.org\/10.1109\/ICCE.2005.1429798 (2005)","DOI":"10.1109\/ICCE.2005.1429798"},{"key":"1992_CR25","unstructured":"Rodriguez, C., Marrese-Taylor, E., Saleh, F. S., et\u00a0al.: Proposal-free temporal moment localization of a natural-language query in video using guided attention. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision. IEEE, Snowmass Village, Aspen, CO, USA, pp. 2464\u20132473 (2020)"},{"key":"1992_CR26","doi-asserted-by":"crossref","unstructured":"Soldan M, Xu M, Qu S, et\u00a0al (2021) Vlg-net: Video-language graph matching network for video grounding. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. IEEE, Montreal, Canada, pp. 3224\u20133234","DOI":"10.1109\/ICCVW54120.2021.00361"},{"key":"1992_CR27","doi-asserted-by":"crossref","unstructured":"Song, Y., Redi, M., Vallmitjana, J., et\u00a0al.: To click or not to click: Automatic selection of beautiful thumbnails from videos. In: Proceedings of the 25th ACM International on Conference on Information and Knowledge Management. Association for Computing Machinery, New York, NY, USA, pp. 659\u2013668 (2016)","DOI":"10.1145\/2983323.2983349"},{"key":"1992_CR28","doi-asserted-by":"publisher","unstructured":"Tong, X., Liu, Q., Zhang, Y., et\u00a0al.: Highlight ranking for sports video browsing. In: Proceedings of the 13th Annual ACM International Conference on Multimedia. Association for Computing Machinery, New York, NY, USA, MULTIMEDIA \u201905, pp. 519\u2013522. https:\/\/doi.org\/10.1145\/1101149.1101266. https:\/\/doi.org\/10.1145\/1101149.1101266 (2005)","DOI":"10.1145\/1101149.1101266"},{"key":"1992_CR29","doi-asserted-by":"crossref","unstructured":"Wang, L., Liu, D., Puri, R., et al.: Learning trailer moments in full-length movies with co-contrastive attention. In: Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XVIII 16, pp. 300\u2013316. Springer, Heidelberg (2020)","DOI":"10.1007\/978-3-030-58523-5_18"},{"issue":"9","key":"1992_CR30","doi-asserted-by":"publisher","first-page":"8896","DOI":"10.1109\/TCSVT.2024.3389024","volume":"34","author":"R Wang","year":"2024","unstructured":"Wang, R., Feng, J., Zhang, F., et al.: Modality-aware heterogeneous graph for joint video moment retrieval and highlight detection. IEEE Trans. Circuits Syst. Video Technol. 34(9), 8896\u20138911 (2024). https:\/\/doi.org\/10.1109\/TCSVT.2024.3389024","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1992_CR31","doi-asserted-by":"crossref","unstructured":"Wei, F., Wang, B., Ge, T., et\u00a0al.: Learning pixel-level distinctions for video highlight detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. IEEE, New Orleans, USA, pp. 3073\u20133082 (2022)","DOI":"10.1109\/CVPR52688.2022.00308"},{"key":"1992_CR32","doi-asserted-by":"crossref","unstructured":"Xie, E., Ding, J., Wang, W., et\u00a0al.: Detco: Unsupervised contrastive learning for object detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. IEEE, Montreal, Canada, pp. 8392\u20138401 (2021)","DOI":"10.1109\/ICCV48922.2021.00828"},{"key":"1992_CR33","doi-asserted-by":"publisher","unstructured":"Xie, J., Xiang, J., Chen, J., et\u00a0al.: C2 am: Contrastive learning of class-agnostic activation map for weakly supervised object localization and semantic segmentation. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, New Orleans, USA, pp. 979\u2013988. https:\/\/doi.org\/10.1109\/CVPR52688.2022.00106 (2022)","DOI":"10.1109\/CVPR52688.2022.00106"},{"key":"1992_CR34","doi-asserted-by":"crossref","unstructured":"Xiong, B., Kalantidis, Y., Ghadiyaram, D., et\u00a0al.: Less is more: Learning highlight detection from video duration. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. IEEE, Long Beach, USA, pp. 1258\u20131267 (2019)","DOI":"10.1109\/CVPR.2019.00135"},{"key":"1992_CR35","doi-asserted-by":"crossref","unstructured":"Xu, M., Wang, H., Ni, B., et\u00a0al.: Cross-category video highlight detection via set-based learning. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. IEEE, Montreal, Canada, pp. 7970\u20137979 (2021)","DOI":"10.1109\/ICCV48922.2021.00787"},{"key":"1992_CR36","doi-asserted-by":"crossref","unstructured":"Xu, Y., Sun, Y., Li, Y., et\u00a0al.: MH-DETR: Video moment and highlight detection with cross-modal transformer. In: 2024 International Joint Conference on Neural Networks (IJCNN). IEEE, Yokohama, Japan, pp. 1\u20138. https:\/\/api.semanticscholar.org\/CorpusID:258427080 (2024a)","DOI":"10.1109\/IJCNN60899.2024.10650814"},{"key":"1992_CR37","doi-asserted-by":"publisher","unstructured":"Xu, Y., Sun, Y., Zhai, B., et\u00a0al.: Multi-modal fusion and query refinement network for video moment retrieval and highlight detection. In: 2024 IEEE International Conference on Multimedia and Expo (ICME), pp. 1\u20136. https:\/\/doi.org\/10.1109\/ICME57554.2024.10687844 (2024b)","DOI":"10.1109\/ICME57554.2024.10687844"},{"key":"1992_CR38","doi-asserted-by":"crossref","unstructured":"Yan, S., Xiong, X., Nagrani, A., et\u00a0al.: UnLoc: A unified framework for video localization tasks. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. IEEE, Paris, France, pp. 13623\u201313633 (2023)","DOI":"10.1109\/ICCV51070.2023.01253"},{"key":"1992_CR39","doi-asserted-by":"crossref","unstructured":"Ye, Q., Shen, X., Gao, Y., et\u00a0al.: Temporal cue guided video highlight detection with low-rank audio-visual fusion. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. IEEE, Montreal, Canada, pp. 7950\u20137959 (2021)","DOI":"10.1109\/ICCV48922.2021.00785"},{"key":"1992_CR40","unstructured":"Yu, S., Cho, J., Yadav, P., et\u00a0al.: Self-chained image-language model for video localization and question answering. In: Proceedings of the 37th International Conference on Neural Information Processing Systems. Curran Associates Inc., Red Hook, NY, USA, NIPS \u201923 (2024)"},{"key":"1992_CR41","doi-asserted-by":"crossref","unstructured":"Zeng, R., Xu, H., Huang, W., et\u00a0al.: Dense regression network for video grounding. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. IEEE, Seattle, USA, pp. 10287\u201310296 (2020)","DOI":"10.1109\/CVPR42600.2020.01030"},{"key":"1992_CR42","doi-asserted-by":"crossref","unstructured":"Zhang, H., Sun, A., Jing, W., et\u00a0al.: Span-based localizing network for natural language video localization. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, pp. 6543\u20136554. https:\/\/www.aclweb.org\/anthology\/2020.acl-main.585 (2020)","DOI":"10.18653\/v1\/2020.acl-main.585"},{"key":"1992_CR43","doi-asserted-by":"crossref","unstructured":"Zhang, K., Chao, W.L., Sha, F., et al.: Video summarization with long short-term memory. In: Computer Vision-ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part VII 14, pp. 766\u2013782. Springer International Publishing, Cham (2016)","DOI":"10.1007\/978-3-319-46478-7_47"},{"issue":"12","key":"1992_CR44","doi-asserted-by":"publisher","first-page":"9073","DOI":"10.1109\/TPAMI.2021.3120745","volume":"44","author":"S Zhang","year":"2022","unstructured":"Zhang, S., Peng, H., Fu, J., et al.: Multi-scale 2D temporal adjacency networks for moment localization with natural language. IEEE Trans. Pattern Anal. Mach. Intell. 44(12), 9073\u20139087 (2022). https:\/\/doi.org\/10.1109\/TPAMI.2021.3120745","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"11","key":"1992_CR45","doi-asserted-by":"publisher","first-page":"11271","DOI":"10.1109\/TCSVT.2024.3409897","volume":"34","author":"S Zhou","year":"2024","unstructured":"Zhou, S., Zhang, F., Wang, R., et al.: Subtask prior-driven optimized mechanism on joint video moment retrieval and highlight detection. IEEE Trans. Circuits Syst. Video Technol. 34(11), 11271\u201311285 (2024). https:\/\/doi.org\/10.1109\/TCSVT.2024.3409897","journal-title":"IEEE Trans. Circuits Syst. Video Technol."}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01992-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-01992-2","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01992-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T07:00:15Z","timestamp":1766127615000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-01992-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,9]]},"references-count":45,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["1992"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-01992-2","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"type":"print","value":"0942-4962"},{"type":"electronic","value":"1432-1882"}],"subject":[],"published":{"date-parts":[[2025,10,9]]},"assertion":[{"value":"19 March 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 August 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 October 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The authors declare no Conflict of interest.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"413"}}