{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,3]],"date-time":"2026-03-03T16:00:25Z","timestamp":1772553625584,"version":"3.50.1"},"reference-count":54,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2024,7,30]],"date-time":"2024-07-30T00:00:00Z","timestamp":1722297600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,7,30]],"date-time":"2024-07-30T00:00:00Z","timestamp":1722297600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1007\/s00530-024-01421-w","type":"journal-article","created":{"date-parts":[[2024,7,30]],"date-time":"2024-07-30T18:32:12Z","timestamp":1722364332000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":17,"title":["Text-centered cross-sample fusion network for multimodal sentiment analysis"],"prefix":"10.1007","volume":"30","author":[{"given":"Qionghao","family":"Huang","sequence":"first","affiliation":[]},{"given":"Jili","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Changqin","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Xiaodi","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Yi","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,30]]},"reference":[{"key":"1421_CR1","doi-asserted-by":"crossref","unstructured":"Shenoy, A., Sardana, A.: Multilogue-net: a context aware rnn for multi-modal emotion detection and sentiment analysis in conversation (2020). arXiv preprint arXiv:2002.08267","DOI":"10.18653\/v1\/2020.challengehml-1.3"},{"issue":"5","key":"1421_CR2","doi-asserted-by":"publisher","first-page":"451","DOI":"10.1007\/s00530-017-0547-8","volume":"25","author":"J Fu","year":"2019","unstructured":"Fu, J., Mao, Q., Tu, J., Zhan, Y.: Multimodal shared features learning for emotion recognition by enhanced sparse local discriminative canonical correlation analysis. Multimed. Syst. 25(5), 451\u2013461 (2019)","journal-title":"Multimed. Syst."},{"key":"1421_CR3","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1016\/j.ins.2021.08.043","volume":"580","author":"Q Huang","year":"2021","unstructured":"Huang, Q., Huang, C., Wang, X., Jiang, F.: Facial expression recognition with grid-wise attention and visual transformer. Inf. Sci. 580, 35\u201354 (2021)","journal-title":"Inf. Sci."},{"issue":"1","key":"1421_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s00530-023-01208-5","volume":"30","author":"Y Luo","year":"2024","unstructured":"Luo, Y., Wu, R., Liu, J., Tang, X.: Balanced sentimental information via multimodal interaction model. Multimed. Syst. 30(1), 1\u20139 (2024)","journal-title":"Multimed. Syst."},{"issue":"13s","key":"1421_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3586075","volume":"55","author":"R Das","year":"2023","unstructured":"Das, R., Singh, T.D.: Multimodal sentiment analysis: a survey of methods, trends, and challenges. ACM Comput. Surv. 55(13s), 1\u201338 (2023)","journal-title":"ACM Comput. Surv."},{"key":"1421_CR6","doi-asserted-by":"publisher","first-page":"102040","DOI":"10.1016\/j.inffus.2023.102040","volume":"102","author":"T Shaik","year":"2023","unstructured":"Shaik, T., Tao, X., Li, L., Xie, H., Vel\u00e1squez, J.D.: A survey of multimodal information fusion for smart healthcare: mapping the journey from data to wisdom. Inf Fusion 102, 102040 (2023)","journal-title":"Inf Fusion"},{"key":"1421_CR7","doi-asserted-by":"crossref","unstructured":"Zadeh, A., Chen, M., Poria, S., Cambria, E., Morency, L.-P.: Tensor fusion network for multimodal sentiment analysis. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp. 1103\u20131114 (2017)","DOI":"10.18653\/v1\/D17-1115"},{"key":"1421_CR8","doi-asserted-by":"crossref","unstructured":"Zadeh, A., Liang, P.P., Mazumder, N., Poria, S., Cambria, E., Morency, L.-P.: Memory fusion network for multi-view sequential learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32, pp. 5634\u20135641 (2018)","DOI":"10.1609\/aaai.v32i1.12021"},{"key":"1421_CR9","doi-asserted-by":"crossref","unstructured":"Tsai, Y.-H.H., Bai, S., Liang, P.P., Kolter, J.Z., Morency, L.-P., Salakhutdinov, R.: Multimodal transformer for unaligned multimodal language sequences. In: Proceedings of the Conference. Association for Computational Linguistics. Meeting, vol. 2019, p. 6558. NIH Public Access (2019)","DOI":"10.18653\/v1\/P19-1656"},{"key":"1421_CR10","first-page":"6616","volume":"33","author":"Z Gan","year":"2020","unstructured":"Gan, Z., Chen, Y.-C., Li, L., Zhu, C., Cheng, Y., Liu, J.: Large-scale adversarial training for vision-and-language representation learning. Adv. Neural Inf .Process. Syst. 33, 6616\u20136628 (2020)","journal-title":"Adv. Neural Inf .Process. Syst."},{"key":"1421_CR11","doi-asserted-by":"crossref","unstructured":"Hazarika, D., Zimmermann, R., Poria, S.: Misa: modality-invariant and-specific representations for multimodal sentiment analysis. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 1122\u20131131 (2020)","DOI":"10.1145\/3394171.3413678"},{"key":"1421_CR12","doi-asserted-by":"crossref","unstructured":"Yu, W., Xu, H., Yuan, Z., Wu, J.: Learning modality-specific representations with self-supervised multi-task learning for multimodal sentiment analysis. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 10790\u201310797 (2021)","DOI":"10.1609\/aaai.v35i12.17289"},{"key":"1421_CR13","doi-asserted-by":"publisher","first-page":"2689","DOI":"10.1109\/TASLP.2022.3192728","volume":"30","author":"Q Chen","year":"2022","unstructured":"Chen, Q., Huang, G., Wang, Y.: The weighted cross-modal attention mechanism with sentiment prediction auxiliary task for multimodal sentiment analysis. IEEE\/ACM Trans. Audio Speech Lang. Process. 30, 2689\u20132695 (2022)","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"1421_CR14","doi-asserted-by":"publisher","first-page":"4909","DOI":"10.1109\/TMM.2022.3183830","volume":"25","author":"D Wang","year":"2022","unstructured":"Wang, D., Liu, S., Wang, Q., Tian, Y., He, L., Gao, X.: Cross-modal enhancement network for multimodal sentiment analysis. IEEE Trans. Multimed. 25, 4909\u20134921 (2022)","journal-title":"IEEE Trans. Multimed."},{"key":"1421_CR15","doi-asserted-by":"publisher","first-page":"424","DOI":"10.1016\/j.inffus.2022.09.025","volume":"91","author":"A Gandhi","year":"2023","unstructured":"Gandhi, A., Adhvaryu, K., Poria, S., Cambria, E., Hussain, A.: Multimodal sentiment analysis: a systematic review of history, datasets, multimodal fusion methods, applications, challenges and future directions. Inf. Fusion 91, 424\u2013444 (2023)","journal-title":"Inf. Fusion"},{"key":"1421_CR16","doi-asserted-by":"crossref","unstructured":"Sun, Z., Sarma, P., Sethares, W., Liang, Y.: Learning relationships between text, audio, and video via deep canonical correlation for multimodal language analysis. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 8992\u20138999 (2020)","DOI":"10.1609\/aaai.v34i05.6431"},{"key":"1421_CR17","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.109259","volume":"136","author":"D Wang","year":"2023","unstructured":"Wang, D., Guo, X., Tian, Y., Liu, J., He, L., Luo, X.: Tetfn: a text enhanced transformer fusion network for multimodal sentiment analysis. Pattern Recognit. 136, 109259 (2023)","journal-title":"Pattern Recognit."},{"key":"1421_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.110502","volume":"269","author":"C Huang","year":"2023","unstructured":"Huang, C., Zhang, J., Wu, X., Wang, Y., Li, M., Huang, X.: Tefna: text-centered fusion network with crossmodal attention for multimodal sentiment analysis. Knowl.-Based Syst. 269, 110502 (2023)","journal-title":"Knowl.-Based Syst."},{"key":"1421_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2023.126836","volume":"560","author":"Y Luo","year":"2023","unstructured":"Luo, Y., Wu, R., Liu, J., Tang, X.: A text guided multi-task learning network for multimodal sentiment analysis. Neurocomputing 560, 126836 (2023)","journal-title":"Neurocomputing"},{"key":"1421_CR20","doi-asserted-by":"crossref","unstructured":"Wei, X., Zhang, T., Li, Y., Zhang, Y., Wu, F.: Multi-modality cross attention network for image and sentence matching. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10941\u201310950 (2020)","DOI":"10.1109\/CVPR42600.2020.01095"},{"issue":"12","key":"1421_CR21","doi-asserted-by":"publisher","first-page":"5412","DOI":"10.1109\/TNNLS.2020.2967597","volume":"31","author":"X Xu","year":"2020","unstructured":"Xu, X., Wang, T., Yang, Y., Zuo, L., Shen, F., Shen, H.T.: Cross-modal attention with semantic consistence for image-text matching. IEEE Trans. Neural Netw. Learn. Syst. 31(12), 5412\u20135425 (2020)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"1421_CR22","first-page":"3354304","volume":"62","author":"Z Chen","year":"2024","unstructured":"Chen, Z., Zhang, C., Zhang, B., He, Y.: Triplet contrastive learning framework with adversarial hard-negative sample generation for multimodal remote sensing images. IEEE Trans. Geosci. Remote Sens. 62, 3354304 (2024)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"1421_CR23","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1016\/j.knosys.2018.07.041","volume":"161","author":"N Majumder","year":"2018","unstructured":"Majumder, N., Hazarika, D., Gelbukh, A., Cambria, E., Poria, S.: Multimodal sentiment analysis using hierarchical fusion with context modeling. Knowl.-Based Syst. 161, 124\u2013133 (2018)","journal-title":"Knowl.-Based Syst."},{"key":"1421_CR24","doi-asserted-by":"crossref","unstructured":"Rahman, W., Hasan, M.K., Lee, S., Zadeh, A., Mao, C., Morency, L.-P., Hoque, E.: Integrating multimodal information in large pretrained transformers. In: Proceedings of the Conference. Association for Computational Linguistics. Meeting, vol. 2020, p. 2359. NIH Public Access (2020)","DOI":"10.18653\/v1\/2020.acl-main.214"},{"key":"1421_CR25","doi-asserted-by":"crossref","unstructured":"Han, W., Chen, H., Poria, S.: Improving multimodal fusion with hierarchical mutual information maximization for multimodal sentiment analysis. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 9180\u20139192 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.723"},{"key":"1421_CR26","doi-asserted-by":"publisher","first-page":"37","DOI":"10.1016\/j.inffus.2022.11.022","volume":"92","author":"K Kim","year":"2023","unstructured":"Kim, K., Park, S.: Aobert: all-modalities-in-one BERT for multimodal sentiment analysis. Inf. Fusion 92, 37\u201345 (2023)","journal-title":"Inf. Fusion"},{"key":"1421_CR27","first-page":"9694","volume":"34","author":"J Li","year":"2021","unstructured":"Li, J., Selvaraju, R., Gotmare, A., Joty, S., Xiong, C., Hoi, S.C.H.: Align before fuse: vision and language representation learning with momentum distillation. Adv. Neural Inf. Process. Syst. 34, 9694\u20139705 (2021)","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"3","key":"1421_CR28","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1109\/MIS.2021.3057757","volume":"36","author":"W Peng","year":"2021","unstructured":"Peng, W., Hong, X., Zhao, G.: Adaptive modality distillation for separable multimodal sentiment analysis. IEEE Intell. Syst. 36(3), 82\u201389 (2021)","journal-title":"IEEE Intell. Syst."},{"key":"1421_CR29","doi-asserted-by":"publisher","first-page":"48410","DOI":"10.1109\/ACCESS.2023.3276932","volume":"11","author":"J Zhang","year":"2023","unstructured":"Zhang, J., Wu, X., Huang, C.: Adamow: multimodal sentiment analysis based on adaptive modality-specific weight fusion network. IEEE Access 11, 48410\u201348420 (2023)","journal-title":"IEEE Access"},{"key":"1421_CR30","doi-asserted-by":"crossref","unstructured":"Ando, A., Masumura, R., Takashima, A., Suzuki, S., Makishima, N., Suzuki, K., Moriya, T., Ashihara, T., Sato, H.: On the use of modality-specific large-scale pre-trained encoders for multimodal sentiment analysis. In: 2022 IEEE Spoken Language Technology Workshop (SLT), pp. 739\u2013746. IEEE (2023)","DOI":"10.1109\/SLT54892.2023.10022548"},{"key":"1421_CR31","doi-asserted-by":"publisher","first-page":"130","DOI":"10.1016\/j.neucom.2021.09.041","volume":"467","author":"B Yang","year":"2022","unstructured":"Yang, B., Shao, B., Wu, L., Lin, X.: Multimodal sentiment analysis with unidirectional modality translation. Neurocomputing 467, 130\u2013137 (2022)","journal-title":"Neurocomputing"},{"key":"1421_CR32","doi-asserted-by":"publisher","first-page":"1424","DOI":"10.1109\/TASLP.2021.3068598","volume":"29","author":"S Mai","year":"2021","unstructured":"Mai, S., Xing, S., Hu, H.: Analyzing multimodal sentiment via acoustic-and visual-LSTM with channel-aware temporal convolution network. IEEE\/ACM Trans. Audio Speech Lang. Process. 29, 1424\u20131437 (2021)","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"1421_CR33","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16x16 words: transformers for image recognition at scale (2020). arXiv preprint arXiv:2010.11929"},{"key":"1421_CR34","unstructured":"Kenton, J.D.M.-W.C., Toutanova, L.K.: Bert: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of NAACL-HLT, pp. 4171\u20134186 (2019)"},{"key":"1421_CR35","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et al: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"issue":"6","key":"1421_CR36","doi-asserted-by":"publisher","first-page":"3151","DOI":"10.1007\/s00530-023-01188-6","volume":"29","author":"C Dixit","year":"2023","unstructured":"Dixit, C., Satapathy, S.M.: A customizable framework for multimodal emotion recognition using ensemble of deep neural network models. Multimed. Syst. 29(6), 3151\u20133168 (2023)","journal-title":"Multimed. Syst."},{"key":"1421_CR37","unstructured":"Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., Stoyanov, V.: Roberta: a robustly optimized BERT pretraining approach. arXiv e-prints, 1907 (2019)"},{"key":"1421_CR38","doi-asserted-by":"crossref","unstructured":"Wu, Y., Lin, Z., Zhao, Y., Qin, B., Zhu, L.-N.: A text-centered shared-private framework via cross-modal prediction for multimodal sentiment analysis. In: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pp. 4730\u20134738 (2021)","DOI":"10.18653\/v1\/2021.findings-acl.417"},{"key":"1421_CR39","first-page":"5998","volume":"30","author":"A Vaswani","year":"2017","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141, Polosukhin, I.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30, 5998\u20136008 (2017)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"1421_CR40","unstructured":"Oord, A.v.d., Li, Y., Vinyals, O.: Representation learning with contrastive predictive coding (2018). arXiv preprint arXiv:1807.03748"},{"key":"1421_CR41","doi-asserted-by":"crossref","unstructured":"Wu, Z., Xiong, Y., Yu, S.X., Lin, D.: Unsupervised feature learning via non-parametric instance discrimination. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3733\u20133742 (2018)","DOI":"10.1109\/CVPR.2018.00393"},{"key":"1421_CR42","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2023.126992","volume":"565","author":"J Huang","year":"2024","unstructured":"Huang, J., Pu, Y., Zhou, D., Cao, J., Gu, J., Zhao, Z., Xu, D.: Dynamic hypergraph convolutional network for multimodal sentiment analysis. Neurocomputing 565, 126992 (2024)","journal-title":"Neurocomputing"},{"issue":"6","key":"1421_CR43","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1109\/MIS.2016.94","volume":"31","author":"A Zadeh","year":"2016","unstructured":"Zadeh, A., Zellers, R., Pincus, E., Morency, L.-P.: Multimodal sentiment intensity analysis in videos: Facial gestures and verbal messages. IEEE Intell. Syst. 31(6), 82\u201388 (2016)","journal-title":"IEEE Intell. Syst."},{"key":"1421_CR44","unstructured":"Zadeh, A.B., Liang, P.P., Poria, S., Cambria, E., Morency, L.-P.: Multimodal language analysis in the wild: Cmu-mosei dataset and interpretable dynamic fusion graph. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 2236\u20132246 (2018)"},{"key":"1421_CR45","doi-asserted-by":"crossref","unstructured":"Hasan, M.K., Rahman, W., Zadeh, A., Zhong, J., Tanveer, M.I., Morency, L.-P., et al.: Ur-funny: a multimodal language dataset for understanding humor (2019). arXiv preprint arXiv:1904.06618","DOI":"10.18653\/v1\/D19-1211"},{"key":"1421_CR46","doi-asserted-by":"crossref","unstructured":"Liu, Z., Shen, Y., Lakshminarasimhan, V.B., Liang, P.P., Zadeh, A.B., Morency, L.-P.: Efficient low-rank multimodal fusion with modality-specific factors. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 2247\u20132256 (2018)","DOI":"10.18653\/v1\/P18-1209"},{"key":"1421_CR47","unstructured":"Tsai, Y.-H.H., Liang, P.P., Zadeh, A., Morency, L.-P., Salakhutdinov, R.: Learning factorized multimodal representations. In: International Conference on Representation Learning, pp. 1\u201320 (2019)"},{"key":"1421_CR48","doi-asserted-by":"publisher","first-page":"107676","DOI":"10.1016\/j.knosys.2021.107676","volume":"235","author":"T Wu","year":"2022","unstructured":"Wu, T., Peng, J., Zhang, W., Zhang, H., Tan, S., Yi, F., Ma, C., Huang, Y.: Video sentiment analysis with bimodal information-augmented multi-head attention. Knowl.-Based Syst. 235, 107676 (2022)","journal-title":"Knowl.-Based Syst."},{"key":"1421_CR49","doi-asserted-by":"publisher","first-page":"549","DOI":"10.1109\/TAFFC.2023.3282410","volume":"15","author":"R Lin","year":"2023","unstructured":"Lin, R., Hu, H.: Multi-task momentum distillation for multimodal sentiment analysis. IEEE Trans. Affect. Comput. 15, 549\u2013565 (2023)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"1421_CR50","doi-asserted-by":"publisher","first-page":"127201","DOI":"10.1016\/j.neucom.2023.127201","volume":"571","author":"Y Fu","year":"2024","unstructured":"Fu, Y., Zhang, Z., Yang, R., Yao, C.: Hybrid cross-modal interaction learning for multimodal sentiment analysis. Neurocomputing 571, 127201 (2024)","journal-title":"Neurocomputing"},{"key":"1421_CR51","doi-asserted-by":"publisher","first-page":"111149","DOI":"10.1016\/j.knosys.2023.111149","volume":"283","author":"H Shi","year":"2024","unstructured":"Shi, H., Pu, Y., Zhao, Z., Huang, J., Zhou, D., Xu, D., Cao, J.: Co-space representation interaction network for multimodal sentiment analysis. Knowl.-Based Syst. 283, 111149 (2024)","journal-title":"Knowl.-Based Syst."},{"key":"1421_CR52","doi-asserted-by":"publisher","first-page":"111346","DOI":"10.1016\/j.knosys.2023.111346","volume":"285","author":"J Huang","year":"2024","unstructured":"Huang, J., Zhou, J., Tang, Z., Lin, J., Chen, C.Y.-C.: Tmbl: transformer-based multimodal binding learning model for multimodal sentiment analysis. Knowl.-Based Syst. 285, 111346 (2024)","journal-title":"Knowl.-Based Syst."},{"key":"1421_CR53","doi-asserted-by":"crossref","unstructured":"Degottex, G., Kane, J., Drugman, T., Raitio, T., Scherer, S.: COVAREP\u2014a collaborative voice analysis repository for speech technologies. In: 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 960\u2013964. IEEE (2014)","DOI":"10.1109\/ICASSP.2014.6853739"},{"key":"1421_CR54","doi-asserted-by":"crossref","unstructured":"Baltru\u0161aitis, T., Robinson, P., Morency, L.-P.: Openface: an open source facial behavior analysis toolkit. In: 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 1\u201310. IEEE (2016)","DOI":"10.1109\/WACV.2016.7477553"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01421-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01421-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01421-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,22]],"date-time":"2024-08-22T08:39:04Z","timestamp":1724315944000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01421-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,30]]},"references-count":54,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["1421"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01421-w","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7,30]]},"assertion":[{"value":"29 March 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 July 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 July 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"228"}}