{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,24]],"date-time":"2026-02-24T17:22:25Z","timestamp":1771953745368,"version":"3.50.1"},"reference-count":60,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2024,10,21]],"date-time":"2024-10-21T00:00:00Z","timestamp":1729468800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,21]],"date-time":"2024-10-21T00:00:00Z","timestamp":1729468800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2025,4]]},"DOI":"10.1007\/s11263-024-02267-5","type":"journal-article","created":{"date-parts":[[2024,10,21]],"date-time":"2024-10-21T15:03:18Z","timestamp":1729522998000},"page":"1808-1833","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["CogCartoon: Towards Practical Story Visualization"],"prefix":"10.1007","volume":"133","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3955-3676","authenticated-orcid":false,"given":"Zhongyang","family":"Zhu","sequence":"first","affiliation":[]},{"given":"Jie","family":"Tang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,21]]},"reference":[{"key":"2267_CR1","doi-asserted-by":"crossref","unstructured":"Ahn, D., Kim, D., Song, G., Kim, S.H., Lee, H., Kang, D., & Choi, J. (2023). Story visualization by online text augmentation with context memory. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 3125\u20133135).","DOI":"10.1109\/ICCV51070.2023.00290"},{"key":"2267_CR2","doi-asserted-by":"crossref","unstructured":"Avrahami, O., Hertz, A., Vinker, Y., Arar, M., Fruchter, S., Fried, O., Cohen-Or, D., & Lischinski, D. (2024). The chosen one: Consistent characters in text-to-image diffusion models. In ACM SIGGRAPH 2024 conference papers (pp. 1\u201312).","DOI":"10.1145\/3641519.3657430"},{"key":"2267_CR3","doi-asserted-by":"crossref","unstructured":"Bao, J., Chen, D., Wen, F., Li, H., & Hua, G. (2017). CVAE-GAN: fine-grained image generation through asymmetric training. In Proceedings of the IEEE international conference on computer vision (pp. 2745\u20132754).","DOI":"10.1109\/ICCV.2017.299"},{"key":"2267_CR4","doi-asserted-by":"crossref","unstructured":"Brooks, T., Holynski, A., & Efros, A.A. (2023). Instructpix2pix: Learning to follow image editing instructions. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 18392\u201318402).","DOI":"10.1109\/CVPR52729.2023.01764"},{"key":"2267_CR5","doi-asserted-by":"crossref","unstructured":"Deng, J., Fan, D., Qiu, X., & Zhou, F. (2023). Improving crowded object detection via copy-paste. In Proceedings of the AAAI conference on artificial intelligence (Vol. 37, pp. 497\u2013505).","DOI":"10.1609\/aaai.v37i1.25124"},{"key":"2267_CR6","first-page":"8780","volume":"34","author":"P Dhariwal","year":"2021","unstructured":"Dhariwal, P., & Nichol, A. (2021). Diffusion models beat GANs on image synthesis. Advances in Neural Information Processing Systems, 34, 8780\u20138794.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2267_CR7","first-page":"19822","volume":"34","author":"M Ding","year":"2021","unstructured":"Ding, M., Yang, Z., Hong, W., Zheng, W., Zhou, C., Yin, D., Lin, J., Zou, X., Shao, Z., Yang, H., Tang, J., et al. (2021). Cogview: Mastering text-to-image generation via transformers. Advances in Neural Information Processing Systems, 34, 19822\u201319835.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2267_CR8","first-page":"16890","volume":"35","author":"M Ding","year":"2022","unstructured":"Ding, M., Zheng, W., Hong, W., & Tang, J. (2022). Cogview2: Faster and better text-to-image generation via hierarchical transformers. Advances in Neural Information Processing Systems, 35, 16890\u201316902.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2267_CR9","doi-asserted-by":"crossref","first-page":"548","DOI":"10.1016\/j.ins.2022.10.083","volume":"614","author":"P Dong","year":"2022","unstructured":"Dong, P., Wu, L., Meng, L., & Meng, X. (2022). HR-PrGAN: High-resolution story visualization with progressive generative adversarial networks. Information Sciences, 614, 548\u2013562.","journal-title":"Information Sciences"},{"key":"2267_CR10","unstructured":"Gal, R., Alaluf, Y., Atzmon, Y., Patashnik, O., Bermano, A.H., Chechik, G., & Cohen-Or, D. (2022). An image is worth one word: Personalizing text-to-image generation using textual inversion. arXiv preprint arXiv:2208.01618."},{"key":"2267_CR11","doi-asserted-by":"crossref","unstructured":"Ghiasi, G., Cui, Y., Srinivas, A., Qian, R., Lin, T.-Y., Cubuk, E.D., Le, Q.V., & Zoph, B. (2021). Simple copy-paste is a strong data augmentation method for instance segmentation. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 2918\u20132928).","DOI":"10.1109\/CVPR46437.2021.00294"},{"key":"2267_CR12","doi-asserted-by":"crossref","unstructured":"Gong, Y., Pang, Y., Cun, X., Xia, M., Chen, H., Wang, L., Zhang, Y., Wang, X., Shan, Y., & Yang, Y. (2023). Talecrafter: Interactive story visualization with multiple characters. arXiv preprint arXiv:2305.18247.","DOI":"10.1145\/3610548.3618184"},{"key":"2267_CR13","unstructured":"Gu, Y., Wang, X., Wu, J.Z., Shi, Y., Chen, Y., Fan, Z., Xiao, W., Zhao, R., Chang, S., Wu, W.Ge, Y., Shan, Y., & Shou, M. Z., et al. (2024). Mix-of-show: Decentralized low-rank adaptation for multi-concept customization of diffusion models. Advances in Neural Information Processing Systems36."},{"key":"2267_CR14","unstructured":"Hertz, A., Mokady, R., Tenenbaum, J., Aberman, K., Pritch, Y., & Cohen-Or, D. (2022). Prompt-to-prompt image editing with cross attention control. arXiv preprint arXiv:2208.01626."},{"key":"2267_CR15","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., & Hochreiter, S. (2017). Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in Neural Information Processing Systems30."},{"key":"2267_CR16","unstructured":"Ho, J., & Salimans, T. (2022). Classifier-free diffusion guidance. arXiv preprint arXiv:2207.12598."},{"key":"2267_CR17","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho, J., Jain, A., & Abbeel, P. (2020). Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33, 6840\u20136851.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2267_CR18","unstructured":"Hui, M., Yang, S., Zhao, B., Shi, Y., Wang, H., Wang, P., Zhou, Y., & Xie, C. (2024). Hq-edit: A high-quality dataset for instruction-based image editing. arXiv preprint arXiv:2404.09990."},{"key":"2267_CR19","unstructured":"Jeong, H., Kwon, G., & Ye, J.C. (2023). Zero-shot generation of coherent storybook from plain text story using diffusion models. arXiv preprint arXiv:2302.03900."},{"key":"2267_CR20","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A.C., Lo, W.-Y.Dollar, P., & Girshick, R., et al. (2023). Segment anything. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 4015\u20134026).","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"2267_CR21","doi-asserted-by":"crossref","unstructured":"Kumari, N., Zhang, B., Zhang, R., Shechtman, E., & Zhu, J.-Y. (2023). Multi-concept customization of text-to-image diffusion. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 1931\u20131941).","DOI":"10.1109\/CVPR52729.2023.00192"},{"key":"2267_CR22","doi-asserted-by":"crossref","unstructured":"Li, B. (2022). Word-level fine-grained story visualization. In European conference on computer vision (pp. 347\u2013362). Springer.","DOI":"10.1007\/978-3-031-20059-5_20"},{"key":"2267_CR23","doi-asserted-by":"crossref","unstructured":"Li, Y., Gan, Z., Shen, Y., Liu, J., Cheng, Y., Wu, Y., Carin, L., Carlson, D., & Gao, J. (2019). StoryGAN: A sequential conditional GAN for story visualization. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 6329\u20136338).","DOI":"10.1109\/CVPR.2019.00649"},{"key":"2267_CR24","doi-asserted-by":"crossref","unstructured":"Li, Y., Gan, Z., Shen, Y., Liu, J., Cheng, Y., Wu, Y., Carin, L., Carlson, D., & Gao, J. (2019). Storygan: A sequential conditional gan for story visualization. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 6329\u20136338).","DOI":"10.1109\/CVPR.2019.00649"},{"key":"2267_CR25","doi-asserted-by":"crossref","unstructured":"Li, B., Qi, X., Lukasiewicz, T., & Torr, P.H. (2020). Manigan: Text-guided image manipulation. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 7880\u20137889).","DOI":"10.1109\/CVPR42600.2020.00790"},{"key":"2267_CR26","doi-asserted-by":"crossref","unstructured":"Li, W., Zhang, P., Zhang, L., Huang, Q., He, X., Lyu, S., & Gao, J. (2019). Object-driven text-to-image synthesis via adversarial training. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 12174\u201312182).","DOI":"10.1109\/CVPR.2019.01245"},{"key":"2267_CR27","volume":"73","author":"C Li","year":"2020","unstructured":"Li, C., Kong, L., & Zhou, Z. (2020). Improved-StoryGAN for sequential images visualization. Journal of Visual Communication and Image Representation, 73, 102956.","journal-title":"Journal of Visual Communication and Image Representation"},{"key":"2267_CR28","unstructured":"Liu, Z., Feng, R., Zhu, K., Zhang, Y., Zheng, K., Liu, Y., Zhao, D., Zhou, J., & Cao, Y. (2023) Cones: Concept neurons in diffusion models for customized generation. arXiv preprint arXiv:2303.05125."},{"key":"2267_CR29","doi-asserted-by":"crossref","unstructured":"Liu, S., Zeng, Z., Ren, T., Li, F., Zhang, H., Yang, J., Li, C., Yang, J., Su, H., Zhu, J. & Zhang, L., et al. (2023). Grounding DINO: Marrying DINO with grounded pre-training for open-set object detection. arXiv preprint arXiv:2303.05499.","DOI":"10.1007\/978-3-031-72970-6_3"},{"key":"2267_CR30","unstructured":"Liu, Z., Zhang, Y., Shen, Y., Zheng, K., Zhu, K., Feng, R., Liu, Y., Zhao, D., Zhou, J., & Cao, Y. (2023). Cones 2: Customizable image synthesis with multiple subjects. arXiv preprint arXiv:2305.19327."},{"key":"2267_CR31","doi-asserted-by":"crossref","unstructured":"Maharana, A., & Bansal, M. (2021). Integrating visuospatial, linguistic and commonsense structure into story visualization. arXiv preprint arXiv:2110.10834.","DOI":"10.18653\/v1\/2021.emnlp-main.543"},{"key":"2267_CR32","doi-asserted-by":"crossref","unstructured":"Maharana, A., Hannan, D., & Bansal, M. (2021). Improving generation and evaluation of visual stories via semantic consistency. arXiv preprint arXiv:2105.10026.","DOI":"10.18653\/v1\/2021.naacl-main.194"},{"key":"2267_CR33","doi-asserted-by":"crossref","unstructured":"Maharana, A., Hannan, D., & Bansal, M. (2022). Storydall-e: Adapting pretrained text-to-image transformers for story continuation. In European conference on computer vision (pp. 70\u201387). Springer.","DOI":"10.1007\/978-3-031-19836-6_5"},{"key":"2267_CR34","doi-asserted-by":"crossref","unstructured":"Mokady, R., Hertz, A., Aberman, K., Pritch, Y., & Cohen-Or, D. (2023). Null-text inversion for editing real images using guided diffusion models. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 6038\u20136047).","DOI":"10.1109\/CVPR52729.2023.00585"},{"key":"2267_CR35","unstructured":"Nichol, A.Q., & Dhariwal, P. (2021). Improved denoising diffusion probabilistic models. In International conference on machine learning (pp. 8162\u20138171). PMLR."},{"key":"2267_CR36","unstructured":"Nichol, A., Dhariwal, P., Ramesh, A., Shyam, P., Mishkin, P., McGrew, B., Sutskever, I., & Chen, M. (2021). Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741."},{"key":"2267_CR37","unstructured":"Pan, X., Qin, P., Li, Y., Xue, H., & Chen, W. (2022). Synthesizing coherent story with auto-regressive latent diffusion models. arXiv preprint arXiv:2211.10950."},{"key":"2267_CR38","doi-asserted-by":"crossref","unstructured":"Parmar, G., Kumar\u00a0Singh, K., Zhang, R., Li, Y., Lu, J., & Zhu, J.-Y. (2023). Zero-shot image-to-image translation. In ACM SIGGRAPH 2023 conference proceedings (pp. 1\u201311).","DOI":"10.1145\/3588432.3591513"},{"key":"2267_CR39","doi-asserted-by":"crossref","unstructured":"Rahman, T., Lee, H.-Y., Ren, J., Tulyakov, S., Mahajan, S., & Sigal, L. (2023). Make-a-story: Visual memory conditioned consistent story generation. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 2493\u20132502).","DOI":"10.1109\/CVPR52729.2023.00246"},{"key":"2267_CR40","unstructured":"Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., & Chen, M. (2022). Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125."},{"key":"2267_CR41","unstructured":"Ramesh, A., Pavlov, M., Goh, G., Gray, S., Voss, C., Radford, A., Chen, M., & Sutskever, I. (2021). Zero-shot text-to-image generation. In International conference on machine learning (pp. 8821\u20138831). PMLR."},{"key":"2267_CR42","unstructured":"Reed, S., Akata, Z., Yan, X., Logeswaran, L., Schiele, B., & Lee, H. (2016). Generative adversarial text to image synthesis. In International conference on machine learning (pp. 1060\u20131069). PMLR."},{"key":"2267_CR43","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., & Ommer, B. (2022). High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 10684\u201310695).","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"2267_CR44","doi-asserted-by":"crossref","unstructured":"Ruan, S., Zhang, Y., Zhang, K., Fan, Y., Tang, F., Liu, Q., & Chen, E. (2021). Dae-gan: Dynamic aspect-aware gan for text-to-image synthesis. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 13960\u201313969).","DOI":"10.1109\/ICCV48922.2021.01370"},{"key":"2267_CR45","doi-asserted-by":"crossref","unstructured":"Ruiz, N., Li, Y., Jampani, V., Pritch, Y., Rubinstein, M., & Aberman, K. (2023). Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 22500\u201322510).","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"2267_CR46","first-page":"36479","volume":"35","author":"C Saharia","year":"2022","unstructured":"Saharia, C., Chan, W., Saxena, S., Li, L., Whang, J., Denton, E. L., Ghasemipour, K., Gontijo Lopes, R., Karagol Ayan, B., Salimans, T., Ho, J., Fleet, D. J., Norouzi, M., et al. (2022). Photorealistic text-to-image diffusion models with deep language understanding. Advances in Neural Information Processing Systems, 35, 36479\u201336494.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2267_CR47","unstructured":"Sanh, V., Debut, L., Chaumond, J., & Wolf, T. (2019). Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108."},{"key":"2267_CR48","unstructured":"Sauer, A., Karras, T., Laine, S., Geiger, A., & Aila, T. (2023). StyleGAN-T: Unlocking the power of GANs for fast large-scale text-to-image synthesis. arXiv preprint arXiv:2301.09515."},{"key":"2267_CR49","doi-asserted-by":"crossref","unstructured":"Song, T., Cao, J., Wang, K., Liu, B., & Zhang, X. (2023). Causal-story: Local causal attention utilizing parameter-efficient tuning for visual story synthesis. arXiv preprint arXiv:2309.09553.","DOI":"10.1109\/ICASSP48485.2024.10446420"},{"key":"2267_CR50","unstructured":"Song, J., Meng, C., & Ermon, S. (2020). Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502."},{"key":"2267_CR51","doi-asserted-by":"crossref","unstructured":"Song, Y.-Z., Rui\u00a0Tam, Z., Chen, H.-J., Lu, H.-H., & Shuai, H.-H. (2020). Character-preserving coherent story visualization. In European conference on computer vision (pp. 18\u201333). Springer.","DOI":"10.1007\/978-3-030-58520-4_2"},{"key":"2267_CR52","unstructured":"Su, S., Guo, L., Gao, L., Shen, H.T., & Song, J. (2023). Make-a-storyboard: A general framework for storyboard with disentangled and merged control. arXiv preprint arXiv:2312.07549."},{"key":"2267_CR53","doi-asserted-by":"crossref","unstructured":"Xiao, G., Yin, T., Freeman, W.T., Durand, F., & Han, S. (2023). Fastcomposer: Tuning-free multi-subject image generation with localized attention. arXiv preprint arXiv:2305.10431.","DOI":"10.1007\/s11263-024-02227-z"},{"key":"2267_CR54","doi-asserted-by":"crossref","unstructured":"Yang, B., Gu, S., Zhang, B., Zhang, T., Chen, X., Sun, X., Chen, D., & Wen, F. (2023). Paint by example: Exemplar-based image editing with diffusion models. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 18381\u201318391).","DOI":"10.1109\/CVPR52729.2023.01763"},{"key":"2267_CR55","doi-asserted-by":"crossref","unstructured":"Yang, B., Gu, S., Zhang, B., Zhang, T., Chen, X., Sun, X., Chen, D., & Wen, F. (2023). Paint by example: Exemplar-based image editing with diffusion models. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 18381\u201318391).","DOI":"10.1109\/CVPR52729.2023.01763"},{"key":"2267_CR56","doi-asserted-by":"crossref","unstructured":"Yoon, J., & Choi, M.-K. (2023). Exploring video frame redundancies for efficient data sampling and annotation in instance segmentation. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 3307\u20133316).","DOI":"10.1109\/CVPRW59228.2023.00333"},{"key":"2267_CR57","doi-asserted-by":"crossref","unstructured":"Yu, C., Zhou, Q., Li, J., Yuan, J., Wang, Z., & Wang, F. (2023). Foundation model drives weakly incremental learning for semantic segmentation. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 23685\u201323694).","DOI":"10.1109\/CVPR52729.2023.02268"},{"key":"2267_CR58","unstructured":"Zeng, A., Liu, X., Du, Z., Wang, Z., Lai, H., Ding, M., Yang, Z., Xu, Y., Zheng, W., Xia, X.Tam, W. L., Ma, Z., Xue, Y., Zhai, J., Chen, W., Zhang, P., Dong, Y., & Tang, J., et al. (2022). GLM-130b: An open bilingual pre-trained model. arXiv preprint arXiv:2210.02414."},{"key":"2267_CR59","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Xie, Y., & Yang, L. (2018). Photographic text-to-image synthesis with a hierarchically-nested adversarial network. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 6199\u20136208).","DOI":"10.1109\/CVPR.2018.00649"},{"key":"2267_CR60","doi-asserted-by":"crossref","unstructured":"Zhang, H., Xu, T., Li, H., Zhang, S., Wang, X., Huang, X., & Metaxas, D.N. (2017). StackGAN: Text to photo-realistic image synthesis with stacked generative adversarial networks. In Proceedings of the IEEE international conference on computer vision (pp. 5907\u20135915).","DOI":"10.1109\/ICCV.2017.629"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-024-02267-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-024-02267-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-024-02267-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,30]],"date-time":"2025-03-30T22:01:59Z","timestamp":1743372119000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-024-02267-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,21]]},"references-count":60,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,4]]}},"alternative-id":["2267"],"URL":"https:\/\/doi.org\/10.1007\/s11263-024-02267-5","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,21]]},"assertion":[{"value":"20 December 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 September 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 October 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}