{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,24]],"date-time":"2026-04-24T19:49:17Z","timestamp":1777060157542,"version":"3.51.4"},"reference-count":56,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2023,12,4]],"date-time":"2023-12-04T00:00:00Z","timestamp":1701648000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,12,4]],"date-time":"2023-12-04T00:00:00Z","timestamp":1701648000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61472220\uff0c61572286"],"award-info":[{"award-number":["61472220\uff0c61572286"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2024,4]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Multi-scale representation provides an effective answer to the scale variation of objects and entities in semantic segmentation. The ability to capture long-range pixel dependency facilitates semantic segmentation. In addition, semantic segmentation necessitates the effective use of pixel-to-pixel similarity in the channel direction to enhance pixel areas. By reviewing the characteristics of earlier successful segmentation models, we discover a number of crucial elements that enhance segmentation model performance, including a robust encoder structure, multi-scale interactions, attention mechanisms, and a robust decoder structure. The attention mechanism of the asymmetric non-local neural network (ANNet) is merged with multi-scale pyramidal modules to accelerate model segmentation while maintaining high accuracy. However, ANNet does not account for the similarity between pixels in the feature map channel direction, making the segmentation accuracy unsatisfactory. As a result, we propose EMSNet, a straightforward convolutional network architecture for semantic segmentation that consists of Integration of enhanced regional module (IERM) and Multi-scale convolution module (MSCM). The IERM module generates weights using four or five-stage feature maps, then fuses the input features with the weights and uses more computation. The similarity of the channel direction feature graphs is also calculated using ANNet\u2019s auxiliary loss function. The MSCM module can more accurately describe the interactions between various channels, capture the interdependencies between feature pixels, and capture the multi-scale context. Experiments prove that we perform well in tests using the benchmark dataset. On Cityscapes test data, we get 82.2% segmentation accuracy. The mIoU in the ADE20k and Pascal VOC datasets are, respectively, 45.58% and 85.46%.<\/jats:p>","DOI":"10.1007\/s40747-023-01279-x","type":"journal-article","created":{"date-parts":[[2023,12,4]],"date-time":"2023-12-04T04:10:52Z","timestamp":1701663052000},"page":"2557-2568","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":14,"title":["Enhanced multi-scale networks for semantic segmentation"],"prefix":"10.1007","volume":"10","author":[{"given":"Tianping","family":"Li","sequence":"first","affiliation":[]},{"given":"Zhaotong","family":"Cui","sequence":"additional","affiliation":[]},{"given":"Yu","family":"Han","sequence":"additional","affiliation":[]},{"given":"Guanxing","family":"Li","sequence":"additional","affiliation":[]},{"given":"Meng","family":"Li","sequence":"additional","affiliation":[]},{"given":"Dongmei","family":"Wei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,4]]},"reference":[{"key":"1279_CR1","doi-asserted-by":"crossref","unstructured":"Zhou B, Zhao H, Puig X, et al (2017) Scene parsing through ADE20K dataset. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR). IEEE, Honolulu, HI, pp 5122\u20135130","DOI":"10.1109\/CVPR.2017.544"},{"issue":"4","key":"1279_CR2","doi-asserted-by":"publisher","first-page":"648","DOI":"10.1109\/TSMC.2016.2623683","volume":"47","author":"Y Li","year":"2016","unstructured":"Li Y, Guo Y, Kao Y, He R (2016) Image piece learning for weakly supervised semantic segmentation. IEEE Trans Syst Man Cybern Syst 47(4):648\u2013659. https:\/\/doi.org\/10.1109\/TSMC.2016.2623683","journal-title":"IEEE Trans Syst Man Cybern Syst"},{"issue":"12","key":"1279_CR3","doi-asserted-by":"publisher","first-page":"25489","DOI":"10.1109\/TITS.2021.3098355","volume":"23","author":"G Gao","year":"2021","unstructured":"Gao G, Xu G, Yu Y et al (2021) MSCFNet: a lightweight network with multi-scale context fusion for real-time semantic segmentation. IEEE Trans Intell Transport Syst 23(12):25489\u201325499. https:\/\/doi.org\/10.1109\/TITS.2021.3098355","journal-title":"IEEE Trans Intell Transport Syst"},{"key":"1279_CR4","doi-asserted-by":"crossref","unstructured":"Teichmann M, Weber M, Zollner M, et al (2018) MultiNet: real-time joint semantic reasoning for autonomous driving. In: 2018 IEEE intelligent vehicles symposium (IV). IEEE, Changshu, pp 1013\u20131020","DOI":"10.1109\/IVS.2018.8500504"},{"key":"1279_CR5","doi-asserted-by":"crossref","unstructured":"Siam M, Elkerdawy S, Jagersand M, Yogamani S (2017) Deep semantic segmentation for automated driving: taxonomy, roadmap and challenges. In: 2017 IEEE 20th international conference on intelligent transportation systems (ITSC). IEEE, Yokohama, pp 1\u20138","DOI":"10.1109\/ITSC.2017.8317714"},{"key":"1279_CR6","doi-asserted-by":"publisher","first-page":"1430","DOI":"10.1109\/JPROC.2003.817125","volume":"91","author":"M Hardens","year":"2003","unstructured":"Hardens M, Szekely G (2003) Enhancing human-computer interaction in medical segmentation. Proc IEEE 91:1430\u20131442. https:\/\/doi.org\/10.1109\/JPROC.2003.817125","journal-title":"Proc IEEE"},{"key":"1279_CR7","unstructured":"Alhaija H A, Mustikovela S K, Mescheder L et al (2017) Augmented reality meets deep learning for car instance segmentation in urban scenes. In: British machine vision conference, vol 1, p 2"},{"key":"1279_CR8","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-00889-5_1","volume-title":"Deep learning in medical image analysis and multimodal learning for clinical decision support","author":"Z Zhou","year":"2018","unstructured":"Zhou Z, Rahman Siddiquee MM, Tajbakhsh N, Liang J (2018) UNet++: a nested U-Net architecture for medical image segmentation. In: Stoyanov D, Taylor Z, Carneiro G et al (eds) Deep learning in medical image analysis and multimodal learning for clinical decision support. Springer International Publishing, Cham, pp 3\u201311"},{"key":"1279_CR9","doi-asserted-by":"publisher","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y Lecun","year":"1998","unstructured":"Lecun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86:2278\u20132324. https:\/\/doi.org\/10.1109\/5.726791","journal-title":"Proc IEEE"},{"key":"1279_CR10","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2021.3084827","author":"Z Li","year":"2021","unstructured":"Li Z, Liu F, Yang W, Peng S, Zhou J (2021) A survey of convolutional neural networks: analysis, applications, and prospects. IEEE Trans Neural Netw Learn Syst. https:\/\/doi.org\/10.1109\/TNNLS.2021.3084827","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"1279_CR11","doi-asserted-by":"crossref","unstructured":"Szegedy C, Liu W, Jia Y et al (2015) Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 1\u20139","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"1279_CR12","doi-asserted-by":"publisher","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556. https:\/\/doi.org\/10.48550\/ARXIV.1409.1556","DOI":"10.48550\/ARXIV.1409.1556"},{"key":"1279_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.rse.2020.112045","volume":"250","author":"Y Li","year":"2020","unstructured":"Li Y, Chen W, Zhang Y et al (2020) Accurate cloud detection in high-resolution remote sensing imagery by weakly supervised deep learning. Remote Sens Environ 250:112045. https:\/\/doi.org\/10.1016\/j.rse.2020.112045","journal-title":"Remote Sens Environ"},{"key":"1279_CR14","doi-asserted-by":"publisher","first-page":"155","DOI":"10.1016\/j.isprsjprs.2019.10.001","volume":"158","author":"C Tao","year":"2019","unstructured":"Tao C, Qi J, Li Y et al (2019) Spatial information inference net: Road extraction using road-specific contextual information. ISPRS J Photogramm Remote Sens 158:155\u2013166. https:\/\/doi.org\/10.1016\/j.isprsjprs.2019.10.001","journal-title":"ISPRS J Photogramm Remote Sens"},{"key":"1279_CR15","doi-asserted-by":"crossref","unstructured":"Long J, Shelhamer E, Darrell T (2015) Fully Convolutional Networks for Semantic Segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3431\u20133440","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"1279_CR16","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"1279_CR17","doi-asserted-by":"publisher","first-page":"173","DOI":"10.1007\/978-3-030-58539-6_11","volume-title":"Computer vision\u2014ECCV 2020","author":"Y Yuan","year":"2020","unstructured":"Yuan Y, Chen X, Wang J (2020) Object-contextual representations for semantic segmentation. In: Vedaldi A, Bischof H, Brox T, Frahm J-M (eds) Computer vision\u2014ECCV 2020. Springer International Publishing, Cham, pp 173\u2013190"},{"key":"1279_CR18","doi-asserted-by":"publisher","unstructured":"Chen L-C, Zhu Y, Papandreou G, et al (2018) Encoder-decoder with Atrous separable convolution for semantic image segmentation. In: Proceedings of the European conference on computer vision (ECCV), pp 801\u2013818. https:\/\/doi.org\/10.48550\/ARXIV.1802.02611","DOI":"10.48550\/ARXIV.1802.02611"},{"key":"1279_CR19","unstructured":"Chen L-C, Papandreou G, Schroff F, Adam H (2017) Rethinking Atrous convolution for semantic image segmentation. arXiv preprint arXiv:1706.05587"},{"key":"1279_CR20","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1109\/TPAMI.2017.2699184","volume":"40","author":"L-C Chen","year":"2018","unstructured":"Chen L-C, Papandreou G, Kokkinos I et al (2018) DeepLab: semantic image segmentation with deep convolutional nets, Atrous convolution, and fully connected CRFs. IEEE Trans Pattern Anal Mach Intell 40:834\u2013848. https:\/\/doi.org\/10.1109\/TPAMI.2017.2699184","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1279_CR21","unstructured":"Badrinarayanan V, Handa A, Cipolla R (2015) SegNet: a deep convolutional encoder-decoder architecture for robust semantic pixel-wise labelling. arXiv preprint arXiv:1505.07293"},{"key":"1279_CR22","doi-asserted-by":"crossref","unstructured":"Zhao H, Shi J, Qi X et al (2017) Pyramid scene parsing network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2881\u20132890","DOI":"10.1109\/CVPR.2017.660"},{"key":"1279_CR23","doi-asserted-by":"crossref","unstructured":"Wang X, Girshick R, Gupta A, He K (2018) Non-local neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 7794\u20137803","DOI":"10.1109\/CVPR.2018.00813"},{"key":"1279_CR24","doi-asserted-by":"crossref","unstructured":"Zhu Z, Xu M, Bai S et al (2019) Asymmetric non-local neural networks for semantic segmentation. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 593\u2013602","DOI":"10.1109\/ICCV.2019.00068"},{"key":"1279_CR25","doi-asserted-by":"publisher","DOI":"10.1007\/s40747-023-01056-w","author":"T Li","year":"2023","unstructured":"Li T, Wei Y, Cui Z et al (2023) Mutually reinforcing non-local neural networks for semantic segmentation. Complex Intell Syst. https:\/\/doi.org\/10.1007\/s40747-023-01056-w","journal-title":"Complex Intell Syst"},{"key":"1279_CR26","doi-asserted-by":"crossref","unstructured":"Fu J, Liu J, Tian H, et al (2019) Dual attention network for scene segmentation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 3146\u20133154","DOI":"10.1109\/CVPR.2019.00326"},{"key":"1279_CR27","volume-title":"PRICAI 2022: trends in artificial intelligence. PRICAI 2022. Lecture notes in computer science","author":"F Dai","year":"2022","unstructured":"Dai F, Zhang S, Liu H, Ma Y, Zhao Q (2022) Global boundary refinement for semantic segmentation via optimal transport. In: Khanna S, Cao J, Bai Q, Xu G (eds) PRICAI 2022: trends in artificial intelligence. PRICAI 2022. Lecture notes in computer science, vol 13631. Springer, Cham"},{"key":"1279_CR28","unstructured":"Dosovitskiy, A, Beyer, L, Kolesnikov, A et al (2020) An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929"},{"key":"1279_CR29","doi-asserted-by":"crossref","unstructured":"Mottaghi R, Chen X, Liu X, et al (2014) The role of context for object detection and semantic segmentation in the wild. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 891\u2013898","DOI":"10.1109\/CVPR.2014.119"},{"key":"1279_CR30","doi-asserted-by":"publisher","first-page":"273","DOI":"10.1016\/j.neucom.2019.07.078","volume":"365","author":"B Zhao","year":"2019","unstructured":"Zhao B, Zhang X, Li Z, Hu X (2019) A multi-scale strategy for deep semantic segmentation with convolutional neural networks. Neurocomputing 365:273\u2013284. https:\/\/doi.org\/10.1016\/j.neucom.2019.07.078","journal-title":"Neurocomputing"},{"key":"1279_CR31","first-page":"234","volume-title":"Medical image computing and computer-assisted intervention\u2014MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-Net: convolutional networks for biomedical image segmentation. In: Navab N, Hornegger J, Wells WM, Frangi AF (eds) Medical image computing and computer-assisted intervention\u2014MICCAI 2015. Springer International Publishing, Cham, pp 234\u2013241"},{"key":"1279_CR32","doi-asserted-by":"crossref","unstructured":"Noh H, Hong S, Han B (2015) Learning deconvolution network for semantic segmentation. In: Proceedings of the IEEE international conference on computer vision, pp 1520\u20131528","DOI":"10.1109\/ICCV.2015.178"},{"key":"1279_CR33","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: International conference on learning representations, pp 1\u201314"},{"key":"1279_CR34","doi-asserted-by":"crossref","unstructured":"Xie S, Girshick R, Dollar P, Tu Z, He K (2017) Aggregated residual transformations for deep neural networks. In: IEEE conf. comput. vis. pattern recog., pp 1492\u20131500","DOI":"10.1109\/CVPR.2017.634"},{"key":"1279_CR35","doi-asserted-by":"crossref","unstructured":"Chen C-F(Richard), Fan Q, Panda R (2021) CrossViT: cross-attention multi-scale vision transformer for image classification. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 357\u2013366","DOI":"10.1109\/ICCV48922.2021.00041"},{"key":"1279_CR36","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y et al (2021) Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 10012\u201310022","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"1279_CR37","doi-asserted-by":"publisher","first-page":"187","DOI":"10.1007\/s41095-021-0229-5","volume":"7","author":"M-H Guo","year":"2021","unstructured":"Guo M-H, Cai J-X, Liu Z-N et al (2021) PCT: point cloud transformer. Comp Vis Media 7:187\u2013199. https:\/\/doi.org\/10.1007\/s41095-021-0229-5","journal-title":"Comp Vis Media"},{"key":"1279_CR38","doi-asserted-by":"crossref","unstructured":"Wang Q, Wu B, Zhu P et al (2020) Supplementary material for \u201cECA-Net: efficient channel attention for deep convolutional neural networks\u201d. In: Proceedings of the 2020 IEEE\/CVF conference on computer vision and pattern recognition. IEEE, Seattle, WA, USA, pp 13\u201319","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"1279_CR39","doi-asserted-by":"crossref","unstructured":"Zhang H, Wu C, Zhang Z et al (2022) ResNeSt: split-attention networks. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 2736\u20132746","DOI":"10.1109\/CVPRW56347.2022.00309"},{"key":"1279_CR40","doi-asserted-by":"crossref","unstructured":"Huang Z, Shi X, Zhang C et al (2022) FlowFormer: a transformer architecture for optical flow. arXiv preprint arXiv:2203.16194","DOI":"10.1007\/978-3-031-19790-1_40"},{"key":"1279_CR41","unstructured":"Yuan Y, Huang L, Guo J et al (2018) OCNet: object context network for scene parsing. arXiv preprint arXiv:1809.00916"},{"key":"1279_CR42","unstructured":"Zhang H, Goodfellow I, Metaxas D, Odena A (2019) Self-attention generative adversarial networks. In: International conference on machine learning. PMLR, pp 7354\u20137363"},{"key":"1279_CR43","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3211006","author":"M-H Guo","year":"2022","unstructured":"Guo M-H, Liu Z-N, Mu T-J, Hu S-M (2022) Beyond self-attention: external attention using two linear layers for visual tasks. IEEE Trans Pattern Anal Mach Intell. https:\/\/doi.org\/10.1109\/TPAMI.2022.3211006","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1279_CR44","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. In: Medical image computing and computer-assisted intervention\u2013MICCAI 2015: 18th international conference, Munich, Germany, October 5\u20139, 2015, proceedings, part III 18. Springer International Publishing, pp 234\u2013241","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"1279_CR45","doi-asserted-by":"crossref","unstructured":"Zhang H, Dana K, Shi J, Zhang Z, Wang X, Tyagi A, Agrawal A (2018) Context encoding for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7151\u20137160","DOI":"10.1109\/CVPR.2018.00747"},{"key":"1279_CR46","doi-asserted-by":"crossref","unstructured":"Gu J, Kwon H, Wang D, Ye W, Li M, Chen YH et al. (2022) Multi-scale high-resolution vision transformer for semantic segmentation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 12094\u201312103","DOI":"10.1109\/CVPR52688.2022.01178"},{"key":"1279_CR47","doi-asserted-by":"publisher","unstructured":"Tao A., Sapra K, Catanzaro B (2020) Hierarchical multi-scale attention for semantic segmentation. arXiv preprint arXiv:2005.10821. https:\/\/doi.org\/10.48550\/arXiv.2005.10821","DOI":"10.48550\/arXiv.2005.10821"},{"issue":"10","key":"1279_CR48","doi-asserted-by":"publisher","first-page":"3349","DOI":"10.1109\/TPAMI.2020.2983686","volume":"43","author":"J Wang","year":"2020","unstructured":"Wang J, Sun K, Cheng T et al (2020) Deep high-resolution representation learning for visual recognition. IEEE Trans Pattern Anal Mach Intell 43(10):3349\u20133364","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1279_CR49","unstructured":"Everingham M, Winn J (2012) The PASCAL visual object classes challenge 2012 (VOC2012) development kit. In: Pattern Anal. Stat. Model. Comput. Learn., Tech. Rep 2007, pp 1\u201345"},{"key":"1279_CR50","doi-asserted-by":"crossref","unstructured":"Cordts M, Omran M, Ramos S, et al (2016) The Cityscapes dataset for semantic urban scene understanding. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3213\u20133223","DOI":"10.1109\/CVPR.2016.350"},{"issue":"3","key":"1279_CR51","doi-asserted-by":"publisher","first-page":"302","DOI":"10.1007\/s11263-018-1140-0","volume":"127","author":"B Zhou","year":"2019","unstructured":"Zhou B, Zhao H, Puig X et al (2019) Semantic understanding of scenes through the ADE20K dataset. Int J Comput Vis 127(3):302\u2013321. https:\/\/doi.org\/10.1007\/s11263-018-1140-0","journal-title":"Int J Comput Vis"},{"issue":"6","key":"1279_CR52","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky A, Sutskever I, Hinton GE (2017) ImageNet classification with deep convolutional neural networks. Commun ACM 60(6):84\u201390. https:\/\/doi.org\/10.1145\/3065386","journal-title":"Commun ACM"},{"key":"1279_CR53","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R et al (2009) ImageNet: a large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition, pp 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"1279_CR54","doi-asserted-by":"crossref","unstructured":"He J, Deng Z, Qiao Y (2019) Dynamic multi-scale filters for semantic segmentation. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 3562\u20133572","DOI":"10.1109\/ICCV.2019.00366"},{"key":"1279_CR55","doi-asserted-by":"crossref","unstructured":"Zheng S, Lu J, Zhao H, Zhu X, Luo Z, Wang Y et al (2021) Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 6881\u20136890","DOI":"10.1109\/CVPR46437.2021.00681"},{"key":"1279_CR56","doi-asserted-by":"crossref","unstructured":"Li X, You A, Zhu Z, Zhao H, Yang M, Yang K et al (2020) Semantic flow for fast and accurate scene parsing. In: Computer vision\u2013ECCV 2020: 16th European conference, Glasgow, UK, August 23\u201328, 2020, proceedings, part I 16, pp 775\u2013793","DOI":"10.1007\/978-3-030-58452-8_45"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-023-01279-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-023-01279-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-023-01279-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,30]],"date-time":"2024-03-30T15:31:16Z","timestamp":1711812676000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-023-01279-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,4]]},"references-count":56,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["1279"],"URL":"https:\/\/doi.org\/10.1007\/s40747-023-01279-x","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,12,4]]},"assertion":[{"value":"30 June 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 November 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 December 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no conflicts of interest in the publication of this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}