{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T17:06:55Z","timestamp":1772644015561,"version":"3.50.1"},"reference-count":61,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2026,1,7]],"date-time":"2026-01-07T00:00:00Z","timestamp":1767744000000},"content-version":"vor","delay-in-days":6,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"name":"Biological Breeding-National Science and Technology Major Project","award":["2023ZD04076)"],"award-info":[{"award-number":["2023ZD04076)"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1007\/s00371-025-04314-9","type":"journal-article","created":{"date-parts":[[2026,1,7]],"date-time":"2026-01-07T07:21:36Z","timestamp":1767770496000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhancing rice breeding efficiency through semi-supervised detection and segmentation of panicles and leaves"],"prefix":"10.1007","volume":"42","author":[{"given":"Yihong","family":"Hu","sequence":"first","affiliation":[]},{"given":"Ling","family":"Xiong","sequence":"additional","affiliation":[]},{"given":"Bowen","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Peiyi","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Changrong","family":"Ye","sequence":"additional","affiliation":[]},{"given":"Gaofeng","family":"Jia","sequence":"additional","affiliation":[]},{"given":"Bingchuan","family":"Tian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,7]]},"reference":[{"key":"4314_CR1","unstructured":"Jian, L.: Professional periodicals of promoting rice sci-tech innovation and industrial development. J. Library Inform. Sci. Agricult. (2005)"},{"key":"4314_CR2","unstructured":"Xu\u00a0Chunchun, C.Z. Ji\u00a0Long: Analysis of china\u2019s rice industry in 2021 and the outlook for 2022. China Rice (2022)"},{"key":"4314_CR3","doi-asserted-by":"crossref","unstructured":"Sharma, S., Vardhan, M.: Advancing precision agriculture: Enhanced weed detection using the optimized yolov8t model. Arab. J. Sci. Eng. 1\u201318 (2024)","DOI":"10.1007\/s13369-024-09419-2"},{"key":"4314_CR4","doi-asserted-by":"crossref","unstructured":"Choudhury, S.D., Bashyam, S., Qiu, Y., Samal, A., Awada, T.: Holistic and component plant phenotyping using temporal image sequence. Plant Methods 14 (2018)","DOI":"10.1186\/s13007-018-0303-x"},{"issue":"02","key":"4314_CR5","doi-asserted-by":"publisher","first-page":"175","DOI":"10.3724\/SP.T.1006.2015.00175","volume":"41","author":"P Ying-Hong","year":"2015","unstructured":"Ying-Hong, P.: Analysis of concepts and categories of plant phenome and phenomics. IActa Agronomica Sinica 41(02), 175 (2015). https:\/\/doi.org\/10.3724\/SP.T.1006.2015.00175","journal-title":"IActa Agronomica Sinica"},{"key":"4314_CR6","doi-asserted-by":"crossref","unstructured":"Devaki, P., Arunachalam, P., Sankar, K.S.A., Prasanna, M.V.: A deep learning approach for yield estimation and phenotype analysis in rice crops. 2021 International Conference on Advancements in Electrical, Electronics, Communication, Computing and Automation (ICAECA), 1\u20136 (2021)","DOI":"10.1109\/ICAECA52838.2021.9675671"},{"key":"4314_CR7","unstructured":"Liao, H., Huang, H.: The application of fcm-based computer image segmentation technology in agricultural production. Service Oriented Comput. Appl. 1\u201311 (2024)"},{"issue":"9","key":"4314_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.11975\/j.issn.1002-6819.2020.09.001","volume":"36","author":"C Haiyan","year":"2020","unstructured":"Haiyan, C., Yueming, Z., Dawei, S., Li, Z., Liang, W., Zhihong, M., Ziyi, L., Yong, H.: Current status and future perspective of the application of deep learning in plant phenotype research. Trans. Chin. Soc. Agricult. Eng. (Transactions of the CSAE) 36(9), 1\u201316 (2020). https:\/\/doi.org\/10.11975\/j.issn.1002-6819.2020.09.001","journal-title":"Trans. Chin. Soc. Agricult. Eng. (Transactions of the CSAE)"},{"issue":"23","key":"4314_CR9","doi-asserted-by":"publisher","first-page":"168","DOI":"10.11975\/j.issn.1002-6819.2015.23.022","volume":"31","author":"D Zhihua","year":"2015","unstructured":"Zhihua, D., Beibei, W., Yuanyuan, W., Yuquan, W., Xiaoliang, Q.: Skeleton extraction algorithm of corn crop rows based on maximum square. Trans. Chin. Soc. Agricult. Eng. (Transactions of the CSAE) 31(23), 168\u2013172 (2015). https:\/\/doi.org\/10.11975\/j.issn.1002-6819.2015.23.022","journal-title":"Trans. Chin. Soc. Agricult. Eng. (Transactions of the CSAE)"},{"issue":"20","key":"4314_CR10","doi-asserted-by":"publisher","first-page":"156","DOI":"10.11975\/j.issn.1002-6819.2022.20.018","volume":"38","author":"S Chenxu","year":"2022","unstructured":"Chenxu, S., Chongyu, Y., Yongchao, X., Sumei, L., Hong, H., Hui, Y., Xianzhong, F.: Algorith for acquiring multi-phenotype parameters of soybean seed based on opencv. Trans. Chin. Soc. Agricult. Eng. (Transactions of the CSAE) 38(20), 156\u2013163 (2022). https:\/\/doi.org\/10.11975\/j.issn.1002-6819.2022.20.018","journal-title":"Trans. Chin. Soc. Agricult. Eng. (Transactions of the CSAE)"},{"issue":"2","key":"4314_CR11","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s42979-024-03520-x","volume":"6","author":"S Sonawane","year":"2025","unstructured":"Sonawane, S., Patil, N.N.: Performance evaluation of modified yolov5 object detectors for crop-weed classification and detection in agriculture images. SN Comput. Sci. 6(2), 1\u201312 (2025)","journal-title":"SN Comput. Sci."},{"issue":"11","key":"4314_CR12","doi-asserted-by":"publisher","first-page":"4037","DOI":"10.1109\/TPAMI.2020.2992393","volume":"43","author":"L Jing","year":"2020","unstructured":"Jing, L., Tian, Y.: Self-supervised visual feature learning with deep neural networks: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 43(11), 4037\u20134058 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4314_CR13","doi-asserted-by":"publisher","first-page":"261","DOI":"10.1007\/s11263-019-01247-4","volume":"128","author":"L Liu","year":"2020","unstructured":"Liu, L., Ouyang, W., Wang, X., Fieguth, P., Chen, J., Liu, X., Pietik\u00e4inen, M.: Deep learning for generic object detection: A survey. Int. J. Comput. Vision 128, 261\u2013318 (2020)","journal-title":"Int. J. Comput. Vision"},{"issue":"7","key":"4314_CR14","first-page":"3523","volume":"44","author":"S Minaee","year":"2021","unstructured":"Minaee, S., Boykov, Y., Porikli, F., Plaza, A., Kehtarnavaz, N., Terzopoulos, D.: Image segmentation using deep learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 44(7), 3523\u20133542 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4314_CR15","doi-asserted-by":"crossref","unstructured":"Misra, I., Maaten, L.v.d.: Self-supervised learning of pretext-invariant representations. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6707\u20136717 (2020)","DOI":"10.1109\/CVPR42600.2020.00674"},{"key":"4314_CR16","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.: Masked autoencoders are scalable vision learners. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16000\u201316009 (2022)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"4314_CR17","doi-asserted-by":"crossref","unstructured":"Girdhar, R., El-Nouby, A., Singh, M., Alwala, K.V., Joulin, A., Misra, I.: Omnimae: Single model masked pretraining on images and videos. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10406\u201310417 (2023)","DOI":"10.1109\/CVPR52729.2023.01003"},{"key":"4314_CR18","unstructured":"Koroteev, M.V.: Bert: a review of applications in natural language processing and understanding (2021). arXiv preprint arXiv:2103.11943"},{"key":"4314_CR19","unstructured":"Zhang, H., Cao, Y.: Understanding the benefits of simclr pre-training in two-layer convolutional neural networks (2024). arXiv preprint arXiv:2409.18685"},{"key":"4314_CR20","unstructured":"Zhang, H., Li, F., Liu, S., Zhang, L., Su, H., Zhu, J., Ni, L.M., Shum, H.-Y.: Dino: Detr with improved denoising anchor boxes for end-to-end object detection (2022). arXiv preprint arXiv:2203.03605"},{"issue":"11","key":"4314_CR21","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial networks. Commun. ACM 63(11), 139\u2013144 (2020)","journal-title":"Commun. ACM"},{"key":"4314_CR22","doi-asserted-by":"crossref","unstructured":"Feng, Y., Shen, A., Hu, J., Liang, Y., Wang, S., Du, J.: Enhancing few-shot learning with integrated data and gan model approaches (2024). arXiv preprint arXiv:2411.16567","DOI":"10.1109\/DSInS64146.2024.10992169"},{"key":"4314_CR23","unstructured":"Shen, F., Ye, H., Zhang, J., Wang, C., Han, X., Yang, W.: Advancing pose-guided image synthesis with progressive conditional diffusion models (2023). arXiv preprint arXiv:2310.06313"},{"key":"4314_CR24","doi-asserted-by":"crossref","unstructured":"Shen, F., Jiang, X., He, X., Ye, H., Wang, C., Du, X., Li, Z., Tang, J.: Imagdressing-v1: Customizable virtual dressing (2024). arXiv preprint arXiv:2407.12705","DOI":"10.1609\/aaai.v39i7.32729"},{"key":"4314_CR25","doi-asserted-by":"crossref","unstructured":"Shen, F., Tang, J.: Imagpose: A unified conditional framework for pose-guided person generation. In: The Thirty-eighth Annual Conference on Neural Information Processing Systems (2024)","DOI":"10.52202\/079017-0202"},{"key":"4314_CR26","doi-asserted-by":"crossref","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: European Conference on Computer Vision, pp. 213\u2013229 (2020). Springer","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"4314_CR27","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580\u2013587 (2014)","DOI":"10.1109\/CVPR.2014.81"},{"key":"4314_CR28","doi-asserted-by":"crossref","unstructured":"Redmon, J.: You only look once: Unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"4314_CR29","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., Berg, A.C.: Ssd: Single shot multibox detector. In: Computer Vision\u2013ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part I 14, pp. 21\u201337 (2016). Springer","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"4314_CR30","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"4314_CR31","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., Dai, J.: Deformable detr: Deformable transformers for end-to-end object detection (2020). ArXiv arXiv:abs\/2010.04159"},{"key":"4314_CR32","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2021","unstructured":"Lin, X., Sun, S., Huang, W., Sheng, B., Li, P., Feng, D.D.: Eapt: efficient attention pyramid transformer for image processing. IEEE Trans. Multimedia 25, 50\u201361 (2021)","journal-title":"IEEE Trans. Multimedia"},{"issue":"4","key":"4314_CR33","doi-asserted-by":"publisher","first-page":"366","DOI":"10.1016\/j.vrih.2022.08.010","volume":"5","author":"X Hu","year":"2023","unstructured":"Hu, X., Bao, X., Wei, G., Li, Z.: Human-pose estimation based on weak supervision. Virtual Reality & Intell. Hardw. 5(4), 366\u2013377 (2023)","journal-title":"Virtual Reality & Intell. Hardw."},{"key":"4314_CR34","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A.C., Lo, W.-Y., et al: Segment anything. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4015\u20134026 (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"4314_CR35","doi-asserted-by":"crossref","unstructured":"Wang, X., Yu, Z., De\u00a0Mello, S., Kautz, J., Anandkumar, A., Shen, C., Alvarez, J.M.: Freesolo: Learning to segment objects without annotations. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14176\u201314186 (2022)","DOI":"10.1109\/CVPR52688.2022.01378"},{"key":"4314_CR36","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-assisted intervention\u2013MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pp. 234\u2013241 (2015). Springer","DOI":"10.1007\/978-3-319-24574-4_28"},{"issue":"4","key":"4314_CR37","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1109\/TPAMI.2017.2699184","volume":"40","author":"L-C Chen","year":"2017","unstructured":"Chen, L.-C., Papandreou, G., Kokkinos, I., Murphy, K., Yuille, A.L.: Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 40(4), 834\u2013848 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4314_CR38","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961\u20132969 (2017)","DOI":"10.1109\/ICCV.2017.322"},{"key":"4314_CR39","doi-asserted-by":"crossref","unstructured":"Cheng, B., Misra, I., Schwing, A.G., Kirillov, A., Girdhar, R.: Masked-attention mask transformer for universal image segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1290\u20131299 (2022)","DOI":"10.1109\/CVPR52688.2022.00135"},{"key":"4314_CR40","doi-asserted-by":"crossref","unstructured":"Li, F., Zhang, H., Xu, H., Liu, S., Zhang, L., Ni, L.M., Shum, H.-Y.: Mask dino: Towards a unified transformer-based framework for object detection and segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3041\u20133050 (2023)","DOI":"10.1109\/CVPR52729.2023.00297"},{"key":"4314_CR41","doi-asserted-by":"crossref","unstructured":"Ali, S.G., Wang, X., Li, P., Li, H., Yang, P., Jung, Y., Qin, J., Kim, J., Sheng, B.: Egdnet: an efficient glomerular detection network for multiple anomalous pathological feature in glomerulonephritis. Visual Compu. 1\u201318 (2024)","DOI":"10.1007\/s00371-024-03570-5"},{"key":"4314_CR42","doi-asserted-by":"publisher","first-page":"880","DOI":"10.1109\/TIP.2021.3136619","volume":"31","author":"A Nazir","year":"2021","unstructured":"Nazir, A., Cheema, M.N., Sheng, B., Li, P., Li, H., Xue, G., Qin, J., Kim, J., Feng, D.D.: Ecsu-net: an embedded clustering sliced u-net coupled with fusing strategy for efficient intervertebral disc segmentation and classification. IEEE Trans. Image Process. 31, 880\u2013893 (2021)","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"4314_CR43","doi-asserted-by":"publisher","first-page":"2201","DOI":"10.1002\/cav.2201","volume":"35","author":"X Zhu","year":"2024","unstructured":"Zhu, X., Yao, X., Zhang, J., Zhu, M., You, L., Yang, X., Zhang, J., Zhao, H., Zeng, D.: Tmsdnet: Transformer with multi-scale dense network for single and multi-view 3d reconstruction. Comput. Animation Virtual Worlds 35(1), 2201 (2024)","journal-title":"Comput. Animation Virtual Worlds"},{"issue":"1","key":"4314_CR44","doi-asserted-by":"publisher","first-page":"2223","DOI":"10.1002\/cav.2223","volume":"35","author":"MA Guerroudji","year":"2024","unstructured":"Guerroudji, M.A., Amara, K., Lichouri, M., Zenati, N., Masmoudi, M.: A 3d visualization-based augmented reality application for brain tumor segmentation. Comput. Animation Virtual Worlds 35(1), 2223 (2024)","journal-title":"Comput. Animation Virtual Worlds"},{"key":"4314_CR45","unstructured":"Shen, F., Du, X., Gao, Y., Yu, J., Cao, Y., Lei, X., Tang, J.: Imagharmony: Controllable image editing with consistent object quantity and layout (2025). arXiv preprint arXiv:2506.01949"},{"key":"4314_CR46","unstructured":"Shen, F., Yu, J., Wang, C., Jiang, X., Du, X., Tang, J.: Imaggarment-1: Fine-grained garment generation for controllable fashion design (2025). arXiv preprint arXiv:2504.13176"},{"key":"4314_CR47","unstructured":"Shen, F., Wang, C., Gao, J., Guo, Q., Dang, J., Tang, J., Chua, T.-S.: Long-term talkingface generation via motion-prior conditional diffusion model (2025). arXiv preprint arXiv:2502.09533"},{"issue":"10","key":"4314_CR48","doi-asserted-by":"publisher","first-page":"083","DOI":"10.1093\/gigascience\/gix083","volume":"6","author":"MP Pound","year":"2017","unstructured":"Pound, M.P., Atkinson, J.A., Townsend, A.J., Wilson, M.H., Griffiths, M., Jackson, A.S., Bulat, A., Tzimiropoulos, G., Wells, D.M., Murchie, E.H., et al.: Deep machine learning provides state-of-the-art performance in image-based plant phenotyping. Gigascience 6(10), 083 (2017)","journal-title":"Gigascience"},{"key":"4314_CR49","unstructured":"Luc, P., et al.: Semantic segmentation using adversarial networks (2016). arXiv preprint arXiv:1611.08408"},{"key":"4314_CR50","unstructured":"Ultralytics: YOLOv5: A PyTorch implementation of the YOLO architecture (2020). https:\/\/github.com\/ultralytics\/yolov5. Accessed: [Insert Date]"},{"key":"4314_CR51","doi-asserted-by":"crossref","unstructured":"Wang, C.-Y., Bochkovskiy, A., Liao, H.-Y.M.: Yolov7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors (2022). arXiv preprint arXiv:2207.02696","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"4314_CR52","unstructured":"Zhu, X.: Semi-supervised learning literature survey. University of Wisconsin-Madison (2005)"},{"key":"4314_CR53","unstructured":"Oquab, M., Darcet, T., Moutakanni, T., Vo, H., Szafraniec, M., Khalidov, V., Fernandez, P., Haziza, D., Massa, F., El-Nouby, A., et al.: Dinov2: Learning robust visual features without supervision (2023). arXiv preprint arXiv:2304.07193"},{"key":"4314_CR54","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N.: An image is worth 16x16 words: Transformers for image recognition at scale. ArXiv abs\/2010.11929 (2020)"},{"key":"4314_CR55","unstructured":"Zhou, J., Wei, C., Wang, H., Shen, W., Xie, C., Yuille, A.L., Kong, T.: ibot: Image bert pre-training with online tokenizer. ArXiv abs\/2111.07832 (2021)"},{"key":"4314_CR56","unstructured":"Caron, M., Misra, I., Mairal, J., Goyal, P., Bojanowski, P., Joulin, A.: Unsupervised learning of visual features by contrasting cluster assignments (2020). arXiv: abs\/2006.09882"},{"key":"4314_CR57","doi-asserted-by":"crossref","unstructured":"Caron, M., Touvron, H., Misra, I., J\u2019egou, H., Mairal, J., Bojanowski, P., Joulin, A.: Emerging properties in self-supervised vision transformers. 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), 9630\u20139640 (2021)","DOI":"10.1109\/ICCV48922.2021.00951"},{"key":"4314_CR58","unstructured":"Sablayrolles, A., Douze, M., Schmid, C., J\u00e9gou, H.: Spreading vectors for similarity search. arXiv: Machine Learning (2018)"},{"key":"4314_CR59","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., Dai, J.: Deformable detr: Deformable transformers for end-to-end object detection (2020). arXiv preprint arXiv:2010.04159"},{"key":"4314_CR60","first-page":"17864","volume":"34","author":"B Cheng","year":"2021","unstructured":"Cheng, B., Schwing, A., Kirillov, A.: Per-pixel classification is not all you need for semantic segmentation. Adv. Neural. Inf. Process. Syst. 34, 17864\u201317875 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"6","key":"4314_CR61","doi-asserted-by":"publisher","first-page":"471","DOI":"10.1016\/j.vrih.2023.06.009","volume":"5","author":"X Tian","year":"2023","unstructured":"Tian, X., Wu, Z., Cao, J., Chen, S., Dong, X.: Ilidviz: an incremental learning-based visual analysis system for network anomaly detection. Virtual Reality & Intelligent Hardw. 5(6), 471\u2013489 (2023)","journal-title":"Virtual Reality & Intelligent Hardw."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04314-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04314-9","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04314-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T13:03:18Z","timestamp":1772629398000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04314-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1]]},"references-count":61,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["4314"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04314-9","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1]]},"assertion":[{"value":"25 February 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 December 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"112"}}