{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T07:50:57Z","timestamp":1767340257889,"version":"3.40.3"},"publisher-location":"Cham","reference-count":57,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031729515"},{"type":"electronic","value":"9783031729522"}],"license":[{"start":{"date-parts":[[2024,10,1]],"date-time":"2024-10-01T00:00:00Z","timestamp":1727740800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,1]],"date-time":"2024-10-01T00:00:00Z","timestamp":1727740800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72952-2_2","type":"book-chapter","created":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T05:02:02Z","timestamp":1727672522000},"page":"18-35","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["Rethinking and\u00a0Improving Visual Prompt Selection for\u00a0In-Context Learning Segmentation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8283-8637","authenticated-orcid":false,"given":"Wei","family":"Suo","sequence":"first","affiliation":[]},{"given":"Lanqing","family":"Lai","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0638-3295","authenticated-orcid":false,"given":"Mengyang","family":"Sun","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7374-8739","authenticated-orcid":false,"given":"Hanwang","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Peng","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2977-8057","authenticated-orcid":false,"given":"Yanning","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,1]]},"reference":[{"key":"2_CR1","doi-asserted-by":"crossref","unstructured":"Agrawal, S., Zhou, C., Lewis, M., Zettlemoyer, L., Ghazvininejad, M.: In-context examples selection for machine translation. arXiv preprint arXiv:2212.02437 (2022)","DOI":"10.18653\/v1\/2023.findings-acl.564"},{"key":"2_CR2","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"427","DOI":"10.1007\/978-3-031-20056-4_25","volume-title":"ECCV 2022","author":"X Wang","year":"2022","unstructured":"Wang, X., Lian, L., Yu, S.X.: Unsupervised selective labeling for more effective semi-supervised learning. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13690, pp. 427\u2013445. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20056-4_25"},{"key":"2_CR3","doi-asserted-by":"crossref","unstructured":"Wang, X., et al.: SegGPT: segmenting everything in context. In: ICCV 2023 (2023)","DOI":"10.1109\/ICCV51070.2023.00110"},{"key":"2_CR4","doi-asserted-by":"crossref","unstructured":"Anderson, P., et al.: Bottom-up and top-down attention for image captioning and visual question answering. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6077\u20136086 (2018)","DOI":"10.1109\/CVPR.2018.00636"},{"issue":"12","key":"2_CR5","doi-asserted-by":"publisher","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","volume":"39","author":"V Badrinarayanan","year":"2017","unstructured":"Badrinarayanan, V., Kendall, A., Cipolla, R.: Segnet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481\u20132495 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2_CR6","unstructured":"Bar, A., Gandelsman, Y., Darrell, T., Globerson, A., Efros, A.: Visual prompting via image inpainting. In: Advances in Neural Information Processing Systems, vol. 35, pp. 25005\u201325017 (2022)"},{"key":"2_CR7","unstructured":"Brown, T., et al.: Language models are few-shot learners. In: Advances in Neural Information Processing Systems, vol. 33, pp. 1877\u20131901 (2020)"},{"key":"2_CR8","unstructured":"Dosovitskiy, A., et\u00a0al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"2_CR9","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","volume":"88","author":"M Everingham","year":"2010","unstructured":"Everingham, M., Van Gool, L., Williams, C.K., Winn, J., Zisserman, A.: The pascal visual object classes (VOC) challenge. Int. J. Comput. Vision 88, 303\u2013338 (2010)","journal-title":"Int. J. Comput. Vision"},{"key":"2_CR10","doi-asserted-by":"crossref","unstructured":"Hariharan, B., Arbel\u00e1ez, P., Bourdev, L., Maji, S., Malik, J.: Semantic contours from inverse detectors. In: 2011 International Conference on Computer Vision, pp. 991\u2013998. IEEE (2011)","DOI":"10.1109\/ICCV.2011.6126343"},{"key":"2_CR11","doi-asserted-by":"crossref","unstructured":"Haussmann, E., et al.: Scalable active learning for object detection. In: 2020 IEEE Intelligent Vehicles Symposium (IV), pp. 1430\u20131435. IEEE (2020)","DOI":"10.1109\/IV47402.2020.9304793"},{"key":"2_CR12","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961\u20132969 (2017)","DOI":"10.1109\/ICCV.2017.322"},{"key":"2_CR13","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"108","DOI":"10.1007\/978-3-031-19818-2_7","volume-title":"ECCV 2022","author":"S Hong","year":"2022","unstructured":"Hong, S., Cho, S., Nam, J., Lin, S., Kim, S.: Cost aggregation with 4D convolutional swin transformer for few-shot segmentation. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13689, pp. 108\u2013126. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19818-2_7"},{"key":"2_CR14","unstructured":"Hongjin, S., et al.: Selective annotation makes language models better few-shot learners. In: The Eleventh International Conference on Learning Representations (2022)"},{"key":"2_CR15","doi-asserted-by":"crossref","unstructured":"Hore, A., Ziou, D.: Image quality metrics: PSNR vs. SSIM. In: 2010 20th International Conference on Pattern Recognition, pp. 2366\u20132369. IEEE (2010)","DOI":"10.1109\/ICPR.2010.579"},{"key":"2_CR16","doi-asserted-by":"publisher","first-page":"14606","DOI":"10.1109\/ACCESS.2020.2964760","volume":"8","author":"H Hu","year":"2020","unstructured":"Hu, H., Li, Z., Li, L., Yang, H., Zhu, H.: Classification of very high-resolution remote sensing imagery using a fully convolutional network with global and local context information enhancements. IEEE Access 8, 14606\u201314619 (2020)","journal-title":"IEEE Access"},{"key":"2_CR17","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"2_CR18","unstructured":"Kirillov, A., et al.: Segment anything. arXiv preprint arXiv:2304.02643 (2023)"},{"key":"2_CR19","unstructured":"Li, X., et al.: Unified demonstration retriever for in-context learning. arXiv preprint arXiv:2305.04320 (2023)"},{"key":"2_CR20","doi-asserted-by":"crossref","unstructured":"Li, Y., Padmanabhan, A., Zhao, P., Wang, Y., Xu, G.H., Netravali, R.: Reducto: on-camera filtering for resource-efficient real-time video analytics. In: Proceedings of the Annual Conference of the ACM Special Interest Group on Data Communication on the Applications, Technologies, Architectures, and Protocols for Computer Communication, pp. 359\u2013376 (2020)","DOI":"10.1145\/3387514.3405874"},{"key":"2_CR21","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"2_CR22","doi-asserted-by":"crossref","unstructured":"Liu, J., Shen, D., Zhang, Y., Dolan, B., Carin, L., Chen, W.: What makes good in-context examples for GPT-$$3 $$? arXiv preprint arXiv:2101.06804 (2021)","DOI":"10.18653\/v1\/2022.deelio-1.10"},{"key":"2_CR23","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3431\u20133440 (2015)","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"2_CR24","doi-asserted-by":"crossref","unstructured":"Luo, G., et al.: Multi-task collaborative network for joint referring expression comprehension and segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10034\u201310043 (2020)","DOI":"10.1109\/CVPR42600.2020.01005"},{"key":"2_CR25","doi-asserted-by":"publisher","unstructured":"Lyu, X., Fang, Y., Tong, B., Li, X., Zeng, T.: Multiscale normalization attention network for water body extraction from remote sensing imagery. Remote Sens. 14(19) (2022). https:\/\/doi.org\/10.3390\/rs14194983. https:\/\/www.mdpi.com\/2072-4292\/14\/19\/4983","DOI":"10.3390\/rs14194983"},{"key":"2_CR26","doi-asserted-by":"crossref","unstructured":"Min, S., et al.: Rethinking the role of demonstrations: what makes in-context learning work? arXiv preprint arXiv:2202.12837 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.759"},{"key":"2_CR27","unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"2_CR28","doi-asserted-by":"publisher","first-page":"33","DOI":"10.1016\/j.image.2017.11.001","volume":"61","author":"R Reisenhofer","year":"2018","unstructured":"Reisenhofer, R., Bosse, S., Kutyniok, G., Wiegand, T.: A HAAR wavelet-based perceptual similarity index for image quality assessment. Signal Process. Image Commun. 61, 33\u201343 (2018)","journal-title":"Signal Process. Image Commun."},{"key":"2_CR29","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"2_CR30","doi-asserted-by":"crossref","unstructured":"Rubin, O., Herzig, J., Berant, J.: Learning to retrieve prompts for in-context learning. arXiv preprint arXiv:2112.08633 (2021)","DOI":"10.18653\/v1\/2022.naacl-main.191"},{"key":"2_CR31","doi-asserted-by":"crossref","unstructured":"Rubin, O., Herzig, J., Berant, J.: Learning to retrieve prompts for in-context learning. In: Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 2655\u20132671 (2022)","DOI":"10.18653\/v1\/2022.naacl-main.191"},{"key":"2_CR32","doi-asserted-by":"crossref","unstructured":"Shaban, A., Bansal, S., Liu, Z., Essa, I., Boots, B.: One-shot learning for semantic segmentation. arXiv preprint arXiv:1709.03410 (2017)","DOI":"10.5244\/C.31.167"},{"key":"2_CR33","doi-asserted-by":"publisher","DOI":"10.1016\/j.apenergy.2019.114199","volume":"259","author":"F Shen","year":"2020","unstructured":"Shen, F., Zhao, L., Du, W., Zhong, W., Qian, F.: Large-scale industrial energy systems optimization under uncertainty: a data-driven robust optimization approach. Appl. Energy 259, 114199 (2020)","journal-title":"Appl. Energy"},{"key":"2_CR34","doi-asserted-by":"crossref","unstructured":"Shi, H., Li, H., Meng, F., Wu, Q.: Key-word-aware network for referring expression image segmentation. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 38\u201354 (2018)","DOI":"10.1007\/978-3-030-01231-1_3"},{"issue":"4","key":"2_CR35","doi-asserted-by":"publisher","first-page":"5558","DOI":"10.1109\/LRA.2020.3007457","volume":"5","author":"L Sun","year":"2020","unstructured":"Sun, L., Yang, K., Hu, X., Hu, W., Wang, K.: Real-time fusion network for RGB-D semantic segmentation incorporating unexpected obstacle detection for road-driving images. IEEE Robot. Autom. Lett. 5(4), 5558\u20135565 (2020)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"2_CR36","doi-asserted-by":"crossref","unstructured":"Sun, M., et al.: An adaptive correlation filtering method for text-based person search. Int. J. Comput. Vision 1\u201316 (2024)","DOI":"10.1007\/s11263-024-02094-8"},{"key":"2_CR37","unstructured":"Sun, Y., Chen, Q., Wang, J., Wang, J., Li, Z.: Exploring effective factors for improving visual in-context learning. arXiv preprint arXiv:2304.04748 (2023)"},{"key":"2_CR38","doi-asserted-by":"crossref","unstructured":"Suo, W., et al.: S3C: semi-supervised VQA natural language explanation via self-critical learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2646\u20132656 (2023)","DOI":"10.1109\/CVPR52729.2023.00260"},{"key":"2_CR39","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"726","DOI":"10.1007\/978-3-031-19833-5_42","volume-title":"Computer Vision \u2013 ECCV 2022","author":"W Suo","year":"2022","unstructured":"Suo, W., et al.: A simple and robust correlation filtering method for text-based person search. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13695, pp. 726\u2013742. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19833-5_42"},{"key":"2_CR40","doi-asserted-by":"publisher","first-page":"854","DOI":"10.1109\/TIP.2022.3227466","volume":"32","author":"W Suo","year":"2022","unstructured":"Suo, W., Sun, M., Wang, P., Zhang, Y., Wu, Q.: Rethinking and improving feature pyramids for one-stage referring expression comprehension. IEEE Trans. Image Process. 32, 854\u2013864 (2022)","journal-title":"IEEE Trans. Image Process."},{"key":"2_CR41","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"638","DOI":"10.1007\/978-3-030-58548-8_37","volume-title":"Computer Vision \u2013 ECCV 2020","author":"AB Vasudevan","year":"2020","unstructured":"Vasudevan, A.B., Dai, D., Van Gool, L.: Semantic object prediction and spatial sound super-resolution with binaural sounds. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12349, pp. 638\u2013655. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58548-8_37"},{"key":"2_CR42","doi-asserted-by":"crossref","unstructured":"Wang, W., et al.: Internimage: exploring large-scale vision foundation models with deformable convolutions. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14408\u201314419 (2023)","DOI":"10.1109\/CVPR52729.2023.01385"},{"key":"2_CR43","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1016\/j.neucom.2019.11.019","volume":"381","author":"X Wang","year":"2020","unstructured":"Wang, X., Ma, H., You, S.: Deep clustering for weakly-supervised semantic segmentation in autonomous driving scenes. Neurocomputing 381, 20\u201328 (2020)","journal-title":"Neurocomputing"},{"key":"2_CR44","doi-asserted-by":"crossref","unstructured":"Wang, X., Wang, W., Cao, Y., Shen, C., Huang, T.: Images speak in images: a generalist painter for in-context visual learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6830\u20136839 (2023)","DOI":"10.1109\/CVPR52729.2023.00660"},{"key":"2_CR45","doi-asserted-by":"crossref","unstructured":"Wang, X., Zhang, X., Cao, Y., Wang, W., Shen, C., Huang, T.: Seggpt: segmenting everything in context. arXiv preprint arXiv:2304.03284 (2023)","DOI":"10.1109\/ICCV51070.2023.00110"},{"issue":"4","key":"2_CR46","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"2_CR47","unstructured":"Waqas\u00a0Zamir, S., et al.: iSAID: a large-scale dataset for instance segmentation in aerial images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 28\u201337 (2019)"},{"key":"2_CR48","unstructured":"Wei, J., et al.: Chain-of-thought prompting elicits reasoning in large language models. In: Advances in Neural Information Processing Systems, vol. 35, pp. 24824\u201324837 (2022)"},{"key":"2_CR49","first-page":"1","volume":"60","author":"X Yao","year":"2021","unstructured":"Yao, X., Cao, Q., Feng, X., Cheng, G., Han, J.: Scale-aware detailed matching for few-shot aerial image semantic segmentation. IEEE Trans. Geosci. Remote Sens. 60, 1\u201311 (2021)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"2_CR50","unstructured":"Ye, J., Wu, Z., Feng, J., Yu, T., Kong, L.: Compositional exemplars for in-context learning. arXiv preprint arXiv:2302.05698 (2023)"},{"key":"2_CR51","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107756","volume":"113","author":"Q Yu","year":"2021","unstructured":"Yu, Q., Gao, Y., Zheng, Y., Zhu, J., Dai, Y., Shi, Y.: Crossover-net: leveraging vertical-horizontal crossover relation for robust medical image segmentation. Pattern Recogn. 113, 107756 (2021)","journal-title":"Pattern Recogn."},{"key":"2_CR52","unstructured":"Yu, Y., Yang, X., Li, Q., Zhou, Y., Da, F., Yan, J.: H2rbox-v2: incorporating symmetry for boosting horizontal box supervised oriented object detection. In: Thirty-Seventh Conference on Neural Information Processing Systems (2023)"},{"key":"2_CR53","doi-asserted-by":"crossref","unstructured":"Zendel, O., Sch\u00f6rghuber, M., Rainer, B., Murschitz, M., Beleznai, C.: Unifying panoptic segmentation for autonomous driving. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 21351\u201321360 (2022)","DOI":"10.1109\/CVPR52688.2022.02066"},{"issue":"8","key":"2_CR54","doi-asserted-by":"publisher","first-page":"2378","DOI":"10.1109\/TIP.2011.2109730","volume":"20","author":"L Zhang","year":"2011","unstructured":"Zhang, L., Zhang, L., Mou, X., Zhang, D.: FSIM: a feature similarity index for image quality assessment. IEEE Trans. Image Process. 20(8), 2378\u20132386 (2011)","journal-title":"IEEE Trans. Image Process."},{"key":"2_CR55","unstructured":"Zhang, Y., Zhou, K., Liu, Z.: What makes good examples for visual in-context learning? arXiv preprint arXiv:2301.13670 (2023)"},{"key":"2_CR56","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1007\/978-3-030-87193-2_2","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Liu, H., Hu, Q.: TransFuse: fusing transformers and CNNs for medical image segmentation. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12901, pp. 14\u201324. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_2"},{"key":"2_CR57","unstructured":"Zhang, Z., Zhang, A., Li, M., Smola, A.: Automatic chain of thought prompting in large language models. arXiv preprint arXiv:2210.03493 (2022)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72952-2_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T05:02:40Z","timestamp":1727672560000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72952-2_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,1]]},"ISBN":["9783031729515","9783031729522"],"references-count":57,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72952-2_2","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,10,1]]},"assertion":[{"value":"1 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}