{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T14:17:47Z","timestamp":1742912267424,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":37,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819786190"},{"type":"electronic","value":"9789819786206"}],"license":[{"start":{"date-parts":[[2024,10,20]],"date-time":"2024-10-20T00:00:00Z","timestamp":1729382400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,20]],"date-time":"2024-10-20T00:00:00Z","timestamp":1729382400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-97-8620-6_21","type":"book-chapter","created":{"date-parts":[[2024,10,19]],"date-time":"2024-10-19T21:02:10Z","timestamp":1729371730000},"page":"304-317","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Deep Noisy Multi-label Learning for\u00a0Robust Cross-Modal Retrieval"],"prefix":"10.1007","author":[{"given":"Ruitao","family":"Pu","sequence":"first","affiliation":[]},{"given":"Dezhong","family":"Peng","sequence":"additional","affiliation":[]},{"given":"Fujun","family":"Hua","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,20]]},"reference":[{"key":"21_CR1","doi-asserted-by":"crossref","unstructured":"Chatfield, K., Simonyan, K., Vedaldi, A., Zisserman, A.: Return of the devil in the details: delving deep into convolutional nets. In: Proceedings of the British Machine Vision Conference 2014. British Machine Vision Association (2014)","DOI":"10.5244\/C.28.6"},{"key":"21_CR2","doi-asserted-by":"crossref","unstructured":"Chua, T.S., Tang, J., Hong, R., Li, H., Luo, Z., Zheng, Y.: Nus-wide: a real-world web image database from national university of Singapore. In: Proceedings of the ACM International Conference on Image and Video Retrieval (2009)","DOI":"10.1145\/1646396.1646452"},{"key":"21_CR3","doi-asserted-by":"crossref","unstructured":"Escalante, H.J., Hern\u00e1ndez, C.A., Gonzalez, J.A., L\u00f3pez-L\u00f3pez, A., Montes, M., Morales, E.F., Sucar, L.E., Villasenor, L., Grubinger, M.: The segmented and annotated IAPR TC-12 benchmark. Comput. Vis. Image Underst. 114(4), 419\u2013428 (2010)","DOI":"10.1016\/j.cviu.2009.03.008"},{"key":"21_CR4","doi-asserted-by":"crossref","unstructured":"Feng, Y., Zhu, H., Peng, D., Peng, X., Hu, P.: RONO: robust discriminative learning with noisy labels for 2D-3D cross-modal retrieval. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11610\u201311619 (2023)","DOI":"10.1109\/CVPR52729.2023.01117"},{"key":"21_CR5","doi-asserted-by":"crossref","unstructured":"Ghosh, A., Kumar, H., Sastry, P.S. Robust loss functions under label noise for deep neural networks. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a031 (2017)","DOI":"10.1609\/aaai.v31i1.10894"},{"key":"21_CR6","doi-asserted-by":"crossref","unstructured":"Hotelling, H.: Relations between two sets of variates. Biometrika 321\u2013377","DOI":"10.1093\/biomet\/28.3-4.321"},{"key":"21_CR7","doi-asserted-by":"crossref","unstructured":"Hu, P., Peng, X., Zhu, H., Zhen, L., Lin, J.: Learning cross-modal retrieval with noisy labels. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5403\u20135413 (2021a)","DOI":"10.1109\/CVPR46437.2021.00536"},{"key":"21_CR8","doi-asserted-by":"crossref","unstructured":"Hu, P., Peng, X., Zhu, H., Zhen, L., Lin, J.: Learning cross-modal retrieval with noisy labels. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5403\u20135413 (2021b)","DOI":"10.1109\/CVPR46437.2021.00536"},{"issue":"3","key":"21_CR9","first-page":"3877","volume":"45","author":"H Peng","year":"2023","unstructured":"Peng, H., Zhu, H., Lin, J., Peng, D., Zhao, Y.-P., Peng, X.: Unsupervised contrastive cross-modal hashing. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 3877\u20133889 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"21_CR10","doi-asserted-by":"crossref","unstructured":"Kuznetsova, A., Rom, H., Alldrin, N., Uijlings, J., Krasin, I., Pont-Tuset, J., Kamali, S., Popov, S., Malloci, M., Kolesnikov, A., Duerig, T., Ferrari, V.: The open images dataset v4: unified image classification, object detection, and visual relationship detection at scale. Int. J. Comput. Vis., Int. J. Comput. Vis. (2020)","DOI":"10.1007\/s11263-020-01316-z"},{"key":"21_CR11","doi-asserted-by":"crossref","unstructured":"Lau, J.H., Baldwin, T.: An empirical evaluation of doc2vec with practical insights into document embedding generation. Cornell University (2016)","DOI":"10.18653\/v1\/W16-1609"},{"key":"21_CR12","unstructured":"Li, J., Socher, R., Hoi, S.C.: Dividemix: learning with noisy labels as semi-supervised learning (2020). arXiv:2002.07394"},{"key":"21_CR13","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft COCO: Common Objects in Context, pp. 740\u2013755 (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"21_CR14","doi-asserted-by":"publisher","first-page":"2851","DOI":"10.1109\/TMM.2022.3152086","volume":"25","author":"Y Liu","year":"2023","unstructured":"Liu, Y., Jianlong, W., Leigang, Q., Gan, T., Yin, J., Nie, L.: Self-supervised correlation learning for cross-modal retrieval. IEEE Trans. Multimedia 25, 2851\u20132863 (2023)","journal-title":"IEEE Trans. Multimedia"},{"key":"21_CR15","doi-asserted-by":"crossref","unstructured":"Lu, Y., He, W.: SELC: self-ensemble label correction improves learning with noisy labels (2022). arXiv:2205.01156","DOI":"10.24963\/ijcai.2022\/455"},{"key":"21_CR16","unstructured":"Lyu, Y., Tsang, I.W.: Curriculum loss: robust learning and generalization against label corruption. Learning (2019)"},{"issue":"3","key":"21_CR17","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1109\/83.988962","volume":"11","author":"AA Nielsen","year":"2002","unstructured":"Nielsen, A.A.: Multiset canonical correlations analysis and multispectral, truly multitemporal remote sensing data. IEEE Trans. Image Process. 11(3), 293\u2013305 (2002)","journal-title":"IEEE Trans. Image Process."},{"key":"21_CR18","doi-asserted-by":"crossref","unstructured":"Patrini, G., Rozza, A., Krishna Menon, A., Nock, R., Qu, L.: Making deep neural networks robust to label noise: a loss correction approach. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.240"},{"key":"21_CR19","unstructured":"Pene, C.O., Ghiassi, A., Younesian, T., Birke, R., Chen, L.Y.: Multi-label gold asymmetric loss correction with single-label regulators (2021). arXiv:2108.02032"},{"key":"21_CR20","doi-asserted-by":"publisher","first-page":"3520","DOI":"10.1109\/TMM.2021.3101642","volume":"24","author":"S Qian","year":"2021","unstructured":"Qian, S., Xue, D., Fang, Q., Changsheng, X.: Adaptive label-aware graph convolutional networks for cross-modal retrieval. IEEE Trans. Multimedia 24, 3520\u20133532 (2021)","journal-title":"IEEE Trans. Multimedia"},{"issue":"4","key":"21_CR21","first-page":"4794","volume":"45","author":"S Qian","year":"2023","unstructured":"Qian, S., Xue, D., Fang, Q., Changsheng, X.: Integrating multi-label contrastive learning with dual adversarial graph neural networks for cross-modal retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 45(4), 4794\u20134811 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"21_CR22","doi-asserted-by":"crossref","unstructured":"Qin, Y., Peng, D., Peng, X., Wang, X., Hu, P.: Deep evidential learning with noisy correspondence for cross-modal retrieval. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 4948\u20134956 (2022)","DOI":"10.1145\/3503161.3547922"},{"key":"21_CR23","unstructured":"Qin, Y., Sun, Y., Peng, D., Zhou, J.T., Peng, X., Hu, P.: Cross-modal active complementary learning with self-refining correspondence. Adv. Neural Inf. Process. Syst. 36 (2023)"},{"key":"21_CR24","doi-asserted-by":"crossref","unstructured":"Qin, Y., Chen, Y., Peng, D., Peng, X., Zhou, J.T., Hu, P.: Noisy-correspondence learning for text-to-image person re-identification. In: IEEE International Conference on Computer Vision and Pattern Recognition (CVPR) (2024)","DOI":"10.1109\/CVPR52733.2024.02568"},{"key":"21_CR25","doi-asserted-by":"crossref","unstructured":"Ranjan, V., Rasiwasia, N., Jawahar, C.V.: Multi-label cross-modal retrieval. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV) (2015)","DOI":"10.1109\/ICCV.2015.466"},{"key":"21_CR26","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: International Conference on Learning Representations,International Conference on Learning Representations (2015)"},{"key":"21_CR27","unstructured":"Song, H., Kim, M., Park, D., Shin, Y., Lee, J.G.: Learning from noisy labels with deep neural networks: a survey. IEEE Trans. Neural Netw. Learn. Syst. 1\u201319 (2022)"},{"key":"21_CR28","doi-asserted-by":"crossref","unstructured":"Sun, Y., Ren, Z., Hu, P., Peng, D., Wang, X.: Hierarchical consensus hashing for cross-modal retrieval. IEEE Trans. Multimedia (2023)","DOI":"10.1109\/TMM.2023.3272169"},{"key":"21_CR29","doi-asserted-by":"crossref","unstructured":"Sun, Y., Dai, J., Ren, Z., Chen, Y., Peng, D., Peng, H.: Dual self-paced cross-modal hashing. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, pp. 15184\u201315192 (2024)","DOI":"10.1609\/aaai.v38i14.29441"},{"key":"21_CR30","doi-asserted-by":"publisher","first-page":"824","DOI":"10.1109\/TMM.2023.3272169","volume":"26","author":"Y Sun","year":"2024","unstructured":"Sun, Y., Ren, Z., Peng, H., Peng, D., Wang, X.: Hierarchical consensus hashing for cross-modal retrieval. IEEE Trans. Multimedia 26, 824\u2013836 (2024). https:\/\/doi.org\/10.1109\/TMM.2023.3272169","journal-title":"IEEE Trans. Multimedia"},{"issue":"7","key":"21_CR31","doi-asserted-by":"publisher","first-page":"4765","DOI":"10.1109\/TCSVT.2021.3136330","volume":"32","author":"Y Wang","year":"2022","unstructured":"Wang, Y., Peng, Y.: Mars: learning modality-agnostic representation for scalable cross-media retrieval. IEEE Trans. Circuits Syst. Video Technol. 32(7), 4765\u20134777 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"21_CR32","doi-asserted-by":"publisher","first-page":"165","DOI":"10.1016\/j.neucom.2018.11.042","volume":"331","author":"W Yiling","year":"2019","unstructured":"Yiling, W., Wang, S., Huang, Q.: Multi-modal semantic autoencoder for cross-modal retrieval. Neurocomputing 331, 165\u2013175 (2019)","journal-title":"Neurocomputing"},{"key":"21_CR33","doi-asserted-by":"crossref","unstructured":"Xu, T., Liu, X., Huang, Z., Guo, D., Hong, R., Wang, M.: Early-learning regularized contrastive learning for cross-modal retrieval with noisy labels. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 629\u2013637 (2022). ISBN 9781450392037","DOI":"10.1145\/3503161.3548066"},{"key":"21_CR34","doi-asserted-by":"crossref","unstructured":"Yang, M., Huang, Z., Hu, P., Li, T., Lv, J., Peng, X.: Learning with twin noisy labels for visible-infrared person re-identification. In: CVPR, pp. 14308\u201314317 (2022)","DOI":"10.1109\/CVPR52688.2022.01391"},{"key":"21_CR35","unstructured":"Yao, J., Wang, J., Tsang, I.W., Zhang, Y., Sun, J., Zhang, C., Zhang, R.: Deep learning from noisy image labels with quality embedding. IEEE Trans. Image Process. (2017)"},{"key":"21_CR36","unstructured":"Zhang, Z., Sabuncu, M.: Generalized cross entropy loss for training deep neural networks with noisy labels. Adv. Neural Inf. Process. Syst. 31 (2018)"},{"key":"21_CR37","doi-asserted-by":"crossref","unstructured":"Zhen, L., Hu, P., Wang, X., Peng, D.: Deep supervised cross-modal retrieval. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10394\u201310403 (2019)","DOI":"10.1109\/CVPR.2019.01064"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-8620-6_21","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,14]],"date-time":"2025-01-14T20:17:47Z","timestamp":1736885867000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-8620-6_21"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,20]]},"ISBN":["9789819786190","9789819786206"],"references-count":37,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-8620-6_21","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,10,20]]},"assertion":[{"value":"20 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Urumqi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2024.prcv.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}