{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T15:11:23Z","timestamp":1769181083576,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":27,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819555666","type":"print"},{"value":"9789819555673","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5567-3_23","type":"book-chapter","created":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:13:51Z","timestamp":1769116431000},"page":"327-341","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Dual Fusion with\u00a0Auxiliary Loss Hashing for\u00a0Cross-Modal Retrieval"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-4387-8326","authenticated-orcid":false,"given":"Shenao","family":"Shao","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0210-2273","authenticated-orcid":false,"given":"Liejun","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0001-2135-5084","authenticated-orcid":false,"given":"Shaochen","family":"Jiang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0002-1694-2222","authenticated-orcid":false,"given":"Beibei","family":"Gao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,23]]},"reference":[{"key":"23_CR1","doi-asserted-by":"crossref","unstructured":"Chua, T.S., et al.: Nus-wide: a real-world web image database from national university of singapore. In: ACM International Conference on Image and Video Retrieval (2009). https:\/\/api.semanticscholar.org\/CorpusID:6483070","DOI":"10.1145\/1646396.1646452"},{"key":"23_CR2","doi-asserted-by":"crossref","unstructured":"Du, B., Wu, C., Huang, Z.: Learning resource allocation and pricing for cloud profit maximization. In: AAAI Conference on Artificial Intelligence (2019). https:\/\/api.semanticscholar.org\/CorpusID:69620644","DOI":"10.1609\/aaai.v33i01.33017570"},{"key":"23_CR3","unstructured":"Grubinger, M., Clough, P.D., M\u00fcller, H., Deselaers, T.: The iapr tc-12 benchmark: A new evaluation resource for visual information systems (2006). https:\/\/api.semanticscholar.org\/CorpusID:18883184"},{"key":"23_CR4","doi-asserted-by":"crossref","unstructured":"Hu, H., Xie, L., Hong, R., Tian, Q.: Creating something from nothing: Unsupervised knowledge distillation for cross-modal hashing. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3120\u20133129 (2020). https:\/\/api.semanticscholar.org\/CorpusID:214743601","DOI":"10.1109\/CVPR42600.2020.00319"},{"issue":"3","key":"23_CR5","first-page":"3877","volume":"45","author":"P Hu","year":"2022","unstructured":"Hu, P.: Unsupervised contrastive cross-modal hashing. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 3877\u20133889 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3","key":"23_CR6","doi-asserted-by":"publisher","first-page":"3877","DOI":"10.1109\/TPAMI.2022.3177356","volume":"45","author":"P Hu","year":"2023","unstructured":"Hu, P.: Unsupervised contrastive cross-modal hashing. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 3877\u20133889 (2023). https:\/\/doi.org\/10.1109\/TPAMI.2022.3177356","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"23_CR7","doi-asserted-by":"crossref","unstructured":"Huiskes, M.J., Lew, M.S.: The mir flickr retrieval evaluation. In: Multimedia Information Retrieval (2008). https:\/\/api.semanticscholar.org\/CorpusID:14040310","DOI":"10.1145\/1460096.1460104"},{"issue":"1","key":"23_CR8","doi-asserted-by":"publisher","first-page":"576","DOI":"10.1109\/TCSVT.2023.3285266","volume":"34","author":"Y Huo","year":"2023","unstructured":"Huo, Y.: Deep semantic-aware proxy hashing for multi-label cross-modal retrieval. IEEE Trans. Circuits Syst. Video Technol. 34(1), 576\u2013589 (2023)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"23_CR9","doi-asserted-by":"crossref","unstructured":"Huo, Y., et al.: Deep neighborhood-aware proxy hashing with uniform distribution constraint for cross-modal retrieval. ACM Trans. Multimedia Comput. Commun. Appl. 20, 1 \u2013 23 (2024). https:\/\/api.semanticscholar.org\/CorpusID:267352535","DOI":"10.1145\/3643639"},{"issue":"11","key":"23_CR10","doi-asserted-by":"publisher","first-page":"5926","DOI":"10.1109\/TKDE.2024.3401050","volume":"36","author":"Y Huo","year":"2024","unstructured":"Huo, Y., Qin, Q., Zhang, W., Huang, L., Nie, J.: Deep hierarchy-aware proxy hashing with self-paced learning for cross-modal retrieval. IEEE Trans. Knowl. Data Eng. 36(11), 5926\u20135939 (2024). https:\/\/doi.org\/10.1109\/TKDE.2024.3401050","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"23_CR11","doi-asserted-by":"crossref","unstructured":"Jiang, Q.Y., Li, W.J.: Deep cross-modal hashing. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3270\u20133278 (2017)","DOI":"10.1109\/CVPR.2017.348"},{"key":"23_CR12","doi-asserted-by":"crossref","unstructured":"Kim, Y.: Convolutional neural networks for sentence classification. In: Conference on Empirical Methods in Natural Language Processing (2014). https:\/\/api.semanticscholar.org\/CorpusID:9672033","DOI":"10.3115\/v1\/D14-1181"},{"key":"23_CR13","doi-asserted-by":"publisher","unstructured":"Li, J., et al.: Lightweight contrastive distilled hashing for online cross-modal retrieval. In: Walsh, T., Shah, J., Kolter, Z. (eds.) AAAI-25, Sponsored by the Association for the Advancement of Artificial Intelligence, February 25 - March 4, 2025, Philadelphia, PA, USA, pp. 4779\u20134787. AAAI Press (2025). https:\/\/doi.org\/10.1609\/AAAI.V39I5.32505","DOI":"10.1609\/AAAI.V39I5.32505"},{"key":"23_CR14","doi-asserted-by":"crossref","unstructured":"Liao, L., Yang, M., Zhang, B.: Deep supervised dual cycle adversarial network for cross-modal retrieval. IEEE Trans. Circuits Syst. Video Tech. 33, 920\u2013934 (2023). https:\/\/api.semanticscholar.org\/CorpusID:252153973","DOI":"10.1109\/TCSVT.2022.3203247"},{"key":"23_CR15","unstructured":"Made, I., Dwi, A.: A new algorithm for data compression optimization. ArXiv abs\/1209.1045 (2012). https:\/\/api.semanticscholar.org\/CorpusID:16027091"},{"key":"23_CR16","unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning (2021). https:\/\/api.semanticscholar.org\/CorpusID:231591445"},{"key":"23_CR17","unstructured":"Schmidt, R.M.: Recurrent Neural Networks (RNNs): a gentle introduction and overview. ArXiv abs\/1912.05911 (2019). https:\/\/api.semanticscholar.org\/CorpusID:209324034"},{"key":"23_CR18","doi-asserted-by":"crossref","unstructured":"Tu, J., Liu, X., Hao, Y., Hong, R., Wang, M.: Two-step discrete hashing for cross-modal retrieval. IEEE Trans. Multimedia, 1\u201312 (2024)","DOI":"10.1145\/3744567"},{"key":"23_CR19","doi-asserted-by":"crossref","unstructured":"Tu, J., Liu, X., Lin, Z., Hong, R., Wang, M.: Differentiable cross-modal hashing via multimodal transformers. In: Proceedings of the 30th ACM International Conference on Multimedia (2022). https:\/\/api.semanticscholar.org\/CorpusID:252783059","DOI":"10.1145\/3503161.3548187"},{"key":"23_CR20","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Neural Information Processing Systems (2017). https:\/\/api.semanticscholar.org\/CorpusID:13756489"},{"key":"23_CR21","doi-asserted-by":"crossref","unstructured":"Wang, G., Cheng, S., Du, A., Zou, Q.: Covariance attention guidance mamba hashing for cross-modal retrieval. Eng. Appl. Artif. Intell. 152, 110777 (2025). https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0952197625007778","DOI":"10.1016\/j.engappai.2025.110777"},{"key":"23_CR22","doi-asserted-by":"publisher","first-page":"255","DOI":"10.1016\/j.neucom.2020.03.019","volume":"400","author":"X Wang","year":"2020","unstructured":"Wang, X., Zou, X., Bakker, E.M., Wu, S.: Self-constraining and attention-based hashing network for bit-scalable cross-modal retrieval. Neurocomputing 400, 255\u2013271 (2020)","journal-title":"Neurocomputing"},{"key":"23_CR23","doi-asserted-by":"crossref","unstructured":"Wu, L., et al.: Deep multi-similarity hashing via label-guided network for cross-modal retrieval. Neurocomputing 616, 128830 (2025). https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0925231224016011","DOI":"10.1016\/j.neucom.2024.128830"},{"key":"23_CR24","doi-asserted-by":"crossref","unstructured":"Xia, X., Dong, G., Li, F., Zhu, L., Ying, X.: When clip meets cross-modal hashing retrieval: a new strong baseline. Inf. Fusion 100, 101968 (2023). https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253523002841","DOI":"10.1016\/j.inffus.2023.101968"},{"key":"23_CR25","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2021.104475","volume":"106","author":"Y Xie","year":"2021","unstructured":"Xie, Y., Zeng, X., Wang, T., Xu, L., Wang, D.: Matching images and texts with multi-head attention network for cross-media hashing retrieval. Eng. Appl. Artif. Intell. 106, 104475 (2021)","journal-title":"Eng. Appl. Artif. Intell."},{"key":"23_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2025.127365","volume":"279","author":"C Zhu","year":"2025","unstructured":"Zhu, C., Qin, Q., Zhang, W., Huang, L.: Deep neighbor-coherence hashing with discriminative sample mining for supervised cross-modal retrieval. Expert Syst. Appl. 279, 127365 (2025). https:\/\/doi.org\/10.1016\/j.eswa.2025.127365","journal-title":"Expert Syst. Appl."},{"key":"23_CR27","doi-asserted-by":"crossref","unstructured":"Zhuo, Y., Li, Y., Hsiao, J., Ho, C., Li, B.: Clip4hashing: unsupervised deep hashing for cross-modal video-text retrieval. In: Proceedings of the 2022 international conference on multimedia retrieval, pp. 158\u2013166 (2022)","DOI":"10.1145\/3512527.3531381"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5567-3_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:13:56Z","timestamp":1769116436000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5567-3_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819555666","9789819555673"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5567-3_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"23 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}