{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T12:46:03Z","timestamp":1770813963686,"version":"3.50.1"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"23","license":[{"start":{"date-parts":[[2023,10,19]],"date-time":"2023-10-19T00:00:00Z","timestamp":1697673600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,10,19]],"date-time":"2023-10-19T00:00:00Z","timestamp":1697673600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["61976057"],"award-info":[{"award-number":["61976057"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62172101"],"award-info":[{"award-number":["62172101"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"the Science and Technology Commission of Shanghai Municipality","award":["21511101000"],"award-info":[{"award-number":["21511101000"]}]},{"name":"the Science and Technology Commission of Shanghai Municipality","award":["22DZ1100101"],"award-info":[{"award-number":["22DZ1100101"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2023,12]]},"DOI":"10.1007\/s10489-023-05028-y","type":"journal-article","created":{"date-parts":[[2023,10,19]],"date-time":"2023-10-19T10:03:30Z","timestamp":1697709810000},"page":"28954-28973","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Deep cross-modal hashing with fine-grained similarity"],"prefix":"10.1007","volume":"53","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6319-5579","authenticated-orcid":false,"given":"Yangdong","family":"Chen","sequence":"first","affiliation":[]},{"given":"Jiaqi","family":"Quan","sequence":"additional","affiliation":[]},{"given":"Yuejie","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Rui","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Tao","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,10,19]]},"reference":[{"issue":"9","key":"5028_CR1","doi-asserted-by":"publisher","first-page":"1095","DOI":"10.1002\/asi.24045","volume":"69","author":"H Ayadi","year":"2018","unstructured":"Ayadi H, Torjmen-Khemakhem M, Daoud M et al (2018) Mf-re-rank: A modality feature-based re-ranking model for medical image retrieval. J Assoc Inf Sci Technol 69(9):1095\u20131108","journal-title":"J Assoc Inf Sci Technol"},{"issue":"9","key":"5028_CR2","doi-asserted-by":"publisher","first-page":"820","DOI":"10.1145\/361573.361582","volume":"15","author":"RH Bartels","year":"1972","unstructured":"Bartels RH, Stewart GW (1972) Solution of the matrix equation ax+ xb= c [f4]. Commun ACM 15(9):820\u2013826","journal-title":"Commun ACM"},{"key":"5028_CR3","doi-asserted-by":"crossref","unstructured":"Boyd S, Parikh N, Chu E (2011) Distributed optimization and statistical learning via the alternating direction method of multipliers. Now Publishers Inc","DOI":"10.1561\/9781601984616"},{"key":"5028_CR4","doi-asserted-by":"crossref","unstructured":"Bronstein MM, Bronstein AM, Michel F, et al (2010) Data fusion through cross-modality metric learning using similarity-sensitive hashing. In: CVPR, pp 3594\u20133601","DOI":"10.1109\/CVPR.2010.5539928"},{"key":"5028_CR5","doi-asserted-by":"crossref","unstructured":"Cao Y, Long M, Wang J, et al (2016) Deep visual-semantic hashing for cross-modal retrieval. In: KDD, pp 1445\u20131454","DOI":"10.1145\/2939672.2939812"},{"key":"5028_CR6","doi-asserted-by":"crossref","unstructured":"Cao Y, Liu B, Long M, et al (2018) Hashgan: Deep learning to hash with pair conditional wasserstein gan. In: CVPR, pp 1287\u20131296","DOI":"10.1109\/CVPR.2018.00140"},{"key":"5028_CR7","doi-asserted-by":"crossref","unstructured":"Cao Z, Long M, Huang C, et al (2018) Transfer adversarial hashing for hamming space retrieval. In: AAAI","DOI":"10.1109\/CVPR.2018.00134"},{"key":"5028_CR8","doi-asserted-by":"crossref","unstructured":"Cao Z, Sun Z, Long M, et al (2018) Deep priority hashing. In: MM, pp 1653\u20131661","DOI":"10.1145\/3240508.3240543"},{"key":"5028_CR9","doi-asserted-by":"crossref","unstructured":"Chatfield K, Simonyan K, Vedaldi A, et al (2014) Return of the devil in the details: Delving deep into convolutional nets. In: BMVC","DOI":"10.5244\/C.28.6"},{"key":"5028_CR10","doi-asserted-by":"crossref","unstructured":"Chen ZD, Wang Y, Li HQ, et al (2019) A two-step cross-modal hashing by exploiting label correlations and preserving similarity in both steps. In: MM, pp 1694\u20131702","DOI":"10.1145\/3343031.3350862"},{"key":"5028_CR11","doi-asserted-by":"crossref","unstructured":"Chua TS, Tang J, Hong R, et al (2009) Nus-wide: a real-world web image database from national university of singapore. In: CIVR, pp 1\u20139","DOI":"10.1145\/1646396.1646452"},{"issue":"8","key":"5028_CR12","doi-asserted-by":"publisher","first-page":"3893","DOI":"10.1109\/TIP.2018.2821921","volume":"27","author":"C Deng","year":"2018","unstructured":"Deng C, Chen Z, Liu X et al (2018) Triplet-based deep hashing network for cross-modal retrieval. IEEE Trans Image Process 27(8):3893\u20133903","journal-title":"IEEE Trans Image Process"},{"key":"5028_CR13","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, et al (2009) Imagenet: A large-scale hierarchical image database. In: CVPR, pp 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"5028_CR14","doi-asserted-by":"crossref","unstructured":"Ding G, Guo Y, Zhou J (2014) Collective matrix factorization hashing for multimodal data. In: CVPR, pp 2075\u20132082","DOI":"10.1109\/CVPR.2014.267"},{"key":"5028_CR15","doi-asserted-by":"publisher","first-page":"108676","DOI":"10.1016\/j.patcog.2022.108676","volume":"128","author":"Y Duan","year":"2022","unstructured":"Duan Y, Chen N, Zhang P et al (2022) Ms2gah: Multi-label semantic supervised graph attention hashing for robust cross-modal retrieval. Pattern Recognit 128:108676","journal-title":"Pattern Recognit"},{"key":"5028_CR16","doi-asserted-by":"crossref","unstructured":"Erin Liong V, Lu J, Tan YP, et al (2017) Cross-modal deep variational hashing. In: ICCV, pp 4077\u20134085","DOI":"10.1109\/ICCV.2017.439"},{"key":"5028_CR17","doi-asserted-by":"crossref","unstructured":"Hendricks LA, Venugopalan S, Rohrbach M, et al (2016) Deep compositional captioning: Describing novel object categories without paired training data. In: CVPR, pp 1\u201310","DOI":"10.1109\/CVPR.2016.8"},{"key":"5028_CR18","doi-asserted-by":"crossref","unstructured":"Hu H, Xie L, Hong R, et al (2020) Creating something from nothing: Unsupervised knowledge distillation for cross-modal hashing. In: CVPR, pp 3123\u20133132","DOI":"10.1109\/CVPR42600.2020.00319"},{"key":"5028_CR19","doi-asserted-by":"crossref","unstructured":"Hu Y, Jin Z, Ren H, et al (2014) Iterative multi-view hashing for cross media indexing. In: MM, pp 527\u2013536","DOI":"10.1145\/2647868.2654906"},{"key":"5028_CR20","doi-asserted-by":"crossref","unstructured":"Huiskes MJ, Lew MS (2008) The mir flickr retrieval evaluation. In: MIR, pp 39\u201343","DOI":"10.1145\/1460096.1460104"},{"key":"5028_CR21","doi-asserted-by":"crossref","unstructured":"Jiang QY, Li WJ (2017) Deep cross-modal hashing. In: CVPR, pp 3232\u20133240","DOI":"10.1109\/CVPR.2017.348"},{"issue":"7","key":"5028_CR22","doi-asserted-by":"publisher","first-page":"3490","DOI":"10.1109\/TIP.2019.2897944","volume":"28","author":"QY Jiang","year":"2019","unstructured":"Jiang QY, Li WJ (2019) Discrete latent factor model for cross-modal hashing. IEEE Trans Image Process 28(7):3490\u20133501","journal-title":"IEEE Trans Image Process"},{"issue":"5","key":"5028_CR23","first-page":"1429","volume":"30","author":"L Jin","year":"2018","unstructured":"Jin L, Li K, Li Z et al (2018) Deep semantic-preserving ordinal hashing for cross-modal similarity search. TNNLS 30(5):1429\u20131440","journal-title":"TNNLS"},{"key":"5028_CR24","unstructured":"Kumar S, Udupa R (2011) Learning hash functions for cross-view similarity search. In: IJCAI, pp 1360\u20131365"},{"key":"5028_CR25","doi-asserted-by":"crossref","unstructured":"Li C, Deng C, Li N, et al (2018) Self-supervised adversarial hashing networks for cross-modal retrieval. In: CVPR, pp 4242\u20134251","DOI":"10.1109\/CVPR.2018.00446"},{"key":"5028_CR26","doi-asserted-by":"crossref","unstructured":"Li H, Zhang C, Jia X, et al (2021) Adaptive label correlation based asymmetric discrete hashing for cross-modal retrieval. IEEE Transactions on Knowledge and Data Engineering","DOI":"10.1109\/TKDE.2021.3102119"},{"issue":"1","key":"5028_CR27","doi-asserted-by":"publisher","first-page":"46","DOI":"10.1002\/asi.24373","volume":"72","author":"W Li","year":"2021","unstructured":"Li W, Zheng Y, Zhang Y et al (2021) Cross-modal retrieval with dual multi-angle self-attention. J Assoc Inf Sci Technol 72(1):46\u201365","journal-title":"J Assoc Inf Sci Technol"},{"key":"5028_CR28","doi-asserted-by":"crossref","unstructured":"Lin G, Shen C, Suter D, et al (2013) A general two-step approach to learning-based hashing. In: ICCV, pp 2552\u20132559","DOI":"10.1109\/ICCV.2013.317"},{"key":"5028_CR29","doi-asserted-by":"crossref","unstructured":"Lin TY, Maire M, Belongie S, et al (2014) Microsoft coco: Common objects in context. In: ECCV, pp 740\u2013755","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"5028_CR30","doi-asserted-by":"crossref","unstructured":"Lin TY, Goyal P, Girshick R, et al (2017) Focal loss for dense object detection. In: ICCV, pp 2980\u20132988","DOI":"10.1109\/ICCV.2017.324"},{"key":"5028_CR31","doi-asserted-by":"crossref","unstructured":"Lin Z, Ding G, Hu M, et al (2015) Semantics-preserving hashing for cross-view retrieval. In: CVPR, pp 3864\u20133872","DOI":"10.1109\/CVPR.2015.7299011"},{"key":"5028_CR32","doi-asserted-by":"crossref","unstructured":"Liu H, Ji R, Wu Y, et al (2017) Cross-modality binary code learning via fusion similarity hashing. In: CVPR, pp 7380\u20137388","DOI":"10.1109\/CVPR.2017.672"},{"issue":"11","key":"5028_CR33","doi-asserted-by":"publisher","first-page":"6306","DOI":"10.1109\/TNNLS.2021.3076684","volume":"33","author":"X Liu","year":"2021","unstructured":"Liu X, Wang X, Ym Cheung (2021) Fddh: Fast discriminative discrete hashing for large-scale cross-modal retrieval. IEEE Trans Neural Netw Learn Syst 33(11):6306\u20136320","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"5028_CR34","doi-asserted-by":"crossref","unstructured":"Luo X, Yin XY, Nie L, et al (2018) Sdmch: Supervised discrete manifold-embedded cross-modal hashing. In: IJCAI, pp 2518\u20132524","DOI":"10.24963\/ijcai.2018\/349"},{"issue":"6","key":"5028_CR35","doi-asserted-by":"publisher","first-page":"2962","DOI":"10.1109\/TIP.2019.2892703","volume":"28","author":"X Luo","year":"2019","unstructured":"Luo X, Zhang PF, Huang Z et al (2019) Discrete hashing with multiple supervision. IEEE Trans Image Process 28(6):2962\u20132975","journal-title":"IEEE Trans Image Process"},{"issue":"12","key":"5028_CR36","doi-asserted-by":"publisher","first-page":"3101","DOI":"10.1109\/TMM.2020.2969792","volume":"22","author":"X Ma","year":"2020","unstructured":"Ma X, Zhang T, Xu C (2020) Multi-level correlation adversarial hashing for cross-modal retrieval. IEEE Trans Multimed 22(12):3101\u20133114","journal-title":"IEEE Trans Multimed"},{"key":"5028_CR37","doi-asserted-by":"publisher","first-page":"986","DOI":"10.1109\/TIP.2020.3038365","volume":"30","author":"M Meng","year":"2020","unstructured":"Meng M, Wang H, Yu J et al (2020) Asymmetric supervised consistent and specific hashing for cross-modal retrieval. IEEE Trans Image Process 30:986\u20131000","journal-title":"IEEE Trans Image Process"},{"key":"5028_CR38","unstructured":"Nie X, Wang B, Li J, et al (2020) Deep multiscale fusion hashing for cross-modal retrieval. TCSVT"},{"key":"5028_CR39","doi-asserted-by":"crossref","unstructured":"Noh H, Seo PH, Han B (2016) Image question answering using convolutional nal network with dynamic parameter prediction. In: CVPR, pp 30\u201338","DOI":"10.1109\/CVPR.2016.11"},{"key":"5028_CR40","unstructured":"Rastegari M, Choi J, Fakhraei S, et al (2013) Predictable dual-view hashing. In: ICML, pp 1328\u20131336"},{"issue":"12","key":"5028_CR41","doi-asserted-by":"publisher","first-page":"3034","DOI":"10.1109\/TPAMI.2018.2789887","volume":"40","author":"F Shen","year":"2018","unstructured":"Shen F, Xu Y, Liu L et al (2018) Unsupervised deep hashing with similarity-adaptive and discrete optimization. IEEE Trans Pattern Anal Mach Intell 40(12):3034\u20133044","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"5028_CR42","doi-asserted-by":"crossref","unstructured":"Shen Y, Liu L, Shao L, et al (2017) Deep binaries: Encoding semantic-rich cues for efficient textual-visual cross retrieval. In: ICCV, pp 4097\u20134106","DOI":"10.1109\/ICCV.2017.441"},{"key":"5028_CR43","doi-asserted-by":"crossref","unstructured":"Song J, Yang Y, Yang Y, et al (2013) Inter-media hashing for large-scale retrieval from heterogeneous data sources. In: SIGMOD, pp 785\u2013796","DOI":"10.1145\/2463676.2465274"},{"key":"5028_CR44","unstructured":"Tu RC, Mao XL, Ma B, et al (2020) Deep cross-modal hashing with hashing functions and unified hash codes jointly learning. IEEE Transactions on Knowledge and Data Engineering"},{"key":"5028_CR45","unstructured":"Wang D, Cui P, Ou M, et al (2015) Deep multimodal hashing with orthogonal regularization. In: IJCAI"},{"key":"5028_CR46","unstructured":"Wang D, Gao X, Wang X, et al (2015) Semantic topic multimodal hashing for cross-media retrieval. In: IJCAI"},{"key":"5028_CR47","doi-asserted-by":"crossref","unstructured":"Wang W, Ooi BC, Yang X, et al (2014) Effective multi-modal retrieval based on stacked auto-encoders. In: VLDB, pp 649\u2013660","DOI":"10.14778\/2732296.2732301"},{"key":"5028_CR48","doi-asserted-by":"crossref","unstructured":"Wang Y, Luo X, Nie L, et al (2020) Batch: A scalable asymmetric discrete cross-modal hashing. IEEE Transactions on Knowledge and Data Engineering","DOI":"10.1109\/TKDE.2020.2974825"},{"issue":"10","key":"5028_CR49","doi-asserted-by":"publisher","first-page":"10064","DOI":"10.1109\/TCYB.2021.3059886","volume":"52","author":"Y Wang","year":"2021","unstructured":"Wang Y, Chen ZD, Luo X et al (2021) Fast cross-modal hashing with global and local similarity embedding. IEEE Trans Cybern 52(10):10064\u201310077","journal-title":"IEEE Trans Cybern"},{"key":"5028_CR50","doi-asserted-by":"crossref","unstructured":"Xu R, Li C, Yan J, et al (2019) Graph convolutional network hashing for cross-modal retrieval. In: IJCAI, pp 982\u2013988","DOI":"10.24963\/ijcai.2019\/138"},{"key":"5028_CR51","doi-asserted-by":"crossref","unstructured":"Yan C, Pang G, Bai X, et al (2019) Deep hashing by discriminating hard examples. In: MM, pp 1535\u20131542","DOI":"10.1145\/3343031.3350927"},{"key":"5028_CR52","doi-asserted-by":"crossref","unstructured":"Yang E, Deng C, Liu W, et al (2017) Pairwise relationship guided deep hashing for cross-modal retrieval. In: AAAI","DOI":"10.1609\/aaai.v31i1.10719"},{"key":"5028_CR53","doi-asserted-by":"crossref","unstructured":"Zhang D, Li WJ (2014) Large-scale supervised multimodal hashing with semantic correlation maximization. In: AAAI","DOI":"10.1609\/aaai.v28i1.8995"},{"issue":"5","key":"5028_CR54","first-page":"5091","volume":"35","author":"Z Zhang","year":"2022","unstructured":"Zhang Z, Luo H, Zhu L et al (2022) Modality-invariant asymmetric networks for cross-modal hashing. IEEE Trans Knowl Data Eng 35(5):5091\u20135104","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"5028_CR55","unstructured":"Zhen Y, Yeung DY (2012) Co-regularized hashing for multimodal data. In: NIPS, p 1376"},{"issue":"11","key":"5028_CR56","doi-asserted-by":"publisher","first-page":"2171","DOI":"10.1109\/TKDE.2019.2913388","volume":"32","author":"C Zheng","year":"2019","unstructured":"Zheng C, Zhu L, Lu X et al (2019) Fast discrete collaborative multi-modal hashing for large-scale multimedia retrieval. IEEE Trans Knowl Data Eng 32(11):2171\u20132184","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"5028_CR57","doi-asserted-by":"crossref","unstructured":"Zhou J, Ding G, Guo Y (2014) Latent semantic sparse hashing for cross-modal similarity search. In: SIGIR, pp 415\u2013424","DOI":"10.1145\/2600428.2609610"},{"key":"5028_CR58","doi-asserted-by":"crossref","unstructured":"Zhu X, Huang Z, Shen HT, et al (2013) Linear cross-modal hashing for efficient multimedia search. In: MM, pp 143\u2013152","DOI":"10.1145\/2502081.2502107"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-023-05028-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-023-05028-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-023-05028-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,29]],"date-time":"2023-11-29T14:24:05Z","timestamp":1701267845000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-023-05028-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,19]]},"references-count":58,"journal-issue":{"issue":"23","published-print":{"date-parts":[[2023,12]]}},"alternative-id":["5028"],"URL":"https:\/\/doi.org\/10.1007\/s10489-023-05028-y","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,10,19]]},"assertion":[{"value":"18 September 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 October 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}},{"value":"The authors declare that they have no potential conflicts of interest with the content of this paper, and there is no research involving human participants or animals in this paper.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical standard"}}]}}