{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,21]],"date-time":"2026-02-21T19:25:43Z","timestamp":1771701943362,"version":"3.50.1"},"reference-count":38,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2023,7,29]],"date-time":"2023-07-29T00:00:00Z","timestamp":1690588800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,7,29]],"date-time":"2023-07-29T00:00:00Z","timestamp":1690588800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Multimed Info Retr"],"published-print":{"date-parts":[[2023,12]]},"DOI":"10.1007\/s13735-023-00288-3","type":"journal-article","created":{"date-parts":[[2023,7,29]],"date-time":"2023-07-29T10:01:56Z","timestamp":1690624916000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Deep adversarial multi-label cross-modal hashing algorithm"],"prefix":"10.1007","volume":"12","author":[{"given":"Xiaohan","family":"Yang","sequence":"first","affiliation":[]},{"given":"Zhen","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Wenhao","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Xinyi","family":"Chang","sequence":"additional","affiliation":[]},{"given":"Nannan","family":"Wu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,7,29]]},"reference":[{"key":"288_CR1","doi-asserted-by":"publisher","first-page":"148","DOI":"10.1016\/j.neucom.2022.02.007","volume":"483","author":"Z Li","year":"2022","unstructured":"Li Z, Lu H, Fu H (2022) Image-text bidirectional learning network based cross-modal retrieval. Neurocomputing 483:148\u2013159","journal-title":"Neurocomputing"},{"key":"288_CR2","doi-asserted-by":"publisher","first-page":"43","DOI":"10.3390\/fi14020043","volume":"14","author":"L Cai","year":"2022","unstructured":"Cai L, Zhu L, Zhang H, Zhu X (2022) DA-GAN: dual attention generative adversarial network for cross-modal retrieval. Future Internet 14:43","journal-title":"Future Internet"},{"key":"288_CR3","doi-asserted-by":"crossref","unstructured":"Zhen L, Hu P, Wang X, Peng D (2019) Deep supervised cross-modal retrieval. In: Proceedings of the computer vision and pattern recognition (CVPR), pp 10394\u201310403","DOI":"10.1109\/CVPR.2019.01064"},{"key":"288_CR4","doi-asserted-by":"crossref","unstructured":"Kan M, Shan S, Zhang H, Lao S, Chen X (2012) Multi-view discriminant analysis. In: Proceedings of the European conference on computer vision, pp 808\u2013821","DOI":"10.1007\/978-3-642-33718-5_58"},{"key":"288_CR5","unstructured":"Shaishav K, Raghavendra U (2011) Learning hash functions for cross-view similarity search. In: Proceedings of the 22nd international joint conference on artificial intelligence, pp 1360\u20131365"},{"key":"288_CR6","unstructured":"Wang D, Gao X, Wang X, He L (2015) Semantic topic multi-modal hashing for cross-media retrieval. In: Proceedings of the international joint conference on artificial intelligence, pp 3890\u20133896"},{"key":"288_CR7","unstructured":"Michael M, Alexander MB, Fabrice MB, Nikos P (2010) Data fusion through cross-modality metric learning using similarity-sensitive hashing. In: Proceedings of the computer vision and pattern recognition (CVPR), pp 3594\u20133601"},{"key":"288_CR8","doi-asserted-by":"crossref","unstructured":"Zhang D, Li W (2014) Large-scale supervised multimodal hashing with semantic correlation maximization. In: Proceedings of the twenty-eighth (AAAI) conference on artificial intelligence, pp 2177\u20132183","DOI":"10.1609\/aaai.v28i1.8995"},{"key":"288_CR9","doi-asserted-by":"crossref","unstructured":"Lin Z, Ding G, Hu M, Wang J (2015) Semantics-preserving hashing for cross-view retrieval. In: Proceedings of the computer vision and pattern recognition (CVPR), pp 3864\u20133872","DOI":"10.1109\/CVPR.2015.7299011"},{"key":"288_CR10","doi-asserted-by":"crossref","unstructured":"Jiang Q, Li W (2017) Deep cross-modal hashing. In: Proceedings of the computer vision and pattern recognition (CVPR), pp 3270\u20133278","DOI":"10.1109\/CVPR.2017.348"},{"key":"288_CR11","doi-asserted-by":"crossref","unstructured":"Chen Z, Yu W, Li C, Nie L, Xu X (2018) Dual deep neural networks cross-modal hashing. In: AAAI, pp 274\u2013281","DOI":"10.1609\/aaai.v32i1.11249"},{"key":"288_CR12","doi-asserted-by":"crossref","unstructured":"Li C, Deng C, Li N, Liu W, Gao X, Tao D (2018 )Self-supervised adversarial hashing networks for cross-modal retrieval. In: Proceedings of the computer vision and pattern recognition (CVPR), pp 4242\u20134251","DOI":"10.1109\/CVPR.2018.00446"},{"key":"288_CR13","doi-asserted-by":"crossref","unstructured":"Yang EK, Deng C, Liu W, Liu X, Tao D, Gao X (2017) Pairwise relationship guided deep hashing for cross-modal retrieval. In: Proceedings of the thirty-first (AAAI) conference on artificial intelligence, pp 1618\u20131625","DOI":"10.1609\/aaai.v31i1.10719"},{"key":"288_CR14","doi-asserted-by":"crossref","unstructured":"Zhan Y, Luo X, Wang Y, Xu X (2020) Supervised hierarchical deep hashing for cross-modal retrieval. In: The 28th international conference on multimedia, pp 3386\u20133394","DOI":"10.1145\/3394171.3413962"},{"key":"288_CR15","doi-asserted-by":"crossref","unstructured":"Song J, Yang Y, Yang Y, Huang Z, Shen HT (2013) Inter-media hashing for large-scale retrieval from heterogeneous data sources. In: Proceedings of the ACM SIGMOD international conference on management of data, pp 785\u2013796","DOI":"10.1145\/2463676.2465274"},{"key":"288_CR16","doi-asserted-by":"crossref","unstructured":"Ding G, Guo Y, Zhou J (2014) Collective matrix factorization hashing for multimodal data. In: Proceedings of the 2014 IEEE conference on computer vision and pattern recognition (CVPR), pp 2083\u20132090","DOI":"10.1109\/CVPR.2014.267"},{"key":"288_CR17","doi-asserted-by":"crossref","unstructured":"Zhang D, Li W (2014) Large-scale supervised multimodal hashing with semantic correlation maximization. In: Proceedings of the twenty-eighth conference on artificial intelligence (AAAI). pp 2177\u20132183","DOI":"10.1609\/aaai.v28i1.8995"},{"key":"288_CR18","doi-asserted-by":"crossref","unstructured":"Lin Z, Ding G, Hu M, Wang J (2015) Semantics-preserving hashing for cross-view retrieval. In: Proceedings of the computer vision and pattern recognition (CVPR), pp 3864\u20133872","DOI":"10.1109\/CVPR.2015.7299011"},{"issue":"2","key":"288_CR19","doi-asserted-by":"publisher","first-page":"489","DOI":"10.1109\/TCYB.2018.2868826","volume":"50","author":"J Zhang","year":"2020","unstructured":"Zhang J, Peng Y, Yuan M (2020) SCH-GAN: semi-supervised cross-modal hashing by generative adversarial network. IEEE Trans Cybern 50(2):489\u2013502","journal-title":"IEEE Trans Cybern"},{"key":"288_CR20","doi-asserted-by":"publisher","first-page":"560","DOI":"10.1109\/TKDE.2020.2987312","volume":"34","author":"R Tu","year":"2022","unstructured":"Tu R, Mao X, Ma B, Hu Y, Yan T, Wei W, Huang H (2022) Deep cross-modal hashing with hashing functions and unified hash codes jointly learning. IEEE Trans Knowl Data Eng 34:560\u2013572","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"288_CR21","first-page":"1","volume":"99","author":"J Lu","year":"2020","unstructured":"Lu J, Tang J, Li Z, Guo J (2020) Deep semantic multimodal hashing network for scalable image-text and video-text retrievals? IEEE Trans Neural Netw Learn Syst 99:1\u201314","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"288_CR22","doi-asserted-by":"crossref","unstructured":"Gu W, Gu X, Gu J, Li B, Xiong Z, Wang W (2019) adversary guided asymmetric hashing for cross-modal retrieval. In: Proceedings of the 2019 on international conference on multimedia retrieval, pp 159\u2013167","DOI":"10.1145\/3323873.3325045"},{"key":"288_CR23","doi-asserted-by":"publisher","DOI":"10.1016\/j.compeleceng.2021.107262","volume":"93","author":"S Chen","year":"2021","unstructured":"Chen S, Wu S, Wang L, Yu Z (2021) Self-attention and adversary learning deep hashing network for cross-modal retrieval. Comput Electr Eng 93:107262","journal-title":"Comput Electr Eng"},{"key":"288_CR24","doi-asserted-by":"publisher","first-page":"1819","DOI":"10.1109\/TKDE.2013.39","volume":"26","author":"M Zhang","year":"2013","unstructured":"Zhang M, Zhou Z (2013) A review on multi-label learning algorithms. IEEE Trans Knowl Data Eng 26:1819\u20131837","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"288_CR25","unstructured":"Henry G, Bernhard P, Michael C (2016) Learning distance metrics for multi-label classification. In: Proceedings of the machine learning. PMLR, pp 318\u2013333"},{"key":"288_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.image.2020.116131","volume":"93","author":"X Zou","year":"2021","unstructured":"Zou X, Wang X, Bakker Erwin M, Wu S (2021) Multi-label semantics preserving based deep cross-modal hashing. Signal Process Image Commun 93:116131","journal-title":"Signal Process Image Commun"},{"key":"288_CR27","doi-asserted-by":"crossref","unstructured":"Bai C, Zeng C, Ma Q, Zhang J, Chen S (2020) Deep adversarial discrete hashing for cross-modal retrieval. In: Proceedings of the 2020 on international conference on multimedia retrieval","DOI":"10.1145\/3372278.3390711"},{"key":"288_CR28","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: Proceedings of the 3rd international conference on learning representations"},{"key":"288_CR29","doi-asserted-by":"publisher","first-page":"2393","DOI":"10.1007\/s11042-021-11543-2","volume":"81","author":"G Siddan","year":"2022","unstructured":"Siddan G, Palraj P (2022) Foetal neurodegenerative disease classification using improved deep ResNet classification based VGG-19 feature extraction network. Multimed Tools Appl 81:2393\u20132408","journal-title":"Multimed Tools Appl"},{"key":"288_CR30","doi-asserted-by":"publisher","first-page":"6916","DOI":"10.1109\/JSTARS.2021.3090085","volume":"14","author":"Y Mu","year":"2021","unstructured":"Mu Y, Ni R, Zhang C, Gong H, Hu T, Li S, Sun Y, Zhang T, Guo Y (2021) A lightweight model of VGG-16 for remote sensing image classification. IEEE J Sel Top Appl Earth Obs Remote Sens 14:6916\u20136922","journal-title":"IEEE J Sel Top Appl Earth Obs Remote Sens"},{"key":"288_CR31","doi-asserted-by":"publisher","first-page":"568","DOI":"10.20965\/jaciii.2020.p0568","volume":"24","author":"C Zhang","year":"2020","unstructured":"Zhang C, Meng D, He J (2020) VGG-16 convolutional neural network-oriented detection of filling flow status of viscous food. J Adv Comput Intell Intell Inf 24:568\u2013575","journal-title":"J Adv Comput Intell Intell Inf"},{"key":"288_CR32","doi-asserted-by":"crossref","unstructured":"Olga R, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M (2015) ImageNet large scale visual recognition challenge. Int J Comput Vis 211\u2013252","DOI":"10.1007\/s11263-015-0816-y"},{"key":"288_CR33","first-page":"36","volume":"2","author":"Z Li","year":"2021","unstructured":"Li Z, Xu X, Dl Zhang et al (2021) Based on deep residual network. CSSE 2:36","journal-title":"CSSE"},{"key":"288_CR34","doi-asserted-by":"crossref","unstructured":"Huiskes MJ, Lew MS (2008) The MIR flickr retrieval evaluation. In: Proceedings of the international conference on multimedia information retrieval, pp 39\u201343","DOI":"10.1145\/1460096.1460104"},{"key":"288_CR35","doi-asserted-by":"crossref","unstructured":"Chua T, Tang J, Hong R, Li H, Luo Z, Zheng Y (2009) NUS-WIDE: a real-world web image database from National University of Singapore. In: Proceedings of the international conference on image and video retrieval","DOI":"10.1145\/1646396.1646452"},{"key":"288_CR36","doi-asserted-by":"crossref","unstructured":"Cao Y, Long M, Wang J (2017) Correlation hashing network for efficient cross-modal retrieval. In: Proceedings of the British machine vision conference","DOI":"10.5244\/C.31.128"},{"key":"288_CR37","doi-asserted-by":"publisher","first-page":"401","DOI":"10.1109\/TCSVT.2020.2974877","volume":"31","author":"X Nie","year":"2021","unstructured":"Nie X, Wang B, Li J, Hao F, Jian M, Yin Y (2021) Deep multiscale fusion hashing for cross-modal retrieval. IEEE Trans Circuits Syst Video Technol 31:401\u2013410","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"288_CR38","doi-asserted-by":"crossref","unstructured":"Wang B, Yang Y, Xu X, Alan H, Shen H (2017) Adversarial cross-modal retrieval. In: Proceedings of the 2017 ACM on multimedia conference. Mountain View, pp 154\u2013162","DOI":"10.1145\/3123266.3123326"}],"container-title":["International Journal of Multimedia Information Retrieval"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13735-023-00288-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13735-023-00288-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13735-023-00288-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,2]],"date-time":"2023-12-02T14:09:47Z","timestamp":1701526187000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13735-023-00288-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,29]]},"references-count":38,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2023,12]]}},"alternative-id":["288"],"URL":"https:\/\/doi.org\/10.1007\/s13735-023-00288-3","relation":{},"ISSN":["2192-6611","2192-662X"],"issn-type":[{"value":"2192-6611","type":"print"},{"value":"2192-662X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,7,29]]},"assertion":[{"value":"27 February 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 May 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 July 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 July 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article. The authors have no competing interests as defined by Springer, or other interests that might be perceived to influence the results and\/or discussion reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The results\/data\/figures in this manuscript have not been published elsewhere, nor are they under consideration (from you or one of your Contributing Authors) by another publisher. We have read the Springer journal policies on author responsibilities and submit this manuscript in accordance with those policies. All of the material is owned by the authors, and\/or no permissions are required.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}],"article-number":"16"}}