{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,16]],"date-time":"2026-01-16T01:40:55Z","timestamp":1768527655597,"version":"3.49.0"},"reference-count":52,"publisher":"Elsevier BV","issue":"2","license":[{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2024,10,23]],"date-time":"2024-10-23T00:00:00Z","timestamp":1729641600000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100005089","name":"Natural Science Foundation of Beijing Municipality","doi-asserted-by":"publisher","award":["Z200002"],"award-info":[{"award-number":["Z200002"]}],"id":[{"id":"10.13039\/501100005089","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003787","name":"Natural Science Foundation of Hebei Province","doi-asserted-by":"publisher","award":["F2018201115"],"award-info":[{"award-number":["F2018201115"]}],"id":[{"id":"10.13039\/501100003787","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2022YFE0196100"],"award-info":[{"award-number":["2022YFE0196100"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["High-Confidence Computing"],"published-print":{"date-parts":[[2025,6]]},"DOI":"10.1016\/j.hcc.2024.100274","type":"journal-article","created":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T16:31:58Z","timestamp":1730305918000},"page":"100274","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":1,"title":["A novel deep high-level concept-mining jointing hashing model for unsupervised cross-modal retrieval"],"prefix":"10.1016","volume":"5","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1726-5534","authenticated-orcid":false,"given":"Chun-Ru","family":"Dong","sequence":"first","affiliation":[]},{"given":"Jun-Yan","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Feng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Qiang","family":"Hua","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7846-0969","authenticated-orcid":false,"given":"Dachuan","family":"Xu","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.hcc.2024.100274_b1","series-title":"Proceedings of the 22th International Joint Conference on Artificial Intelligence","article-title":"Learning hash functions for cross-view similarity search","author":"Kumar","year":"2011"},{"key":"10.1016\/j.hcc.2024.100274_b2","first-page":"5","article-title":"Unsupervised deep hashing via binary latent factor models for large-scale cross-modal retrieval","volume":"vol. 1(3)","author":"Wu","year":"2018"},{"issue":"1","key":"10.1016\/j.hcc.2024.100274_b3","article-title":"Unsupervised generative adversarial cross-modal hashing","volume":"32","author":"Zhang","year":"2018","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"10.1016\/j.hcc.2024.100274_b4","series-title":"Deep cross-modal hashing via margin-dynamic-softmax loss","author":"Tu","year":"2020"},{"key":"10.1016\/j.hcc.2024.100274_b5","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"3864","article-title":"Semantics-preserving hashing for cross-view retrieval","author":"Lin","year":"2015"},{"key":"10.1016\/j.hcc.2024.100274_b6","series-title":"Proceedings of the 26th ACM International Conference on Multimedia","first-page":"852","article-title":"Multi-scale correlation for sequential cross-modal hashing learning","author":"Ye","year":"2018"},{"key":"10.1016\/j.hcc.2024.100274_b7","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"3232","article-title":"Deep cross-modal hashing","author":"Jiang","year":"2017"},{"key":"10.1016\/j.hcc.2024.100274_b8","series-title":"Correlation hashing network for efficient cross-modal retrieval","author":"Cao","year":"2016"},{"key":"10.1016\/j.hcc.2024.100274_b9","doi-asserted-by":"crossref","first-page":"3626","DOI":"10.1109\/TIP.2020.2963957","article-title":"Multi-task consistency-preserving adversarial hashing for cross-modal retrieval","volume":"29","author":"Xie","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.hcc.2024.100274_b10","article-title":"Deep cross-modal hashing with hashing functions and unified hash codes jointly learning","author":"Tu","year":"2020","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"10.1016\/j.hcc.2024.100274_b11","series-title":"Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","first-page":"1445","article-title":"Deep visual-semantic hashing for cross-modal retrieval","author":"Cao","year":"2016"},{"issue":"7","key":"10.1016\/j.hcc.2024.100274_b12","doi-asserted-by":"crossref","first-page":"3157","DOI":"10.1109\/TIP.2016.2564638","article-title":"Supervised matrix factorization hashing for cross-modal retrieval","volume":"25","author":"Tang","year":"2016","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.hcc.2024.100274_b13","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"3027","article-title":"Deep joint-semantics reconstructing hashing for large-scale unsupervised cross-modal retrieval","author":"Su","year":"2019"},{"key":"10.1016\/j.hcc.2024.100274_b14","series-title":"Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval","first-page":"1379","article-title":"Joint-modal distribution-based similarity hashing for large-scale unsupervised deep cross-modal retrieval","author":"Liu","year":"2020"},{"key":"10.1016\/j.hcc.2024.100274_b15","series-title":"Proceedings of the 2020 International Conference on Multimedia Retrieval","first-page":"44","article-title":"Deep semantic-alignment hashing for unsupervised cross-modal retrieval","author":"Yang","year":"2020"},{"issue":"5","key":"10.1016\/j.hcc.2024.100274_b16","first-page":"4626","article-title":"Deep graph-neighbor coherence preserving network for unsupervised cross-modal hashing","volume":"35","author":"Yu","year":"2021","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"10.1016\/j.hcc.2024.100274_b17","series-title":"Proceedings of the 25th ACM International Conference on Multimedia","first-page":"1398","article-title":"Deep binary reconstruction for cross-modal hashing","author":"Li","year":"2017"},{"issue":"3","key":"10.1016\/j.hcc.2024.100274_b18","first-page":"3877","article-title":"Unsupervised contrastive cross-modal hashing","volume":"45","author":"Hu","year":"2022","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.hcc.2024.100274_b19","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"2075","article-title":"Collective matrix factorization hashing for multimodal data","author":"Ding","year":"2014"},{"key":"10.1016\/j.hcc.2024.100274_b20","first-page":"25","article-title":"Self-supervised multimodal versatile networks","volume":"33","author":"Alayrac","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.hcc.2024.100274_b21","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"16000","article-title":"Masked autoencoders are scalable vision learners","author":"He","year":"2022"},{"key":"10.1016\/j.hcc.2024.100274_b22","series-title":"2022 IEEE International Conference on Multimedia and Expo","first-page":"1","article-title":"NHFNET: A non-homogeneous fusion network for multimodal sentiment analysis","author":"Fu","year":"2022"},{"key":"10.1016\/j.hcc.2024.100274_b23","series-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018"},{"key":"10.1016\/j.hcc.2024.100274_b24","series-title":"An image is worth 16x16 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020"},{"key":"10.1016\/j.hcc.2024.100274_b25","series-title":"Vl-bert: Pre-training of generic visual-linguistic representations","author":"Su","year":"2019"},{"key":"10.1016\/j.hcc.2024.100274_b26","article-title":"ViLBERT: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks","volume":"vol. 32","author":"Lu","year":"2019"},{"key":"10.1016\/j.hcc.2024.100274_b27","first-page":"8024","article-title":"Pytorch: An imperative style, high-performance deep learning library","volume":"vol. 32","author":"Paszke","year":"2019"},{"key":"10.1016\/j.hcc.2024.100274_b28","series-title":"Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation","author":"Li","year":"2022"},{"key":"10.1016\/j.hcc.2024.100274_b29","series-title":"Visualbert: A simple and performant baseline for vision and language","author":"Li","year":"2019"},{"key":"10.1016\/j.hcc.2024.100274_b30","series-title":"International Conference on Machine Learning","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","author":"Radford","year":"2021"},{"key":"10.1016\/j.hcc.2024.100274_b31","series-title":"Lxmert: Learning cross-modality encoder representations from transformers","author":"Tan","year":"2019"},{"key":"10.1016\/j.hcc.2024.100274_b32","series-title":"Proceedings of the 37th International ACM SIGIR Conference on Research & Development in Information Retrieval","first-page":"415","article-title":"Latent semantic sparse hashing for cross-modal similarity search","author":"Zhou","year":"2014"},{"issue":"11","key":"10.1016\/j.hcc.2024.100274_b33","doi-asserted-by":"crossref","first-page":"3507","DOI":"10.1109\/TKDE.2020.2974825","article-title":"BATCH: A scalable asymmetric discrete cross-modal hashing","volume":"33","author":"Wang","year":"2020","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"10.1016\/j.hcc.2024.100274_b34","series-title":"Proceedings of the 2013 ACM SIGMOD International Conference on Management of Data","first-page":"785","article-title":"Inter-media hashing for large-scale retrieval from heterogeneous data sources","author":"Song","year":"2013"},{"issue":"01","key":"10.1016\/j.hcc.2024.100274_b35","first-page":"176","article-title":"Coupled CycleGAN: Unsupervised hashing network for cross-modal retrieval","volume":"33","author":"Li","year":"2019","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"issue":"6","key":"10.1016\/j.hcc.2024.100274_b36","doi-asserted-by":"crossref","first-page":"84","DOI":"10.1145\/3065386","article-title":"ImageNet classification with deep convolutional neural networks","volume":"60","author":"Krizhevsky","year":"2017","journal-title":"Commun. ACM"},{"key":"10.1016\/j.hcc.2024.100274_b37","first-page":"26","article-title":"A general framework for parallel distributed processing","volume":"vol. 1(45\u201376)","author":"Rumelhart","year":"1986"},{"key":"10.1016\/j.hcc.2024.100274_b38","doi-asserted-by":"crossref","first-page":"43","DOI":"10.1007\/s13042-010-0001-0","article-title":"Understanding bag-of-words model: a statistical framework","volume":"1","author":"Zhang","year":"2010","journal-title":"Int. J. Mach. Learn. Cybern."},{"key":"10.1016\/j.hcc.2024.100274_b39","first-page":"24286","article-title":"One loss for all: Deep hashing with a single cosine similarity based learning objective","volume":"34","author":"Hoe","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"4","key":"10.1016\/j.hcc.2024.100274_b40","doi-asserted-by":"crossref","first-page":"824","DOI":"10.1109\/TPAMI.2013.225","article-title":"Multimodal similarity-preserving hashing","volume":"36","author":"Masci","year":"2013","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"1","key":"10.1016\/j.hcc.2024.100274_b41","article-title":"Pairwise relationship guided deep hashing for cross-modal retrieval","volume":"31","author":"Yang","year":"2017","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"10.1016\/j.hcc.2024.100274_b42","series-title":"Proceedings of the European Conference on Computer Vision","first-page":"354","article-title":"Deep hash distillation for image retrieval","author":"Jang","year":"2022"},{"key":"10.1016\/j.hcc.2024.100274_b43","series-title":"International Conference on Machine Learning","first-page":"1597","article-title":"A simple framework for contrastive learning of visual representations","author":"Chen","year":"2020"},{"issue":"1","key":"10.1016\/j.hcc.2024.100274_b44","doi-asserted-by":"crossref","first-page":"15:1","DOI":"10.1145\/3532624","article-title":"A survey on deep hashing methods","volume":"17","author":"Luo","year":"2023","journal-title":"ACM Trans. Knowl. Discov. Data"},{"key":"10.1016\/j.hcc.2024.100274_b45","series-title":"Proceedings of the 30th International Joint Conference on Artificial Intelligence","first-page":"959","article-title":"Unsupervised hashing with contrastive information bottleneck","author":"Qiu","year":"2021"},{"key":"10.1016\/j.hcc.2024.100274_b46","series-title":"IEEE International Conference on Acoustics, Speech and Signal Processing","first-page":"4463","article-title":"Unsupervised contrastive hashing for cross-modal retrieval in remote sensing","author":"Mikriukov","year":"2022"},{"key":"10.1016\/j.hcc.2024.100274_b47","series-title":"Proceedings of the IEEE International Conference on Computer Vision","first-page":"5608","article-title":"HashNet: Deep learning to hash by continuation","author":"Cao","year":"2017"},{"key":"10.1016\/j.hcc.2024.100274_b48","series-title":"Proceedings of the ACM International Conference on Image and Video Retrieval","first-page":"1","article-title":"Nus-wide: A real-world web image database from national university of Singapore","author":"Chua","year":"2009"},{"key":"10.1016\/j.hcc.2024.100274_b49","series-title":"Proceedings of the 1st ACM International Conference on Multimedia Information Retrieval","first-page":"39","article-title":"The MIR Flickr retrieval evaluation","author":"Huiskes","year":"2008"},{"key":"10.1016\/j.hcc.2024.100274_b50","series-title":"CCF Conference on Computer Supported Cooperative Work and Social Computing","first-page":"431","article-title":"Self-auxiliary hashing for unsupervised cross modal retrieval","author":"Xu","year":"2022"},{"key":"10.1016\/j.hcc.2024.100274_b51","series-title":"Computer Vision\u2013ECCV","first-page":"740","article-title":"Microsoft coco: Common objects in context","author":"Lin","year":"2014"},{"issue":"3","key":"10.1016\/j.hcc.2024.100274_b52","doi-asserted-by":"crossref","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","article-title":"ImageNet large scale visual recognition challenge","volume":"115","author":"Russakovsky","year":"2015","journal-title":"Int. J. Comput. Vis."}],"container-title":["High-Confidence Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S2667295224000771?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S2667295224000771?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,11,8]],"date-time":"2025-11-08T23:09:16Z","timestamp":1762643356000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S2667295224000771"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6]]},"references-count":52,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,6]]}},"alternative-id":["S2667295224000771"],"URL":"https:\/\/doi.org\/10.1016\/j.hcc.2024.100274","relation":{},"ISSN":["2667-2952"],"issn-type":[{"value":"2667-2952","type":"print"}],"subject":[],"published":{"date-parts":[[2025,6]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"A novel deep high-level concept-mining jointing hashing model for unsupervised cross-modal retrieval","name":"articletitle","label":"Article Title"},{"value":"High-Confidence Computing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.hcc.2024.100274","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2024 The Author(s). Published by Elsevier B.V. on behalf of Shandong University.","name":"copyright","label":"Copyright"}],"article-number":"100274"}}