{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T14:08:39Z","timestamp":1743084519040,"version":"3.40.3"},"publisher-location":"Cham","reference-count":19,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031441974"},{"type":"electronic","value":"9783031441981"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-44198-1_24","type":"book-chapter","created":{"date-parts":[[2023,9,21]],"date-time":"2023-09-21T08:02:34Z","timestamp":1695283354000},"page":"281-293","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Global-to-Contextual Shared Semantic Learning for\u00a0Fine-Grained Vision-Language Alignment"],"prefix":"10.1007","author":[{"given":"Min","family":"Zheng","sequence":"first","affiliation":[]},{"given":"Chunpeng","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Jiaqi","family":"Qin","sequence":"additional","affiliation":[]},{"given":"Weiwei","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Ming","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Long","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Fei","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,22]]},"reference":[{"key":"24_CR1","doi-asserted-by":"crossref","unstructured":"Chen, S., Zhao, Y., Jin, Q., Wu, Q.: Fine-grained video-text retrieval with hierarchical graph reasoning. In: CVPR, pp. 10635\u201310644 (2020)","DOI":"10.1109\/CVPR42600.2020.01065"},{"key":"24_CR2","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. ArXiv abs\/1810.04805 (2019)"},{"key":"24_CR3","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. In: International Conference on Learning Representations abs\/2010.11929 (2020)"},{"key":"24_CR4","unstructured":"Fei-Fei, L., Fergus, R., Perona, P.: Learning generative visual models from few training examples: an incremental Bayesian approach tested on 101 object categories. In: CVPR Workshop, p. 178 (2004)"},{"key":"24_CR5","doi-asserted-by":"crossref","unstructured":"Gu, J., Cai, J., Joty, S.R., Niu, L., Wang, G.: Look, imagine and match: improving textual-visual cross-modal retrieval with generative models. In: CVPR, pp. 7181\u20137189 (2018)","DOI":"10.1109\/CVPR.2018.00750"},{"key":"24_CR6","doi-asserted-by":"crossref","unstructured":"Hadfield, S., Wang, Z., Bryan O\u2019Gorman, B., Rieffel, E.G., Venturelli, D., Biswas, R.: From the quantum approximate optimization algorithm to a quantum alternating operator ansatz. Algorithms 12(2) (2019)","DOI":"10.3390\/a12020034"},{"key":"24_CR7","doi-asserted-by":"publisher","first-page":"1047","DOI":"10.1109\/TCYB.2018.2879846","volume":"50","author":"X Huang","year":"2017","unstructured":"Huang, X., Peng, Y., Yuan, M.: MHTN: modal-adversarial hybrid transfer network for cross-modal retrieval. IEEE Trans. Cybern. 50, 1047\u20131059 (2017)","journal-title":"IEEE Trans. Cybern."},{"key":"24_CR8","doi-asserted-by":"publisher","first-page":"2081","DOI":"10.1109\/TSG.2013.2261565","volume":"4","author":"JY Joo","year":"2013","unstructured":"Joo, J.Y., Ili\u0107, M.: Multi-layered optimization of demand resources using Lagrange dual decomposition. IEEE Trans. Smart Grid 4, 2081\u20132088 (2013)","journal-title":"IEEE Trans. Smart Grid"},{"key":"24_CR9","doi-asserted-by":"crossref","unstructured":"Lee, K.H., Chen, X.D., Hua, G., Hu, H., He, X.: Stacked cross attention for image-text matching. ArXiv abs\/1803.08024 (2018)","DOI":"10.1007\/978-3-030-01225-0_13"},{"key":"24_CR10","doi-asserted-by":"crossref","unstructured":"Mandal, D., Chaudhury, K.N., Biswas, S.: Generalized semantic preserving hashing for N-label cross-modal retrieval. In: CVPR, pp. 2633\u20132641 (2017)","DOI":"10.1109\/CVPR.2017.282"},{"key":"24_CR11","unstructured":"Peng, Y., Huang, X., Qi, J.: Cross-media shared representation by hierarchical learning with multiple deep networks. In: IJCAI, pp. 3846\u20133853 (2016)"},{"key":"24_CR12","doi-asserted-by":"crossref","unstructured":"Reed, S., Akata, Z., Lee, H., Schiele, B.: Learning deep representations of fine-grained visual descriptions. In: CVPR, pp. 49\u201358 (2016)","DOI":"10.1109\/CVPR.2016.13"},{"key":"24_CR13","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: ICLR, pp. 1\u201314 (2015)"},{"key":"24_CR14","unstructured":"Wah, C., Branson, S., Welinder, P., Perona, P., Belongie, S.: The caltech-UCSD birds-200-2011 dataset. Technical report (2011)"},{"key":"24_CR15","doi-asserted-by":"crossref","unstructured":"Wang, B., Yang, Y., Xu, X., Hanjalic, A., Shen, H.T.: Adversarial cross-modal retrieval. In: ACM International Conference on Multimedia, pp. 154\u2013162 (2017)","DOI":"10.1145\/3123266.3123326"},{"key":"24_CR16","doi-asserted-by":"crossref","unstructured":"Wu, P., He, X., Tang, M., Lv, Y., Liu, J.: HANet: hierarchical alignment networks for video-text retrieval. In: ACM MM (2021)","DOI":"10.1145\/3474085.3475515"},{"key":"24_CR17","doi-asserted-by":"crossref","unstructured":"He, X., Peng, Y., Xie, L.: A new benchmark and approach for fine-grained cross-media retrieva. In: ACM International Conference on Multimedia, pp. 1740\u20131748 (2019)","DOI":"10.1145\/3343031.3350974"},{"key":"24_CR18","first-page":"965","volume":"24","author":"X Zhai","year":"2014","unstructured":"Zhai, X., Peng, Y., Xiao, J.: Learning cross-media joint representation with sparse and semisupervised regularization. IEEE TCSVT 24, 965\u2013978 (2014)","journal-title":"IEEE TCSVT"},{"key":"24_CR19","doi-asserted-by":"publisher","first-page":"643","DOI":"10.1109\/LSP.2021.3065595","volume":"28","author":"M Zheng","year":"2021","unstructured":"Zheng, M., Wang, W., Li, Q.: Fine-grained image-text retrieval via discriminative latent space learning. IEEE Signal Process. Lett. 28, 643\u2013647 (2021)","journal-title":"IEEE Signal Process. Lett."}],"container-title":["Lecture Notes in Computer Science","Artificial Neural Networks and Machine Learning \u2013 ICANN 2023"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-44198-1_24","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,28]],"date-time":"2023-11-28T22:13:13Z","timestamp":1701209593000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-44198-1_24"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031441974","9783031441981"],"references-count":19,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-44198-1_24","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"22 September 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICANN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Neural Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Heraklion","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Greece","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 September 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"32","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icann2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/e-nns.org\/icann2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"easyacademia.org","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"947","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"426","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"22","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.4","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"type of other papers accepted  : 9 Abstract","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}