{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,24]],"date-time":"2025-08-24T01:39:58Z","timestamp":1755999598676,"version":"3.40.3"},"publisher-location":"Cham","reference-count":27,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031533013"},{"type":"electronic","value":"9783031533020"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-53302-0_12","type":"book-chapter","created":{"date-parts":[[2024,1,28]],"date-time":"2024-01-28T09:02:09Z","timestamp":1706432529000},"page":"161-174","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Mining Landmark Images for\u00a0Scene Reconstruction from\u00a0Weakly Annotated Video Collections"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8153-6840","authenticated-orcid":false,"given":"Helmut","family":"Neuschmied","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2442-4900","authenticated-orcid":false,"given":"Werner","family":"Bailer","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,29]]},"reference":[{"key":"12_CR1","doi-asserted-by":"crossref","unstructured":"Boiarov, A., Tyantov, E.: Large scale landmark recognition via deep metric learning. In: Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 169\u2013178 (2019)","DOI":"10.1145\/3357384.3357956"},{"key":"12_CR2","doi-asserted-by":"crossref","unstructured":"Brodersen, K.H., Ong, C.S., Stephan, K.E., Buhmann, J.M.: The balanced accuracy and its posterior distribution. In: 2010 20th International Conference on Pattern Recognition, pp. 3121\u20133124. IEEE (2010)","DOI":"10.1109\/ICPR.2010.764"},{"key":"12_CR3","unstructured":"Caimotti, E., Montagnuolo, M., Messina, A.: An efficient visual search engine for cultural broadcast archives. In: AI* CH@ AI* IA, pp. 1\u20138 (2017)"},{"key":"12_CR4","doi-asserted-by":"crossref","unstructured":"Chen, X., Gupta, A.: Webly supervised learning of convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), December 2015","DOI":"10.1109\/ICCV.2015.168"},{"issue":"2","key":"12_CR5","doi-asserted-by":"publisher","first-page":"44","DOI":"10.1109\/MMUL.2018.2873844","volume":"26","author":"LY Duan","year":"2019","unstructured":"Duan, L.Y., et al.: Compact descriptors for video analysis: the emerging mpeg standard. IEEE Multimedia 26(2), 44\u201354 (2019). https:\/\/doi.org\/10.1109\/MMUL.2018.2873844","journal-title":"IEEE Multimedia"},{"issue":"6","key":"12_CR6","doi-asserted-by":"publisher","first-page":"381","DOI":"10.1145\/358669.358692","volume":"24","author":"MA Fischler","year":"1981","unstructured":"Fischler, M.A., Bolles, R.C.: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 24(6), 381\u2013395 (1981). https:\/\/doi.org\/10.1145\/358669.358692","journal-title":"Commun. ACM"},{"key":"12_CR7","first-page":"17136","volume":"34","author":"M G\u00f6sgens","year":"2021","unstructured":"G\u00f6sgens, M., Zhiyanov, A., Tikhonov, A., Prokhorenkova, L.: Good classification measures and how to find them. Adv. Neural. Inf. Process. Syst. 34, 17136\u201317147 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"12_CR8","unstructured":"Li, K., et al.: Learning from weakly-labeled web videos via exploring sub-concepts. CoRR abs\/2101.03713 (2021). https:\/\/arxiv.org\/abs\/2101.03713"},{"key":"12_CR9","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"12_CR10","unstructured":"Mezuman, E., Weiss, Y.: Learning about canonical views from internet image collections. In: Pereira, F., Burges, C., Bottou, L., Weinberger, K. (eds.) Advances in Neural Information Processing Systems, vol. 25. Curran Associates, Inc. (2012)"},{"key":"12_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"12_CR12","doi-asserted-by":"crossref","unstructured":"Noh, H., Araujo, A., Sim, J., Weyand, T., Han, B.: Large-scale image retrieval with attentive deep local features. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3456\u20133465 (2017)","DOI":"10.1109\/ICCV.2017.374"},{"key":"12_CR13","doi-asserted-by":"publisher","unstructured":"Perd\u2019och, M., Chum, O., Matas, J.: Efficient representation of local geometry for large scale object retrieval. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 9\u201316 (2009). https:\/\/doi.org\/10.1109\/CVPR.2009.5206529","DOI":"10.1109\/CVPR.2009.5206529"},{"key":"12_CR14","doi-asserted-by":"crossref","unstructured":"Perronnin, F., Liu, Y., Renders, J.M.: A family of contextual measures of similarity between distributions with application to image retrieval. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 2358\u20132365. IEEE (2009)","DOI":"10.1109\/CVPR.2009.5206505"},{"key":"12_CR15","doi-asserted-by":"crossref","unstructured":"Philbin, J., Chum, O., Isard, M., Sivic, J., Zisserman, A.: Object retrieval with large vocabularies and fast spatial matching. In: 2007 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1\u20138. IEEE (2007)","DOI":"10.1109\/CVPR.2007.383172"},{"key":"12_CR16","doi-asserted-by":"crossref","unstructured":"Philbin, J., Chum, O., Isard, M., Sivic, J., Zisserman, A.: Lost in quantization: improving particular object retrieval in large scale image databases. In: 2008 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1\u20138. IEEE (2008)","DOI":"10.1109\/CVPR.2008.4587635"},{"key":"12_CR17","doi-asserted-by":"crossref","unstructured":"Razali, M.N.B., Tony, E.O.N., Ibrahim, A.A.A., Hanapi, R., Iswandono, Z.: Landmark recognition model for smart tourism using lightweight deep learning and linear discriminant analysis. Int. J. Adv. Comput. Sci. Appl. (2023). https:\/\/api.semanticscholar.org\/CorpusID:257386803","DOI":"10.14569\/IJACSA.2023.0140225"},{"key":"12_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"349","DOI":"10.1007\/978-3-030-05710-7_29","volume-title":"MultiMedia Modeling","author":"L Rossetto","year":"2019","unstructured":"Rossetto, L., Schuldt, H., Awad, G., Butt, A.A.: V3C \u2013 a research video collection. In: Kompatsiaris, I., Huet, B., Mezaris, V., Gurrin, C., Cheng, W.-H., Vrochidis, S. (eds.) MMM 2019. LNCS, vol. 11295, pp. 349\u2013360. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-05710-7_29"},{"key":"12_CR19","doi-asserted-by":"publisher","first-page":"8135","DOI":"10.1109\/TNNLS.2022.3152527","volume":"34","author":"H Song","year":"2022","unstructured":"Song, H., Kim, M., Park, D., Shin, Y., Lee, J.G.: Learning from noisy labels with deep neural networks: a survey. IEEE Trans. Neural Netw. Learn. Syst. 34, 8135\u20138153 (2022)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"12_CR20","unstructured":"Tan, M., Le, Q.: EfficientNet: rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning, pp. 6105\u20136114. PMLR (2019)"},{"key":"12_CR21","doi-asserted-by":"crossref","unstructured":"Torii, A., Arandjelovi\u0107, R., Sivic, J., Okutomi, M., Pajdla, T.: 24\/7 place recognition by view synthesis. In: CVPR (2015)","DOI":"10.1109\/CVPR.2015.7298790"},{"key":"12_CR22","doi-asserted-by":"publisher","unstructured":"Torii, A., Sivic, J., Pajdla, T., Okutomi, M.: Visual place recognition with repetitive structures. In: Proceedings \/ CVPR, IEEE Computer Society Conference on Computer Vision and Pattern Recognition, June 2013, vol. 37, pp. 883\u2013890. IEEE Computer Society Conference on Computer Vision and Pattern Recognition (2013). https:\/\/doi.org\/10.1109\/CVPR.2013.119","DOI":"10.1109\/CVPR.2013.119"},{"key":"12_CR23","doi-asserted-by":"crossref","unstructured":"Weyand, T., Araujo, A., Cao, B., Sim, J.: Google landmarks dataset v2-a large-scale benchmark for instance-level recognition and retrieval. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2575\u20132584 (2020)","DOI":"10.1109\/CVPR42600.2020.00265"},{"key":"12_CR24","unstructured":"Yang, M., Cui, C., Xue, X., Ren, H., Wei, K.: 2nd place solution to google landmark retrieval 2020 (2022)"},{"key":"12_CR25","doi-asserted-by":"crossref","unstructured":"Yokoo, S., Ozaki, K., Simo-Serra, E., Iizuka, S.: Two-stage discriminative re-ranking for large-scale landmark retrieval. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 1012\u20131013 (2020)","DOI":"10.1109\/CVPRW50498.2020.00514"},{"key":"12_CR26","unstructured":"Yu, F.: BDD100K: a large-scale diverse driving video database. BAIR (Berkeley Artificial Intelligence Research) (2018). https:\/\/bair.berkeley.edu\/blog\/2018\/05\/30\/bdd"},{"key":"12_CR27","unstructured":"Zhuang, P., Wang, Y., Qiao, Y.: Learning attentive pairwise interaction for fine-grained classification. CoRR abs\/2002.10191 (2020). https:\/\/arxiv.org\/abs\/2002.10191"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-53302-0_12","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,7]],"date-time":"2024-03-07T11:57:30Z","timestamp":1709812650000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-53302-0_12"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031533013","9783031533020"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-53302-0_12","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"29 January 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Amsterdam","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"The Netherlands","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 January 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 February 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ConfTool Pro","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"297","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"112","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"38% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}