{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T09:48:21Z","timestamp":1743155301970,"version":"3.40.3"},"publisher-location":"Cham","reference-count":15,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031533013"},{"type":"electronic","value":"9783031533020"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-53302-0_23","type":"book-chapter","created":{"date-parts":[[2024,1,28]],"date-time":"2024-01-28T09:02:09Z","timestamp":1706432529000},"page":"288-294","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Few-Shot Object Detection as\u00a0a\u00a0Service: Facilitating Training and\u00a0Deployment for\u00a0Domain Experts"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2442-4900","authenticated-orcid":false,"given":"Werner","family":"Bailer","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8189-8566","authenticated-orcid":false,"given":"Mihai","family":"Dogariu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7728-0640","authenticated-orcid":false,"given":"Bogdan","family":"Ionescu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7113-0038","authenticated-orcid":false,"given":"Hannes","family":"Fassold","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,29]]},"reference":[{"doi-asserted-by":"crossref","unstructured":"Afrasiyabi, A., Larochelle, H., Lalonde, J.F., Gagn\u00e9, C.: Matching feature sets for few-shot image classification. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9014\u20139024 (2022)","key":"23_CR1","DOI":"10.1109\/CVPR52688.2022.00881"},{"key":"23_CR2","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"445","DOI":"10.1007\/978-3-030-98355-0_37","volume-title":"MultiMedia Modeling","author":"W Bailer","year":"2022","unstructured":"Bailer, W.: Making few-shot object detection simpler and less frustrating. In: THorn\u00f3r J\u00f3nsson, B., et al. (eds.) MMM 2022. LNCS, vol. 13142, pp. 445\u2013451. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-030-98355-0_37"},{"doi-asserted-by":"crossref","unstructured":"Bar, A., et al.: DETReg: unsupervised pretraining with region priors for object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14605\u201314615 (2022)","key":"23_CR3","DOI":"10.1109\/CVPR52688.2022.01420"},{"key":"23_CR4","doi-asserted-by":"publisher","first-page":"229","DOI":"10.1016\/j.neucom.2021.04.001","volume":"449","author":"M Carranza-Garc\u00eda","year":"2021","unstructured":"Carranza-Garc\u00eda, M., Lara-Ben\u00edtez, P., Garc\u00eda-Guti\u00e9rrez, J., Riquelme, J.C.: Enhancing object detection for autonomous driving by optimizing anchor generation and addressing class imbalance. Neurocomputing 449, 229\u2013244 (2021)","journal-title":"Neurocomputing"},{"unstructured":"Casado-Garc\u00eda, \u00c1., Heras, J.: Ensemble methods for object detection. In: ECAI 2020, pp. 2688\u20132695. IOS Press (2020)","key":"23_CR5"},{"unstructured":"Deleu, T., W\u00fcrfl, T., Samiei, M., Cohen, J.P., Bengio, Y.: TorchMeta: a meta-learning library for PyTorch (2019). https:\/\/arxiv.org\/abs\/1909.06576, https:\/\/github.com\/tristandeleu\/pytorch-meta","key":"23_CR6"},{"doi-asserted-by":"crossref","unstructured":"Dvornik, N., Schmid, C., Mairal, J.: Diversity with cooperation: ensemble methods for few-shot classification. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3723\u20133731 (2019)","key":"23_CR7","DOI":"10.1109\/ICCV.2019.00382"},{"doi-asserted-by":"crossref","unstructured":"Gassen, M., et al.: I3: interactive iterative improvement for few-shot action segmentation. In: 2023 32nd IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), Busan, South Korea (2023)","key":"23_CR8","DOI":"10.1109\/RO-MAN57019.2023.10309435"},{"doi-asserted-by":"crossref","unstructured":"Lee, J., Lee, S.K., Yang, S.I.: An ensemble method of CNN models for object detection. In: 2018 International Conference on Information and Communication Technology Convergence (ICTC), pp. 898\u2013901. IEEE (2018)","key":"23_CR9","DOI":"10.1109\/ICTC.2018.8539396"},{"doi-asserted-by":"publisher","unstructured":"Lin, T.Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) Computer Vision-ECCV 2014: 13th European Conference, Zurich, Switzerland, 6\u201312 September 2014, Proceedings, Part V 13, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48","key":"23_CR10","DOI":"10.1007\/978-3-319-10602-1_48"},{"doi-asserted-by":"crossref","unstructured":"Liu, F., et al.: Integrally migrating pre-trained transformer encoder-decoders for visual object detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6825\u20136834 (2023)","key":"23_CR11","DOI":"10.1109\/ICCV51070.2023.00628"},{"doi-asserted-by":"publisher","unstructured":"Miyata, S., Chang, C.M., Igarashi, T.: Trafne: a training framework for non-expert annotators with auto validation and expert feedback. In: Degen, H., Ntoa, S. (eds.) International Conference on Human-Computer Interaction, pp. 475\u2013494. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-05643-7_31","key":"23_CR12","DOI":"10.1007\/978-3-031-05643-7_31"},{"doi-asserted-by":"publisher","unstructured":"Tian, Y., Wang, Y., Krishnan, D., Tenenbaum, J.B., Isola, P.: Rethinking few-shot image classification: a good embedding is all you need? In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M. (eds.) Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, 23\u201328 August 2020, Proceedings, Part XIV 16, pp. 266\u2013282. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58568-6_16","key":"23_CR13","DOI":"10.1007\/978-3-030-58568-6_16"},{"unstructured":"Wang, X., Huang, T., Gonzalez, J., Darrell, T., Yu, F.: Frustratingly simple few-shot object detection. In: International Conference on Machine Learning, pp. 9919\u20139928. PMLR (2020)","key":"23_CR14"},{"issue":"3","key":"23_CR15","first-page":"3090","volume":"45","author":"Y Xiao","year":"2022","unstructured":"Xiao, Y., Lepetit, V., Marlet, R.: Few-shot object detection and viewpoint estimation for objects in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 3090\u20133106 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-53302-0_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,7]],"date-time":"2024-03-07T11:59:58Z","timestamp":1709812798000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-53302-0_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031533013","9783031533020"],"references-count":15,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-53302-0_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"29 January 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Amsterdam","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"The Netherlands","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 January 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 February 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ConfTool Pro","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"297","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"112","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"38% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}