{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T03:48:10Z","timestamp":1769831290913,"version":"3.49.0"},"publisher-location":"Cham","reference-count":43,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030586034","type":"print"},{"value":"9783030586041","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-58604-1_15","type":"book-chapter","created":{"date-parts":[[2020,11,2]],"date-time":"2020-11-02T22:02:49Z","timestamp":1604354569000},"page":"237-253","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":29,"title":["Attentive Prototype Few-Shot Learning with Capsule Network-Based Embedding"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9618-8965","authenticated-orcid":false,"given":"Fangyu","family":"Wu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0212-2365","authenticated-orcid":false,"given":"Jeremy S.","family":"Smith","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7038-3789","authenticated-orcid":false,"given":"Wenjin","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Chaoyi","family":"Pang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5762-5763","authenticated-orcid":false,"given":"Bailing","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,11,3]]},"reference":[{"key":"15_CR1","unstructured":"Arik, S.O., Pfister, T.: Attention-based prototypical learning towards interpretable, confident and robust deep neural networks. arXiv preprint arXiv:1902.06292 (2019)"},{"key":"15_CR2","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 248\u2013255 (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"15_CR3","doi-asserted-by":"crossref","unstructured":"Fe-Fei, L., et al.: A Bayesian approach to unsupervised one-shot learning of object categories. In: IEEE International Conference on Computer Vision (ICCV), pp. 1134\u20131141 (2003)","DOI":"10.1109\/ICCV.2003.1238476"},{"issue":"4","key":"15_CR4","doi-asserted-by":"publisher","first-page":"594","DOI":"10.1109\/TPAMI.2006.79","volume":"28","author":"L Fei-Fei","year":"2006","unstructured":"Fei-Fei, L., Fergus, R., Perona, P.: One-shot learning of object categories. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 28(4), 594\u2013611 (2006)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI)"},{"key":"15_CR5","unstructured":"Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: Proceedings of the 34th International Conference on Machine Learning (ICML), pp. 1126\u20131135 (2017)"},{"key":"15_CR6","unstructured":"Fukunaga, K.: Introduction to Statistical Pattern Recognition. Elsevier, New York (2013)"},{"key":"15_CR7","doi-asserted-by":"crossref","unstructured":"Gao, T., Han, X., Liu, Z., Sun, M.: Hybrid attention-based prototypical networks for noisy few-shot relation classification. In: AAAI Conference on Artificial Intelligence (AAAI) (2019)","DOI":"10.1609\/aaai.v33i01.33016407"},{"key":"15_CR8","doi-asserted-by":"crossref","unstructured":"Gidaris, S., Komodakis, N.: Dynamic few-shot visual learning without forgetting. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4367\u20134375 (2018)","DOI":"10.1109\/CVPR.2018.00459"},{"key":"15_CR9","doi-asserted-by":"crossref","unstructured":"Gidaris, S., Komodakis, N.: Generating classification weights with GNN denoising autoencoders for few-shot learning. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/CVPR.2019.00011"},{"key":"15_CR10","unstructured":"Hermans, A., Beyer, L., Leibe, B.: In defense of the triplet loss for person re-identification. arXiv preprint arXiv:1703.07737 (2017)"},{"key":"15_CR11","doi-asserted-by":"publisher","first-page":"44","DOI":"10.1007\/978-3-642-21735-7_6","volume-title":"International Conference on Artificial Neural Networks","author":"GE Hinton","year":"2011","unstructured":"Hinton, G.E., Krizhevsky, A., Wang, S.D.: Transforming auto-encoders. In: Honkela, T., Duch, W., Girolami, M., Kaski, S. (eds.) ICANN 2011, vol. 6791, pp. 44\u201351. Springer, Berlin, Heidelberg (2011). https:\/\/doi.org\/10.1007\/978-3-642-21735-7_6"},{"key":"15_CR12","doi-asserted-by":"crossref","unstructured":"Kim, J., Kim, T., Kim, S., Yoo, C.D.: Edge-labeling graph neural network for few-shot learning. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11\u201320 (2019)","DOI":"10.1109\/CVPR.2019.00010"},{"key":"15_CR13","unstructured":"Kosiorek, A.R., Sabour, S., Teh, Y.W., Hinton, G.E.: Stacked capsule autoencoders. arXiv preprint arXiv:1906.06818 (2019)"},{"key":"15_CR14","unstructured":"Krizhevsky, A., Hinton, G., et al.: Learning multiple layers of features from tiny images. Technical report, Citeseer (2009)"},{"key":"15_CR15","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems (NIPS), pp. 1097\u20131105 (2012)"},{"key":"15_CR16","doi-asserted-by":"crossref","unstructured":"Lee, K., Maji, S., Ravichandran, A., Soatto, S.: Meta-learning with differentiable convex optimization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019","DOI":"10.1109\/CVPR.2019.01091"},{"key":"15_CR17","unstructured":"Lenssen, J.E., Fey, M., Libuschewski, P.: Group equivariant capsule networks. In: Advances in Neural Information Processing Systems (NIPS), pp. 8844\u20138853 (2018)"},{"key":"15_CR18","doi-asserted-by":"crossref","unstructured":"Li, H., Eigen, D., Dodge, S., Zeiler, M., Wang, X.: Finding task-relevant features for few-shot learning by category traversal. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1\u201310 (2019)","DOI":"10.1109\/CVPR.2019.00009"},{"key":"15_CR19","unstructured":"Li, Z., Zhou, F., Chen, F., Li, H.: Meta-SQD: learning to learn quickly for few-shot learning. arXiv preprint arXiv:1707.09835 (2017)"},{"key":"15_CR20","unstructured":"Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9, 2579\u20132605 (2008)"},{"key":"15_CR21","unstructured":"Mishra, N., Rohaninejad, M., Chen, X., Abbeel, P.: A simple neural attentive meta-learner. arXiv preprint arXiv:1707.03141 (2017)"},{"issue":"10","key":"15_CR22","doi-asserted-by":"publisher","first-page":"1151","DOI":"10.1016\/j.patrec.2005.12.016","volume":"27","author":"Y Mitani","year":"2006","unstructured":"Mitani, Y., Hamamoto, Y.: A local mean-based nonparametric classifier. Pattern Recogn. Lett. 27(10), 1151\u20131159 (2006)","journal-title":"Pattern Recogn. Lett."},{"key":"15_CR23","unstructured":"Nichol, A., Achiam, J., Schulman, J.: On first-order meta-learning algorithms. arXiv preprint arXiv:1803.02999 (2018)"},{"key":"15_CR24","unstructured":"Oreshkin, B., L\u00f3pez, P.R., Lacoste, A.: Tadam: task dependent adaptive metric for improved few-shot learning. In: Advances in Neural Information Processing Systems (NIPS), pp. 721\u2013731 (2018)"},{"key":"15_CR25","unstructured":"Peng, H., et al.: Hierarchical taxonomy-aware and attentional graph capsule RCNNs for large-scale multi-label text classification. arXiv preprint arXiv:1906.04898 (2019)"},{"key":"15_CR26","doi-asserted-by":"crossref","unstructured":"Qiao, S., Liu, C., Shen, W., Yuille, A.L.: Few-shot image recognition by predicting parameters from activations. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7229\u20137238 (2018)","DOI":"10.1109\/CVPR.2018.00755"},{"key":"15_CR27","doi-asserted-by":"crossref","unstructured":"Rajasegaran, J., Jayasundara, V., Jayasekara, S., Jayasekara, H., Seneviratne, S., Rodrigo, R.: Deepcaps: going deeper with capsule networks. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10725\u201310733 (2019)","DOI":"10.1109\/CVPR.2019.01098"},{"key":"15_CR28","unstructured":"Rawlinson, D., Ahmed, A., Kowadlo, G.: Sparse unsupervised capsules generalize better. arXiv preprint arXiv:1804.06094 (2018)"},{"key":"15_CR29","unstructured":"Ren, M., et al.: Meta-learning for semi-supervised few-shot classification. In: International Conference on Learning Representations (ICLR) (2018)"},{"issue":"3","key":"15_CR30","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., et al.: Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 115(3), 211\u2013252 (2015)","journal-title":"Int. J. Comput. Vis."},{"key":"15_CR31","unstructured":"Rusu, A.A., Rao, D., Sygnowski, J., Vinyals, O., Pascanu, R., Osindero, S., Hadsell, R.: Meta-learning with latent embedding optimization. In: International Conference on Learning Representations (ICLR) (2018)"},{"key":"15_CR32","unstructured":"Sabour, S., Frosst, N., Hinton, G.E.: Dynamic routing between capsules. In: Advances in Neural Information Processing Systems (NIPS), pp. 3856\u20133866 (2017)"},{"key":"15_CR33","doi-asserted-by":"crossref","unstructured":"Schroff, F., Kalenichenko, D., Philbin, J.: Facenet: a unified embedding for face recognition and clustering. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 815\u2013823 (2015)","DOI":"10.1109\/CVPR.2015.7298682"},{"key":"15_CR34","unstructured":"Snell, J., Swersky, K., Zemel, R.: Prototypical networks for few-shot learning. In: Advances in Neural Information Processing Systems (NIPS), pp. 4077\u20134087 (2017)"},{"key":"15_CR35","unstructured":"Sohn, K.: Improved deep metric learning with multi-class n-pair loss objective. In: Advances in Neural Information Processing Systems (NIPS), pp. 1857\u20131865 (2016)"},{"key":"15_CR36","doi-asserted-by":"crossref","unstructured":"Sun, Q., Liu, Y., Chua, T.S., Schiele, B.: Meta-transfer learning for few-shot learning. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 403\u2013412 (2019)","DOI":"10.1109\/CVPR.2019.00049"},{"key":"15_CR37","doi-asserted-by":"crossref","unstructured":"Sung, F., Yang, Y., Zhang, L., Xiang, T., Torr, P.H., Hospedales, T.M.: Learning to compare: Relation network for few-shot learning. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1199\u20131208 (2018)","DOI":"10.1109\/CVPR.2018.00131"},{"key":"15_CR38","doi-asserted-by":"crossref","unstructured":"Szegedy, C., et al.: Going deeper with convolutions. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1\u20139 (2015)","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"15_CR39","unstructured":"Vinyals, O., Blundell, C., Lillicrap, T., Wierstra, D., et al.: Matching networks for one shot learning. In: Advances in Neural Information Processing Systems (NIPS), pp. 3630\u20133638 (2016)"},{"key":"15_CR40","doi-asserted-by":"crossref","unstructured":"Xia, C., Zhang, C., Yan, X., Chang, Y., Yu, P.S.: Zero-shot user intent detection via capsule neural networks. arXiv preprint arXiv:1809.00385 (2018)","DOI":"10.18653\/v1\/D18-1348"},{"issue":"5","key":"15_CR41","doi-asserted-by":"publisher","first-page":"494","DOI":"10.3390\/rs11050494","volume":"11","author":"W Zhang","year":"2019","unstructured":"Zhang, W., Tang, P., Zhao, L.: Remote sensing image scene classification using CNN-capsNET. Remote Sens. 11(5), 494 (2019)","journal-title":"Remote Sens."},{"issue":"1","key":"15_CR42","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1002\/ima.22291","volume":"29","author":"X Zhang","year":"2019","unstructured":"Zhang, X., Zhao, S.G.: Cervical image classification based on image segmentation preprocessing and a CapsNET network model. Int. J. Imaging Syst. Technol. 29(1), 19\u201328 (2019)","journal-title":"Int. J. Imaging Syst. Technol."},{"key":"15_CR43","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Birdal, T., Deng, H., Tombari, F.: 3D point capsule networks. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1009\u20131018 (2019)","DOI":"10.1109\/CVPR.2019.00110"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2020"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-58604-1_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T00:11:53Z","timestamp":1730506313000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-58604-1_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030586034","9783030586041"],"references-count":43,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-58604-1_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"3 November 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Glasgow","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 August 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2020.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"OpenReview","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5025","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1360","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"27% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually due to the COVID-19 pandemic.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}