{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T18:12:11Z","timestamp":1775067131933,"version":"3.50.1"},"publisher-location":"Cham","reference-count":60,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030585679","type":"print"},{"value":"9783030585686","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-58568-6_16","type":"book-chapter","created":{"date-parts":[[2020,11,12]],"date-time":"2020-11-12T14:04:57Z","timestamp":1605189897000},"page":"266-282","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":571,"title":["Rethinking Few-Shot Image Classification: A Good Embedding is All You Need?"],"prefix":"10.1007","author":[{"given":"Yonglong","family":"Tian","sequence":"first","affiliation":[]},{"given":"Yue","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Dilip","family":"Krishnan","sequence":"additional","affiliation":[]},{"given":"Joshua B.","family":"Tenenbaum","sequence":"additional","affiliation":[]},{"given":"Phillip","family":"Isola","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,11,13]]},"reference":[{"key":"16_CR1","unstructured":"Machine learning in python. https:\/\/scikit-learn.org\/stable\/"},{"key":"16_CR2","unstructured":"Allen, K., Shelhamer, E., Shin, H., Tenenbaum, J.: Infinite mixture prototypes for few-shot learning. In: ICML (2019)"},{"key":"16_CR3","unstructured":"Bertinetto, L., Henriques, J.F., Torr, P.H., Vedaldi, A.: Meta-learning with differentiable closed-form solvers. arXiv preprint arXiv:1805.08136 (2018)"},{"key":"16_CR4","doi-asserted-by":"crossref","unstructured":"Bucilu\u01ce, C., Caruana, R., Niculescu-Mizil, A.: Model compression. In: SIGKDD (2006)","DOI":"10.1145\/1150402.1150464"},{"key":"16_CR5","unstructured":"Chen, W.Y., Liu, Y.C., Kira, Z., Wang, Y.C., Huang, J.B.: A closer look at few-shot classification. In: ICLR (2019)"},{"key":"16_CR6","unstructured":"Chen, Y., Wang, X., Liu, Z., Xu, H., Darrell, T.: A new meta-baseline for few-shot learning. ArXiv abs\/2003.04390 (2020)"},{"key":"16_CR7","doi-asserted-by":"crossref","unstructured":"Clark, K., Luong, M.T., Manning, C.D., Le, Q.V.: Bam! born-again multi-task networks for natural language understanding. In: ACL (2019)","DOI":"10.18653\/v1\/P19-1595"},{"key":"16_CR8","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: CVPR09 (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"16_CR9","unstructured":"Dhillon, G.S., Chaudhari, P., Ravichandran, A., Soatto, S.: A baseline for few-shot image classification. In: ICLR (2020)"},{"key":"16_CR10","doi-asserted-by":"crossref","unstructured":"Dvornik, N., Schmid, C., Mairal, J.: Diversity with cooperation: ensemble methods for few-shot classification. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00382"},{"key":"16_CR11","unstructured":"Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: ICML (2017)"},{"key":"16_CR12","unstructured":"Furlanello, T., Lipton, Z.C., Tschannen, M., Itti, L., Anandkumar, A.: Born-again neural networks. In: ICML (2018)"},{"key":"16_CR13","doi-asserted-by":"crossref","unstructured":"Gan, C., Gong, B., Liu, K., Su, H., Guibas, L.J.: Geometry guided convolutional neural networks for self-supervised video representation learning. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00586"},{"key":"16_CR14","doi-asserted-by":"crossref","unstructured":"Gan, C., Zhao, H., Chen, P., Cox, D., Torralba, A.: Self-supervised moving vehicle tracking with stereo sound. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00715"},{"key":"16_CR15","doi-asserted-by":"crossref","unstructured":"Gidaris, S., Komodakis, N.: Dynamic few-shot visual learning without forgetting. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00459"},{"key":"16_CR16","doi-asserted-by":"crossref","unstructured":"Hao, F., He, F., Cheng, J., Wang, L., Cao, J., Tao, D.: Collect and select: semantic alignment metric learning for few-shot learning. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00855"},{"key":"16_CR17","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.B.: Momentum contrast for unsupervised visual representation learning. ArXiv abs\/1911.05722 (2019)","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"16_CR18","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"16_CR19","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. In: NIPS Deep Learning and Representation Learning Workshop (2015)"},{"key":"16_CR20","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"16_CR21","unstructured":"Huang, S., Tao, D.: All you need is a good representation: A multi-level and classifier-centric representation for few-shot learning. ArXiv abs\/1911.12476 (2019)"},{"key":"16_CR22","doi-asserted-by":"crossref","unstructured":"Jamal, M.A., Qi, G.J.: Task agnostic meta-learning for few-shot learning. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01199"},{"key":"16_CR23","unstructured":"Koch, G., Zemel, R., Salakhutdinov, R.: Siamese neural networks for one-shot image recognition. In: ICML Deep Learning Workshop (2015)"},{"issue":"6266","key":"16_CR24","doi-asserted-by":"publisher","first-page":"1332","DOI":"10.1126\/science.aab3050","volume":"350","author":"BM Lake","year":"2015","unstructured":"Lake, B.M., Salakhutdinov, R., Tenenbaum, J.B.: Human-level concept learning through probabilistic program induction. Science 350(6266), 1332\u20131338 (2015)","journal-title":"Science"},{"key":"16_CR25","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1016\/j.cobeha.2019.04.007","volume":"29","author":"BM Lake","year":"2019","unstructured":"Lake, B.M., Salakhutdinov, R., Tenenbaum, J.B.: The Omniglot challenge: a 3-year progress report. Curr. Opin. Behav. Sci. 29, 97\u2013104 (2019)","journal-title":"Curr. Opin. Behav. Sci."},{"key":"16_CR26","doi-asserted-by":"crossref","unstructured":"Lee, K., Maji, S., Ravichandran, A., Soatto, S.: Meta-learning with differentiable convex optimization. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01091"},{"key":"16_CR27","doi-asserted-by":"crossref","unstructured":"Li, A., Luo, T., Xiang, T., Huang, W., Wang, L.: Few-shot learning with global class representations. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00981"},{"key":"16_CR28","doi-asserted-by":"crossref","unstructured":"Li, H., Eigen, D., Dodge, S., Zeiler, M., Wang, X.: Finding task-relevant features for few-shot learning by category traversal. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00009"},{"key":"16_CR29","unstructured":"Mishra, N., Rohaninejad, M., Chen, X., Abbeel, P.: A simple neural attentive meta-learner. arXiv preprint arXiv:1707.03141 (2017)"},{"key":"16_CR30","unstructured":"Mobahi, H., Farajtabar, M., Bartlett, P.L.: Self-distillation amplifies regularization in hilbert space. arXiv preprint arXiv:2002.05715 (2020)"},{"key":"16_CR31","unstructured":"Munkhdalai, T., Yuan, X., Mehri, S., Trischler, A.: Rapid adaptation with conditionally shifted neurons. arXiv preprint arXiv:1712.09926 (2017)"},{"key":"16_CR32","unstructured":"Nichol, A., Achiam, J., Schulman, J.: On first-order meta-learning algorithms. ArXiv abs\/1803.02999 (2018)"},{"key":"16_CR33","unstructured":"Oreshkin, B., L\u00f3pez, P.R., Lacoste, A.: Tadam: task dependent adaptive metric for improved few-shot learning. In: NIPS (2018)"},{"key":"16_CR34","doi-asserted-by":"crossref","unstructured":"Peng, Z., Li, Z., Zhang, J., Li, Y., Qi, G.J., Tang, J.: Few-shot image recognition with knowledge transfer. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00053"},{"key":"16_CR35","doi-asserted-by":"crossref","unstructured":"Qiao, L., Shi, Y., Li, J., Wang, Y., Huang, T., Tian, Y.: Transductive episodic-wise adaptive metric for few-shot learning. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00370"},{"key":"16_CR36","doi-asserted-by":"crossref","unstructured":"Qiao, S., Liu, C., Shen, W., Yuille, A.L.: Few-shot image recognition by predicting parameters from activations. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00755"},{"key":"16_CR37","unstructured":"Raghu, A., Raghu, M., Bengio, S., Vinyals, O.: Rapid learning or feature reuse? towards understanding the effectiveness of maml. arXiv preprint arXiv:1909.09157 (2019)"},{"key":"16_CR38","unstructured":"Ravi, S., Larochelle, H.: Optimization as a model for few-shot learning. In: ICLR (2017)"},{"key":"16_CR39","doi-asserted-by":"crossref","unstructured":"Ravichandran, A., Bhotika, R., Soatto, S.: Few-shot learning with embedded class models and shot-free meta training. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00042"},{"key":"16_CR40","unstructured":"Ren, M., et al.: Meta-learning for semi-supervised few-shot classification. In: ICLR (2018)"},{"key":"16_CR41","unstructured":"Rusu, A.A., Rao, D., Sygnowski, J., Vinyals, O., Pascanu, R., Osindero, S., Hadsell, R.: Meta-learning with latent embedding optimization. In: ICLR (2019)"},{"key":"16_CR42","unstructured":"Scott, T., Ridgeway, K., Mozer, M.C.: Adapted deep embeddings: a synthesis of methods for k-shot inductive transfer learning. In: NIPS (2018)"},{"key":"16_CR43","unstructured":"Snell, J., Swersky, K., Zemel, R.: Prototypical networks for few-shot learning. In: NIPS (2017)"},{"key":"16_CR44","doi-asserted-by":"crossref","unstructured":"Sun, Q., Liu, Y., Chua, T.S., Schiele, B.: Meta-transfer learning for few-shot learning. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00049"},{"key":"16_CR45","doi-asserted-by":"crossref","unstructured":"Sung, F., Yang, Y., Zhang, L., Xiang, T., Torr, P.H., Hospedales, T.M.: Learning to compare: Relation network for few-shot learning. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00131"},{"key":"16_CR46","doi-asserted-by":"crossref","unstructured":"Tian, Y., Krishnan, D., Isola, P.: Contrastive multiview coding. arXiv preprint arXiv:1906.05849 (2019)","DOI":"10.1007\/978-3-030-58621-8_45"},{"key":"16_CR47","unstructured":"Tian, Y., Krishnan, D., Isola, P.: Contrastive representation distillation. arXiv preprint arXiv:1910.10699 (2019)"},{"key":"16_CR48","unstructured":"Tian, Y., Sun, C., Poole, B., Krishnan, D., Schmid, C., Isola, P.: What makes for good views for contrastive learning? arXiv preprint arXiv:2005.10243 (2020)"},{"key":"16_CR49","unstructured":"Triantafillou, E., Zemel, R.S., Urtasun, R.: Few-shot learning through an information retrieval lens. In: NIPS (2017)"},{"key":"16_CR50","unstructured":"Triantafillou, E., et al.: Meta-dataset: a dataset of datasets for learning to learn from few examples. arXiv preprint arXiv:1903.03096 (2019)"},{"key":"16_CR51","unstructured":"Vinyals, O., Blundell, C., Lillicrap, T., kavukcuoglu, K., Wierstra, D.: Matching networks for one shot learning. In: NIPS (2016)"},{"key":"16_CR52","doi-asserted-by":"crossref","unstructured":"Wang, Y.X., Girshick, R.B., Hebert, M., Hariharan, B.: Low-shot learning from imaginary data. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00760"},{"key":"16_CR53","first-page":"244","volume":"29","author":"YX Wang","year":"2016","unstructured":"Wang, Y.X., Hebert, M.: Learning from small sample sets by combining unsupervised meta-training with CNNs. Adv. Neural Inform. Process. Syst. 29, 244\u2013252 (2016)","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"16_CR54","doi-asserted-by":"crossref","unstructured":"Wang, Y., Hebert, M.: Learning to learn: model regression networks for easy small sample learning. In: ECCV (2016)","DOI":"10.1007\/978-3-319-46466-4_37"},{"key":"16_CR55","unstructured":"Weng, L.: Meta-learning: Learning to learn fast. lilianweng.github.io\/lil-log (2018). http:\/\/lilianweng.github.io\/lil-log\/2018\/11\/29\/meta-learning.html"},{"key":"16_CR56","doi-asserted-by":"crossref","unstructured":"Wu, Z., Xiong, Y., Yu, S.X., Lin, D.: Unsupervised feature learning via non-parametric instance discrimination. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00393"},{"key":"16_CR57","doi-asserted-by":"crossref","unstructured":"Wu, Z., Li, Y., Guo, L., Jia, K.: Parn: position-aware relation networks for few-shot learning. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00676"},{"key":"16_CR58","unstructured":"Ye, H.J., Hu, H., Zhan, D.C., Sha, F.: Learning embedding adaptation for few-shot learning. CoRR abs\/1812.03664 (2018)"},{"key":"16_CR59","doi-asserted-by":"crossref","unstructured":"Yim, J., Joo, D., Bae, J., Kim, J.: A gift from knowledge distillation: fast optimization, network minimization and transfer learning. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.754"},{"key":"16_CR60","doi-asserted-by":"crossref","unstructured":"Zhang, J., Zhao, C., Ni, B., Xu, M., Yang, X.: Variational few-shot learning. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00177"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2020"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-58568-6_16","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,12]],"date-time":"2024-11-12T00:19:56Z","timestamp":1731370796000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-58568-6_16"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030585679","9783030585686"],"references-count":60,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-58568-6_16","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"13 November 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Glasgow","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 August 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2020.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"OpenReview","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5025","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1360","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"27% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually due to the COVID-19 pandemic. From the ECCV Workshops 249 full papers, 18 short papers, and 21 further contributions were published out of a total of 467 submissions.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}