{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T11:12:46Z","timestamp":1772881966078,"version":"3.50.1"},"reference-count":68,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2022,2,21]],"date-time":"2022-02-21T00:00:00Z","timestamp":1645401600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,2,21]],"date-time":"2022-02-21T00:00:00Z","timestamp":1645401600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["IIS-1901030"],"award-info":[{"award-number":["IIS-1901030"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CCF-1231216"],"award-info":[{"award-number":["CCF-1231216"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CCF-1231216"],"award-info":[{"award-number":["CCF-1231216"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["IIS-1901030"],"award-info":[{"award-number":["IIS-1901030"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CCF-1231216"],"award-info":[{"award-number":["CCF-1231216"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003393","name":"Fujitsu","doi-asserted-by":"publisher","award":["Contract No. 40008819"],"award-info":[{"award-number":["Contract No. 40008819"]}],"id":[{"id":"10.13039\/501100003393","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003393","name":"Fujitsu","doi-asserted-by":"publisher","award":["Contract No. 40008819"],"award-info":[{"award-number":["Contract No. 40008819"]}],"id":[{"id":"10.13039\/501100003393","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Google Faculty Research Award"},{"name":"MIT-Sensetime Alliance on Artificial Intelligence"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Nat Mach Intell"],"DOI":"10.1038\/s42256-021-00437-5","type":"journal-article","created":{"date-parts":[[2022,2,23]],"date-time":"2022-02-23T00:02:40Z","timestamp":1645574560000},"page":"146-153","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":25,"title":["When and how convolutional neural networks generalize to out-of-distribution category\u2013viewpoint combinations"],"prefix":"10.1038","volume":"4","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8957-0004","authenticated-orcid":false,"given":"Spandan","family":"Madan","sequence":"first","affiliation":[]},{"given":"Timothy","family":"Henry","sequence":"additional","affiliation":[]},{"given":"Jamell","family":"Dozier","sequence":"additional","affiliation":[]},{"given":"Helen","family":"Ho","sequence":"additional","affiliation":[]},{"given":"Nishchal","family":"Bhandari","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3376-2779","authenticated-orcid":false,"given":"Tomotake","family":"Sasaki","sequence":"additional","affiliation":[]},{"given":"Fr\u00e9do","family":"Durand","sequence":"additional","affiliation":[]},{"given":"Hanspeter","family":"Pfister","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4656-3485","authenticated-orcid":false,"given":"Xavier","family":"Boix","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,2,21]]},"reference":[{"key":"437_CR1","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S. and Sun, J. Deep residual learning for image recognition. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 770\u2013778 (IEEE, 2016).","DOI":"10.1109\/CVPR.2016.90"},{"key":"437_CR2","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J. & Wojna, Z. Rethinking the inception architecture for computer vision. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 2818\u20132826 (IEEE, 2016).","DOI":"10.1109\/CVPR.2016.308"},{"key":"437_CR3","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L. & Weinberger, K. Q. Densely connected convolutional networks. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 4700\u20134708 (IEEE, 2017).","DOI":"10.1109\/CVPR.2017.243"},{"key":"437_CR4","doi-asserted-by":"crossref","unstructured":"Su, H., Qi, C. R., Li, Y. & Guibas, L. J. Render for CNN: Viewpoint estimation in images using CNNs trained with rendered 3D model views. In Proc. IEEE International Conference on Computer Vision 2686\u20132694 (IEEE, 2015).","DOI":"10.1109\/ICCV.2015.308"},{"key":"437_CR5","doi-asserted-by":"crossref","unstructured":"Massa, F., Marlet, R. & Aubry, M. Crafting a multi-task CNN for viewpoint estimation. In Proc. British Machine Vision Conference 91.1\u201391.12 (BMVA, 2016).","DOI":"10.5244\/C.30.91"},{"key":"437_CR6","unstructured":"Elhoseiny, M., El-Gaaly, T., Bakry, A. & Elgammal, A. A comparative analysis and study of multiview CNN models for joint object categorization and pose estimation. In Proc. International Conference on Machine Learning 888\u2013897 (PMLR, 2016)."},{"key":"437_CR7","doi-asserted-by":"crossref","unstructured":"Mahendran, S., Ali, H. & Vidal, R. Convolutional networks for object category and 3D pose estimation from 2D images. In Proc. European Conference on Computer Vision Workshops 698\u2013715 (Springer, 2018).","DOI":"10.1007\/978-3-030-11009-3_43"},{"key":"437_CR8","doi-asserted-by":"crossref","unstructured":"Afifi, A. J., Hellwich, O. & Soomro, T. A. Simultaneous object classification and viewpoint estimation using deep multi-task convolutional neural network. In Proc. International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications 177\u2013184 (2018).","DOI":"10.5220\/0006544001770184"},{"key":"437_CR9","unstructured":"Engstrom, L., Tran, B., Tsipras, D., Schmidt, L. & Madry, A. Exploring the landscape of spatial robustness. In Proc. International Conference on Machine Learning 1802\u20131811 (PMLR, 2019)."},{"key":"437_CR10","first-page":"1","volume":"20","author":"A Azulay","year":"2019","unstructured":"Azulay, A. & Weiss, Y. Why do deep convolutional networks generalize so poorly to small image transformations? J. Mach. Learn. Res. 20, 1\u201325 (2019).","journal-title":"J. Mach. Learn. Res."},{"key":"437_CR11","unstructured":"Srivastava, S., Ben-Yosef, G. & Boix, X. Minimal images in deep neural networks: fragile object recognition in natural images. In Proc. International Conference on Learning Representations (2019)."},{"key":"437_CR12","doi-asserted-by":"crossref","unstructured":"Alcorn, M. A. et al. Strike (with) a pose: neural networks are easily fooled by strange poses of familiar objects. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 4845\u20134854 (IEEE, 2019).","DOI":"10.1109\/CVPR.2019.00498"},{"key":"437_CR13","first-page":"9448","volume":"32","author":"A Barbu","year":"2019","unstructured":"Barbu, A. et al. ObjectNet: a large-scale bias-controlled dataset for pushing the limits of object recognition models. Adv. Neural Inf. Process. Syst. 32, 9448\u20139458 (2019).","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"437_CR14","doi-asserted-by":"crossref","unstructured":"Tulsiani, S. & Malik, J. Viewpoints and keypoints. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 1510\u20131519 (IEEE, 2015).","DOI":"10.1109\/CVPR.2015.7298758"},{"key":"437_CR15","doi-asserted-by":"crossref","unstructured":"Xiang, Y., Schmidt, T., Narayanan, V. & Fox, D. PoseCNN: a convolutional neural network for 6D object pose estimation in cluttered scenes. In Proc. Robotics: Science and Systems (2018).","DOI":"10.15607\/RSS.2018.XIV.019"},{"key":"437_CR16","unstructured":"Manhardt, F. et al. CPS++: improving class-level 6D pose and shape estimation from monocular images with self-supervised learning. Preprint at https:\/\/arxiv.org\/abs\/2003.05848 (2020)."},{"key":"437_CR17","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1023\/A:1007379606734","volume":"28","author":"R Caruana","year":"1997","unstructured":"Caruana, R. Multitask learning. Mach. Learn. 28, 41\u201375 (1997).","journal-title":"Mach. Learn."},{"key":"437_CR18","doi-asserted-by":"publisher","first-page":"4972","DOI":"10.1364\/AO.26.004972","volume":"26","author":"CL Giles","year":"1987","unstructured":"Giles, C. L. & Maxwell, T. Learning, invariance, and generalization in high-order neural networks. Appl. Optics 26, 4972\u20134978 (1987).","journal-title":"Appl. Optics"},{"key":"437_CR19","first-page":"215","volume":"10","author":"M Riesenhuber","year":"1998","unstructured":"Riesenhuber, M. & Poggio, T. Just one view: Invariances in inferotemporal cell tuning. Adv. Neural Inf. Process. Syst. 10, 215\u2013221 (1998).","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"437_CR20","first-page":"646","volume":"22","author":"I Goodfellow","year":"2009","unstructured":"Goodfellow, I., Lee, H., Le, Q. V., Saxe, A. & Ng, A. Y. Measuring invariances in deep networks. Adv. Neural Inf. Process. Syst. 22, 646\u2013654 (2009).","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"437_CR21","first-page":"1947","volume":"19","author":"A Achille","year":"2018","unstructured":"Achille, A. & Soatto, S. Emergence of invariance and disentanglement in deep representations. J. Mach. Learn. Res. 19, 1947\u20131980 (2018).","journal-title":"J. Mach. Learn. Res."},{"key":"437_CR22","doi-asserted-by":"crossref","unstructured":"Borji, A., Izadi, S. & Itti, L. iLab-20M: a large-scale controlled object dataset to investigate deep learning. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 2221\u20132230 (IEEE, 2016).","DOI":"10.1109\/CVPR.2016.244"},{"key":"437_CR23","unstructured":"Visual variation learning for object recognition. Jatuporn Toy Leksut https:\/\/bmobear.github.io\/projects\/viva\/ (2016)."},{"key":"437_CR24","doi-asserted-by":"crossref","unstructured":"LeCun, Y., Bottou, L., Bengio, Y. & Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 86, 2278\u20132324 (1998).","DOI":"10.1109\/5.726791"},{"key":"437_CR25","unstructured":"The MNIST Database of Handwritten Digits (accessed 13 January 2022); http:\/\/yann.lecun.com\/exdb\/mnist\/"},{"key":"437_CR26","doi-asserted-by":"crossref","unstructured":"Xiang, Y., Mottaghi, R. & Savarese, S. Beyond pascal: a benchmark for 3D object detection in the wild. In Proc. IEEE Winter Conference on Applications of Computer Vision 75\u201382 (IEEE, 2014).","DOI":"10.1109\/WACV.2014.6836101"},{"key":"437_CR27","doi-asserted-by":"crossref","unstructured":"Caesar, H. et al. nuScenes: a multimodal dataset for autonomous driving. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 11618\u201311628 (IEEE, 2020).","DOI":"10.1109\/CVPR42600.2020.01164"},{"key":"437_CR28","unstructured":"Min, J., Lee, J., Ponce, J. & Cho, M. Spair-71k: a large-scale benchmark for semantic correspondence. Preprint at https:\/\/arxiv.org\/abs\/1908.10543 (2019)."},{"key":"437_CR29","doi-asserted-by":"crossref","unstructured":"Larochelle, H., Erhan, D., Courville, A., Bergstra, J. & Bengio, Y. An empirical evaluation of deep architectures on problems with many factors of variation. In Proc. 24th International Conference on Machine Learning 473\u2013480 (PMLR, 2007).","DOI":"10.1145\/1273496.1273556"},{"key":"437_CR30","doi-asserted-by":"crossref","unstructured":"Krause, J., Stark, M., Deng, J. & Fei-Fei, L. 3D object representations for fine-grained categorization. In Proc. 4th International IEEE Workshop on 3D Representation and Recognition 554\u2013561 (IEEE, 2013).","DOI":"10.1109\/ICCVW.2013.77"},{"key":"437_CR31","doi-asserted-by":"crossref","unstructured":"Ozuysal, M., Lepetit, V. & Fua, P. Pose estimation for category specific multiview object localization. In Proc. IEEE Conference on Computer Vision and Pattern Recognition 778\u2013785 (IEEE, 2009).","DOI":"10.1109\/CVPR.2009.5206633"},{"key":"437_CR32","doi-asserted-by":"crossref","unstructured":"Qiu, W. & Yuille, A. UnrealCV: connecting computer vision to Unreal Engine. In Proc. European Conference on Computer Vision 909\u2013916 (Springer, 2016).","DOI":"10.1007\/978-3-319-49409-8_75"},{"key":"437_CR33","unstructured":"Dosovitskiy, A., Ros, G., Codevilla, F., Lopez, A. & Koltun, V. CARLA: an open urban driving simulator. In Proc. Annual Conference on Robot Learning 1\u201316 (2017)."},{"key":"437_CR34","doi-asserted-by":"crossref","unstructured":"Zhang, Y. et al. Physically-based rendering for indoor scene understanding using convolutional neural networks. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 5287\u20135295 (IEEE, 2017).","DOI":"10.1109\/CVPR.2017.537"},{"key":"437_CR35","doi-asserted-by":"crossref","unstructured":"Halder, S. S., Lalonde, J.-F. & de Charette, R. Physics-based rendering for improving robustness to rain. In Proc. IEEE\/CVF International Conference on Computer Vision 10203\u201310212 (IEEE, 2019).","DOI":"10.1109\/ICCV.2019.01030"},{"key":"437_CR36","doi-asserted-by":"crossref","unstructured":"Divon, G. & Tal, A. Viewpoint estimation\u2014insights & model. In Proc. European Conference on Computer Vision 252\u2013268 (Springer, 2018).","DOI":"10.1007\/978-3-030-01264-9_16"},{"key":"437_CR37","unstructured":"Mueller, P. et al. Esri CityEngine\u2014A 3D City Modeling Software for Urban Design, Visual Effects, and VR\/AR (Esri R&D Center Zurich, 2020); http:\/\/www.esri.com\/cityengine"},{"key":"437_CR38","unstructured":"Blender\u2014A 3D Modelling and Rendering Package (Blender Foundation, Stichting Blender Foundation, 2020); http:\/\/www.blender.org"},{"key":"437_CR39","doi-asserted-by":"crossref","unstructured":"Savarese, S. Fei-Fei, L. 3D generic object categorization, localization and pose estimation. In 2007 IEEE 11th International Conference on Computer Vision 1\u20138 (IEEE, 2007).","DOI":"10.1109\/ICCV.2007.4408987"},{"key":"437_CR40","doi-asserted-by":"crossref","unstructured":"Ghodrati, A., Pedersoli, M. & Tuytelaars, T. Is 2D information enough for viewpoint estimation? In Proc. British Machine Vision Conference (BMVA, 2014).","DOI":"10.5244\/C.28.19"},{"key":"437_CR41","doi-asserted-by":"crossref","unstructured":"Tulsiani, S., Carreira, J. & Malik, J. Pose induction for novel object categories. In Proc. IEEE International Conference on Computer Vision 64\u201372 (IEEE, 2015).","DOI":"10.1109\/ICCV.2015.16"},{"key":"437_CR42","unstructured":"Penedones, H., Collobert, R., Fleuret, F. & Grangier, D. Improving Object Classification Using Pose Information Technical Report Idiap-RR-30-2012 (Idiap Research Institute, 2012)."},{"key":"437_CR43","doi-asserted-by":"crossref","unstructured":"Zhao, J. & Itti, L. Improved deep learning of object category using pose information. In Proc. IEEE Winter Conference on Applications of Computer Vision 550\u2013559 (IEEE, 2017).","DOI":"10.1109\/WACV.2017.67"},{"key":"437_CR44","doi-asserted-by":"crossref","unstructured":"Li, C., Bai, J. & Hager, G. D. A unified framework for multi-view multi-class object pose estimation. In Proc. European Conference on Computer Vision 254\u2013269 (Springer, 2018).","DOI":"10.1007\/978-3-030-01270-0_16"},{"key":"437_CR45","doi-asserted-by":"crossref","unstructured":"Grabner, A., Roth, P. M. & Lepetit, V. 3D pose estimation and 3D model retrieval for objects in the wild. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 3022\u20133031 (IEEE, 2018).","DOI":"10.1109\/CVPR.2018.00319"},{"key":"437_CR46","first-page":"41","volume":"9","author":"E Bricolo","year":"1997","unstructured":"Bricolo, E., Poggio, T. & Logothetis, N. K. 3D object recognition: a model of view-tuned neurons. Adv. Neural Inf. Process. Syst. 9, 41\u201347 (1997).","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"437_CR47","doi-asserted-by":"crossref","unstructured":"Poggio, T. & Anselmi, F. Visual Cortex and Deep Networks: Learning Invariant Representations (MIT Press, 2016).","DOI":"10.7551\/mitpress\/10177.001.0001"},{"key":"437_CR48","doi-asserted-by":"publisher","first-page":"4700","DOI":"10.1523\/JNEUROSCI.13-11-04700.1993","volume":"13","author":"BA Olshausen","year":"1993","unstructured":"Olshausen, B. A., Anderson, C. H. & Van Essen, D. C. A neurobiological model of visual attention and invariant pattern recognition based on dynamic routing of information. J. Neurosci. 13, 4700\u20134719 (1993).","journal-title":"J. Neurosci."},{"key":"437_CR49","doi-asserted-by":"publisher","first-page":"1102","DOI":"10.1038\/nature03687","volume":"435","author":"RQ Quiroga","year":"2005","unstructured":"Quiroga, R. Q., Reddy, L., Kreiman, G., Koch, C. & Fried, I. Invariant visual representation by single neurons in the human brain. Nature 435, 1102\u20131107 (2005).","journal-title":"Nature"},{"key":"437_CR50","doi-asserted-by":"publisher","first-page":"12978","DOI":"10.1523\/JNEUROSCI.0179-10.2010","volume":"30","author":"NC Rust","year":"2010","unstructured":"Rust, N. C. & DiCarlo, J. J. Selectivity and tolerance (invariance) both increase as visual information propagates from cortical area V4 to IT. J. Neurosci. 30, 12978\u201312995 (2010).","journal-title":"J. Neurosci."},{"key":"437_CR51","doi-asserted-by":"crossref","unstructured":"Zeiler, M. D. & Fergus, R. Visualizing and understanding convolutional networks. In Proc. European Conference on Computer Vision 818\u2013833 (Springer, 2014).","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"437_CR52","unstructured":"Simonyan, K., Vedaldi, A. & Zisserman, A. Deep inside convolutional networks: visualising image classification models and saliency maps. In Proc. International Conference on Learning Representations Workshop (2014)."},{"key":"437_CR53","unstructured":"Zhou, B., Khosla, A., Lapedriza, A., Oliva, A. & Torralba, A. Object detectors emerge in deep scene CNNs. In Proc. International Conference on Learning Representations (2015)."},{"key":"437_CR54","doi-asserted-by":"crossref","unstructured":"Bau, D., Zhou, B., Khosla, A., Oliva, A. & Torralba, A. Network dissection: quantifying interpretability of deep visual representations. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 6541\u20136549 (IEEE, 2017).","DOI":"10.1109\/CVPR.2017.354"},{"key":"437_CR55","doi-asserted-by":"crossref","unstructured":"Oquab, M., Bottou, L., Laptev, I. & Sivic, J. Is object localization for free? Weakly-supervised learning with convolutional neural networks. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 685\u2013694 (IEEE, 2015).","DOI":"10.1109\/CVPR.2015.7298668"},{"key":"437_CR56","unstructured":"Morcos, A. S., Barrett, D. G. T., Rabinowitz, N. C. & Botvinick, M. On the importance of single directions for generalization. In Proc. International Conference on Learning Representations (2018)."},{"key":"437_CR57","unstructured":"Zhou, B., Sun, Y., Bau, D. & Torralba, A. Revisiting the importance of individual units in CNNs via ablation. Preprint at https:\/\/arxiv.org\/abs\/1806.02891 (2018)."},{"key":"437_CR58","doi-asserted-by":"publisher","first-page":"297","DOI":"10.1038\/s41593-018-0310-2","volume":"22","author":"GR Yang","year":"2019","unstructured":"Yang, G. R., Joglekar, M. R., Song, H. F., Newsome, W. T. & Wang, X.-J. Task representations in neural networks trained to perform many cognitive tasks. Nat. Neurosci. 22, 297\u2013306 (2019).","journal-title":"Nat. Neurosci."},{"key":"437_CR59","doi-asserted-by":"crossref","unstructured":"Torralba, A. & Efros, A. A. Unbiased look at dataset bias. In Proc. IEEE Conference on Computer Vision and Pattern Recognition 1521\u20131528 (IEEE, 2011).","DOI":"10.1109\/CVPR.2011.5995347"},{"key":"437_CR60","unstructured":"Kingma, D. P. & Ba, J. Adam: a method for stochastic optimization. Preprint at https:\/\/arxiv.org\/abs\/1412.6980 (2014)."},{"key":"437_CR61","unstructured":"Standley, T. et al. Which tasks should be learned together in multi-task learning? In Proc. International Conference on Machine Learning (PMLR, 2020)."},{"key":"437_CR62","doi-asserted-by":"crossref","unstructured":"Shin, D., Fowlkes, C. C. & Hoiem, D. Pixels, voxels, and views: a study of shape representations for single view 3D object shape prediction. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 3061\u20133069 (IEEE, 2018).","DOI":"10.1109\/CVPR.2018.00323"},{"key":"437_CR63","doi-asserted-by":"crossref","unstructured":"Xie, S., Girshick, R., Doll\u00e1r, P., Tu, Z. & He, K. Aggregated residual transformations for deep neural networks. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 1492\u20131500 (IEEE, 2017).","DOI":"10.1109\/CVPR.2017.634"},{"key":"437_CR64","doi-asserted-by":"crossref","unstructured":"Zagoruyko, S. & Komodakis, N. Wide residual networks. In Proc. British Machine Vision Conference 87.1\u201387.12 (BMVA, 2016).","DOI":"10.5244\/C.30.87"},{"key":"437_CR65","doi-asserted-by":"crossref","unstructured":"Nakkiran, P. et al. Deep double descent: where bigger models and more data hurt. In Proc. International Conference on Learning Representations (2020).","DOI":"10.1088\/1742-5468\/ac3a74"},{"key":"437_CR66","doi-asserted-by":"crossref","unstructured":"Casper, S. et al. Frivolous units: wider networks are not really that wide. In Proc. Association for the Advancement of Artificial Intelligence (2021).","DOI":"10.1609\/aaai.v35i8.16853"},{"key":"437_CR67","unstructured":"Cohen, T. S., Geiger, M., K\u00f6hler, J. & Welling, M. Spherical CNNs. In Proc. International Conference on Learning Representations (2018)."},{"key":"437_CR68","unstructured":"Cohen, T. S., Weiler, M., Kicanaoglu, B. & Welling, M. Gauge equivariant convolutional networks and the Icosahedral CNN. In Proc. International Conference on Machine Learning 1321\u20131330 (PMLR, 2019)."}],"container-title":["Nature Machine Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.nature.com\/articles\/s42256-021-00437-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s42256-021-00437-5","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s42256-021-00437-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,11,25]],"date-time":"2022-11-25T09:16:41Z","timestamp":1669367801000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.nature.com\/articles\/s42256-021-00437-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,2,21]]},"references-count":68,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2022,2]]}},"alternative-id":["437"],"URL":"https:\/\/doi.org\/10.1038\/s42256-021-00437-5","relation":{},"ISSN":["2522-5839"],"issn-type":[{"value":"2522-5839","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,2,21]]},"assertion":[{"value":"11 February 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 December 2021","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 February 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"This study received funding from Fujitsu Laboratories. The funder through T.S. was involved in conception of the experiment, writing this article and supervising the study. All other authors declare no competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}