{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,6]],"date-time":"2026-05-06T15:09:05Z","timestamp":1778080145090,"version":"3.51.4"},"publisher-location":"Cham","reference-count":45,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030585792","type":"print"},{"value":"9783030585808","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-58580-8_4","type":"book-chapter","created":{"date-parts":[[2020,12,2]],"date-time":"2020-12-02T07:03:09Z","timestamp":1606892589000},"page":"53-69","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":92,"title":["A Simple Way to Make Neural Networks Robust Against Diverse Image Corruptions"],"prefix":"10.1007","author":[{"given":"Evgenia","family":"Rusak","sequence":"first","affiliation":[]},{"given":"Lukas","family":"Schott","sequence":"additional","affiliation":[]},{"given":"Roland S.","family":"Zimmermann","sequence":"additional","affiliation":[]},{"given":"Julian","family":"Bitterwolf","sequence":"additional","affiliation":[]},{"given":"Oliver","family":"Bringmann","sequence":"additional","affiliation":[]},{"given":"Matthias","family":"Bethge","sequence":"additional","affiliation":[]},{"given":"Wieland","family":"Brendel","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,12,3]]},"reference":[{"key":"4_CR1","unstructured":"Azulay, A., Weiss, Y.: Why do deep convolutional networks generalize so poorly to small image transformations? (2018)"},{"issue":"1\u20132","key":"4_CR2","doi-asserted-by":"publisher","first-page":"57","DOI":"10.1016\/S0004-3702(01)00129-1","volume":"134","author":"M Campbell","year":"2002","unstructured":"Campbell, M., Hoane Jr., A.J., Hsu, F.: Deep blue. Artif. Intell. 134(1\u20132), 57\u201383 (2002). https:\/\/doi.org\/10.1016\/S0004-3702(01)00129-1","journal-title":"Artif. Intell."},{"key":"4_CR3","doi-asserted-by":"crossref","unstructured":"Cubuk, E.D., Zoph, B., Mane, D., Vasudevan, V., Le, Q.V.: AutoAugment: Learning augmentation policies from data. arXiv preprint arXiv:1805.09501 (2018)","DOI":"10.1109\/CVPR.2019.00020"},{"issue":"2","key":"4_CR4","doi-asserted-by":"crossref","first-page":"193","DOI":"10.1111\/j.2517-6161.1984.tb01290.x","volume":"46","author":"PJ Diggle","year":"1984","unstructured":"Diggle, P.J., Gratton, R.J.: Monte Carlo methods of inference for implicit statistical models. J. Roy. Stat. Soc.: Ser. B (Methodol.) 46(2), 193\u2013212 (1984)","journal-title":"J. Roy. Stat. Soc.: Ser. B (Methodol.)"},{"key":"4_CR5","unstructured":"Dodge, S.F., Karam, L.J.: A study and comparison of human and deep learning recognition performance under visual distortions. CoRR abs\/1705.02498 (2017). http:\/\/arxiv.org\/abs\/1705.02498"},{"key":"4_CR6","unstructured":"Engstrom, L., Ilyas, A., Athalye, A.: Evaluating and understanding the robustness of adversarial logit pairing. CoRR abs\/1807.10272 (2018). https:\/\/arxiv.org\/abs\/1807.10272"},{"key":"4_CR7","unstructured":"Engstrom, L., Tsipras, D., Schmidt, L., Madry, A.: A rotation and a translation suffice: fooling CNNs with simple transformations. In: ICML (2019)"},{"key":"4_CR8","unstructured":"Ford, N., Gilmer, J., Carlini, N., Cubuk, D.: Adversarial examples are a natural consequence of test error in noise. In: ICML (2019)"},{"key":"4_CR9","unstructured":"Geirhos, R., Rubisch, P., Michaelis, C., Bethge, M., Wichmann, F.A., Brendel, W.: ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In: International Conference on Learning Representations (2019). https:\/\/openreview.net\/forum?id=Bygh9j09KX"},{"key":"4_CR10","unstructured":"Geirhos, R., Temme, C.R.M., Rauber, J., Sch\u00fctt, H.H., Bethge, M., Wichmann, F.A.: Generalisation in humans and deep neural networks. In: Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 31, pp. 7538\u20137550. Curran Associates, Inc. (2018). http:\/\/papers.nips.cc\/paper\/7982-generalisation-in-humans-and-deep-neural-networks.pdf"},{"key":"4_CR11","unstructured":"Gilmer, J., et al.: Adversarial spheres. CoRR abs\/1801.02774 (2018). http:\/\/arxiv.org\/abs\/1801.02774"},{"key":"4_CR12","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"4_CR13","unstructured":"Hendrycks, D., Dietterich, T.: Benchmarking neural network robustness to common corruptions and perturbations. In: International Conference on Learning Representations (2019). https:\/\/openreview.net\/forum?id=HJz6tiCqYm"},{"key":"4_CR14","unstructured":"Hendrycks, D., Mu, N., Cubuk, E.D., Zoph, B., Gilmer, J., Lakshminarayanan, B.: AugMix: a simple data processing method to improve robustness and uncertainty. In: International Conference on Learning Representations (2020). https:\/\/openreview.net\/forum?id=S1gmrxHFvB"},{"key":"4_CR15","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Weinberger, K.Q.: Densely connected convolutional networks. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"4_CR16","unstructured":"Jordan, M., Manoj, N., Goel, S., Dimakis, A.G.: Quantifying perceptual distortion of adversarial examples. arXiv preprint arXiv:1902.08265 (2019)"},{"key":"4_CR17","unstructured":"Kang, D., Sun, Y., Brown, T., Hendrycks, D., Steinhardt, J.: Transfer of adversarial robustness between perturbation types. CoRR abs\/1905.01034 (2019). http:\/\/arxiv.org\/abs\/1905.01034"},{"key":"4_CR18","unstructured":"Lee, J., Won, T., Hong, K.: Compounding the performance improvements of assembled techniques in a convolutional neural network. arXiv preprint arXiv:2001.06268 (2020)"},{"key":"4_CR19","unstructured":"Lopes, R.G., Yin, D., Poole, B., Gilmer, J., Cubuk, E.D.: Improving robustness without sacrificing accuracy with patch Gaussian augmentation. CoRR abs\/1906.02611 (2019). http:\/\/arxiv.org\/abs\/1906.02611"},{"key":"4_CR20","unstructured":"Madry, A., Makelov, A., Schmidt, L., Tsipras, D., Vladu, A.: Towards deep learning models resistant to adversarial attacks. arXiv preprint arXiv:1706.06083 (2017)"},{"key":"4_CR21","unstructured":"Madry, A., Makelov, A., Schmidt, L., Tsipras, D., Vladu, A.: Towards deep learning models resistant to adversarial attacks. In: International Conference on Learning Representations (2018). https:\/\/openreview.net\/forum?id=rJzIBfZAb"},{"key":"4_CR22","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1007\/978-3-030-01216-8_12","volume-title":"Computer Vision \u2013 ECCV 2018","author":"D Mahajan","year":"2018","unstructured":"Mahajan, D., et al.: Exploring the limits of weakly supervised pretraining. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11206, pp. 185\u2013201. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01216-8_12"},{"key":"4_CR23","doi-asserted-by":"crossref","unstructured":"Marcel, S., Rodriguez, Y.: Torchvision the machine-vision package of torch. In: ACM International Conference on Multimedia (2010)","DOI":"10.1145\/1873951.1874254"},{"issue":"239","key":"4_CR24","first-page":"2","volume":"2014","author":"D Merkel","year":"2014","unstructured":"Merkel, D.: Docker: lightweight Linux containers for consistent development and deployment. Linux J. 2014(239), 2 (2014)","journal-title":"Linux J."},{"key":"4_CR25","unstructured":"Michaelis, C., et al.: Benchmarking robustness in object detection: Autonomous driving when winter is coming. arXiv preprint arXiv:1907.07484 (2019)"},{"key":"4_CR26","doi-asserted-by":"crossref","unstructured":"Miko\u0142ajczyk, A., Grochowski, M.: Data augmentation for improving deep learning in image classification problem. In: 2018 International Interdisciplinary PhD Workshop (IIPhDW), pp. 117\u2013122 (2018)","DOI":"10.1109\/IIPHDW.2018.8388338"},{"issue":"7540","key":"4_CR27","doi-asserted-by":"publisher","first-page":"529","DOI":"10.1038\/nature14236","volume":"518","author":"V Mnih","year":"2015","unstructured":"Mnih, V., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529 (2015)","journal-title":"Nature"},{"key":"4_CR28","unstructured":"Mohamed, S., Lakshminarayanan, B.: Learning in implicit generative models. arXiv preprint arXiv:1610.03483 (2016)"},{"key":"4_CR29","unstructured":"Mu, N., Gilmer, J.: MNIST-C: A robustness benchmark for computer vision. arXiv preprint arXiv:1906.02337 (2019)"},{"key":"4_CR30","unstructured":"OpenAI: OpenAI Five. https:\/\/blog.openai.com\/openai-five\/ (2018)"},{"key":"4_CR31","unstructured":"Paszke, A., et al.: Automatic differentiation in PyTorch. In: NIPS Autodiff Workshop (2017)"},{"key":"4_CR32","unstructured":"Rauber, J., Bethge, M.: Fast differentiable clipping-aware normalization and rescaling. arXiv preprint arXiv:2007.07677 (2020). https:\/\/github.com\/jonasrauber\/clipping-aware-rescaling"},{"key":"4_CR33","doi-asserted-by":"crossref","unstructured":"Rony, J., Hafemann, L.G., Oliveira, L.S., Ayed, I.B., Sabourin, R., Granger, E.: Decoupling direction and norm for efficient gradient-based L2 adversarial attacks and defenses. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4322\u20134330 (2019)","DOI":"10.1109\/CVPR.2019.00445"},{"key":"4_CR34","unstructured":"Russakovsky, O., et al.: ImageNet large scale visual recognition challenge. CoRR abs\/1409.0575 (2014). http:\/\/arxiv.org\/abs\/1409.0575"},{"key":"4_CR35","unstructured":"Schott, L., Rauber, J., Bethge, M., Brendel, W.: Towards the first adversarially robust neural network model on MNIST. In: International Conference on Learning Representations (2019). https:\/\/openreview.net\/forum?id=S1EHOsC9tX"},{"key":"4_CR36","unstructured":"Shafahi, A., et al.: Adversarial training for free! arXiv preprint arXiv:1904.12843 (2019)"},{"key":"4_CR37","unstructured":"Shafahi, A., Najibi, M., Xu, Z., Dickerson, J.P., Davis, L.S., Goldstein, T.: Universal adversarial training. CoRR abs\/1811.11304 (2018). http:\/\/arxiv.org\/abs\/1811.11304"},{"key":"4_CR38","doi-asserted-by":"publisher","first-page":"354","DOI":"10.1038\/nature24270","volume":"550","author":"D Silver","year":"2017","unstructured":"Silver, D., et al.: Mastering the game of go without human knowledge. Nature 550, 354\u2013359 (2017)","journal-title":"Nature"},{"key":"4_CR39","unstructured":"Szegedy, C., et al.: Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199 (2013)"},{"key":"4_CR40","doi-asserted-by":"crossref","unstructured":"Tram\u00e8r, F., Boneh, D.: Adversarial training and robustness for multiple perturbations. In: NeurIPS (2019). http:\/\/arxiv.org\/abs\/1904.13000","DOI":"10.1145\/3319535.3354222"},{"key":"4_CR41","doi-asserted-by":"publisher","first-page":"261","DOI":"10.1038\/s41592-019-0686-2","volume":"17","author":"P Virtanen","year":"2020","unstructured":"Virtanen, P., et al.: SciPy 1.0: fundamental algorithms for scientific computing in Python. Nat. Meth. 17, 261\u2013272 (2020). https:\/\/doi.org\/10.1038\/s41592-019-0686-2","journal-title":"Nat. Meth."},{"key":"4_CR42","doi-asserted-by":"crossref","unstructured":"Xie, C., Wu, Y., van der Maaten, L., Yuille, A.L., He, K.: Feature denoising for improving adversarial robustness. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00059"},{"key":"4_CR43","doi-asserted-by":"crossref","unstructured":"Xie, Q., Hovy, E., Luong, M.T., Le, Q.V.: Self-training with noisy student improves ImageNet classification. arXiv preprint arXiv:1911.04252 (2019)","DOI":"10.1109\/CVPR42600.2020.01070"},{"key":"4_CR44","doi-asserted-by":"crossref","unstructured":"Xiong, W., et al.: Achieving human parity in conversational speech recognition. In: IEEE\/ACM Transactions on Audio, Speech, and Language Processing (2016)","DOI":"10.1109\/TASLP.2017.2756440"},{"key":"4_CR45","unstructured":"Zhang, R.: Making convolutional networks shift-invariant again. In: ICML (2019)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2020"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-58580-8_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T00:04:07Z","timestamp":1733097847000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-58580-8_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030585792","9783030585808"],"references-count":45,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-58580-8_4","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"3 December 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Glasgow","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 August 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2020.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"OpenReview","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5025","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1360","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"27% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually due to the COVID-19 pandemic. From the ECCV Workshops 249 full papers, 18 short papers, and 21 further contributions were published out of a total of 467 submissions.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}