{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T12:23:41Z","timestamp":1768998221287,"version":"3.49.0"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030755485","type":"print"},{"value":"9783030755492","type":"electronic"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-75549-2_25","type":"book-chapter","created":{"date-parts":[[2021,4,29]],"date-time":"2021-04-29T07:06:18Z","timestamp":1619679978000},"page":"307-319","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":17,"title":["CLIP: Cheap Lipschitz Training of Neural Networks"],"prefix":"10.1007","author":[{"given":"Leon","family":"Bungert","sequence":"first","affiliation":[]},{"given":"Ren\u00e9","family":"Raab","sequence":"additional","affiliation":[]},{"given":"Tim","family":"Roith","sequence":"additional","affiliation":[]},{"given":"Leo","family":"Schwinn","sequence":"additional","affiliation":[]},{"given":"Daniel","family":"Tenbrinck","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,4,30]]},"reference":[{"key":"25_CR1","unstructured":"Anil, C., Lucas, J., Grosse, R.B.: Sorting out Lipschitz function approximation. In: ICML, vol. 97, pp. 291\u2013301. PMLR (2019)"},{"key":"25_CR2","doi-asserted-by":"crossref","unstructured":"Anzengruber, S.W., Ramlau, R.: Morozov\u2019s discrepancy principle for Tikhonov-type functionals with nonlinear operators. Inverse Probl. 26(2), 025001 (2009)","DOI":"10.1088\/0266-5611\/26\/2\/025001"},{"key":"25_CR3","doi-asserted-by":"publisher","first-page":"4688","DOI":"10.1109\/TSP.2020.3014611","volume":"68","author":"S Aziznejad","year":"2020","unstructured":"Aziznejad, S., Gupta, H., Campos, J., Unser, M.: Deep neural networks with trainable activations and controlled Lipschitz constant. IEEE Trans. Signal Process. 68, 4688\u20134699 (2020)","journal-title":"IEEE Trans. Signal Process."},{"key":"25_CR4","doi-asserted-by":"crossref","unstructured":"Bungert, L., Burger, M.: Solution paths of variational regularization methods for inverse problems. Inverse Probl. 35(10), 105012 (2019)","DOI":"10.1088\/1361-6420\/ab1d71"},{"key":"25_CR5","doi-asserted-by":"crossref","unstructured":"Bungert, L., Burger, M., Korolev, Y., Sch\u00f6nlieb, C.B.: Variational regularisation for inverse problems with imperfect forward operators and general noise models. Inverse Probl. 36(12), 125014 (2020)","DOI":"10.1088\/1361-6420\/abc531"},{"key":"25_CR6","doi-asserted-by":"publisher","unstructured":"Burger, M., Osher, S.: A guide to the TV zoo. In: Level Set and PDE Based Reconstruction Methods in Imaging, vol. 2090, pp. 1\u201370. Springer, Cham (2013). https:\/\/doi.org\/10.1007\/978-3-319-01712-9_1","DOI":"10.1007\/978-3-319-01712-9_1"},{"issue":"2","key":"25_CR7","doi-asserted-by":"publisher","first-page":"529","DOI":"10.1137\/19M1272780","volume":"2","author":"PL Combettes","year":"2020","unstructured":"Combettes, P.L., Pesquet, J.C.: Lipschitz certificates for layered network structures driven by averaged activation operators. SIAM J. Math. Data Sci. 2(2), 529\u2013557 (2020)","journal-title":"SIAM J. Math. Data Sci."},{"key":"25_CR8","unstructured":"Fazlyab, M., Robey, A., Hassani, H., Morari, M., Pappas, G.: Efficient and accurate estimation of Lipschitz constants for deep neural networks. In: NeurIPS (2019)"},{"key":"25_CR9","unstructured":"Goodfellow, I.J., Shlens, J., Szegedy, C.: Explaining and harnessing adversarial examples. In: ICLR (2015)"},{"key":"25_CR10","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10994-020-05929-w","volume":"110","author":"H Gouk","year":"2020","unstructured":"Gouk, H., Frank, E., Pfahringer, B., Cree, M.J.: Regularisation of neural networks by enforcing Lipschitz continuity. Mach. Learn. 110, 1\u201324 (2020). https:\/\/doi.org\/10.1007\/s10994-020-05929-w","journal-title":"Mach. Learn."},{"key":"25_CR11","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"16","DOI":"10.1007\/978-3-030-13453-2_2","volume-title":"ECML PKDD 2018 Workshops","author":"T Huster","year":"2019","unstructured":"Huster, T., Chiang, C.-Y.J., Chadha, R.: Limitations of the Lipschitz constant as a defense against adversarial examples. In: Alzate, C., et al. (eds.) ECML PKDD 2018. LNCS (LNAI), vol. 11329, pp. 16\u201329. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-13453-2_2"},{"key":"25_CR12","unstructured":"Krishnan, V., Makdah, A.A.A., Pasqualetti, F.: Lipschitz bounds and provably robust training by Laplacian smoothing. arXiv preprint arXiv:2006.03712 (2020)"},{"key":"25_CR13","unstructured":"Krizhevsky, A.: Learning multiple layers of features from tiny images. Technical report (2009)"},{"issue":"11","key":"25_CR14","doi-asserted-by":"publisher","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y LeCun","year":"1998","unstructured":"LeCun, Y., Bottou, L., Bengio, Y., Haffner, P., et al.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278\u20132324 (1998)","journal-title":"Proc. IEEE"},{"key":"25_CR15","doi-asserted-by":"crossref","unstructured":"Liang, Y., Huang, D.: Large norms of CNN layers do not hurt adversarial robustness. arXiv preprint arXiv:2009.08435 (2020)","DOI":"10.1609\/aaai.v35i10.17039"},{"key":"25_CR16","unstructured":"Madry, A., Makelov, A., Schmidt, L., Tsipras, D., Vladu, A.: Towards deep learning models resistant to adversarial attacks. In: ICLR (2018)"},{"key":"25_CR17","unstructured":"Oberman, A.M., Calder, J.: Lipschitz regularized deep neural networks converge and generalize. arXiv preprint arXiv:1808.09540 (2018)"},{"key":"25_CR18","unstructured":"van den Oord, A., et al.: WaveNet: a generative model for raw audio. In: The 9th ISCA Speech Synthesis Workshop, p. 125 (2016)"},{"issue":"2","key":"25_CR19","doi-asserted-by":"publisher","first-page":"460","DOI":"10.1137\/040605412","volume":"4","author":"S Osher","year":"2005","unstructured":"Osher, S., Burger, M., Goldfarb, D., Xu, J., Yin, W.: An iterative regularization method for total variation-based image restoration. Multiscale Model Sim. 4(2), 460\u2013489 (2005)","journal-title":"Multiscale Model Sim."},{"key":"25_CR20","unstructured":"Roth, K., Kilcher, Y., Hofmann, T.: Adversarial training is a form of data-dependent operator norm regularization. In: NeurIPS (2019)"},{"key":"25_CR21","unstructured":"Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016)"},{"key":"25_CR22","unstructured":"Scaman, K., Virmaux, A.: Lipschitz regularity of deep neural networks: analysis and efficient estimation. In: NeurIPS (2018)"},{"key":"25_CR23","unstructured":"Schwinn, L., Raab, R., Eskofier, B.: Towards rapid and robust adversarial training with one-step attacks. arXiv preprint arXiv:2002.10097 (2020)"},{"key":"25_CR24","unstructured":"Shafahi, A., et al.: Adversarial training for free! In: NeurIPS, pp. 3353\u20133364 (2019)"},{"key":"25_CR25","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9781107298019","volume-title":"Understanding Machine Learning: From Theory to Algorithms","author":"S Shalev-Shwartz","year":"2014","unstructured":"Shalev-Shwartz, S., Ben-David, S.: Understanding Machine Learning: From Theory to Algorithms. Cambridge University Press, New York (2014)"},{"key":"25_CR26","unstructured":"Szegedy, C., et al.: Intriguing properties of neural networks. In: International Conference on Learning Representations (2014)"},{"key":"25_CR27","unstructured":"Terj\u00e9k, D.: Adversarial Lipschitz regularization. arXiv preprint arXiv:1907.05681 (2019)"},{"key":"25_CR28","unstructured":"Xiao, H., Rasul, K., Vollgraf, R.: Fashion-MNIST: a novel image dataset for benchmarking machine learning algorithms (2017)"},{"issue":"3","key":"25_CR29","doi-asserted-by":"publisher","first-page":"1738","DOI":"10.1109\/TIT.2019.2961812","volume":"66","author":"D Zou","year":"2019","unstructured":"Zou, D., Balan, R., Singh, M.: On Lipschitz bounds of general convolutional neural networks. IEEE Trans. Inf. Theory 66(3), 1738\u20131759 (2019)","journal-title":"IEEE Trans. Inf. Theory"}],"container-title":["Lecture Notes in Computer Science","Scale Space and Variational Methods in Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-75549-2_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,7]],"date-time":"2024-03-07T14:16:48Z","timestamp":1709821008000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-75549-2_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030755485","9783030755492"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-75549-2_25","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"30 April 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"SSVM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Scale Space and Variational Methods in Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 May 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 May 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"scalespace2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ssvm2021.sciencesconf.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"64","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"70% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}