{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T22:33:08Z","timestamp":1761172388816,"version":"build-2065373602"},"publisher-location":"Cham","reference-count":33,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032032805","type":"print"},{"value":"9783032032812","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,10,23]],"date-time":"2025-10-23T00:00:00Z","timestamp":1761177600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,23]],"date-time":"2025-10-23T00:00:00Z","timestamp":1761177600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-03281-2_10","type":"book-chapter","created":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T04:57:44Z","timestamp":1761109064000},"page":"140-155","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Spend More to\u00a0Save More (SM2): An Energy and\u00a0Hardware-Aware Implementation of\u00a0Successive Halving for\u00a0Sustainable Hyperparameter Optimization"],"prefix":"10.1007","author":[{"given":"Daniel","family":"Gei\u00dfler","sequence":"first","affiliation":[]},{"given":"Mengxi","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Sungho","family":"Suh","sequence":"additional","affiliation":[]},{"given":"Paul","family":"Lukowicz","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,23]]},"reference":[{"key":"10_CR1","unstructured":"Anthony, L.F.W., Kanding, B., Selvan, R.: Carbontracker: tracking and predicting the carbon footprint of training deep learning models. arXiv preprint arXiv:2007.03051 (2020)"},{"key":"10_CR2","doi-asserted-by":"crossref","unstructured":"Awad, N., Mallik, N., Hutter, F.: Dehb: evolutionary hyperband for scalable, robust and efficient hyperparameter optimization. arXiv preprint arXiv:2105.09821 (2021)","DOI":"10.24963\/ijcai.2021\/296"},{"key":"10_CR3","unstructured":"Bergstra, J., Bardenet, R., Bengio, Y., K\u00e9gl, B.: Algorithms for hyper-parameter optimization. In: Advances in Neural Information Processing Systems, vol. 24 (2011)"},{"key":"10_CR4","unstructured":"Bergstra, J., Bengio, Y.: Random search for hyper-parameter optimization. J. Mach. Learn. Res. 13(2) (2012)"},{"key":"10_CR5","doi-asserted-by":"crossref","unstructured":"Bergstra, J., Yamins, D., Cox, D.D., et\u00a0al.: Hyperopt: a python library for optimizing the hyperparameters of machine learning algorithms. In: Proceedings of the 12th Python in Science Conference, vol.\u00a013, p.\u00a020. Citeseer (2013)","DOI":"10.25080\/Majora-8b375195-003"},{"key":"10_CR6","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1023\/A:1015059928466","volume":"1","author":"HG Beyer","year":"2002","unstructured":"Beyer, H.G., Schwefel, H.P.: Evolution strategies-a comprehensive introduction. Nat. Comput. 1, 3\u201352 (2002)","journal-title":"Nat. Comput."},{"key":"10_CR7","unstructured":"Bischl, B., et\u00a0al.: Hyperparameter optimization: Foundations, algorithms, best practices and open challenges. arxiv 2021. arXiv preprint arXiv:2107.05847 (2021)"},{"key":"10_CR8","doi-asserted-by":"crossref","unstructured":"Bochinski, E., Senst, T., Sikora, T.: Hyper-parameter optimization for convolutional neural network committees based on evolutionary algorithms. In: 2017 IEEE International Conference on Image Processing (ICIP), pp. 3924\u20133928. IEEE (2017)","DOI":"10.1109\/ICIP.2017.8297018"},{"key":"10_CR9","doi-asserted-by":"crossref","unstructured":"Budennyy, S.A., et\u00a0al.: Eco2ai: carbon emissions tracking of machine learning models as the first step towards sustainable AI. In: Doklady Mathematics, vol.\u00a0106, pp. S118\u2013S128. Springer, Cham (2022)","DOI":"10.1134\/S1064562422060230"},{"key":"10_CR10","unstructured":"Code-Carbon: Code carbon (2023). https:\/\/github.com\/mlco2\/codecarbon"},{"key":"10_CR11","unstructured":"Cloud-carbon footprint: Cloud carbon footprint (2023). https:\/\/github.com\/cloud-carbon-footprint\/cloud-carbon-footprint"},{"key":"10_CR12","doi-asserted-by":"crossref","unstructured":"Frey, N.C., et al.: Energy-aware neural architecture selection and hyperparameter optimization. In: 2022 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW), pp. 732\u2013741. IEEE (2022)","DOI":"10.1109\/IPDPSW55747.2022.00125"},{"key":"10_CR13","doi-asserted-by":"crossref","unstructured":"Gei\u00dfler, D., Zhou, B., Liu, M., Suh, S., Lukowicz, P.: The power of training: how different neural network setups influence the energy demand. In: International Conference on Architecture of Computing Systems, pp. 33\u201347. Springer, Cham (2024)","DOI":"10.1007\/978-3-031-66146-4_3"},{"key":"10_CR14","doi-asserted-by":"publisher","unstructured":"Hebrail, G., Berard, A.: Individual Household Electric Power Consumption. UCI Machine Learning Repository (2006). https:\/\/doi.org\/10.24432\/C58K54","DOI":"10.24432\/C58K54"},{"issue":"1","key":"10_CR15","first-page":"10039","volume":"21","author":"P Henderson","year":"2020","unstructured":"Henderson, P., Hu, J., Romoff, J., Brunskill, E., Jurafsky, D., Pineau, J.: Towards the systematic reporting of the energy and carbon footprints of machine learning. J. Mach. Learn. Res. 21(1), 10039\u201310081 (2020)","journal-title":"J. Mach. Learn. Res."},{"key":"10_CR16","unstructured":"Jomaa, H.S., Grabocka, J., Schmidt-Thieme, L.: Hyp-RL: hyperparameter optimization by reinforcement learning. arXiv preprint arXiv:1906.11527 (2019)"},{"key":"10_CR17","unstructured":"Krizhevsky, A., Hinton, G., et\u00a0al.: Learning multiple layers of features from tiny images (2009)"},{"key":"10_CR18","unstructured":"Lacoste, A., Luccioni, A., Schmidt, V., Dandres, T.: Quantifying the carbon emissions of machine learning. arXiv preprint arXiv:1910.09700 (2019)"},{"issue":"12","key":"10_CR19","doi-asserted-by":"publisher","first-page":"2100707","DOI":"10.1002\/advs.202100707","volume":"8","author":"L Lannelongue","year":"2021","unstructured":"Lannelongue, L., Grealey, J., Inouye, M.: Green algorithms: quantifying the carbon footprint of computation. Adv. Sci. 8(12), 2100707 (2021)","journal-title":"Adv. Sci."},{"issue":"11","key":"10_CR20","doi-asserted-by":"publisher","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y LeCun","year":"1998","unstructured":"LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278\u20132324 (1998)","journal-title":"Proc. IEEE"},{"key":"10_CR21","doi-asserted-by":"crossref","unstructured":"Lee, H., Lee, G., Kim, J., Cho, S., Kim, D., Yoo, D.: Improving multi-fidelity optimization with a recurring learning rate for hyperparameter tuning. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 2309\u20132318 (2023)","DOI":"10.1109\/WACV56688.2023.00234"},{"key":"10_CR22","first-page":"230","volume":"2","author":"L Li","year":"2020","unstructured":"Li, L., et al.: A system for massively parallel hyperparameter tuning. Proc. Mach. Learn. Syst. 2, 230\u2013246 (2020)","journal-title":"Proc. Mach. Learn. Syst."},{"issue":"185","key":"10_CR23","first-page":"1","volume":"18","author":"L Li","year":"2018","unstructured":"Li, L., Jamieson, K., DeSalvo, G., Rostamizadeh, A., Talwalkar, A.: Hyperband: a novel bandit-based approach to hyperparameter optimization. J. Mach. Learn. Res. 18(185), 1\u201352 (2018)","journal-title":"J. Mach. Learn. Res."},{"key":"10_CR24","unstructured":"Ludvigsen, K.G.A.: The carbon footprint of GPT-4 (2023). https:\/\/towardsdatascience.com\/the-carbon-footprint-of-gpt-4-d6c676eb21ae. Accessed 22 Feb 2025"},{"key":"10_CR25","unstructured":"Merity, S., Xiong, C., Bradbury, J., Socher, R.: Pointer sentinel mixture models. arXiv:1609.07843 (2017). https:\/\/arxiv.org\/abs\/1609.07843"},{"issue":"2","key":"10_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2636342","volume":"47","author":"S Mittal","year":"2014","unstructured":"Mittal, S., Vetter, J.S.: A survey of methods for analyzing and improving GPU energy efficiency. ACM Comput. Surv. (CSUR) 47(2), 1\u201323 (2014)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"10_CR27","doi-asserted-by":"crossref","unstructured":"Smith, L.N.: Cyclical learning rates for training neural networks. In: 2017 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 464\u2013472. IEEE (2017)","DOI":"10.1109\/WACV.2017.58"},{"key":"10_CR28","unstructured":"Smith, L.N.: A disciplined approach to neural network hyper-parameters: Part 1\u2013learning rate, batch size, momentum, and weight decay. arXiv preprint arXiv:1803.09820 (2018)"},{"key":"10_CR29","unstructured":"Snoek, J., Larochelle, H., Adams, R.P.: Practical Bayesian optimization of machine learning algorithms. In: Advances in Neural Information Processing Systems, vol. 25 (2012)"},{"key":"10_CR30","doi-asserted-by":"publisher","first-page":"427","DOI":"10.1613\/jair.1.14340","volume":"77","author":"T Tornede","year":"2023","unstructured":"Tornede, T., Tornede, A., Hanselle, J., Mohr, F., Wever, M., H\u00fcllermeier, E.: Towards green automated machine learning: status quo and future directions. J. Artif. Intell. Res. 77, 427\u2013457 (2023)","journal-title":"J. Artif. Intell. Res."},{"key":"10_CR31","unstructured":"Wistuba, M., Kadra, A., Grabocka, J.: Supervising the multi-fidelity race of hyperparameter configurations. In: Advances in Neural Information Processing Systems, vol. 35, pp. 13470\u201313484 (2022)"},{"key":"10_CR32","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2022.110160","volume":"260","author":"J Wu","year":"2023","unstructured":"Wu, J., Liu, X., Chen, S.: Hyperparameter optimization through context-based meta-reinforcement learning with task-aware representation. Knowl.-Based Syst. 260, 110160 (2023)","journal-title":"Knowl.-Based Syst."},{"key":"10_CR33","unstructured":"You, J., Chung, J.W., Chowdhury, M.: Zeus: understanding and optimizing $$\\{$$GPU$$\\}$$ energy consumption of $$\\{$$DNN$$\\}$$ training. In: 20th USENIX Symposium on Networked Systems Design and Implementation (NSDI 2023), pp. 119\u2013139 (2023)"}],"container-title":["Lecture Notes in Computer Science","Architecture of Computing Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-03281-2_10","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T04:57:59Z","timestamp":1761109079000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-03281-2_10"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,23]]},"ISBN":["9783032032805","9783032032812"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-03281-2_10","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,23]]},"assertion":[{"value":"23 October 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ARCS","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Architecture of Computing Systems","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kiel","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Germany","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 April 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 April 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"38","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"arcs2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/arcs-conference.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}