{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T22:51:07Z","timestamp":1743029467487,"version":"3.40.3"},"publisher-location":"Cham","reference-count":36,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031530241"},{"type":"electronic","value":"9783031530258"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-53025-8_27","type":"book-chapter","created":{"date-parts":[[2024,1,31]],"date-time":"2024-01-31T20:02:12Z","timestamp":1706731332000},"page":"391-408","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["BHO-MA: Bayesian Hyperparameter Optimization with\u00a0Multi-objective Acquisition"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3807-5425","authenticated-orcid":false,"given":"Vedat","family":"Dogan","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6218-9158","authenticated-orcid":false,"given":"Steven","family":"Prestwich","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,1]]},"reference":[{"key":"27_CR1","unstructured":"Balandat, M., et al.: BoTorch: programmable Bayesian optimization in PyTorch. CoRR abs\/1910.06403 (2019). http:\/\/arxiv.org\/abs\/1910.06403"},{"key":"27_CR2","unstructured":"Bergstra, J., Bardenet, R., Bengio, Y., K\u00e9gl, B.: Algorithms for hyper-parameter optimization. In: Shawe-Taylor, J., Zemel, R., Bartlett, P., Pereira, F., Weinberger, K. (eds.) Advances in Neural Information Processing Systems, vol. 24. Curran Associates, Inc. (2011). https:\/\/proceedings.neurips.cc\/paper\/2011\/file\/86e8f7ab32cfd12577bc2619bc635690-Paper.pdf"},{"key":"27_CR3","first-page":"281","volume":"13","author":"J Bergstra","year":"2012","unstructured":"Bergstra, J., Bengio, Y.: Random search for hyper-parameter optimization. J. Mach. Learn. Res. 13, 281\u2013305 (2012)","journal-title":"J. Mach. Learn. Res."},{"key":"27_CR4","doi-asserted-by":"publisher","first-page":"89497","DOI":"10.1109\/ACCESS.2020.2990567","volume":"8","author":"J Blank","year":"2020","unstructured":"Blank, J., Deb, K.: pymoo: Multi-objective optimization in Python. IEEE Access 8, 89497\u201389509 (2020)","journal-title":"IEEE Access"},{"key":"27_CR5","doi-asserted-by":"publisher","unstructured":"Blume, S., Benedens, T., Schramm, D.: Hyperparameter optimization techniques for designing software sensors based on artificial neural networks. Sensors 21(24), 8435 (2021). https:\/\/doi.org\/10.3390\/s21248435. https:\/\/www.mdpi.com\/1424-8220\/21\/24\/8435","DOI":"10.3390\/s21248435"},{"key":"27_CR6","doi-asserted-by":"publisher","unstructured":"Bockrath, S., Lorentz, V., Pruckner, M.: State of health estimation of lithium-ion batteries with a temporal convolutional neural network using partial load profiles. Appl. Energy 329, 120307 (2023). https:\/\/doi.org\/10.1016\/j.apenergy.2022.120307. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0306261922015641","DOI":"10.1016\/j.apenergy.2022.120307"},{"key":"27_CR7","unstructured":"Bohdal, O., Yang, Y., Hospedales, T.: EvoGrad: efficient gradient-based meta-learning and hyperparameter optimization. In: Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W. (eds.) Advances in Neural Information Processing Systems, vol. 34, pp. 22234\u201322246. Curran Associates, Inc. (2021). https:\/\/proceedings.neurips.cc\/paper\/2021\/file\/bac49b876d5dfc9cd169c22ef5178ca7-Paper.pdf"},{"key":"27_CR8","doi-asserted-by":"publisher","unstructured":"B\u00fchlmann, P.: Bagging, boosting and ensemble methods. In: Gentle, J., H\u00e4rdle, W., Mori, Y. (eds.) Handbook of Computational Statistics. Springer Handbooks of Computational Statistics, pp. 985\u20131022. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-21551-3_33","DOI":"10.1007\/978-3-642-21551-3_33"},{"key":"27_CR9","doi-asserted-by":"publisher","unstructured":"Chen, S., Wu, J., Liu, X.: EMORL: effective multi-objective reinforcement learning method for hyperparameter optimization. Eng. Appl. Artif. Intell. 104, 104315 (2021). https:\/\/doi.org\/10.1016\/j.engappai.2021.104315. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0952197621001639","DOI":"10.1016\/j.engappai.2021.104315"},{"issue":"2","key":"27_CR10","doi-asserted-by":"publisher","first-page":"182","DOI":"10.1109\/4235.996017","volume":"6","author":"K Deb","year":"2002","unstructured":"Deb, K., Pratap, A., Agarwal, S., Meyarivan, T.: A fast and elitist multiobjective genetic algorithm: NSGA-ii. IEEE Trans. Evol. Comput. 6(2), 182\u2013197 (2002). https:\/\/doi.org\/10.1109\/4235.996017","journal-title":"IEEE Trans. Evol. Comput."},{"key":"27_CR11","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2019\/6278908","volume":"2019","author":"N DeCastro-Garc\u00eda","year":"2019","unstructured":"DeCastro-Garc\u00eda, N., Mu\u00f1oz Casta\u00f1eda, A.L., Escudero Garc\u00eda, D., Carriegos, M.V., S\u00e1nchez Lasheras, F.: Effect of the sampling of a dataset in the hyperparameter optimization phase over the efficiency of a machine learning algorithm. CompLex 2019, 1\u201316 (2019). https:\/\/doi.org\/10.1155\/2019\/6278908","journal-title":"CompLex"},{"key":"27_CR12","doi-asserted-by":"crossref","unstructured":"Eggensperger, K., Hutter, F., Hoos, H.H., Leyton-Brown, K.: Efficient benchmarking of hyperparameter optimizers via surrogates. In: Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, AAAI 2015, pp. 1114\u20131120. AAAI Press (2015)","DOI":"10.1609\/aaai.v29i1.9375"},{"key":"27_CR13","doi-asserted-by":"publisher","unstructured":"Han, J., Gondro, C., Reid, K., Steibel, J.: Heuristic hyperparameter optimization of deep learning models for genomic prediction. G3-Genes Genomes Genet. 11 (2021). https:\/\/doi.org\/10.1093\/g3journal\/jkab032","DOI":"10.1093\/g3journal\/jkab032"},{"key":"27_CR14","unstructured":"Hoffman, M., Brochu, E., de Freitas, N.: Portfolio allocation for Bayesian optimization. In: Proceedings of the Twenty-Seventh Conference on Uncertainty in Artificial Intelligence, UAI 2011, pp. 327\u2013336. AUAI Press, Arlington (2011)"},{"issue":"9","key":"27_CR15","doi-asserted-by":"publisher","first-page":"5149","DOI":"10.1109\/TPAMI.2021.3079209","volume":"44","author":"T Hospedales","year":"2022","unstructured":"Hospedales, T., Antoniou, A., Micaelli, P., Storkey, A.: Meta-learning in neural networks: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 44(9), 5149\u20135169 (2022). https:\/\/doi.org\/10.1109\/TPAMI.2021.3079209","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"27_CR16","series-title":"The Springer Series on Challenges in Machine Learning","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-05318-5","volume-title":"Automated Machine Learning","year":"2019","unstructured":"Hutter, F., Kotthoff, L., Vanschoren, J. (eds.): Automated Machine Learning. TSSCML, Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-05318-5"},{"key":"27_CR17","doi-asserted-by":"publisher","unstructured":"Injadat, M., Salo, F., Nassif, A.B., Essex, A., Shami, A.: Bayesian optimization with machine learning algorithms towards anomaly detection. In: 2018 IEEE Global Communications Conference (GLOBECOM), pp. 1\u20136 (2018). https:\/\/doi.org\/10.1109\/GLOCOM.2018.8647714","DOI":"10.1109\/GLOCOM.2018.8647714"},{"key":"27_CR18","doi-asserted-by":"publisher","unstructured":"Jordan, M.I., Mitchell, T.M.: Machine learning: trends, perspectives, and prospects. Science 349(6245), 255\u2013260 (2015). https:\/\/doi.org\/10.1126\/science.aaa8415. https:\/\/www.science.org\/doi\/abs\/10.1126\/science.aaa8415","DOI":"10.1126\/science.aaa8415"},{"key":"27_CR19","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1115\/1.3653121","volume":"86","author":"HJ Kushner","year":"1963","unstructured":"Kushner, H.J.: A new method of locating the maximum point of an arbitrary multipeak curve in the presence of noise. J. Basic Eng. 86, 97\u2013106 (1963)","journal-title":"J. Basic Eng."},{"key":"27_CR20","unstructured":"Li, L., Jamieson, K.G., DeSalvo, G., Rostamizadeh, A., Talwalkar, A.: Efficient hyperparameter optimization and infinitely many armed bandits. CoRR abs\/1603.06560 (2016). http:\/\/arxiv.org\/abs\/1603.06560"},{"key":"27_CR21","unstructured":"Lindauer, M., Feurer, M., Eggensperger, K., Biedenkapp, A., Hutter, F.: Towards assessing the impact of Bayesian optimization\u2019s own hyperparameters. CoRR abs\/1908.06674 (2019). http:\/\/arxiv.org\/abs\/1908.06674"},{"key":"27_CR22","doi-asserted-by":"publisher","unstructured":"Lorenzo, P.R., Nalepa, J., Kawulok, M., Ramos, L.S., Pastor, J.R.: Particle swarm optimization for hyper-parameter selection in deep neural networks. In: Proceedings of the Genetic and Evolutionary Computation Conference, GECCO 2017, pp. 481\u2013488. Association for Computing Machinery, New York (2017). https:\/\/doi.org\/10.1145\/3071178.3071208","DOI":"10.1145\/3071178.3071208"},{"key":"27_CR23","first-page":"2825","volume":"12","author":"F Pedregosa","year":"2011","unstructured":"Pedregosa, F., et al.: Scikit-learn: machine learning in Python. J. Mach. Learn. Res. 12, 2825\u20132830 (2011)","journal-title":"J. Mach. Learn. Res."},{"key":"27_CR24","unstructured":"Raghu, A., Lorraine, J., Kornblith, S., McDermott, M., Duvenaud, D.K.: Meta-learning to improve pre-training. In: Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W. (eds.) Advances in Neural Information Processing Systems, vol. 34, pp. 23231\u201323244. Curran Associates, Inc. (2021). https:\/\/proceedings.neurips.cc\/paper\/2021\/file\/c3810d4a9513b028fc0f2a83cb6d7b50-Paper.pdf"},{"key":"27_CR25","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"63","DOI":"10.1007\/978-3-540-28650-9_4","volume-title":"Advanced Lectures on Machine Learning","author":"CE Rasmussen","year":"2004","unstructured":"Rasmussen, C.E.: Gaussian processes in machine learning. In: Bousquet, O., von Luxburg, U., R\u00e4tsch, G. (eds.) ML -2003. LNCS (LNAI), vol. 3176, pp. 63\u201371. Springer, Heidelberg (2004). https:\/\/doi.org\/10.1007\/978-3-540-28650-9_4"},{"key":"27_CR26","unstructured":"Shah, A., Ghahramani, Z.: Parallel predictive entropy search for batch global optimization of expensive objective functions. In: Cortes, C., Lawrence, N., Lee, D., Sugiyama, M., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 28. Curran Associates, Inc. (2015). https:\/\/proceedings.neurips.cc\/paper\/2015\/file\/57c0531e13f40b91b3b0f1a30b529a1d-Paper.pdf"},{"key":"27_CR27","unstructured":"Srinivas, N., Krause, A., Kakade, S.M., Seeger, M.W.: Gaussian process bandits without regret: an experimental design approach. CoRR abs\/0912.3995 (2009). http:\/\/arxiv.org\/abs\/0912.3995"},{"key":"27_CR28","doi-asserted-by":"publisher","unstructured":"Tsai, C.W., Fang, Z.Y.: An effective hyperparameter optimization algorithm for DNN to predict passengers at a metro station. ACM Trans. Internet Technol. 21(2) (2021). https:\/\/doi.org\/10.1145\/3410156","DOI":"10.1145\/3410156"},{"key":"27_CR29","doi-asserted-by":"publisher","unstructured":"Vulpe-Grigora\u015fi, A., Grigore, O.: Convolutional neural network hyperparameters optimization for facial emotion recognition. In: 2021 12th International Symposium on Advanced Topics in Electrical Engineering (ATEE), pp. 1\u20135 (2021). https:\/\/doi.org\/10.1109\/ATEE52255.2021.9425073","DOI":"10.1109\/ATEE52255.2021.9425073"},{"key":"27_CR30","doi-asserted-by":"publisher","unstructured":"Wada, T., Hino, H.: Bayesian optimization for multi-objective optimization and multi-point search (2019). https:\/\/doi.org\/10.48550\/ARXIV.1905.02370","DOI":"10.48550\/ARXIV.1905.02370"},{"key":"27_CR31","doi-asserted-by":"publisher","unstructured":"Wang, Z., Hutter, F., Zoghi, M., Matheson, D., de Freitas, N.: Bayesian optimization in a billion dimensions via random embeddings (2013). https:\/\/doi.org\/10.48550\/ARXIV.1301.1942. https:\/\/arxiv.org\/abs\/1301.1942","DOI":"10.48550\/ARXIV.1301.1942"},{"key":"27_CR32","doi-asserted-by":"crossref","unstructured":"Wicaksono, A.S., Supianto, A.A.: Hyper parameter optimization using the genetic algorithm on machine learning methods for online news popularity prediction. Int. J. Adv. Comput. Sci. Appl. 9(12) (2018)","DOI":"10.14569\/IJACSA.2018.091238"},{"key":"27_CR33","doi-asserted-by":"publisher","unstructured":"Wu, J., Chen, X.Y., Zhang, H., Xiong, L.D., Lei, H., Deng, S.H.: Hyperparameter optimization for machine learning models based on Bayesian optimization b. J. Electron. Sci. Technol. 17(1), 26\u201340 (2019). https:\/\/doi.org\/10.11989\/JEST.1674-862X.80904120. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1674862X19300047","DOI":"10.11989\/JEST.1674-862X.80904120"},{"key":"27_CR34","doi-asserted-by":"publisher","unstructured":"Wu, J., Frazier, P.I.: The parallel knowledge gradient method for batch Bayesian optimization (2016). https:\/\/doi.org\/10.48550\/ARXIV.1606.04414. https:\/\/arxiv.org\/abs\/1606.04414","DOI":"10.48550\/ARXIV.1606.04414"},{"key":"27_CR35","doi-asserted-by":"publisher","first-page":"295","DOI":"10.1016\/J.NEUCOM.2020.07.061","volume":"415","author":"L Yang","year":"2020","unstructured":"Yang, L., Shami, A.: On hyperparameter optimization of machine learning algorithms: theory and practice. Neurocomputing 415, 295\u2013316 (2020). https:\/\/doi.org\/10.1016\/J.NEUCOM.2020.07.061","journal-title":"Neurocomputing"},{"key":"27_CR36","unstructured":"Zhang, B., et al.: On the importance of hyperparameter optimization for model-based reinforcement learning. In: Banerjee, A., Fukumizu, K. (eds.) Proceedings of The 24th International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, 13\u201315 April 2021, vol. 130, pp. 4015\u20134023. PMLR (2021). https:\/\/proceedings.mlr.press\/v130\/zhang21n.html"}],"container-title":["Communications in Computer and Information Science","Optimization, Learning Algorithms and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-53025-8_27","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,31]],"date-time":"2024-01-31T20:16:22Z","timestamp":1706732182000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-53025-8_27"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031530241","9783031530258"],"references-count":36,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-53025-8_27","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"1 February 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"OL2A","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Optimization, Learning Algorithms and Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Ponta Delgada","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Portugal","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 September 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ol2a2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ol2a.ipb.pt\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"162","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"66","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"41% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}