{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,5]],"date-time":"2026-04-05T20:41:38Z","timestamp":1775421698442,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":58,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,6,26]],"date-time":"2021-06-26T00:00:00Z","timestamp":1624665600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,6,26]]},"DOI":"10.1145\/3449639.3459292","type":"proceedings-article","created":{"date-parts":[[2021,6,21]],"date-time":"2021-06-21T17:51:58Z","timestamp":1624297918000},"page":"323-331","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["Regularized evolutionary population-based training"],"prefix":"10.1145","author":[{"given":"Jason","family":"Liang","sequence":"first","affiliation":[{"name":"Cognizant AI Labs"}]},{"given":"Santiago","family":"Gonzalez","sequence":"additional","affiliation":[{"name":"The Univ. of Texas at Austin"}]},{"given":"Hormoz","family":"Shahrzad","sequence":"additional","affiliation":[{"name":"The Univ. of Texas at Austin"}]},{"given":"Risto","family":"Miikkulainen","sequence":"additional","affiliation":[{"name":"The Univ. of Texas at Austin"}]}],"member":"320","published-online":{"date-parts":[[2021,6,26]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473","author":"Bahdanau Dzmitry","year":"2014","unstructured":"Dzmitry Bahdanau , Kyunghyun Cho , and Yoshua Bengio . 2014. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 ( 2014 ). Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2014. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 (2014)."},{"key":"e_1_3_2_1_2_1","first-page":"998","article-title":"Metareg: Towards domain generalization using meta-regularization","volume":"31","author":"Balaji Yogesh","year":"2018","unstructured":"Yogesh Balaji , Swami Sankaranarayanan , and Rama Chellappa . 2018 . Metareg: Towards domain generalization using meta-regularization . Advances in Neural Information Processing Systems 31 (2018), 998 -- 1008 . Yogesh Balaji, Swami Sankaranarayanan, and Rama Chellappa. 2018. Metareg: Towards domain generalization using meta-regularization. Advances in Neural Information Processing Systems 31 (2018), 998--1008.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_3_1","volume-title":"Genetic programming: an introduction","author":"Banzhaf Wolfgang","unstructured":"Wolfgang Banzhaf , Peter Nordin , Robert E Keller , and Frank D Francone . 1998. Genetic programming: an introduction . Vol. 1 . Morgan Kaufmann San Francisco . Wolfgang Banzhaf, Peter Nordin, Robert E Keller, and Frank D Francone. 1998. Genetic programming: an introduction. Vol. 1. Morgan Kaufmann San Francisco."},{"key":"e_1_3_2_1_4_1","volume-title":"Visualising Basins of Attraction for the Cross-Entropy and the Squared Error Neural Network Loss Functions. arXiv preprint arXiv:1901.02302","author":"Bosman Anna Sergeevna","year":"2019","unstructured":"Anna Sergeevna Bosman , Andries Engelbrecht , and Mard\u00e9 Helbig . 2019. Visualising Basins of Attraction for the Cross-Entropy and the Squared Error Neural Network Loss Functions. arXiv preprint arXiv:1901.02302 ( 2019 ). Anna Sergeevna Bosman, Andries Engelbrecht, and Mard\u00e9 Helbig. 2019. Visualising Basins of Attraction for the Cross-Entropy and the Squared Error Neural Network Loss Functions. arXiv preprint arXiv:1901.02302 (2019)."},{"key":"e_1_3_2_1_5_1","volume-title":"Autoaugment: Learning augmentation policies from data. arXiv preprint arXiv:1805.09501","author":"Cubuk Ekin D","year":"2018","unstructured":"Ekin D Cubuk , Barret Zoph , Dandelion Mane , Vijay Vasudevan , and Quoc V Le . 2018 . Autoaugment: Learning augmentation policies from data. arXiv preprint arXiv:1805.09501 (2018). Ekin D Cubuk, Barret Zoph, Dandelion Mane, Vijay Vasudevan, and Quoc V Le. 2018. Autoaugment: Learning augmentation policies from data. arXiv preprint arXiv:1805.09501 (2018)."},{"key":"e_1_3_2_1_6_1","volume-title":"Improved regularization of convolutional neural networks with cutout. arXiv preprint arXiv:1708.04552","author":"DeVries Terrance","year":"2017","unstructured":"Terrance DeVries and Graham W Taylor . 2017. Improved regularization of convolutional neural networks with cutout. arXiv preprint arXiv:1708.04552 ( 2017 ). Terrance DeVries and Graham W Taylor. 2017. Improved regularization of convolutional neural networks with cutout. arXiv preprint arXiv:1708.04552 (2017)."},{"key":"e_1_3_2_1_7_1","first-page":"10677","article-title":"Time Matters in Regularizing Deep Networks: Weight Decay and Data Augmentation Affect Early Learning Dynamics, Matter Little Near Convergence","volume":"32","author":"Golatkar Aditya Sharad","year":"2019","unstructured":"Aditya Sharad Golatkar , Alessandro Achille , and Stefano Soatto . 2019 . Time Matters in Regularizing Deep Networks: Weight Decay and Data Augmentation Affect Early Learning Dynamics, Matter Little Near Convergence . In Advances in Neural Information Processing Systems 32. 10677 -- 10687 . Aditya Sharad Golatkar, Alessandro Achille, and Stefano Soatto. 2019. Time Matters in Regularizing Deep Networks: Weight Decay and Data Augmentation Affect Early Learning Dynamics, Matter Little Near Convergence. In Advances in Neural Information Processing Systems 32. 10677--10687.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_8_1","volume-title":"Genetic algorithms and machine learning. Machine learning 3, 2","author":"Goldberg David E","year":"1988","unstructured":"David E Goldberg and John H Holland . 1988. Genetic algorithms and machine learning. Machine learning 3, 2 ( 1988 ), 95--99. David E Goldberg and John H Holland. 1988. Genetic algorithms and machine learning. Machine learning 3, 2 (1988), 95--99."},{"key":"e_1_3_2_1_9_1","volume-title":"Improved Training Speed, Accuracy, and Data Utilization Through Loss Function Optimization. arXiv preprint arXiv:1905.11528","author":"Gonzalez Santiago","year":"2019","unstructured":"Santiago Gonzalez and Risto Miikkulainen . 2019. Improved Training Speed, Accuracy, and Data Utilization Through Loss Function Optimization. arXiv preprint arXiv:1905.11528 ( 2019 ). Santiago Gonzalez and Risto Miikkulainen. 2019. Improved Training Speed, Accuracy, and Data Utilization Through Loss Function Optimization. arXiv preprint arXiv:1905.11528 (2019)."},{"key":"e_1_3_2_1_10_1","volume-title":"Evolving Loss Functions With Multivariate Taylor Polynomial Parameterizations. arXiv preprint arXiv:2002.00059","author":"Gonzalez Santiago","year":"2020","unstructured":"Santiago Gonzalez and Risto Miikkulainen . 2020. Evolving Loss Functions With Multivariate Taylor Polynomial Parameterizations. arXiv preprint arXiv:2002.00059 ( 2020 ). Santiago Gonzalez and Risto Miikkulainen. 2020. Evolving Loss Functions With Multivariate Taylor Polynomial Parameterizations. arXiv preprint arXiv:2002.00059 (2020)."},{"key":"e_1_3_2_1_11_1","unstructured":"Ian Goodfellow Jean Pouget-Abadie Mehdi Mirza Bing Xu David Warde-Farley Sherjil Ozair Aaron Courville and Yoshua Bengio. 2014. Generative adversarial nets. In Advances in Neural Information Processing Systems. 2672--2680.  Ian Goodfellow Jean Pouget-Abadie Mehdi Mirza Bing Xu David Warde-Farley Sherjil Ozair Aaron Courville and Yoshua Bengio. 2014. Generative adversarial nets. In Advances in Neural Information Processing Systems. 2672--2680."},{"key":"e_1_3_2_1_12_1","volume-title":"Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572","author":"Goodfellow Ian J","year":"2014","unstructured":"Ian J Goodfellow , Jonathon Shlens , and Christian Szegedy . 2014. Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572 ( 2014 ). Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy. 2014. Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572 (2014)."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICEC.1996.542381"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.123"},{"key":"e_1_3_2_1_15_1","volume-title":"Deep Residual Learning for Image Recognition. IEEE Conference on Computer Vision and Pattern Recognition (CVPR)","author":"He Kaiming","year":"2016","unstructured":"Kaiming He , Xiangyu Zhang , Shaoqing Ren , and Jian Sun . 2016 . Deep Residual Learning for Image Recognition. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016), 770--778. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep Residual Learning for Image Recognition. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016), 770--778."},{"key":"e_1_3_2_1_16_1","volume-title":"Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531","author":"Hinton Geoffrey","year":"2015","unstructured":"Geoffrey Hinton , Oriol Vinyals , and Jeff Dean . 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 ( 2015 ). Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)."},{"key":"e_1_3_2_1_17_1","volume-title":"Population based augmentation: Efficient learning of augmentation policy schedules. arXiv preprint arXiv:1905.05393","author":"Ho Daniel","year":"2019","unstructured":"Daniel Ho , Eric Liang , Ion Stoica , Pieter Abbeel , and Xi Chen . 2019. Population based augmentation: Efficient learning of augmentation policy schedules. arXiv preprint arXiv:1905.05393 ( 2019 ). Daniel Ho, Eric Liang, Ion Stoica, Pieter Abbeel, and Xi Chen. 2019. Population based augmentation: Efficient learning of augmentation policy schedules. arXiv preprint arXiv:1905.05393 (2019)."},{"key":"e_1_3_2_1_18_1","volume-title":"OpenAI Jonathan Ho, and Pieter Abbeel","author":"Houthooft Rein","year":"2018","unstructured":"Rein Houthooft , Yuhua Chen , Phillip Isola , Bradly Stadie , Filip Wolski , OpenAI Jonathan Ho, and Pieter Abbeel . 2018 . Evolved policy gradients. In Advances in Neural Information Processing Systems . 5400--5409. Rein Houthooft, Yuhua Chen, Phillip Isola, Bradly Stadie, Filip Wolski, OpenAI Jonathan Ho, and Pieter Abbeel. 2018. Evolved policy gradients. In Advances in Neural Information Processing Systems. 5400--5409."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.243"},{"key":"e_1_3_2_1_20_1","volume-title":"International conference on machine learning. PMLR, 448--456","author":"Ioffe Sergey","year":"2015","unstructured":"Sergey Ioffe and Christian Szegedy . 2015 . Batch normalization: Accelerating deep network training by reducing internal covariate shift . In International conference on machine learning. PMLR, 448--456 . Sergey Ioffe and Christian Szegedy. 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International conference on machine learning. PMLR, 448--456."},{"key":"e_1_3_2_1_21_1","unstructured":"Max Jaderberg Valentin Dalibard Simon Osindero Wojciech M Czarnecki Jeff Donahue Ali Razavi Oriol Vinyals Tim Green Iain Dunning Karen Simonyan etal 2017. Population based training of neural networks. arXiv preprint arXiv:1711.09846 (2017).  Max Jaderberg Valentin Dalibard Simon Osindero Wojciech M Czarnecki Jeff Donahue Ali Razavi Oriol Vinyals Tim Green Iain Dunning Karen Simonyan et al. 2017. Population based training of neural networks. arXiv preprint arXiv:1711.09846 (2017)."},{"key":"e_1_3_2_1_22_1","volume-title":"On loss functions for deep neural networks in classification. arXiv preprint arXiv:1702.05659","author":"Janocha Katarzyna","year":"2017","unstructured":"Katarzyna Janocha and Wojciech Marian Czarnecki . 2017. On loss functions for deep neural networks in classification. arXiv preprint arXiv:1702.05659 ( 2017 ). Katarzyna Janocha and Wojciech Marian Czarnecki. 2017. On loss functions for deep neural networks in classification. arXiv preprint arXiv:1702.05659 (2017)."},{"key":"e_1_3_2_1_23_1","volume-title":"Self-knowledge distillation: A simple way for better generalization. arXiv preprint arXiv:2006.12000","author":"Kim Kyungyul","year":"2020","unstructured":"Kyungyul Kim , ByeongMoon Ji , Doyoung Yoon , and Sangheum Hwang . 2020. Self-knowledge distillation: A simple way for better generalization. arXiv preprint arXiv:2006.12000 ( 2020 ). Kyungyul Kim, ByeongMoon Ji, Doyoung Yoon, and Sangheum Hwang. 2020. Self-knowledge distillation: A simple way for better generalization. arXiv preprint arXiv:2006.12000 (2020)."},{"key":"e_1_3_2_1_24_1","unstructured":"Aaron Klein Stefan Falkner Simon Bartels Philipp Hennig and Frank Hutter. 2017. Fast bayesian optimization of machine learning hyperparameters on large datasets. In Artificial Intelligence and Statistics. PMLR 528--536.  Aaron Klein Stefan Falkner Simon Bartels Philipp Hennig and Frank Hutter. 2017. Fast bayesian optimization of machine learning hyperparameters on large datasets. In Artificial Intelligence and Statistics. PMLR 528--536."},{"key":"e_1_3_2_1_25_1","unstructured":"Alex Krizhevsky and Geoffrey Hinton. 2009. Learning multiple layers of features from tiny images.  Alex Krizhevsky and Geoffrey Hinton. 2009. Learning multiple layers of features from tiny images."},{"key":"e_1_3_2_1_26_1","volume-title":"Advances in Neural Information Processing Systems 25","author":"Krizhevsky Alex","unstructured":"Alex Krizhevsky , Ilya Sutskever , and Geoffrey E Hinton . 2012. ImageNet Classification with Deep Convolutional Neural Networks . In Advances in Neural Information Processing Systems 25 , F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger (Eds.). Curran Associates, Inc. , 1097--1105. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems 25, F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger (Eds.). Curran Associates, Inc., 1097--1105."},{"key":"e_1_3_2_1_27_1","volume-title":"Regularization for deep learning: A taxonomy. arXiv preprint arXiv:1710.10686","author":"Kuka\u010dka Jan","year":"2017","unstructured":"Jan Kuka\u010dka , Vladimir Golkov , and Daniel Cremers . 2017. Regularization for deep learning: A taxonomy. arXiv preprint arXiv:1710.10686 ( 2017 ). Jan Kuka\u010dka, Vladimir Golkov, and Daniel Cremers. 2017. Regularization for deep learning: A taxonomy. arXiv preprint arXiv:1710.10686 (2017)."},{"key":"e_1_3_2_1_28_1","volume-title":"Deep learning. Nature 521, 7553","author":"LeCun Yann","year":"2015","unstructured":"Yann LeCun , Yoshua Bengio , and Geoffrey Hinton . 2015. Deep learning. Nature 521, 7553 ( 2015 ), 436. Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. 2015. Deep learning. Nature 521, 7553 (2015), 436."},{"key":"e_1_3_2_1_29_1","unstructured":"Joel Lehman and Kenneth O Stanley. 2008. Exploiting open-endedness to solve problems through the search for novelty.. In ALIFE. Citeseer 329--336.  Joel Lehman and Kenneth O Stanley. 2008. Exploiting open-endedness to solve problems through the search for novelty.. In ALIFE. Citeseer 329--336."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.5555\/3122009.3242042"},{"key":"e_1_3_2_1_31_1","volume-title":"Darts: Differentiable architecture search. arXiv preprint arXiv:1806.09055","author":"Liu Hanxiao","year":"2018","unstructured":"Hanxiao Liu , Karen Simonyan , and Yiming Yang . 2018 . Darts: Differentiable architecture search. arXiv preprint arXiv:1806.09055 (2018). Hanxiao Liu, Karen Simonyan, and Yiming Yang. 2018. Darts: Differentiable architecture search. arXiv preprint arXiv:1806.09055 (2018)."},{"key":"e_1_3_2_1_32_1","volume-title":"CMA-ES for hyperparameter optimization of deep neural networks. arXiv preprint arXiv:1604.07269","author":"Loshchilov Ilya","year":"2016","unstructured":"Ilya Loshchilov and Frank Hutter . 2016. CMA-ES for hyperparameter optimization of deep neural networks. arXiv preprint arXiv:1604.07269 ( 2016 ). Ilya Loshchilov and Frank Hutter. 2016. CMA-ES for hyperparameter optimization of deep neural networks. arXiv preprint arXiv:1604.07269 (2016)."},{"key":"e_1_3_2_1_33_1","volume-title":"NSGA-NET: a multi-objective genetic algorithm for neural architecture search. arXiv preprint arXiv:1810.03522","author":"Lu Zhichao","year":"2018","unstructured":"Zhichao Lu , Ian Whalen , Vishnu Boddeti , Yashesh Dhebar , Kalyanmoy Deb , Erik Goodman , and Wolfgang Banzhaf . 2018. NSGA-NET: a multi-objective genetic algorithm for neural architecture search. arXiv preprint arXiv:1810.03522 ( 2018 ). Zhichao Lu, Ian Whalen, Vishnu Boddeti, Yashesh Dhebar, Kalyanmoy Deb, Erik Goodman, and Wolfgang Banzhaf. 2018. NSGA-NET: a multi-objective genetic algorithm for neural architecture search. arXiv preprint arXiv:1810.03522 (2018)."},{"key":"e_1_3_2_1_34_1","volume-title":"International Conference on Machine Learning. 2113--2122","author":"Maclaurin Dougal","year":"2015","unstructured":"Dougal Maclaurin , David Duvenaud , and Ryan Adams . 2015 . Gradient-based hyperparameter optimization through reversible learning . In International Conference on Machine Learning. 2113--2122 . Dougal Maclaurin, David Duvenaud, and Ryan Adams. 2015. Gradient-based hyperparameter optimization through reversible learning. In International Conference on Machine Learning. 2113--2122."},{"key":"e_1_3_2_1_35_1","volume-title":"Implicit self-regularization in deep neural networks: Evidence from random matrix theory and implications for learning. arXiv preprint arXiv:1810.01075","author":"Martin Charles H","year":"2018","unstructured":"Charles H Martin and Michael W Mahoney . 2018. Implicit self-regularization in deep neural networks: Evidence from random matrix theory and implications for learning. arXiv preprint arXiv:1810.01075 ( 2018 ). Charles H Martin and Michael W Mahoney. 2018. Implicit self-regularization in deep neural networks: Evidence from random matrix theory and implications for learning. arXiv preprint arXiv:1810.01075 (2018)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"crossref","unstructured":"Risto Miikkulainen Jason Liang Elliot Meyerson Aditya Rawal Daniel Fink Olivier Francon Bala Raju Hormoz Shahrzad Arshak Navruzyan Nigel Duffy etal 2019. Evolving deep neural networks. In Artificial Intelligence in the Age of Neural Networks and Brain Computing. Elsevier 293--312.  Risto Miikkulainen Jason Liang Elliot Meyerson Aditya Rawal Daniel Fink Olivier Francon Bala Raju Hormoz Shahrzad Arshak Navruzyan Nigel Duffy et al. 2019. Evolving deep neural networks. In Artificial Intelligence in the Age of Neural Networks and Brain Computing. Elsevier 293--312.","DOI":"10.1016\/B978-0-12-815480-9.00015-3"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"crossref","unstructured":"Volodymyr Mnih Koray Kavukcuoglu David Silver Andrei A Rusu Joel Veness Marc G Bellemare Alex Graves Martin Riedmiller Andreas K Fidjeland Georg Ostrovski etal 2015. Human-level control through deep reinforcement learning. Nature 518 7540 (2015) 529.  Volodymyr Mnih Koray Kavukcuoglu David Silver Andrei A Rusu Joel Veness Marc G Bellemare Alex Graves Martin Riedmiller Andreas K Fidjeland Georg Ostrovski et al. 2015. Human-level control through deep reinforcement learning. Nature 518 7540 (2015) 529.","DOI":"10.1038\/nature14236"},{"key":"e_1_3_2_1_38_1","volume-title":"A simple weight decay can improve generalization. Advances in neural information processing systems 4","author":"Moody John","year":"1995","unstructured":"John Moody , Stephen Hanson , Anders Krogh , and John A Hertz . 1995. A simple weight decay can improve generalization. Advances in neural information processing systems 4 , 1995 (1995), 950--957. John Moody, Stephen Hanson, Anders Krogh, and John A Hertz. 1995. A simple weight decay can improve generalization. Advances in neural information processing systems 4, 1995 (1995), 950--957."},{"key":"e_1_3_2_1_39_1","volume-title":"Illuminating search spaces by mapping elites. arXiv preprint arXiv:1504.04909","author":"Mouret Jean-Baptiste","year":"2015","unstructured":"Jean-Baptiste Mouret and Jeff Clune . 2015. Illuminating search spaces by mapping elites. arXiv preprint arXiv:1504.04909 ( 2015 ). Jean-Baptiste Mouret and Jeff Clune. 2015. Illuminating search spaces by mapping elites. arXiv preprint arXiv:1504.04909 (2015)."},{"key":"e_1_3_2_1_40_1","volume-title":"When does label smoothing help? arXiv preprint arXiv:1906.02629","author":"M\u00fcller Rafael","year":"2019","unstructured":"Rafael M\u00fcller , Simon Kornblith , and Geoffrey Hinton . 2019. When does label smoothing help? arXiv preprint arXiv:1906.02629 ( 2019 ). Rafael M\u00fcller, Simon Kornblith, and Geoffrey Hinton. 2019. When does label smoothing help? arXiv preprint arXiv:1906.02629 (2019)."},{"key":"e_1_3_2_1_41_1","unstructured":"Yuval Netzer Tao Wang Adam Coates Alessandro Bissacco Bo Wu and Andrew Y Ng. 2011. Reading digits in natural images with unsupervised feature learning.  Yuval Netzer Tao Wang Adam Coates Alessandro Bissacco Bo Wu and Andrew Y Ng. 2011. Reading digits in natural images with unsupervised feature learning."},{"key":"e_1_3_2_1_42_1","volume-title":"Efficient neural architecture search via parameter sharing. arXiv preprint arXiv:1802.03268","author":"Pham Hieu","year":"2018","unstructured":"Hieu Pham , Melody Y Guan , Barret Zoph , Quoc V Le , and Jeff Dean . 2018. Efficient neural architecture search via parameter sharing. arXiv preprint arXiv:1802.03268 ( 2018 ). Hieu Pham, Melody Y Guan, Barret Zoph, Quoc V Le, and Jeff Dean. 2018. Efficient neural architecture search via parameter sharing. arXiv preprint arXiv:1802.03268 (2018)."},{"key":"e_1_3_2_1_43_1","volume-title":"Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434","author":"Radford Alec","year":"2015","unstructured":"Alec Radford , Luke Metz , and Soumith Chintala . 2015. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434 ( 2015 ). Alec Radford, Luke Metz, and Soumith Chintala. 2015. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434 (2015)."},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33014780"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1162\/isal_a_00113"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-39958-0_14"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/ABLAZE.2015.7154916"},{"key":"e_1_3_2_1_48_1","volume-title":"Cyclical learning rates for training neural networks. In 2017 IEEE winter conference on applications of computer vision (WACV)","author":"Smith Leslie N","unstructured":"Leslie N Smith . 2017. Cyclical learning rates for training neural networks. In 2017 IEEE winter conference on applications of computer vision (WACV) . IEEE , 464--472. Leslie N Smith. 2017. Cyclical learning rates for training neural networks. In 2017 IEEE winter conference on applications of computer vision (WACV). IEEE, 464--472."},{"key":"e_1_3_2_1_49_1","unstructured":"Jasper Snoek Hugo Larochelle and Ryan P Adams. 2012. Practical bayesian optimization of machine learning algorithms. In Advances in neural information processing systems. 2951--2959.  Jasper Snoek Hugo Larochelle and Ryan P Adams. 2012. Practical bayesian optimization of machine learning algorithms. In Advances in neural information processing systems. 2951--2959."},{"key":"e_1_3_2_1_50_1","volume-title":"International conference on machine learning. 2171--2180","author":"Snoek Jasper","year":"2015","unstructured":"Jasper Snoek , Oren Rippel , Kevin Swersky , Ryan Kiros , Nadathur Satish , Narayanan Sundaram , Mostofa Patwary , Mr Prabhat , and Ryan Adams . 2015 . Scalable bayesian optimization using deep neural networks . In International conference on machine learning. 2171--2180 . Jasper Snoek, Oren Rippel, Kevin Swersky, Ryan Kiros, Nadathur Satish, Narayanan Sundaram, Mostofa Patwary, Mr Prabhat, and Ryan Adams. 2015. Scalable bayesian optimization using deep neural networks. In International conference on machine learning. 2171--2180."},{"key":"e_1_3_2_1_51_1","volume-title":"Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research 15, 1","author":"Srivastava Nitish","year":"2014","unstructured":"Nitish Srivastava , Geoffrey Hinton , Alex Krizhevsky , Ilya Sutskever , and Ruslan Salakhutdinov . 2014. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research 15, 1 ( 2014 ), 1929--1958. Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. 2014. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research 15, 1 (2014), 1929--1958."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-018-0006-z"},{"key":"e_1_3_2_1_53_1","volume-title":"A genetic algorithm tutorial. Statistics and computing 4, 2","author":"Whitley Darrell","year":"1994","unstructured":"Darrell Whitley . 1994. A genetic algorithm tutorial. Statistics and computing 4, 2 ( 1994 ), 65--85. Darrell Whitley. 1994. A genetic algorithm tutorial. Statistics and computing 4, 2 (1994), 65--85."},{"key":"e_1_3_2_1_54_1","volume-title":"How does learning rate decay help modern neural networks? arXiv preprint arXiv:1908.01878","author":"You Kaichao","year":"2019","unstructured":"Kaichao You , Mingsheng Long , Jianmin Wang , and Michael I Jordan . 2019. How does learning rate decay help modern neural networks? arXiv preprint arXiv:1908.01878 ( 2019 ). Kaichao You, Mingsheng Long, Jianmin Wang, and Michael I Jordan. 2019. How does learning rate decay help modern neural networks? arXiv preprint arXiv:1908.01878 (2019)."},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00612"},{"key":"e_1_3_2_1_56_1","volume-title":"Wide residual networks. arXiv preprint arXiv:1605.07146","author":"Zagoruyko Sergey","year":"2016","unstructured":"Sergey Zagoruyko and Nikos Komodakis . 2016. Wide residual networks. arXiv preprint arXiv:1605.07146 ( 2016 ). Sergey Zagoruyko and Nikos Komodakis. 2016. Wide residual networks. arXiv preprint arXiv:1605.07146 (2016)."},{"key":"e_1_3_2_1_57_1","volume-title":"Neural architecture search with reinforcement learning. arXiv preprint arXiv:1611.01578","author":"Zoph Barret","year":"2016","unstructured":"Barret Zoph and Quoc V Le. 2016. Neural architecture search with reinforcement learning. arXiv preprint arXiv:1611.01578 ( 2016 ). Barret Zoph and Quoc V Le. 2016. Neural architecture search with reinforcement learning. arXiv preprint arXiv:1611.01578 (2016)."},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00907"}],"event":{"name":"GECCO '21: Genetic and Evolutionary Computation Conference","location":"Lille France","acronym":"GECCO '21","sponsor":["SIGEVO ACM Special Interest Group on Genetic and Evolutionary Computation"]},"container-title":["Proceedings of the Genetic and Evolutionary Computation Conference"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3449639.3459292","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3449639.3459292","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T22:01:57Z","timestamp":1750197717000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3449639.3459292"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,6,26]]},"references-count":58,"alternative-id":["10.1145\/3449639.3459292","10.1145\/3449639"],"URL":"https:\/\/doi.org\/10.1145\/3449639.3459292","relation":{},"subject":[],"published":{"date-parts":[[2021,6,26]]},"assertion":[{"value":"2021-06-26","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}