{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,29]],"date-time":"2024-10-29T12:35:46Z","timestamp":1730205346154,"version":"3.28.0"},"reference-count":37,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,6,30]],"date-time":"2024-06-30T00:00:00Z","timestamp":1719705600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,6,30]],"date-time":"2024-06-30T00:00:00Z","timestamp":1719705600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,6,30]]},"DOI":"10.1109\/cec60901.2024.10612098","type":"proceedings-article","created":{"date-parts":[[2024,8,8]],"date-time":"2024-08-08T17:55:15Z","timestamp":1723139715000},"page":"1-8","source":"Crossref","is-referenced-by-count":0,"title":["High-Throughput Multi-Objective Bayesian Optimization using Gradients"],"prefix":"10.1109","author":[{"given":"Yiming","family":"Yao","sequence":"first","affiliation":[{"name":"City University of Hong Kong (Dongguan),Dongguan,China,523000"}]},{"given":"Fei","family":"Liu","sequence":"additional","affiliation":[{"name":"City University of Hong Kong,Department of Computer Science,Hong Kong SAR,China"}]},{"given":"Qingfu","family":"Zhang","sequence":"additional","affiliation":[{"name":"City University of Hong Kong,Department of Computer Science,Hong Kong SAR,China"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Gaussian processes for machine learning","volume":"2","author":"Williams","year":"2006","journal-title":"MIT press Cambridge, MA"},{"key":"ref2","first-page":"2113","article-title":"Gradient-based hyper-parameter optimization through reversible learning","author":"Maclaurin","journal-title":"International conference on machine learning"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.2514\/2.2412"},{"key":"ref4","article-title":"Do we need \u201charmless\u201d bayesian optimization and \u201cfirst-order\u201d bayesian optimization","author":"Ahmed","year":"2016","journal-title":"NIPS Bayes Opt"},{"key":"ref5","article-title":"Bayesian optimization with gradients","volume":"30","author":"Wu","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref6","article-title":"A batched scalable multi-objective bayesian optimization algorithm","author":"Lin","year":"2018","journal-title":"arXiv preprint"},{"key":"ref7","first-page":"I","article-title":"Moea\/d with gradient-enhanced kriging for expensive multiobjective optimization","author":"Liu","year":"2022","journal-title":"Natural Computing"},{"key":"ref8","first-page":"5213","article-title":"Inducing point allocation for sparse gaussian processes in high-throughput bayesian optimisation","author":"Moss","journal-title":"International Conference on Artificial Intelligence and Statistics. PMLR"},{"journal-title":"Large-batch, iteration-efficient neural bayesian design optimization","year":"2023","author":"Ansari","key":"ref9"},{"key":"ref10","first-page":"133","article-title":"Paral-lelised bayesian optimisation via thompson sampling","author":"Kandasamy","journal-title":"International Conference on Artificial Intelligence and Statistics. PMLR"},{"key":"ref11","first-page":"1470","article-title":"Parallel and distributed thompson sampling for large-scale accelerated exploration of chemical space","author":"Hernandez-Lobato","journal-title":"International conference on machine learning"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1038\/nature17439"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2021.107909"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.5555\/1046920.1194909"},{"key":"ref15","article-title":"Sparse gaussian processes using pseudo-inputs","volume":"18","author":"Snelson","year":"2005","journal-title":"Advances in neural information processing systems"},{"key":"ref16","first-page":"567","article-title":"Variational learning of inducing variables in sparse gaussian processes","volume-title":"Artificial intelligence and statistics","author":"Titsias","year":"2009"},{"key":"ref17","article-title":"Gaussian processes for big data","author":"Hensman","year":"2013","journal-title":"arXiv preprint"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2019.2957109"},{"key":"ref19","article-title":"Sparse gaussian processes for bayesian optimization","volume-title":"UAl","author":"Mclntire","year":"2016"},{"key":"ref20","first-page":"207","article-title":"Deep gaussian processes","volume-title":"Artificial intelligence and statistics","author":"Damianou","year":"2013"},{"key":"ref21","first-page":"6442","article-title":"Scaling gaussian processes with derivative information using variational infer-ence","volume":"34","author":"Padidar","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/TEVC.2009.2033671"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/TEVC.2007.892759"},{"key":"ref24","first-page":"4702","article-title":"Parametric gaussian process regressors","author":"Jankowiak","journal-title":"International Conference on Machine Learning"},{"key":"ref25","article-title":"Gaussian process optimization in the bandit setting: No regret and experimental design","author":"Srinivas","year":"2009","journal-title":"arXiv preprint"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4615-5563-6"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1115\/DETC2003\/DAC-48760"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1162\/106365600568202"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TEVC.2008.925798"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/CEC.2002.1007032"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2020.106078"},{"key":"ref32","article-title":"Gpytorch: Blackbox matrix-matrix gaussian process inference with gpu acceleration","volume":"31","author":"Gardner","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref33","article-title":"Pytorch: An imperative style, high-performance deep learning library","volume":"32","author":"Paszke","year":"2019","journal-title":"Advances in neural information processing systems"},{"key":"ref34","article-title":"Adam: A method for stochastic optimization","author":"Kingma","year":"2014","journal-title":"arXiv preprint"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2990567"},{"key":"ref36","article-title":"Evolutionary pareto set learning with structure constraints","author":"Lin","year":"2023","journal-title":"arXiv preprint"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-27250-9_18"}],"event":{"name":"2024 IEEE Congress on Evolutionary Computation (CEC)","start":{"date-parts":[[2024,6,30]]},"location":"Yokohama, Japan","end":{"date-parts":[[2024,7,5]]}},"container-title":["2024 IEEE Congress on Evolutionary Computation (CEC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10609966\/10611750\/10612098.pdf?arnumber=10612098","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,11]],"date-time":"2024-08-11T04:12:56Z","timestamp":1723349576000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10612098\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,30]]},"references-count":37,"URL":"https:\/\/doi.org\/10.1109\/cec60901.2024.10612098","relation":{},"subject":[],"published":{"date-parts":[[2024,6,30]]}}}