{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T03:44:54Z","timestamp":1775274294356,"version":"3.50.1"},"reference-count":71,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,1,8]],"date-time":"2024-01-08T00:00:00Z","timestamp":1704672000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,1,8]],"date-time":"2024-01-08T00:00:00Z","timestamp":1704672000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/100016818","name":"UT-Battelle","doi-asserted-by":"publisher","award":["DE-AC05-00OR22725"],"award-info":[{"award-number":["DE-AC05-00OR22725"]}],"id":[{"id":"10.13039\/100016818","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100012338","name":"Alan Turing Institute","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100012338","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Mach Learn"],"published-print":{"date-parts":[[2024,3]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Mixtures of experts have become an indispensable tool for flexible modelling in a supervised learning context, allowing not only the mean function but the entire density of the output to change with the inputs. Sparse Gaussian processes (GP) have shown promise as a leading candidate for the experts in such models, and in this article, we propose to design the gating network for selecting the experts from such mixtures of sparse GPs using a deep neural network (DNN). Furthermore, a fast one pass algorithm called Cluster\u2013Classify\u2013Regress (CCR) is leveraged to approximate the maximum a posteriori (MAP) estimator extremely quickly. This powerful combination of model and algorithm together delivers a novel method which is flexible, robust, and extremely efficient. In particular, the method is able to outperform competing methods in terms of accuracy and uncertainty quantification. The cost is competitive on low-dimensional and small data sets, but is<jats:italic>significantly lower for higher-dimensional and big data sets<\/jats:italic>. Iteratively maximizing the distribution of experts given allocations and allocations given experts does not provide significant improvement, which indicates that the algorithm achieves a good approximation to the local MAP estimator very fast. This insight can be useful also in the context of other mixture of experts models.<\/jats:p>","DOI":"10.1007\/s10994-023-06491-x","type":"journal-article","created":{"date-parts":[[2024,1,8]],"date-time":"2024-01-08T18:02:35Z","timestamp":1704736955000},"page":"1483-1508","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Fast deep mixtures of Gaussian process experts"],"prefix":"10.1007","volume":"113","author":[{"given":"Clement","family":"Etienam","sequence":"first","affiliation":[]},{"given":"Kody J. H.","family":"Law","sequence":"additional","affiliation":[]},{"given":"Sara","family":"Wade","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0674-3721","authenticated-orcid":false,"given":"Vitaly","family":"Zankin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,8]]},"reference":[{"key":"6491_CR1","unstructured":"Ambrogioni, L., G\u00fc\u00e7l\u00fc, U., van Gerven, M.A., & Maris, E. (2017). The kernel mixture network: A nonparametric method for conditional density estimation of continuous random variables. arXiv preprint arXiv:1705.07111 ."},{"issue":"4","key":"6491_CR2","doi-asserted-by":"publisher","first-page":"825","DOI":"10.1111\/j.1467-9868.2008.00663.x","volume":"70","author":"S Banerjee","year":"2008","unstructured":"Banerjee, S., Gelfand, A. E., Finley, A. O., & Sang, H. (2008). Gaussian predictive process models for large spatial data sets. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 70(4), 825\u2013848.","journal-title":"Journal of the Royal Statistical Society: Series B (Statistical Methodology)"},{"issue":"5","key":"6491_CR3","doi-asserted-by":"publisher","first-page":"1793","DOI":"10.1063\/1.872848","volume":"5","author":"G Bateman","year":"1998","unstructured":"Bateman, G., Kritz, A. H., Kinsey, J. E., Redd, A. J., & Weiland, J. (1998). Predicting temperature and density profiles in tokamaks. Physics of Plasmas, 5(5), 1793\u20131799.","journal-title":"Physics of Plasmas"},{"key":"6491_CR4","first-page":"1533","volume":"29","author":"M Bauer","year":"2016","unstructured":"Bauer, M., van der Wilk, M., & Rasmussen, C. E. (2016). Understanding probabilistic sparse Gaussian process approximations. Advances in Neural Information Processing Systems, 29, 1533\u20131541.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"6491_CR5","doi-asserted-by":"publisher","first-page":"491","DOI":"10.3934\/fods.2019020","volume":"1","author":"DE Bernholdt","year":"2019","unstructured":"Bernholdt, D. E., Cianciosa, M. R., Green, D. L., Park, J. M., Law, K. J., & Etienam, C. (2019). Cluster, classify, regress: A general method for learning discontinuous functions. Foundations of Data Science, 1, 491.","journal-title":"Foundations of Data Science"},{"key":"6491_CR6","unstructured":"Bishop, C. M. (1994). Mixture density networks. Technical Report, Aston University."},{"key":"6491_CR7","first-page":"1","volume":"18","author":"TD Bui","year":"2017","unstructured":"Bui, T. D., Yan, J., & Turner, R. E. (2017). A unifying framework for sparse Gaussian process approximation using power expectation propagation. Journal of Machine Learning Research, 18, 1\u201372.","journal-title":"Journal of Machine Learning Research"},{"key":"6491_CR8","first-page":"1","volume":"21","author":"DR Burt","year":"2020","unstructured":"Burt, D. R., Rasmussen, C. E., & van der Wilk, M. (2020). Convergence of sparse variational inference in Gaussian processes regression. Journal of Machine Learning Research, 21, 1\u201363.","journal-title":"Journal of Machine Learning Research"},{"key":"6491_CR9","unstructured":"Cao, Y., & Fleet, D. J. (2014). Generalized product of experts for automatic and principled fusion of Gaussian process predictions. CoRR."},{"issue":"2","key":"6491_CR10","doi-asserted-by":"publisher","first-page":"1099","DOI":"10.1214\/17-AAP1326","volume":"28","author":"S Chatterjee","year":"2018","unstructured":"Chatterjee, S., & Diaconis, P. (2018). The sample size required in importance sampling. The Annals of Applied Probability, 28(2), 1099\u20131135.","journal-title":"The Annals of Applied Probability"},{"key":"6491_CR11","doi-asserted-by":"crossref","unstructured":"Chen, Z., Ma, J., & Zhou, Y. (2014). A precise hard-cut EM algorithm for mixtures of Gaussian processes. In International conference on intelligent computing (pp. 68\u201375).","DOI":"10.1007\/978-3-319-09339-0_7"},{"issue":"1","key":"6491_CR12","doi-asserted-by":"publisher","first-page":"266","DOI":"10.1214\/09-AOAS285","volume":"4","author":"HA Chipman","year":"2010","unstructured":"Chipman, H. A., George, E. I., & McCulloch, R. E. (2010). BART: Bayesian additive regression trees. The Annals of Applied Statistics, 4(1), 266\u2013298.","journal-title":"The Annals of Applied Statistics"},{"key":"6491_CR13","unstructured":"Chollet, F., et\u00a0al. (2015). Keras. https:\/\/keras.io"},{"issue":"1","key":"6491_CR14","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1111\/j.1467-9868.2007.00633.x","volume":"70","author":"N Cressie","year":"2008","unstructured":"Cressie, N., & Johannesson, G. (2008). Fixed rank kriging for very large spatial data sets. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 70(1), 209\u2013226.","journal-title":"Journal of the Royal Statistical Society: Series B (Statistical Methodology)"},{"key":"6491_CR15","unstructured":"Damianou, A., & Lawrence, N.D. (2013). Deep gaussian processes. In Artificial intelligence and statistics (pp. 207\u2013215)."},{"key":"6491_CR16","unstructured":"Daskalakis, C., Dellaportas, P., & Panos, A. (2020). Faster Gaussian processes via deep embeddings."},{"key":"6491_CR17","unstructured":"Deisenroth, M., & Ng, J.W. (2015). Distributed Gaussian processes. In International conference on machine learning (pp. 1481\u20131490)."},{"key":"6491_CR18","unstructured":"Durrande, N., Adam, V., Bordeaux, L., Eleftheriadis, S., & Hensman, J. (2019). Banded matrix operators for Gaussian Markov models in the automatic differentiation era. In 22nd International conference on artificial intelligence and statistics (Vol. 89, pp. 780\u20132789)."},{"key":"6491_CR19","unstructured":"Ester, M., Kriegel, H.-P., Sander, J., & Xu, X. (1996). A density-based algorithm for discovering clusters in large spatial databases with noise. Knowledge discovery in databases (Vol. 96, pp. 226\u2013231)."},{"key":"6491_CR20","unstructured":"Gadd, C. W., Wade, S., & Boukouvalas, A. (2020). Enriched mixtures of Gaussian process experts. In 23rd International conference on artificial intelligence and statistics."},{"key":"6491_CR21","volume-title":"Handbook of mixture analysis","author":"IC Gormley","year":"2019","unstructured":"Gormley, I. C., & Fr\u00fchwirth-Schnatter, S. (2019). Mixtures of experts models. In S. Fr\u00fchwirth-Schnatter, G. Celeux, & C. P. Robert (Eds.), Handbook of mixture analysis. Chapman and Hall\/CRC."},{"key":"6491_CR22","unstructured":"GPy (since 2012). GPy: A gaussian process framework in python. http:\/\/github.com\/SheffieldML\/GPy."},{"issue":"483","key":"6491_CR23","doi-asserted-by":"publisher","first-page":"1119","DOI":"10.1198\/016214508000000689","volume":"103","author":"RB Gramacy","year":"2008","unstructured":"Gramacy, R. B., & Lee, H. K. H. (2008). Bayesian treed Gaussian process models with an application to computer modeling. Journal of the American Statistical Association, 103(483), 1119\u20131130.","journal-title":"Journal of the American Statistical Association"},{"key":"6491_CR24","unstructured":"Greenwald, J. (2016). Major next steps for fusion energy based on the spherical tokamak design."},{"key":"6491_CR25","doi-asserted-by":"crossref","unstructured":"Grigorievskiy, A., Lawrence, N., & S\u00e4rkk\u00e4, S. (2017). Parallelizable sparse inverse formulation Gaussian processes (SpInGP). In: 2017 IEEE 27th international workshop on machine learning for signal processing (MLSP) (pp. 1\u20136).","DOI":"10.1109\/MLSP.2017.8168130"},{"issue":"1","key":"6491_CR26","first-page":"100","volume":"28","author":"JA Hartigan","year":"1979","unstructured":"Hartigan, J. A., & Wong, M. A. (1979). Algorithm as 136: A k-means clustering algorithm. Journal of the Royal Statistical Society Series C (Applied Statistics), 28(1), 100\u2013108.","journal-title":"Journal of the Royal Statistical Society Series C (Applied Statistics)"},{"issue":"3","key":"6491_CR27","doi-asserted-by":"publisher","first-page":"398","DOI":"10.1007\/s13253-018-00348-w","volume":"24","author":"MJ Heaton","year":"2019","unstructured":"Heaton, M. J., Datta, A., Finley, A. O., Furrer, R., Guinness, J., Guhaniyogi, R., & Zammit-Mangion, A. (2019). A case study competition among methods for analyzing large spatial data. Journal of Agricultural, Biological and Environmental Statistics, 24(3), 398\u2013425. https:\/\/doi.org\/10.1007\/s13253-018-00348-w","journal-title":"Journal of Agricultural, Biological and Environmental Statistics"},{"key":"6491_CR28","doi-asserted-by":"crossref","unstructured":"Higdon, D. (2002). Space and space-time modeling using process convolutions. In Quantitative methods for current environmental issues (pp. 37\u201356). Springer.","DOI":"10.1007\/978-1-4471-0657-9_2"},{"key":"6491_CR29","unstructured":"Huang, W., Zhao, D., Sun, F., Liu, H., & Chang, E. (2015). Scalable Gaussian process regression using deep neural networks. In Twenty-fourth international joint conference on artificial intelligence."},{"key":"6491_CR30","unstructured":"Iwata, T., & Ghahramani, Z. (2017). Improving output uncertainty estimation and generalization in deep learning via neural network Gaussian processes. arXiv preprint arXiv:1707.05922"},{"issue":"1","key":"6491_CR31","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1162\/neco.1991.3.1.79","volume":"3","author":"RA Jacobs","year":"1991","unstructured":"Jacobs, R. A., Jordan, M. I., Nowlan, S. J., & Hinton, G. E. (1991). Adaptive mixtures of local experts. Neural Computation, 3(1), 79\u201387.","journal-title":"Neural Computation"},{"issue":"5A","key":"6491_CR32","doi-asserted-by":"publisher","first-page":"A459","DOI":"10.1088\/0741-3335\/44\/5A\/351","volume":"44","author":"G Janeschitz","year":"2002","unstructured":"Janeschitz, G., Pacher, G., Zolotukhin, O., Pereverzev, G., Pacher, H., Igitkhanov, Y., & Sugihara, M. (2002). A 1-d predictive model for energy and particle transport in h-mode. Plasma Physics and Controlled Fusion, 44(5A), A459.","journal-title":"Plasma Physics and Controlled Fusion"},{"key":"6491_CR33","unstructured":"Jankowiak, M., Pleiss, G., Gardner, J. (2020a). Deep sigma point processes. In Conference on uncertainty in artificial intelligence (pp. 789\u2013798)."},{"key":"6491_CR34","unstructured":"Jankowiak, M., Pleiss, G., Gardner, J. (2020b). Parametric gaussian process regressors. In International conference on machine learning (pp. 4702\u20134712)."},{"key":"6491_CR35","doi-asserted-by":"crossref","unstructured":"Jin, H., Song, Q., Hu, X. (2019). Auto-keras: An efficient neural architecture search system. IN Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining (pp. 1946\u20131956).","DOI":"10.1145\/3292500.3330648"},{"issue":"2","key":"6491_CR36","doi-asserted-by":"publisher","first-page":"181","DOI":"10.1162\/neco.1994.6.2.181","volume":"6","author":"MI Jordan","year":"1994","unstructured":"Jordan, M. I., & Jacobs, R. A. (1994). Hierarchical mixtures of experts and the EM algorithm. Neural Computation, 6(2), 181\u2013214.","journal-title":"Neural Computation"},{"key":"6491_CR37","unstructured":"Kingma, D.P., & Ba, J. (2014). Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980"},{"issue":"1","key":"6491_CR38","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1016\/0262-8856(84)90040-4","volume":"2","author":"J Kittler","year":"1984","unstructured":"Kittler, J., & F\u00f6glein, J. (1984). Contextual classification of multispectral pixel data. Image and Vision Computing, 2(1), 13\u201329.","journal-title":"Image and Vision Computing"},{"issue":"6","key":"6491_CR39","doi-asserted-by":"publisher","first-page":"2381","DOI":"10.1063\/1.871261","volume":"2","author":"M Kotschenreuther","year":"1995","unstructured":"Kotschenreuther, M., Dorland, W., Beer, M., & Hammett, G. (1995). Quantitative predictions of tokamak energy confinement from first-principles simulations with kinetic effects. Physics of Plasmas, 2(6), 2381\u20132389.","journal-title":"Physics of Plasmas"},{"issue":"4","key":"6491_CR40","doi-asserted-by":"publisher","first-page":"1145","DOI":"10.1162\/neco.2008.12-06-421","volume":"21","author":"K Kurihara","year":"2009","unstructured":"Kurihara, K., & Welling, M. (2009). Bayesian k-means as a \u201cmaximization-expectation\u2019\u2019 algorithm. Neural Computation, 21(4), 1145\u20131172.","journal-title":"Neural Computation"},{"issue":"4","key":"6491_CR41","doi-asserted-by":"publisher","first-page":"423","DOI":"10.1111\/j.1467-9868.2011.00777.x","volume":"73","author":"F Lindgren","year":"2011","unstructured":"Lindgren, F., Rue, H., & Lindstr\u00f6m, J. (2011). An explicit link between Gaussian fields and Gaussian Markov random fields: The stochastic partial differential equation approach. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 73(4), 423\u2013498.","journal-title":"Journal of the Royal Statistical Society: Series B (Statistical Methodology)"},{"key":"6491_CR42","volume-title":"Mixture models: Inference and applications to clustering","author":"GJ McLachlan","year":"1988","unstructured":"McLachlan, G. J., & Basford, K. E. (1988). Mixture models: Inference and applications to clustering (Vol. 38). M. Dekker."},{"key":"6491_CR43","unstructured":"Meeds, E., & Osindero, S. (2006). An alternative infinite mixture of Gaussian process experts. In Advances in neural information processing systems (pp. 883\u2013890)."},{"key":"6491_CR44","doi-asserted-by":"crossref","unstructured":"Monterrubio-G\u00f3mez, K., Roininen, L., Wade, S., Damoulas, T., & Girolami, M. (2020). Posterior inference for sparse hierarchical non-stationary models. Computational Statistics & Data Analysis.","DOI":"10.1016\/j.csda.2020.106954"},{"key":"6491_CR45","doi-asserted-by":"crossref","unstructured":"Nguyen, A., Yosinski, J., & Clune, J. (2015). Deep neural networks are easily fooled: High confidence predictions for unrecognizable images. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 427\u2013436).","DOI":"10.1109\/CVPR.2015.7298640"},{"key":"6491_CR46","unstructured":"Nguyen, T., & Bonilla, E. (2014). Fast allocation of Gaussian process experts. International conference on machine learning (pp. 145\u2013153)."},{"key":"6491_CR47","unstructured":"Nieman, D., Szabo, B., & van Zanten, H. (2021). Contraction rates for sparse variational approximations in Gaussian process regression. arXiv preprint arXiv:2109.10755"},{"key":"6491_CR48","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.cpc.2016.12.018","volume":"214","author":"JM Park","year":"2017","unstructured":"Park, J. M., Murakami, M., John, H. S., Lao, L. L., Chu, M., & Prater, R. (2017). An efficient transport solver for tokamak plasmas. Computer Physics Communications, 214, 1\u20135.","journal-title":"Computer Physics Communications"},{"key":"6491_CR49","doi-asserted-by":"crossref","unstructured":"Pope, C. A., Gosling, J. P., Barber, S., Johnson, J. S., Yamaguchi, T., Feingold, G., & Blackwell, P. G. (2019). Gaussian process modeling of heterogeneity and discontinuities using voronoi tessellations. Technometrics 1\u201320.","DOI":"10.1080\/00401706.2019.1692696"},{"key":"6491_CR50","first-page":"1939","volume":"6","author":"J Qui\u00f1onero-Candela","year":"2005","unstructured":"Qui\u00f1onero-Candela, J., & Rasmussen, C. E. (2005). A unifying view of sparse approximate Gaussian process regression. Journal of Machine Learning Research, 6, 1939\u20131959.","journal-title":"Journal of Machine Learning Research"},{"key":"6491_CR51","unstructured":"Rasmussen, C. E., & Ghahramani, Z. (2002). Infinite mixtures of Gaussian process experts. Advances in neural information processing systems (pp. 881\u2013888)."},{"key":"6491_CR52","unstructured":"Rigon, T., & Durante, D. (2017). Tractable Bayesian density regression via logit stick-breaking priors. arXiv preprint arXiv:1701.02969"},{"key":"6491_CR53","unstructured":"Rossi, S., Heinonen, M., Bonilla, E., Shen, Z., & Filippone, M. (2021). Sparse gaussian processes revisited: Bayesian approaches to inducing-variable approximations. In Banerjee, A., & Fukumizu, K. (Eds.), Proceedings of the 24th international conference on artificial intelligence and statistics (Vol. 130, pp. 1837\u20131845). PMLR."},{"key":"6491_CR54","unstructured":"Salimbeni, H., Cheng, C.-A., Boots, B., & Deisenroth, M. (2018). Orthogonally decoupled variational gaussian processes. In Advances in neural information processing systems (Vol. 31)."},{"key":"6491_CR55","unstructured":"Seeger, M. W., Williams, C. K. I., & Lawrence, N. D. (2003). Fast forward selection to speed up sparse Gaussian process regression. In Bishop, C. M. & Frey, B. J. (Eds.), Proceedings of the ninth international workshop on artificial intelligence and statistics (Vol. R4, pp. 254\u2013261). PMLR. Retrieved from https:\/\/proceedings.mlr.press\/r4\/seeger03a.html (Reissued by PMLR on 01 April 2021)."},{"issue":"1","key":"6491_CR56","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1111\/j.2517-6161.1985.tb01327.x","volume":"47","author":"BW Silverman","year":"1985","unstructured":"Silverman, B. W. (1985). Some aspects of the spline smoothing approach to non-parametric regression curve fitting. Journal of the Royal Statistical Society: Series B (Methodological), 47(1), 1\u201321.","journal-title":"Journal of the Royal Statistical Society: Series B (Methodological)"},{"key":"6491_CR57","unstructured":"Snelson, E., & Ghahramani, Z. (2006). Sparse Gaussian processes using pseudo-inputs. In Advances in neural information processing systems (pp. 1257\u20131264)."},{"issue":"5","key":"6491_CR58","doi-asserted-by":"publisher","first-page":"055909","DOI":"10.1063\/1.2436852","volume":"14","author":"G Staebler","year":"2007","unstructured":"Staebler, G., Kinsey, J., & Waltz, R. (2007). A theory-based transport model with comprehensive physics. Physics of Plasmas, 14(5), 055909.","journal-title":"Physics of Plasmas"},{"issue":"87","key":"6491_CR59","first-page":"1","volume":"20","author":"B Szab\u00f3","year":"2019","unstructured":"Szab\u00f3, B., & Van Zanten, H. (2019). An asymptotic analysis of distributed nonparametric methods. Journal of Machine Learning Research, 20(87), 1\u201330.","journal-title":"Journal of Machine Learning Research"},{"key":"6491_CR60","unstructured":"Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I., & Fergus, R. (2013). Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199"},{"key":"6491_CR61","doi-asserted-by":"crossref","unstructured":"Szymanski, L., & McCane, B. (2012). Deep, super-narrow neural network is a universal classifier. In The 2012 international joint conference on neural networks (IJCNN) (pp. 1\u20138).","DOI":"10.1109\/IJCNN.2012.6252513"},{"key":"6491_CR62","unstructured":"Titsias, M. (2009). Variational learning of inducing variables in sparse Gaussian processes. In Artificial intelligence and statistics (pp. 567\u2013574)."},{"key":"6491_CR63","unstructured":"Trapp, M., Peharz, R., Pernkopf, F., & Rasmussen, C. E. (2020). Deep structured mixtures of Gaussian processes. In International conference on artificial intelligence and statistics (pp. 2251\u20132261)."},{"issue":"11","key":"6491_CR64","doi-asserted-by":"publisher","first-page":"2719","DOI":"10.1162\/089976600300014908","volume":"12","author":"V Tresp","year":"2000","unstructured":"Tresp, V. (2000). A Bayesian committee machine. Neural Computation, 12(11), 2719\u20132741.","journal-title":"Neural Computation"},{"key":"6491_CR65","unstructured":"Tresp, V. (2001). Mixtures of Gaussian processes. In Advances in neural information processing systems (pp. 654\u2013660)."},{"issue":"7","key":"6491_CR66","doi-asserted-by":"publisher","first-page":"2482","DOI":"10.1063\/1.872228","volume":"4","author":"R Waltz","year":"1997","unstructured":"Waltz, R., Staebler, G., Dorland, W., Hammett, G., Kotschenreuther, M., & Konings, J. (1997). A gyro-Landau-fluid transport model. Physics of Plasmas, 4(7), 2482\u20132496.","journal-title":"Physics of Plasmas"},{"key":"6491_CR67","volume-title":"Gaussian processes for machine learning","author":"CK Williams","year":"2006","unstructured":"Williams, C. K., & Rasmussen, C. E. (2006). Gaussian processes for machine learning. MIT Press."},{"key":"6491_CR68","unstructured":"Wilson, A. G., Hu, Z., Salakhutdinov, R., & Xing, E. P. (2016). Deep kernel learning. In Artificial intelligence and statistics (pp. 370\u2013378)."},{"key":"6491_CR69","unstructured":"Xu, L., Jordan, M. I., & Hinton, G. E. (1995). An alternative model for mixtures of experts. In Advances in neural information processing systems (pp. 633\u2013 640)."},{"key":"6491_CR70","unstructured":"Yuan, C., & Neubauer, C. (2009). Variational mixture of Gaussian process experts. In Advances in neural information processing systems (pp. 1897\u20131904)."},{"key":"6491_CR71","first-page":"1","volume":"20","author":"MM Zhang","year":"2019","unstructured":"Zhang, M. M., & Williamson, S. A. (2019). Embarrassingly parallel inference for Gaussian processes. Journal of Machine Learning Research, 20, 1\u201326.","journal-title":"Journal of Machine Learning Research"}],"container-title":["Machine Learning"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10994-023-06491-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10994-023-06491-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10994-023-06491-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,7]],"date-time":"2024-11-07T18:04:45Z","timestamp":1731002685000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10994-023-06491-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,8]]},"references-count":71,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,3]]}},"alternative-id":["6491"],"URL":"https:\/\/doi.org\/10.1007\/s10994-023-06491-x","relation":{},"ISSN":["0885-6125","1573-0565"],"issn-type":[{"value":"0885-6125","type":"print"},{"value":"1573-0565","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,1,8]]},"assertion":[{"value":"28 October 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 October 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 November 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 January 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical standard"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"Not applicable.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}]}}