{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,5]],"date-time":"2024-09-05T12:13:40Z","timestamp":1725538420947},"publisher-location":"Berlin, Heidelberg","reference-count":16,"publisher":"Springer Berlin Heidelberg","isbn-type":[{"type":"print","value":"9783642049200"},{"type":"electronic","value":"9783642049217"}],"license":[{"start":{"date-parts":[[2009,1,1]],"date-time":"2009-01-01T00:00:00Z","timestamp":1230768000000},"content-version":"unspecified","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2009]]},"DOI":"10.1007\/978-3-642-04921-7_36","type":"book-chapter","created":{"date-parts":[[2009,9,22]],"date-time":"2009-09-22T07:46:25Z","timestamp":1253605585000},"page":"350-359","source":"Crossref","is-referenced-by-count":3,"title":["Efficient Hold-Out for Subset of Regressors"],"prefix":"10.1007","author":[{"given":"Tapio","family":"Pahikkala","sequence":"first","affiliation":[]},{"given":"Hanna","family":"Suominen","sequence":"additional","affiliation":[]},{"given":"Jorma","family":"Boberg","sequence":"additional","affiliation":[]},{"given":"Tapio","family":"Salakoski","sequence":"additional","affiliation":[]}],"member":"297","reference":[{"key":"36_CR1","unstructured":"Rifkin, R.: Everything Old Is New Again: A Fresh Look at Historical Approaches in Machine Learning. Ph.D thesis, Massachusetts Institute of Technology (2002)"},{"key":"36_CR2","first-page":"515","volume-title":"Proceedings of the Fifteenth International Conference on Machine Learning","author":"C. Saunders","year":"1998","unstructured":"Saunders, C., Gammerman, A., Vovk, V.: Ridge regression learning algorithm in dual variables. In: Proceedings of the Fifteenth International Conference on Machine Learning, pp. 515\u2013521. Morgan Kaufmann Publishers Inc., San Francisco (1998)"},{"issue":"3","key":"36_CR3","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1023\/A:1018628609742","volume":"9","author":"J.A.K. Suykens","year":"1999","unstructured":"Suykens, J.A.K., Vandewalle, J.: Least squares support vector machine classifiers. Neural Processing Letters\u00a09(3), 293\u2013300 (1999)","journal-title":"Neural Processing Letters"},{"key":"36_CR4","doi-asserted-by":"crossref","DOI":"10.7551\/mitpress\/3206.001.0001","volume-title":"Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning)","author":"C.E. Rasmussen","year":"2005","unstructured":"Rasmussen, C.E., Williams, C.K.I.: Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning). The MIT Press, Cambridge (2005)"},{"issue":"2","key":"36_CR5","doi-asserted-by":"publisher","first-page":"133","DOI":"10.1007\/s10994-008-5082-6","volume":"74","author":"T. Pahikkala","year":"2009","unstructured":"Pahikkala, T., Pyysalo, S., Boberg, J., J\u00e4rvinen, J., Salakoski, T.: Matrix representations, linear transformations, and kernels for disambiguation in natural language. Machine Learning\u00a074(2), 133\u2013158 (2009)","journal-title":"Machine Learning"},{"key":"36_CR6","unstructured":"Pahikkala, T., Tsivtsivadze, E., Airola, A., Boberg, J., Salakoski, T.: Learning to rank with pairwise regularized least-squares. In: Joachims, T., Li, H., Liu, T.Y., Zhai, C. (eds.) SIGIR 2007 Workshop on Learning to Rank for Information Retrieval, pp. 27\u201333 (2007)"},{"issue":"1","key":"36_CR7","doi-asserted-by":"publisher","first-page":"129","DOI":"10.1007\/s10994-008-5097-z","volume":"75","author":"T. Pahikkala","year":"2009","unstructured":"Pahikkala, T., Tsivtsivadze, E., Airola, A., J\u00e4rvinen, J., Boberg, J.: An efficient algorithm for learning to rank from preference graphs. Machine Learning\u00a075(1), 129\u2013165 (2009)","journal-title":"Machine Learning"},{"key":"36_CR8","first-page":"911","volume-title":"Proceedings of the Seventeenth International Conference on Machine Learning","author":"A.J. Smola","year":"2000","unstructured":"Smola, A.J., Sch\u00f6lkopf, B.: Sparse greedy matrix approximation for machine learning. In: Langley, P. (ed.) Proceedings of the Seventeenth International Conference on Machine Learning, pp. 911\u2013918. Morgan Kaufmann, San Francisco (2000)"},{"issue":"10","key":"36_CR9","doi-asserted-by":"publisher","first-page":"1467","DOI":"10.1016\/j.neunet.2004.07.002","volume":"17","author":"G.C. Cawley","year":"2004","unstructured":"Cawley, G.C., Talbot, N.L.C.: Fast exact leave-one-out cross-validation of sparse least-squares support vector machines. Neural Networks\u00a017(10), 1467\u20131475 (2004)","journal-title":"Neural Networks"},{"key":"36_CR10","unstructured":"Pahikkala, T., Boberg, J., Salakoski, T.: Fast n-fold cross-validation for regularized least-squares. In: Honkela, T., Raiko, T., Kortela, J., Valpola, H. (eds.) Proceedings of the Ninth Scandinavian Conference on Artificial Intelligence (SCAI 2006), Espoo, Finland, Otamedia, pp. 83\u201390 (2006)"},{"issue":"8","key":"36_CR11","doi-asserted-by":"publisher","first-page":"2154","DOI":"10.1016\/j.patcog.2006.12.015","volume":"40","author":"S. An","year":"2007","unstructured":"An, S., Liu, W., Venkatesh, S.: Fast cross-validation algorithms for least squares support vector machine and kernel ridge regression. Pattern Recognition\u00a040(8), 2154\u20132162 (2007)","journal-title":"Pattern Recognition"},{"key":"36_CR12","unstructured":"Rifkin, R., Lippert, R.: Notes on regularized least squares. Technical Report MIT-CSAIL-TR-2007-025, Massachusetts Institute of Technology (2007)"},{"key":"36_CR13","unstructured":"Suominen, H., Pahikkala, T., Salakoski, T.: Critical points in assessing learning performance via cross-validation. In: Honkela, T., P\u00f6ll\u00e4, M., Paukkeri, M.S., Simula, O. (eds.) Proceedings of the 2nd International and Interdisciplinary Conference on Adaptive Knowledge Representation and Reasoning (AKRR 2008), Helsinki University of Technology, pp. 9\u201322 (2008)"},{"key":"36_CR14","first-page":"1939","volume":"6","author":"J. Qui\u00f1onero-Candela","year":"2005","unstructured":"Qui\u00f1onero-Candela, J., Rasmussen, C.E.: A unifying view of sparse approximate gaussian process regression. Journal of Machine Learning Research\u00a06, 1939\u20131959 (2005)","journal-title":"Journal of Machine Learning Research"},{"key":"36_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"416","DOI":"10.1007\/3-540-44581-1_27","volume-title":"Computational Learning Theory","author":"B. Sch\u00f6lkopf","year":"2001","unstructured":"Sch\u00f6lkopf, B., Herbrich, R., Smola, A.J.: A generalized representer theorem. In: Helmbold, D., Williamson, R. (eds.) COLT 2001 and EuroCOLT 2001. LNCS, vol.\u00a02111, pp. 416\u2013426. Springer, Heidelberg (2001)"},{"key":"36_CR16","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511810817","volume-title":"Matrix Analysis","author":"R. Horn","year":"1985","unstructured":"Horn, R., Johnson, C.R.: Matrix Analysis. Cambridge University Press, Cambridge (1985)"}],"container-title":["Lecture Notes in Computer Science","Adaptive and Natural Computing Algorithms"],"original-title":[],"link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-642-04921-7_36","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2019,3,9]],"date-time":"2019-03-09T18:54:16Z","timestamp":1552157656000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-642-04921-7_36"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2009]]},"ISBN":["9783642049200","9783642049217"],"references-count":16,"URL":"https:\/\/doi.org\/10.1007\/978-3-642-04921-7_36","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2009]]}}}