{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T10:28:00Z","timestamp":1743157680688,"version":"3.40.3"},"publisher-location":"Cham","reference-count":24,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030039271"},{"type":"electronic","value":"9783030039288"}],"license":[{"start":{"date-parts":[[2018,1,1]],"date-time":"2018-01-01T00:00:00Z","timestamp":1514764800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2018,1,1]],"date-time":"2018-01-01T00:00:00Z","timestamp":1514764800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2018]]},"DOI":"10.1007\/978-3-030-03928-8_13","type":"book-chapter","created":{"date-parts":[[2018,11,8]],"date-time":"2018-11-08T11:21:01Z","timestamp":1541676061000},"page":"159-169","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Analysis of Encoder Representations as Features Using Sparse Autoencoders in Gradient Boosting and Ensemble Tree Models"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4272-2848","authenticated-orcid":false,"given":"Luis","family":"Aguilar","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1555-0748","authenticated-orcid":false,"given":"L. Antonio","family":"Aguilar","sequence":"additional","affiliation":[]}],"member":"297","reference":[{"key":"13_CR1","doi-asserted-by":"publisher","unstructured":"Mart\u00ednez-Romo, J.C., Luna-rosas, F.J., Mora-gonz\u00e1lez, M., De Luna-ortega, C.A.: Optimal feature generation with genetic algorithms and FLDR in a restricted-vocabulary speech recognition system. In: Bio-Inspired Computational Algorithms and Their Applications, pp. 235\u2013262 (2012). https:\/\/doi.org\/10.5772\/36135","DOI":"10.5772\/36135"},{"key":"13_CR2","doi-asserted-by":"publisher","unstructured":"Cheng, W., Kasneci, G., Graepel, T., Stern, D., Herbrich, R.: Automated feature generation from structured knowledge. In: Proceedings of the 20th ACM International Conference on Information and Knowledge Management, CIKM 2011, p. 1395 (2011). https:\/\/doi.org\/10.1145\/2063576.2063779","DOI":"10.1145\/2063576.2063779"},{"key":"13_CR3","doi-asserted-by":"publisher","unstructured":"Katz, G., Shin, E.C.R., Song, D.: ExploreKit: automatic feature generation and selection. In: Proceedings - IEEE 16th International Conference on Data Mining (ICDM), pp. 979\u2013984 (2016). https:\/\/doi.org\/10.1109\/ICDM.2016.0123","DOI":"10.1109\/ICDM.2016.0123"},{"key":"13_CR4","doi-asserted-by":"publisher","first-page":"533","DOI":"10.1038\/323533a0","volume":"323","author":"DE Rumelhart","year":"1986","unstructured":"Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning representations by back-propagating errors. Nature 323, 533\u2013536 (1986). https:\/\/doi.org\/10.1038\/323533a0","journal-title":"Nature"},{"key":"13_CR5","unstructured":"Ng, A.: Sparse autoencoder. In: CS294A Lecture Notes, pp. 1\u201319 (2011). http:\/\/web.stanford.edu\/class\/cs294a\/sae\/sparseAutoencoderNotes.pdf"},{"key":"13_CR6","unstructured":"Kingma, D.P., Welling, M.: Auto-encoding variational Bayes, pp. 1\u201314 (2013). https:\/\/arxiv.org\/abs\/1312.6114"},{"key":"13_CR7","doi-asserted-by":"publisher","unstructured":"Vincent, P., Larochelle, H., Bengio, Y., Manzagol, P.-A.: Extracting and composing robust features with denoising autoencoders. In: Proceedings of 25th Annual International Conference on Machine Learning, ICML 2008, pp. 1096\u20131103 (2008). https:\/\/doi.org\/10.1145\/1390156.1390294","DOI":"10.1145\/1390156.1390294"},{"key":"13_CR8","unstructured":"Baldi, P.: Autoencoders, unsupervised learning, and deep architectures. In: Guyon, I., Dror, G., Lemaire, V., Taylor, G.W., Silver, D.L. (eds.) ICML Unsupervised and Transfer Learning, pp. 37\u201350 (2012). JMLR.org"},{"key":"13_CR9","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"208","DOI":"10.1007\/978-3-642-40994-3_14","volume-title":"Machine Learning and Knowledge Discovery in Databases","author":"W Yu","year":"2013","unstructured":"Yu, W., Zeng, G., Luo, P., Zhuang, F., He, Q., Shi, Z.: Embedding with autoencoder regularization. In: Blockeel, H., Kersting, K., Nijssen, S., \u017delezn\u00fd, F. (eds.) ECML PKDD 2013. LNCS (LNAI), vol. 8190, pp. 208\u2013223. Springer, Heidelberg (2013). https:\/\/doi.org\/10.1007\/978-3-642-40994-3_14"},{"key":"13_CR10","unstructured":"Bosch, N., Paquette, L.: Unsupervised deep autoencoders for feature extraction with educational data. In: Deep Learning with Educational Data Workshop at the 10th International Conference on Educational Data Mining (2017)"},{"key":"13_CR11","doi-asserted-by":"publisher","unstructured":"Meng, Q., Catchpoole, D., Skillicom, D., Kennedy, P.J.: Relational autoencoder for feature extraction. In: Proceedings of International Joint Conference Neural Networks, May 2017, pp. 364\u2013371 (2017). https:\/\/doi.org\/10.1109\/ijcnn.2017.7965877","DOI":"10.1109\/ijcnn.2017.7965877"},{"key":"13_CR12","unstructured":"DeVries, T., Taylor, G.W.: Dataset augmentation in feature space, pp. 1\u201312 (2017). https:\/\/arxiv.org\/abs\/1702.05538v1"},{"key":"13_CR13","doi-asserted-by":"publisher","unstructured":"Yousefi-azar, M., Varadharajan, V., Hamey, L., Tupakula, U.: Autoencoder-based feature learning for cyber security applications. In: International Joint Conference on Neural Networks 2017 (IJCNN), pp. 3854\u20133861 (2017). https:\/\/doi.org\/10.1109\/IJCNN.2017.7966342","DOI":"10.1109\/IJCNN.2017.7966342"},{"key":"13_CR14","doi-asserted-by":"publisher","unstructured":"Bengio, Y., Courville, A., Vincent, P.: Representation learning: a review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 35, 1798\u20131828 (2013). https:\/\/doi.org\/10.1109\/TPAMI.2013.50","DOI":"10.1109\/TPAMI.2013.50"},{"key":"13_CR15","unstructured":"Makhzani, A., Frey, B.: k-sparse autoencoders (2013). https:\/\/arxiv.org\/abs\/1312.5663"},{"key":"13_CR16","doi-asserted-by":"publisher","unstructured":"Ju, Y., Guo, J., Liu, S.: A deep learning method combined sparse autoencoder with SVM. In: 2015 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery, pp. 257\u2013260. IEEE (2015). https:\/\/doi.org\/10.1109\/CyberC.2015.39","DOI":"10.1109\/CyberC.2015.39"},{"key":"13_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"419","DOI":"10.1007\/978-3-319-59126-1_35","volume-title":"Image Analysis. SCIA 2017","author":"M Kampffmeyer","year":"2017","unstructured":"Kampffmeyer, M., L\u00f8kse, S., Bianchi, F.M., Jenssen, R., Livi, L.: Deep kernelized autoencoders. In: Sharma, P., Bianchi, F. (eds.) Image Analysis. SCIA 2017. LNCS, vol. 10269, pp. 419\u2013430. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-59126-1_35"},{"key":"13_CR18","first-page":"2825","volume":"12","author":"F Pedregosa","year":"2011","unstructured":"Pedregosa, F., et al.: Scikit-learn: machine learning in Python. J. Mach. Learn. Res. 12, 2825\u20132830 (2011)","journal-title":"J. Mach. Learn. Res."},{"key":"13_CR19","unstructured":"Chollet, F.: Keras. GitHub Repos (2015). https:\/\/keras.io\/"},{"key":"13_CR20","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1186\/s13040-017-0154-4","volume":"10","author":"RS Olson","year":"2017","unstructured":"Olson, R.S., La Cava, W., Orzechowski, P., Urbanowicz, R.J., Moore, J.H.: PMLB: a large benchmark suite for machine learning evaluation and comparison. BioData Min. 10, 36 (2017). https:\/\/doi.org\/10.1186\/s13040-017-0154-4","journal-title":"BioData Min."},{"key":"13_CR21","first-page":"3148","volume":"30","author":"G Ke","year":"2017","unstructured":"Ke, G., Meng, Q., Wang, T., Chen, W., Ma, W., Liu, T.-Y.: LightGBM: a highly efficient gradient boosting decision tree. Adv. Neural. Inf. Process. Syst. 30, 3148\u20133156 (2017)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"13_CR22","unstructured":"Dorogush, A.V., Ershov, V., Yandex, A.G.: CatBoost: gradient boosting with categorical features support. In: Workshop on ML System, NIPS 2017, pp. 1\u20137 (2017)"},{"key":"13_CR23","doi-asserted-by":"publisher","first-page":"349","DOI":"10.4310\/SII.2009.v2.n3.a8","volume":"2","author":"T Hastie","year":"2009","unstructured":"Hastie, T., Rosset, S., Zhu, J., Zou, H.: Multi-class AdaBoost. Stat. Interface 2, 349\u2013360 (2009). https:\/\/doi.org\/10.4310\/SII.2009.v2.n3.a8","journal-title":"Stat. Interface"},{"key":"13_CR24","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1023\/A:1010933404324","volume":"45","author":"L Breiman","year":"2001","unstructured":"Breiman, L.: Random forests. Mach. Learn. 45, 5\u201332 (2001). https:\/\/doi.org\/10.1023\/A:1010933404324","journal-title":"Mach. Learn."}],"container-title":["Lecture Notes in Computer Science","Advances in Artificial Intelligence \u2013 IBERAMIA 2018"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-03928-8_13","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,19]],"date-time":"2024-10-19T18:03:34Z","timestamp":1729361014000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-03928-8_13"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018]]},"ISBN":["9783030039271","9783030039288"],"references-count":24,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-03928-8_13","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2018]]},"assertion":[{"value":"IBERAMIA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Ibero-American Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Trujillo","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Peru","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2018","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 November 2018","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 November 2018","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iberamia2018","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.iberamia.org\/iberamia\/iberamia2018\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"92","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"41","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}}]}}