{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T21:26:43Z","timestamp":1742938003237,"version":"3.40.3"},"publisher-location":"Cham","reference-count":32,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031434235"},{"type":"electronic","value":"9783031434242"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-43424-2_6","type":"book-chapter","created":{"date-parts":[[2023,9,17]],"date-time":"2023-09-17T20:37:24Z","timestamp":1694983044000},"page":"86-101","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Robust Classification of\u00a0High-Dimensional Data Using Data-Adaptive Energy Distance"],"prefix":"10.1007","author":[{"given":"Jyotishka","family":"Ray Choudhury","sequence":"first","affiliation":[]},{"given":"Aytijhya","family":"Saha","sequence":"additional","affiliation":[]},{"given":"Sarbojit","family":"Roy","sequence":"additional","affiliation":[]},{"given":"Subhajit","family":"Dutta","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,18]]},"reference":[{"issue":"4","key":"6_CR1","doi-asserted-by":"publisher","first-page":"433","DOI":"10.1002\/wics.101","volume":"2","author":"H Abdi","year":"2010","unstructured":"Abdi, H., Williams, L.J.: Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2(4), 433\u2013459 (2010)","journal-title":"Wiley Interdiscip. Rev. Comput. Stat."},{"key":"6_CR2","doi-asserted-by":"publisher","first-page":"420","DOI":"10.1007\/3-540-44503-X_27","volume-title":"Database Theory","author":"CC Aggarwal","year":"2001","unstructured":"Aggarwal, C.C., Hinneburg, A., Keim, D.A.: On the surprising behavior of distance metrics in high dimensional space. In: Van den Bussche, J., Vianu, V. (eds.) Database Theory, pp. 420\u2013434. Springer, Heidelberg (2001). https:\/\/doi.org\/10.1007\/3-540-44503-X_27"},{"issue":"1","key":"6_CR3","doi-asserted-by":"publisher","first-page":"190","DOI":"10.1016\/S0047-259X(03)00079-4","volume":"88","author":"L Baringhaus","year":"2004","unstructured":"Baringhaus, L., Franz, C.: On a new multivariate two-sample test. J. Multivar. Anal. 88(1), 190\u2013206 (2004)","journal-title":"J. Multivar. Anal."},{"key":"6_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"217","DOI":"10.1007\/3-540-49257-7_15","volume-title":"Database Theory \u2014 ICDT\u201999","author":"K Beyer","year":"1999","unstructured":"Beyer, K., Goldstein, J., Ramakrishnan, R., Shaft, U.: When is \u201cnearest neighbor\u2019\u2019 meaningful? In: Beeri, C., Buneman, P. (eds.) ICDT 1999. LNCS, vol. 1540, pp. 217\u2013235. Springer, Heidelberg (1999). https:\/\/doi.org\/10.1007\/3-540-49257-7_15"},{"key":"6_CR5","doi-asserted-by":"crossref","DOI":"10.1093\/oso\/9780198538493.001.0001","volume-title":"Neural Networks for Pattern Recognition","author":"CM Bishop","year":"1995","unstructured":"Bishop, C.M., et al.: Neural Networks for Pattern Recognition. Oxford University Press, Oxford (1995)"},{"key":"6_CR6","doi-asserted-by":"publisher","first-page":"160","DOI":"10.1016\/j.jmva.2013.09.004","volume":"123","author":"M Biswas","year":"2014","unstructured":"Biswas, M., Ghosh, A.K.: A nonparametric two-sample test applicable to high dimensional data. J. Multivar. Anal. 123, 160\u2013171 (2014)","journal-title":"J. Multivar. Anal."},{"key":"6_CR7","volume-title":"Introduction to Strong Mixing Conditions","author":"RC Bradley","year":"2007","unstructured":"Bradley, R.C.: Introduction to Strong Mixing Conditions. Kendrick Press, Heber City (2007)"},{"issue":"3","key":"6_CR8","doi-asserted-by":"publisher","first-page":"273","DOI":"10.1007\/BF00994018","volume":"20","author":"C Cortes","year":"1995","unstructured":"Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20(3), 273\u2013297 (1995)","journal-title":"Mach. Learn."},{"issue":"1","key":"6_CR9","doi-asserted-by":"publisher","first-page":"21","DOI":"10.1109\/TIT.1967.1053964","volume":"13","author":"T Cover","year":"1967","unstructured":"Cover, T., Hart, P.: Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 13(1), 21\u201327 (1967)","journal-title":"IEEE Trans. Inf. Theory"},{"key":"6_CR10","unstructured":"Dau, H.A., et al.: Hexagon-ML: The UCR Time Series Classification Archive (2018)"},{"key":"6_CR11","doi-asserted-by":"crossref","unstructured":"Deegalla, S., Bostrom, H.: Reducing high-dimensional data by principal component analysis vs. random projection for nearest neighbor classification. In: 2006 5th International Conference on Machine Learning and Applications (ICMLA 2006), pp. 245\u2013250. IEEE (2006)","DOI":"10.1109\/ICMLA.2006.43"},{"key":"6_CR12","doi-asserted-by":"crossref","unstructured":"Fix, E., Hodges, J.L.: Discriminatory analysis. Nonparametric discrimination: consistency properties. Int. Stat. Rev.\/Revue Internationale de Statistique 57(3), 238\u2013247 (1989)","DOI":"10.2307\/1403797"},{"key":"6_CR13","doi-asserted-by":"crossref","unstructured":"Francois, D., Wertz, V., Verleysen, M.: The concentration of fractional distances. IEEE Trans. Knowl. Data Eng. 19(7), 873\u2013886 (2007)","DOI":"10.1109\/TKDE.2007.1037"},{"issue":"3","key":"6_CR14","doi-asserted-by":"publisher","first-page":"427","DOI":"10.1111\/j.1467-9868.2005.00510.x","volume":"67","author":"P Hall","year":"2005","unstructured":"Hall, P., Marron, J.S., Neeman, A.: Geometric representation of high dimension, low sample size data. J. Roy. Stat. Soc. Ser. B (Stat. Methodol.) 67(3), 427\u2013444 (2005)","journal-title":"J. Roy. Stat. Soc. Ser. B (Stat. Methodol.)"},{"key":"6_CR15","doi-asserted-by":"publisher","DOI":"10.1007\/978-0-387-84858-7","volume-title":"The Elements of Statistical Learning: Data Mining, Inference, and Prediction","author":"T Hastie","year":"2009","unstructured":"Hastie, T., Tibshirani, R., Friedman, J.H., Friedman, J.H.: The Elements of Statistical Learning: Data Mining, Inference, and Prediction, vol. 2. Springer, New York (2009). https:\/\/doi.org\/10.1007\/978-0-387-84858-7"},{"key":"6_CR16","doi-asserted-by":"crossref","unstructured":"Hinton, G.E., Salakhutdinov, R.R.: Reducing the dimensionality of data with neural networks. Science 313(5786), 504\u2013507 (2006)","DOI":"10.1126\/science.1127647"},{"issue":"4\u20135","key":"6_CR17","doi-asserted-by":"publisher","first-page":"411","DOI":"10.1016\/S0893-6080(00)00026-5","volume":"13","author":"A Hyv\u00e4rinen","year":"2000","unstructured":"Hyv\u00e4rinen, A., Oja, E.: Independent component analysis: algorithms and applications. Neural Netw. 13(4\u20135), 411\u2013430 (2000)","journal-title":"Neural Netw."},{"issue":"6","key":"6_CR18","doi-asserted-by":"publisher","first-page":"3417","DOI":"10.1214\/19-AOS1936","volume":"48","author":"I Kim","year":"2020","unstructured":"Kim, I., Balakrishnan, S., Wasserman, L.: Robust multivariate nonparametric tests via projection averaging. Ann. Stat. 48(6), 3417\u20133441 (2020)","journal-title":"Ann. Stat."},{"key":"6_CR19","unstructured":"Li, Z., Zhang, Y.: On a projective ensemble approach to two sample test for equality of distributions. In: Proceedings of the 37th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 119, pp. 6020\u20136027. PMLR (2020)"},{"issue":"10","key":"6_CR20","doi-asserted-by":"publisher","first-page":"1427","DOI":"10.1016\/j.camwa.2012.09.011","volume":"65","author":"V Pestov","year":"2013","unstructured":"Pestov, V.: Is the k-NN classifier in high dimensions affected by the curse of dimensionality? Comput. Math. Appl. 65(10), 1427\u20131437 (2013)","journal-title":"Comput. Math. Appl."},{"key":"6_CR21","unstructured":"R Core Team: R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria (2022)"},{"key":"6_CR22","unstructured":"Roy, S., Ray Choudhury, J., Dutta, S.: On some fast and robust classifiers for high dimension, low sample size data. In: Proceedings of The 25th International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 151, pp. 9943\u20139968. PMLR (2022)"},{"issue":"14","key":"6_CR23","first-page":"1","volume":"23","author":"S Roy","year":"2022","unstructured":"Roy, S., Sarkar, S., Dutta, S., Ghosh, A.K.: On generalizations of some distance based classifiers for HDLSS data. J. Mach. Learn. Res. 23(14), 1\u201341 (2022)","journal-title":"J. Mach. Learn. Res."},{"key":"6_CR24","doi-asserted-by":"crossref","unstructured":"Roy, S., Sarkar, S., Dutta, S., Ghosh, A.K.: On exact feature screening in ultrahigh-dimensional binary classification (2023)","DOI":"10.1080\/10618600.2023.2270656"},{"issue":"5","key":"6_CR25","doi-asserted-by":"publisher","first-page":"2263","DOI":"10.1214\/13-AOS1140","volume":"41","author":"D Sejdinovic","year":"2013","unstructured":"Sejdinovic, D., Sriperumbudur, B., Gretton, A., Fukumizu, K.: Equivalence of distance-based and RKHS-based statistics in hypothesis testing. Ann. Stat. 41(5), 2263\u20132291 (2013)","journal-title":"Ann. Stat."},{"key":"6_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.108828","volume":"130","author":"L Shen","year":"2022","unstructured":"Shen, L., Er, M.J., Yin, Q.: Classification for high-dimension low-sample size data. Pattern Recogn. 130, 108828 (2022)","journal-title":"Pattern Recogn."},{"issue":"1","key":"6_CR27","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TKDE.2011.181","volume":"25","author":"Q Song","year":"2011","unstructured":"Song, Q., Ni, J., Wang, G.: A fast clustering-based feature subset selection algorithm for high-dimensional data. IEEE Trans. Knowl. Data Eng. 25(1), 1\u201314 (2011)","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"6_CR28","first-page":"1249","volume":"5","author":"G Szekely","year":"2004","unstructured":"Szekely, G., Rizzo, M.: Testing for equal distributions in high dimension. InterStat 5, 1249\u20131272 (2004)","journal-title":"InterStat"},{"issue":"2","key":"6_CR29","doi-asserted-by":"publisher","first-page":"599","DOI":"10.1007\/s00180-017-0777-4","volume":"34","author":"SI Tsukada","year":"2019","unstructured":"Tsukada, S.I.: High dimensional two-sample test based on the inter-point distance. Comput. Stat. 34(2), 599\u2013615 (2019)","journal-title":"Comput. Stat."},{"key":"6_CR30","doi-asserted-by":"crossref","unstructured":"Wang, W., Huang, Y., Wang, Y., Wang, L.: Generalized autoencoder: a neural network framework for dimensionality reduction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 490\u2013497 (2014)","DOI":"10.1109\/CVPRW.2014.79"},{"key":"6_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2019.107030","volume":"97","author":"Q Yin","year":"2020","unstructured":"Yin, Q., Adeli, E., Shen, L., Shen, D.: Population-guided large margin classifier for high-dimension low-sample-size problems. Pattern Recogn. 97, 107030 (2020)","journal-title":"Pattern Recogn."},{"issue":"2","key":"6_CR32","doi-asserted-by":"publisher","first-page":"301","DOI":"10.1111\/j.1467-9868.2005.00503.x","volume":"67","author":"H Zou","year":"2005","unstructured":"Zou, H., Hastie, T.: Regularization and variable selection via the elastic net. J. Roy. Stat. Soc. Ser. B (Stat. Methodol.) 67(2), 301\u2013320 (2005)","journal-title":"J. Roy. Stat. Soc. Ser. B (Stat. Methodol.)"}],"container-title":["Lecture Notes in Computer Science","Machine Learning and Knowledge Discovery in Databases: Research Track"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-43424-2_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,21]],"date-time":"2023-12-21T23:08:21Z","timestamp":1703200101000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-43424-2_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031434235","9783031434242"],"references-count":32,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-43424-2_6","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"18 September 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECML PKDD","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Joint European Conference on Machine Learning and Knowledge Discovery in Databases","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Turin","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 September 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 September 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ecml2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/2023.ecmlpkdd.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"829","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"196","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"24% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.63","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4.5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Applied Data Science Track: 239 submissions, 58 accepted papers; Demo Track: 31 submissions, 16 accepted papers.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}