{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T03:36:07Z","timestamp":1760240167040,"version":"build-2065373602"},"reference-count":22,"publisher":"MDPI AG","issue":"3","license":[{"start":{"date-parts":[[2019,3,14]],"date-time":"2019-03-14T00:00:00Z","timestamp":1552521600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001691","name":"Japan Society for the Promotion of Science","doi-asserted-by":"publisher","award":["17K12650","16H02789"],"award-info":[{"award-number":["17K12650","16H02789"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Entropy"],"abstract":"<jats:p>Statistical inference is considered for variables of interest, called primary variables, when auxiliary variables are observed along with the primary variables. We consider the setting of incomplete data analysis, where some primary variables are not observed. Utilizing a parametric model of joint distribution of primary and auxiliary variables, it is possible to improve the estimation of parametric model for the primary variables when the auxiliary variables are closely related to the primary variables. However, the estimation accuracy reduces when the auxiliary variables are irrelevant to the primary variables. For selecting useful auxiliary variables, we formulate the problem as model selection, and propose an information criterion for predicting primary variables by leveraging auxiliary variables. The proposed information criterion is an asymptotically unbiased estimator of the Kullback\u2013Leibler divergence for complete data of primary variables under some reasonable conditions. We also clarify an asymptotic equivalence between the proposed information criterion and a variant of leave-one-out cross validation. Performance of our method is demonstrated via a simulation study and a real data example.<\/jats:p>","DOI":"10.3390\/e21030281","type":"journal-article","created":{"date-parts":[[2019,3,15]],"date-time":"2019-03-15T04:12:09Z","timestamp":1552623129000},"page":"281","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["An Information Criterion for Auxiliary Variable Selection in Incomplete Data Analysis"],"prefix":"10.3390","volume":"21","author":[{"given":"Shinpei","family":"Imori","sequence":"first","affiliation":[{"name":"Graduate School of Science, Hiroshima University, Hiroshima 739-8526, Japan"},{"name":"RIKEN Center for Advanced Intelligence Project, Tokyo 103-0027, Japan"}]},{"given":"Hidetoshi","family":"Shimodaira","sequence":"additional","affiliation":[{"name":"Graduate School of Informatics, Kyoto University, Kyoto 606-8501, Japan"},{"name":"RIKEN Center for Advanced Intelligence Project, Tokyo 103-0027, Japan"}]}],"member":"1968","published-online":{"date-parts":[[2019,3,14]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1111\/1467-9868.00054","article-title":"Predicting multivariate responses in multiple linear regression","volume":"59","author":"Breiman","year":"1997","journal-title":"J. R. Stat. Soc. Ser. B Stat. Methodol."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"25","DOI":"10.1023\/A:1008815025242","article-title":"Coaching variables for regression and classification","volume":"8","author":"Tibshirani","year":"1998","journal-title":"Stat. Comput."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"41","DOI":"10.1023\/A:1007379606734","article-title":"Multitask learning","volume":"28","author":"Caruana","year":"1997","journal-title":"Mach. Learn."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"34","DOI":"10.1002\/sam.11256","article-title":"Improving inference of Gaussian mixtures using auxiliary variables","volume":"8","author":"Mercatanti","year":"2015","journal-title":"Stat. Anal. Data Min."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"79","DOI":"10.1214\/aoms\/1177729694","article-title":"On information and sufficiency","volume":"22","author":"Kullback","year":"1951","journal-title":"Ann. Math. Stat."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"716","DOI":"10.1109\/TAC.1974.1100705","article-title":"A new look at the statistical model identification","volume":"19","author":"Akaike","year":"1974","journal-title":"IEEE Trans. Autom. Control"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"45","DOI":"10.1093\/biomet\/68.1.45","article-title":"An optimal selection of regression variables","volume":"68","author":"Shibata","year":"1981","journal-title":"Biometrika"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"415","DOI":"10.1007\/BF02480998","article-title":"Asymptotic mean efficiency of a selection of regression variables","volume":"35","author":"Shibata","year":"1983","journal-title":"Ann. Inst. Stat. Math."},{"key":"ref_9","first-page":"12","article-title":"Distribution of information statistics and criteria for adequacy of models","volume":"153","author":"Takeuchi","year":"1976","journal-title":"Math. Sci."},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Cheeseman, P., and Oldford, R.W. (1994). A new criterion for selecting models from partially observed data. Selecting Models from Data, Springer.","DOI":"10.1007\/978-1-4612-2660-4"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"45","DOI":"10.1016\/S0378-3758(97)00115-8","article-title":"An Akaike information criterion for model selection in the presence of incomplete data","volume":"67","author":"Cavanaugh","year":"1998","journal-title":"J. Stat. Plan. Inference"},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1111\/j.2517-6161.1977.tb01600.x","article-title":"Maximum likelihood from incomplete data via the EM algorithm","volume":"39","author":"Dempster","year":"1977","journal-title":"J. R. Stat. Soc. Ser. B Methodol."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"421","DOI":"10.1007\/s10463-016-0592-7","article-title":"An information criterion for model selection with missing data via complete-data divergence","volume":"70","author":"Shimodaira","year":"2018","journal-title":"Ann. Inst. Stat. Math."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"44","DOI":"10.1111\/j.2517-6161.1977.tb01603.x","article-title":"An asymptotic equivalence of choice of model by cross-validation and Akaike\u2019s criterion","volume":"39","author":"Stone","year":"1977","journal-title":"J. R. Stat. Soc. Ser. B Methodol."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"361","DOI":"10.1111\/1467-9876.00240","article-title":"Using auxiliary data for parameter estimation with non-ignorably missing outcomes","volume":"50","author":"Ibrahim","year":"2001","journal-title":"J. R. Stat. Soc. Ser. C Appl. Stat."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"1","DOI":"10.2307\/1912526","article-title":"Maximum likelihood estimation of misspecified models","volume":"50","author":"White","year":"1982","journal-title":"Econometrica"},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"227","DOI":"10.1016\/S0378-3758(00)00115-4","article-title":"Improving predictive inference under covariate shift by weighting the log-likelihood function","volume":"90","author":"Shimodaira","year":"2000","journal-title":"J. Stat. Plan. Inference"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"111","DOI":"10.1111\/j.2517-6161.1974.tb00994.x","article-title":"Cross-validatory choice and assessment of statistical predictions","volume":"36","author":"Stone","year":"1974","journal-title":"J. R. Stat. Soc. Ser. B Methodol."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"40","DOI":"10.1214\/09-SS054","article-title":"A survey of cross-validation procedures for model selection","volume":"4","author":"Arlot","year":"2010","journal-title":"Stat. Surv."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"1965","DOI":"10.1016\/j.jmva.2005.10.009","article-title":"Bias correction of cross-validation criterion based on Kullback\u2013Leibler information under a general condition","volume":"97","author":"Yanagihara","year":"2006","journal-title":"J. Multivar. Anal."},{"key":"ref_21","unstructured":"Dua, D., and Karra Taniskidou, E. (2017). UCI Machine Learning Repository, University of California, School of Information and Computer Science."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"1345","DOI":"10.1109\/TKDE.2009.191","article-title":"A survey on transfer learning","volume":"22","author":"Pan","year":"2010","journal-title":"IEEE Trans. Knowl. Data Eng."}],"container-title":["Entropy"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1099-4300\/21\/3\/281\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T12:38:52Z","timestamp":1760186332000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1099-4300\/21\/3\/281"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,3,14]]},"references-count":22,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2019,3]]}},"alternative-id":["e21030281"],"URL":"https:\/\/doi.org\/10.3390\/e21030281","relation":{},"ISSN":["1099-4300"],"issn-type":[{"type":"electronic","value":"1099-4300"}],"subject":[],"published":{"date-parts":[[2019,3,14]]}}}