{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T07:00:59Z","timestamp":1773212459700,"version":"3.50.1"},"reference-count":32,"publisher":"Institute of Electronics, Information and Communications Engineers (IEICE)","issue":"11","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEICE Trans. Inf. &amp; Syst."],"published-print":{"date-parts":[[2015]]},"DOI":"10.1587\/transinf.2015edp7069","type":"journal-article","created":{"date-parts":[[2015,10,31]],"date-time":"2015-10-31T22:13:16Z","timestamp":1446329596000},"page":"1906-1915","source":"Crossref","is-referenced-by-count":6,"title":["Penalized AdaBoost: Improving the Generalization Error of Gentle AdaBoost through a Margin Distribution"],"prefix":"10.1587","volume":"E98.D","author":[{"given":"Shuqiong","family":"WU","sequence":"first","affiliation":[{"name":"Department of Computational Intelligence and Systems Science, Tokyo Institute of Technology"}]},{"given":"Hiroshi","family":"NAGAHASHI","sequence":"additional","affiliation":[{"name":"Imaging Science and Engineering Laboratory, Tokyo Institute of Technology"}]}],"member":"532","reference":[{"key":"1","unstructured":"[1] Y. Freund and R.E. Schapire, \u201cA short introduction to boosting,\u201d Journal of Japanese Society for Artificial Intelligence, vol.14, no.5, pp.771-780, 1999."},{"key":"2","doi-asserted-by":"crossref","unstructured":"[2] P. Viola and M. Jones, \u201cRapid object detection using a boosted cascade of simple features,\u201d pp.I-511-I-518, 2001.","DOI":"10.1109\/CVPR.2001.990517"},{"key":"3","doi-asserted-by":"crossref","unstructured":"[3] R.E. Schapire, Y. Freund, P. Bartlett, and W.S. Lee, \u201cBoosting the margin: A new explanation for the effectiveness of voting methods,\u201d The Annals of Statistics, vol.26, no.5, pp.1651-1686, 1998.","DOI":"10.1214\/aos\/1024691352"},{"key":"4","doi-asserted-by":"crossref","unstructured":"[4] R.E. Schapire and Y. Singer, \u201cImproved boosting algorithms using confidence-rated predictions,\u201d Machine Learning, vol.37, no.3, pp.297-336, 1999.","DOI":"10.1023\/A:1007614523901"},{"key":"5","doi-asserted-by":"crossref","unstructured":"[5] J. Friedman, T. Hastie, and R. Tibshirani, \u201cAdditive logistic regression: A statistical view of boosting,\u201d The Annals of Statistics, vol.28, no.2, pp.337-374, 2000.","DOI":"10.1214\/aos\/1016218222"},{"key":"6","doi-asserted-by":"crossref","unstructured":"[6] E. Grossmann, \u201cAdatree: Boosting a weak classifier into a decision tree,\u201d In IEEE Workshop on CVPR, pp.105-112, 2004.","DOI":"10.1109\/CVPR.2004.296"},{"key":"7","doi-asserted-by":"crossref","unstructured":"[7] B. Paul, G. Athithan, and M.N. Murty, \u201cSpeeding up AdaBoost classifier with random projection,\u201d in 7th International Conference on Advances in Pattern Recognition, pp.251-254, 2009.","DOI":"10.1109\/ICAPR.2009.67"},{"key":"8","doi-asserted-by":"crossref","unstructured":"[8] C. Sun, J. Hu, and K.-M. Lam, \u201cFeature subset selection for efficient adaboost training,\u201d ICME, pp.1-6, 2011.","DOI":"10.1109\/ICME.2011.6011905"},{"key":"9","unstructured":"[9] M. Seyedhosseini, A.R.C. Paiva, and T. Tasdizen, \u201cFast adaboost training using weighted novelty selection,\u201d Proc. International Conference on Neural Networks, pp.1245-1250, 2011."},{"key":"10","doi-asserted-by":"crossref","unstructured":"[10] Y. Freund, \u201cAn adaptive version of the boost by majority algorithm,\u201d Proc. Twelfth Annual Conference on Computational Learning Theory, pp.102-113, 1999.","DOI":"10.1145\/307400.307419"},{"key":"11","doi-asserted-by":"crossref","unstructured":"[11] A.J. Ferreira and M.T. Figueiredo, \u201cBoosting algorithms: A review of methods, theory, and applications,\u201d in Ensemble Machine Learning: Methods and Applications, pp.35-85, 2012.","DOI":"10.1007\/978-1-4419-9326-7_2"},{"key":"12","unstructured":"[12] C. Domingo and O. Watanabe, \u201cMadaboost: a modification of adaboost,\u201d 13th Annual Conference on Computational Learning Theory, pp.180-189, 2000."},{"key":"13","unstructured":"[13] R. Servedio, \u201cSmooth boosting and learning with malicious noise,\u201d Journal of Machine Learning Research, vol.4, pp.633-648, 2003."},{"key":"14","unstructured":"[14] J.K. Bradley and R.E. Schapire, \u201cFilterboost: Regression and classification on large datasets,\u201d NIPS, pp.185-192, 2007."},{"key":"15","doi-asserted-by":"crossref","unstructured":"[15] Y. Sun, J. Li, and W. Hager, \u201cTwo new regularized adaboost algorithms,\u201d in International Conference on Machine Learning and Applications, pp.41-48, 2004.","DOI":"10.1109\/ICMLA.2004.1383492"},{"key":"16","doi-asserted-by":"crossref","unstructured":"[16] O.T.R. Gunnar and R.M. Klaus, \u201cSoft margins for adaboost,\u201d Machine Learning, vol.42, no.3, pp.287-320, 2001.","DOI":"10.1023\/A:1007618119488"},{"key":"17","unstructured":"[17] Y. Freund, \u201cA more robust boosting algorithm,\u201d preprint available at http:\/\/arxiv.org\/abs\/0905.2138, 2009."},{"key":"18","doi-asserted-by":"crossref","unstructured":"[18] P.K. Mallapragada, R. Jin, A.K. Jain, and Y. Liu, \u201cSemiboost: boosting for semi-supervised learning,\u201d IEEE Trans. Pattern Anal. Mach. Intell., vol.31, no.11, pp.2000-2014, 2009.","DOI":"10.1109\/TPAMI.2008.235"},{"key":"19","doi-asserted-by":"crossref","unstructured":"[19] K. Chen and S. Wang, \u201cSemi-supervised learning via regularized boosting working on multiple semi-supervised assumptions,\u201d IEEE Trans. Pattern Anal. Mach. Intell., vol.33, no.1, pp.129-143, 2011.","DOI":"10.1109\/TPAMI.2010.92"},{"key":"20","doi-asserted-by":"crossref","unstructured":"[20] S.Z. Li and Z. Zhang, \u201cFloatboost learning and statistical face detection,\u201d IEEE Trans. Pattern Anal. Mach. Intell., vol.26, no.9, pp.1112-1123, 2004.","DOI":"10.1109\/TPAMI.2004.68"},{"key":"21","doi-asserted-by":"crossref","unstructured":"[21] A. Demiriz, K.P. Bennett, and J.S. Taylor, \u201cLinear programming boosting via column generation,\u201d Machine Learning, vol.46, no.1-3, pp.225-254, 2002.","DOI":"10.1023\/A:1012470815092"},{"key":"22","unstructured":"[22] H. Li and C. Shen, \u201cBoosting the minimum margin: LPBoost vs. AdaBoost,\u201d Digital Image Computing: Techniques and Applications (DICTA), pp.533-539, 2008."},{"key":"23","unstructured":"[23] A. Vezhnevets and V. Vezhnevets, \u201cModest adaboost-teaching adaboost to generalize better,\u201d Graphicon, vol.12, no.5, pp.987-997, 2005."},{"key":"24","unstructured":"[24] Y.Z.J. Thongkam, G. Xu, and F. Huang, \u201cBreast cancer survivability via adaboost algorithms,\u201d Australian Workshop on Health Data and Knowledge Management, pp.55-64, 2008."},{"key":"25","doi-asserted-by":"crossref","unstructured":"[25] Y. Lu, Q. Tian, and T. Huang, \u201cInteractive boosting for image classification,\u201d Proceedings of the 7th International Conference on Multiple Classifier Systems, pp.180-189, 2007.","DOI":"10.1007\/978-3-540-72523-7_19"},{"key":"26","doi-asserted-by":"crossref","unstructured":"[26] J.J. Rodr\u00edguez and J. Maudes, \u201cBoosting recombined weak classifiers,\u201d Pattern Recognition Letters, vol.29, no.8, pp.1049-1059, 2008.","DOI":"10.1016\/j.patrec.2007.06.019"},{"key":"27","unstructured":"[27] M. Warmuth, K. Glocer, and G. Ratsch, \u201cBoosting algorithms for maximizing the soft margin,\u201d Adavances in NIPS, pp.1-8, 2007."},{"key":"28","doi-asserted-by":"crossref","unstructured":"[28] M.K. Warmuth, K.A. Glocer, and S.V.N. Vishwanathan, \u201cEntropy regularized lpboost,\u201d Proceedings of the 19th International Conference on Algorithmic Learning Theory, pp.256-271, 2008.","DOI":"10.1007\/978-3-540-87987-9_23"},{"key":"29","unstructured":"[29] D. Hegazy and J. Denzler, \u201cPerformance comparison and evaluation of adaboost and softboost algorithms on generic object recognition,\u201d World Academy of Science, Engineering and Technology, pp.70-74, 2008."},{"key":"30","doi-asserted-by":"crossref","unstructured":"[30] S. Wu and H. Nagahashi, \u201cA new method for solving overfitting problem of gentle adaboost,\u201d SPIE 9069, Fifth International Conference on Graphic and Image Processing, pp.1-6, 2014.","DOI":"10.1117\/12.2050093"},{"key":"31","unstructured":"[31] K. Bache and M. Lichman, \u201cUci machine learning repository,\u201d http:\/\/archive.ics.uci.edu\/ml\/, pp.Irvine, CA: University of California, School of Information and Computer Science, 2013."},{"key":"32","unstructured":"[32] A. Vezhnevets and V. Vezhnevets, \u201cGml adaboost matlab toolbox manual,\u201d http:\/\/graphics.cs.msu.ru\/ru\/science\/research\/machinelearning\/adaboosttoolbox, pp.1-12."}],"container-title":["IEICE Transactions on Information and Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transinf\/E98.D\/11\/E98.D_2015EDP7069\/_pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,31]],"date-time":"2025-05-31T07:49:56Z","timestamp":1748677796000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transinf\/E98.D\/11\/E98.D_2015EDP7069\/_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2015]]},"references-count":32,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2015]]}},"URL":"https:\/\/doi.org\/10.1587\/transinf.2015edp7069","relation":{},"ISSN":["0916-8532","1745-1361"],"issn-type":[{"value":"0916-8532","type":"print"},{"value":"1745-1361","type":"electronic"}],"subject":[],"published":{"date-parts":[[2015]]}}}