{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,6]],"date-time":"2025-12-06T05:04:52Z","timestamp":1764997492231,"version":"3.40.3"},"publisher-location":"Cham","reference-count":123,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031198052"},{"type":"electronic","value":"9783031198069"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19806-9_41","type":"book-chapter","created":{"date-parts":[[2022,10,19]],"date-time":"2022-10-19T23:11:54Z","timestamp":1666221114000},"page":"717-738","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Learning with\u00a0Noisy Labels by\u00a0Efficient Transition Matrix Estimation to\u00a0Combat Label Miscorrection"],"prefix":"10.1007","author":[{"given":"Seong Min","family":"Kye","sequence":"first","affiliation":[]},{"given":"Kwanghee","family":"Choi","sequence":"additional","affiliation":[]},{"given":"Joonyoung","family":"Yi","sequence":"additional","affiliation":[]},{"given":"Buru","family":"Chang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,20]]},"reference":[{"unstructured":"Arpit, D., et al.: A closer look at memorization in deep networks. In: International Conference on Machine Learning, pp. 233\u2013242. PMLR (2017)","key":"41_CR1"},{"unstructured":"Azadi, S., Feng, J., Jegelka, S., Darrell, T.: Auxiliary image regularization for deep CNNs with noisy labels. arXiv preprint arXiv:1511.07069 (2015)","key":"41_CR2"},{"unstructured":"Bahri, D., Jiang, H., Gupta, M.: Deep k-NN for noisy labels. In: International Conference on Machine Learning, pp. 540\u2013550. PMLR (2020)","key":"41_CR3"},{"unstructured":"Bartlett, P., Foster, D.J., Telgarsky, M.: Spectrally-normalized margin bounds for neural networks. arXiv preprint arXiv:1706.08498 (2017)","key":"41_CR4"},{"issue":"473","key":"41_CR5","doi-asserted-by":"publisher","first-page":"138","DOI":"10.1198\/016214505000000907","volume":"101","author":"PL Bartlett","year":"2006","unstructured":"Bartlett, P.L., Jordan, M.I., McAuliffe, J.D.: Convexity, classification, and risk bounds. J. Am. Stat. Assoc. 101(473), 138\u2013156 (2006)","journal-title":"J. Am. Stat. Assoc."},{"unstructured":"Bartlett, P.L., Mendelson, S.: Rademacher and Gaussian complexities: risk bounds and structural results. J. Mach. Learn. Res. 3(Nov), 463\u2013482 (2002)","key":"41_CR6"},{"doi-asserted-by":"crossref","unstructured":"Bekker, A.J., Goldberger, J.: Training deep neural-networks based on unreliable labels. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2682\u20132686. IEEE (2016)","key":"41_CR7","DOI":"10.1109\/ICASSP.2016.7472164"},{"unstructured":"Berthon, A., Han, B., Niu, G., Liu, T., Sugiyama, M.: Confidence scores make instance-dependent label-noise learning possible. arXiv preprint arXiv:2001.03772 (2020)","key":"41_CR8"},{"doi-asserted-by":"crossref","unstructured":"Boucheron, S., Lugosi, G., Massart, P.: Concentration Inequalities: A Nonasymptotic Theory of Independence. Oxford University Press, Oxford (2013)","key":"41_CR9","DOI":"10.1093\/acprof:oso\/9780199535255.001.0001"},{"unstructured":"Cao, K., Chen, Y., Lu, J., Arechiga, N., Gaidon, A., Ma, T.: Heteroskedastic and imbalanced deep learning with adaptive regularization. arXiv preprint arXiv:2006.15766 (2020)","key":"41_CR10"},{"unstructured":"Chang, H.S., Learned-Miller, E., McCallum, A.: Active bias: training more accurate neural networks by emphasizing high variance samples. arXiv preprint arXiv:1704.07433 (2017)","key":"41_CR11"},{"doi-asserted-by":"crossref","unstructured":"Charikar, M., Steinhardt, J., Valiant, G.: Learning from untrusted data. In: Proceedings of the 49th Annual ACM SIGACT Symposium on Theory of Computing, pp. 47\u201360 (2017)","key":"41_CR12","DOI":"10.1145\/3055399.3055491"},{"key":"41_CR13","doi-asserted-by":"publisher","first-page":"321","DOI":"10.1613\/jair.953","volume":"16","author":"NV Chawla","year":"2002","unstructured":"Chawla, N.V., Bowyer, K.W., Hall, L.O., Kegelmeyer, W.P.: Smote: synthetic minority over-sampling technique. J. Artif. Intell. Res. 16, 321\u2013357 (2002)","journal-title":"J. Artif. Intell. Res."},{"unstructured":"Chen, P., Liao, B.B., Chen, G., Zhang, S.: Understanding and utilizing deep neural networks trained with noisy labels. In: International Conference on Machine Learning, pp. 1062\u20131070. PMLR (2019)","key":"41_CR14"},{"doi-asserted-by":"crossref","unstructured":"Chen, P., Ye, J., Chen, G., Zhao, J., Heng, P.A.: Robustness of accuracy metric and its inspirations in learning with noisy labels. arXiv preprint arXiv:2012.04193 (2020)","key":"41_CR15","DOI":"10.1609\/aaai.v35i13.17364"},{"unstructured":"Cheng, H., Zhu, Z., Li, X., Gong, Y., Sun, X., Liu, Y.: Learning with instance-dependent label noise: a sample sieve approach. In: ICLR (2021)","key":"41_CR16"},{"unstructured":"Cheng, J., Liu, T., Ramamohanarao, K., Tao, D.: Learning with bounded instance and label-dependent label noise. In: International Conference on Machine Learning, pp. 1789\u20131799. PMLR (2020)","key":"41_CR17"},{"doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255. IEEE (2009)","key":"41_CR18","DOI":"10.1109\/CVPR.2009.5206848"},{"unstructured":"Drory, A., Avidan, S., Giryes, R.: How do neural networks overcome label noise. arXiv Preprint (2018)","key":"41_CR19"},{"unstructured":"Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: International Conference on Machine Learning, pp. 1126\u20131135. PMLR (2017)","key":"41_CR20"},{"issue":"4","key":"41_CR21","doi-asserted-by":"publisher","first-page":"681","DOI":"10.1007\/s11023-020-09548-1","volume":"30","author":"L Floridi","year":"2020","unstructured":"Floridi, L., Chiriatti, M.: GPT-3: its nature, scope, limits, and consequences. Mind. Mach. 30(4), 681\u2013694 (2020)","journal-title":"Mind. Mach."},{"doi-asserted-by":"crossref","unstructured":"Ghosh, A., Kumar, H., Sastry, P.: Robust loss functions under label noise for deep neural networks. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31 (2017)","key":"41_CR22","DOI":"10.1609\/aaai.v31i1.10894"},{"doi-asserted-by":"crossref","unstructured":"Ghosh, A., Lan, A.: Do we really need gold samples for sample weighting under label noise? In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 3922\u20133931 (2021)","key":"41_CR23","DOI":"10.1109\/WACV48630.2021.00397"},{"doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580\u2013587 (2014)","key":"41_CR24","DOI":"10.1109\/CVPR.2014.81"},{"unstructured":"Goldberger, J., Ben-Reuven, E.: Training deep neural-networks using a noise adaptation layer (2016)","key":"41_CR25"},{"unstructured":"Golowich, N., Rakhlin, A., Shamir, O.: Size-independent sample complexity of neural networks. In: Conference On Learning Theory, pp. 297\u2013299. PMLR (2018)","key":"41_CR26"},{"unstructured":"Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press, Cambridge (2016). http:\/\/www.deeplearningbook.org","key":"41_CR27"},{"doi-asserted-by":"crossref","unstructured":"Guan, M., Gulshan, V., Dai, A., Hinton, G.: Who said what: modeling individual labelers improves classification. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32 (2018)","key":"41_CR28","DOI":"10.1609\/aaai.v32i1.11756"},{"unstructured":"Guo, C., Pleiss, G., Sun, Y., Weinberger, K.Q.: On calibration of modern neural networks. In: International Conference on Machine Learning, pp. 1321\u20131330. PMLR (2017)","key":"41_CR29"},{"unstructured":"Guo, J., Gong, M., Liu, T., Zhang, K., Tao, D.: LTF: a label transformation framework for correcting label shift. In: International Conference on Machine Learning, pp. 3843\u20133853. PMLR (2020)","key":"41_CR30"},{"unstructured":"Han, B., et al.: SIGUA: forgetting may make learning with noisy labels more robust. In: International Conference on Machine Learning, pp. 4006\u20134016. PMLR (2020)","key":"41_CR31"},{"unstructured":"Han, B., et al.: Masking: a new perspective of noisy supervision. arXiv preprint arXiv:1805.08193 (2018)","key":"41_CR32"},{"unstructured":"Han, B., et al.: Co-teaching: robust training of deep neural networks with extremely noisy labels. arXiv preprint arXiv:1804.06872 (2018)","key":"41_CR33"},{"doi-asserted-by":"crossref","unstructured":"Han, J., Luo, P., Wang, X.: Deep self-learning from noisy labels. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5138\u20135147 (2019)","key":"41_CR34","DOI":"10.1109\/ICCV.2019.00524"},{"unstructured":"Han, K., Wang, Y., Xu, Y., Xu, C., Wu, E., Xu, C.: Training binary neural networks through learning with noisy supervision. In: International Conference on Machine Learning, pp. 4017\u20134026. PMLR (2020)","key":"41_CR35"},{"unstructured":"Harutyunyan, H., Reing, K., Ver Steeg, G., Galstyan, A.: Improving generalization by controlling label-noise information in neural network weights. In: International Conference on Machine Learning, pp. 4071\u20134081. PMLR (2020)","key":"41_CR36"},{"doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","key":"41_CR37","DOI":"10.1109\/CVPR.2016.90"},{"unstructured":"Hendrycks, D., Lee, K., Mazeika, M.: Using pre-training can improve model robustness and uncertainty. In: International Conference on Machine Learning, pp. 2712\u20132721. PMLR (2019)","key":"41_CR38"},{"unstructured":"Hendrycks, D., Mazeika, M., Kadavath, S., Song, D.: Using self-supervised learning can improve model robustness and uncertainty. arXiv preprint arXiv:1906.12340 (2019)","key":"41_CR39"},{"unstructured":"Hendrycks, D., Mazeika, M., Wilson, D., Gimpel, K.: Using trusted data to train deep networks on labels corrupted by severe noise. Adv. Neural Inf. Process. Syst. 31 (2018)","key":"41_CR40"},{"doi-asserted-by":"crossref","unstructured":"Hong, Y., Han, S., Choi, K., Seo, S., Kim, B., Chang, B.: Disentangling label distribution for long-tailed visual recognition. arXiv preprint arXiv:2012.00321 (2020)","key":"41_CR41","DOI":"10.1109\/CVPR46437.2021.00656"},{"unstructured":"Hu, W., Li, Z., Yu, D.: Simple and effective regularization methods for training on noisily labeled data with generalization guarantee. arXiv preprint arXiv:1905.11368 (2019)","key":"41_CR42"},{"unstructured":"Huang, L., Zhang, C., Zhang, H.: Self-adaptive training: beyond empirical risk minimization. Adv. Neural Inf. Process. Syst. 33 (2020)","key":"41_CR43"},{"unstructured":"Jiang, L., Zhou, Z., Leung, T., Lif, L.J., Fei-Fei, L.: MentorNet: learning data-driven curriculum for very deep neural networks on corrupted labels. In: International Conference on Machine Learning, pp. 2304\u20132313. PMLR (2018)","key":"41_CR44"},{"unstructured":"Jiang, Z., et al.: An information fusion approach to learning with instance-dependent label noise. In: International Conference on Learning Representations (2022). https:\/\/openreview.net\/forum?id=ecH2FKaARUp","key":"41_CR45"},{"doi-asserted-by":"crossref","unstructured":"Jindal, I., Nokleby, M., Chen, X.: Learning deep networks from noisy labels with dropout regularization. In: 2016 IEEE 16th International Conference on Data Mining (ICDM), pp. 967\u2013972. IEEE (2016)","key":"41_CR46","DOI":"10.1109\/ICDM.2016.0121"},{"unstructured":"Kim, T., Ko, J., Choi, J., Yun, S.Y., et al.: Fine samples for learning with noisy labels. Adv. Neural Inf. Process. Syst. 34 (2021)","key":"41_CR47"},{"unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)","key":"41_CR48"},{"unstructured":"Krizhevsky, A., Hinton, G., et al.: Learning multiple layers of features from tiny images (2009)","key":"41_CR49"},{"doi-asserted-by":"crossref","unstructured":"Lee, K.H., He, X., Zhang, L., Yang, L.: CleanNet: transfer learning for scalable image classifier training with label noise. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5447\u20135456 (2018)","key":"41_CR50","DOI":"10.1109\/CVPR.2018.00571"},{"unstructured":"Li, D., Chen, C., Liu, W., Lu, T., Gu, N., Chu, S.M.: Mixture-rank matrix approximation for collaborative filtering. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 477\u2013485 (2017)","key":"41_CR51"},{"unstructured":"Li, D., Chen, C., Lv, Q., Yan, J., Shang, L., Chu, S.: Low-rank matrix approximation with stability. In: International Conference on Machine Learning, pp. 295\u2013303. PMLR (2016)","key":"41_CR52"},{"unstructured":"Li, J., Socher, R., Hoi, S.C.: DivideMix: learning with noisy labels as semi-supervised learning. arXiv preprint arXiv:2002.07394 (2020)","key":"41_CR53"},{"doi-asserted-by":"crossref","unstructured":"Li, J., Wong, Y., Zhao, Q., Kankanhalli, M.S.: Learning to learn from noisy labeled data. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5051\u20135059 (2019)","key":"41_CR54","DOI":"10.1109\/CVPR.2019.00519"},{"unstructured":"Li, M., Soltanolkotabi, M., Oymak, S.: Gradient descent with early stopping is provably robust to label noise for overparameterized neural networks. In: International Conference on Artificial Intelligence and Statistics, pp. 4313\u20134324. PMLR (2020)","key":"41_CR55"},{"unstructured":"Li, X., Liu, T., Han, B., Niu, G., Sugiyama, M.: Provably end-to-end label-noise learning without anchor points. In: International Conference on Machine Learning. PMLR (2021)","key":"41_CR56"},{"doi-asserted-by":"crossref","unstructured":"Lienen, J., H\u00fcllermeier, E.: From label smoothing to label relaxation. In: Proceedings of the 35th AAAI Conference on Artificial Intelligence, AAAI, Online, 2\u20139 February 2021. AAAI Press (2021)","key":"41_CR57","DOI":"10.1609\/aaai.v35i10.17041"},{"doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Doll\u00e1r, P.: Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2980\u20132988 (2017)","key":"41_CR58","DOI":"10.1109\/ICCV.2017.324"},{"unstructured":"Liu, S., Niles-Weed, J., Razavian, N., Fernandez-Granda, C.: Early-learning regularization prevents memorization of noisy labels. arXiv preprint arXiv:2007.00151 (2020)","key":"41_CR59"},{"issue":"3","key":"41_CR60","doi-asserted-by":"publisher","first-page":"447","DOI":"10.1109\/TPAMI.2015.2456899","volume":"38","author":"T Liu","year":"2015","unstructured":"Liu, T., Tao, D.: Classification with noisy labels by importance reweighting. IEEE Trans. Pattern Anal. Mach. Intell. 38(3), 447\u2013461 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"unstructured":"Liu, Y., Guo, H.: Peer loss functions: learning from noisy labels without knowing noise rates. In: International Conference on Machine Learning, pp. 6226\u20136236. PMLR (2020)","key":"41_CR61"},{"unstructured":"Lukasik, M., Bhojanapalli, S., Menon, A., Kumar, S.: Does label smoothing mitigate label noise? In: International Conference on Machine Learning, pp. 6448\u20136458. PMLR (2020)","key":"41_CR62"},{"unstructured":"Ma, X., Huang, H., Wang, Y., Romano, S., Erfani, S., Bailey, J.: Normalized loss functions for deep learning with noisy labels. In: International Conference on Machine Learning, pp. 6543\u20136553. PMLR (2020)","key":"41_CR63"},{"doi-asserted-by":"crossref","unstructured":"Ma, X., Wang, Y., Houle, M.E., Zhou, S., Erfani, S., Xia, S., Wijewickrema, S., Bailey, J.: Dimensionality-driven learning with noisy labels. In: International Conference on Machine Learning. pp. 3355\u20133364. PMLR (2018)","key":"41_CR64","DOI":"10.1109\/CVPR.2018.00906"},{"unstructured":"Menon, A.K., Rawat, A.S., Reddi, S.J., Kumar, S.: Can gradient clipping mitigate label noise? (2020)","key":"41_CR65"},{"unstructured":"Menon, A.K., Van Rooyen, B., Natarajan, N.: Learning from binary labels with instance-dependent corruption. arXiv preprint arXiv:1605.00751 (2016)","key":"41_CR66"},{"unstructured":"Mirzasoleiman, B., Cao, K., Leskovec, J.: Coresets for robust training of deep neural networks against noisy labels. Adv. Neural Inf. Process. Syst. 33 (2020)","key":"41_CR67"},{"unstructured":"Mnih, V., Hinton, G.E.: Learning to label aerial images from noisy data. In: Proceedings of the 29th International Conference on Machine Learning (ICML-12), pp. 567\u2013574 (2012)","key":"41_CR68"},{"unstructured":"Mohri, M., Rostamizadeh, A., Talwalkar, A.: Foundations of Machine Learning. MIT Press, Cambridge (2018)","key":"41_CR69"},{"issue":"2","key":"41_CR70","doi-asserted-by":"publisher","first-page":"517","DOI":"10.1090\/S0002-9939-1990-1013975-0","volume":"109","author":"SJ Montgomery-Smith","year":"1990","unstructured":"Montgomery-Smith, S.J.: The distribution of Rademacher sums. Proc. Am. Math. Soc. 109(2), 517\u2013522 (1990)","journal-title":"Proc. Am. Math. Soc."},{"unstructured":"Natarajan, N., Dhillon, I.S., Ravikumar, P., Tewari, A.: Learning with noisy labels. In: NIPS, vol. 26, pp. 1196\u20131204 (2013)","key":"41_CR71"},{"unstructured":"Neyshabur, B., Bhojanapalli, S., Srebro, N.: A PAC-Bayesian approach to spectrally-normalized margin bounds for neural networks. arXiv preprint arXiv:1707.09564 (2017)","key":"41_CR72"},{"doi-asserted-by":"crossref","unstructured":"Nishi, K., Ding, Y., Rich, A., H\u00f6llerer, T.: Augmentation strategies for learning with noisy labels. arXiv preprint arXiv:2103.02130 (2021)","key":"41_CR73","DOI":"10.1109\/CVPR46437.2021.00793"},{"doi-asserted-by":"crossref","unstructured":"Ortego, D., Arazo, E., Albert, P., O\u2019Connor, N.E., McGuinness, K.: Multi-objective interpolation training for robustness to label noise. arXiv preprint arXiv:2012.04462 (2020)","key":"41_CR74","DOI":"10.1109\/CVPR46437.2021.00654"},{"unstructured":"Patrini, G., Nielsen, F., Nock, R., Carioni, M.: Loss factorization, weakly supervised learning and label noise robustness. In: International Conference on Machine Learning, pp. 708\u2013717. PMLR (2016)","key":"41_CR75"},{"doi-asserted-by":"crossref","unstructured":"Patrini, G., Rozza, A., Krishna Menon, A., Nock, R., Qu, L.: Making deep neural networks robust to label noise: a loss correction approach. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1944\u20131952 (2017)","key":"41_CR76","DOI":"10.1109\/CVPR.2017.240"},{"unstructured":"Pleiss, G., Zhang, T., Elenberg, E.R., Weinberger, K.Q.: Identifying mislabeled data using the area under the margin ranking. arXiv preprint arXiv:2001.10528 (2020)","key":"41_CR77"},{"unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)","key":"41_CR78"},{"unstructured":"Reed, S., Lee, H., Anguelov, D., Szegedy, C., Erhan, D., Rabinovich, A.: Training deep neural networks on noisy labels with bootstrapping. arXiv preprint arXiv:1412.6596 (2014)","key":"41_CR79"},{"unstructured":"Ren, M., Zeng, W., Yang, B., Urtasun, R.: Learning to reweight examples for robust deep learning. In: International Conference on Machine Learning, pp. 4334\u20134343. PMLR (2018)","key":"41_CR80"},{"doi-asserted-by":"crossref","unstructured":"Rodrigues, F., Pereira, F.: Deep learning from crowds. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32 (2018)","key":"41_CR81","DOI":"10.1609\/aaai.v32i1.11506"},{"unstructured":"Scott, C.: A rate of convergence for mixture proportion estimation, with application to learning from noisy labels. In: Artificial Intelligence and Statistics, pp. 838\u2013846. PMLR (2015)","key":"41_CR82"},{"key":"41_CR83","doi-asserted-by":"publisher","first-page":"958","DOI":"10.1214\/12-EJS699","volume":"6","author":"C Scott","year":"2012","unstructured":"Scott, C., et al.: Calibrated asymmetric surrogate losses. Electron. J. Stat. 6, 958\u2013992 (2012)","journal-title":"Electron. J. Stat."},{"doi-asserted-by":"crossref","unstructured":"Shrivastava, A., Gupta, A., Girshick, R.: Training region-based object detectors with online hard example mining. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 761\u2013769 (2016)","key":"41_CR84","DOI":"10.1109\/CVPR.2016.89"},{"unstructured":"Shu, J., Xie, Q., Yi, L., Zhao, Q., Zhou, S., Xu, Z., Meng, D.: Meta-weight-net: learning an explicit mapping for sample weighting. arXiv preprint arXiv:1902.07379 (2019)","key":"41_CR85"},{"unstructured":"Song, H., Kim, M., Lee, J.G.: Selfie: refurbishing unclean samples for robust deep learning. In: International Conference on Machine Learning, pp. 5907\u20135915. PMLR (2019)","key":"41_CR86"},{"unstructured":"Song, H., Kim, M., Park, D., Lee, J.G.: How does early stopping help generalization against label noise? arXiv preprint arXiv:1911.08059 (2019)","key":"41_CR87"},{"unstructured":"Sukhbaatar, S., Bruna, J., Paluri, M., Bourdev, L., Fergus, R.: Training convolutional networks with noisy labels. arXiv preprint arXiv:1406.2080 (2014)","key":"41_CR88"},{"doi-asserted-by":"crossref","unstructured":"Taigman, Y., Yang, M., Ranzato, M., Wolf, L.: DeepFace: closing the gap to human-level performance in face verification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1701\u20131708 (2014)","key":"41_CR89","DOI":"10.1109\/CVPR.2014.220"},{"doi-asserted-by":"crossref","unstructured":"Tanaka, D., Ikami, D., Yamasaki, T., Aizawa, K.: Joint optimization framework for learning with noisy labels. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5552\u20135560 (2018)","key":"41_CR90","DOI":"10.1109\/CVPR.2018.00582"},{"doi-asserted-by":"crossref","unstructured":"Tanno, R., Saeedi, A., Sankaranarayanan, S., Alexander, D.C., Silberman, N.: Learning from noisy labels by regularized estimation of annotator confusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11244\u201311253 (2019)","key":"41_CR91","DOI":"10.1109\/CVPR.2019.01150"},{"unstructured":"Thekumparampil, K.K., Khetan, A., Lin, Z., Oh, S.: Robustness of conditional GANs to noisy labels. arXiv preprint arXiv:1811.03205 (2018)","key":"41_CR92"},{"unstructured":"Thulasidasan, S., Bhattacharya, T., Bilmes, J., Chennupati, G., Mohd-Yusof, J.: Combating label noise in deep learning using abstention. arXiv preprint arXiv:1905.10964 (2019)","key":"41_CR93"},{"unstructured":"Van Rooyen, B., Menon, A.K., Williamson, R.C.: Learning with symmetric label noise: the importance of being unhinged. arXiv preprint arXiv:1505.07634 (2015)","key":"41_CR94"},{"key":"41_CR95","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4757-3264-1","volume-title":"The Nature of Statistical Learning Theory","author":"V Vapnik","year":"2013","unstructured":"Vapnik, V.: The Nature of Statistical Learning Theory. Springer, New York (2013). https:\/\/doi.org\/10.1007\/978-1-4757-3264-1"},{"issue":"5","key":"41_CR96","doi-asserted-by":"publisher","first-page":"988","DOI":"10.1109\/72.788640","volume":"10","author":"VN Vapnik","year":"1999","unstructured":"Vapnik, V.N.: An overview of statistical learning theory. IEEE Trans. Neural Netw. 10(5), 988\u2013999 (1999)","journal-title":"IEEE Trans. Neural Netw."},{"doi-asserted-by":"crossref","unstructured":"Veit, A., Alldrin, N., Chechik, G., Krasin, I., Gupta, A., Belongie, S.: Learning from noisy large-scale datasets with minimal supervision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 839\u2013847 (2017)","key":"41_CR97","DOI":"10.1109\/CVPR.2017.696"},{"unstructured":"Vinyals, O., Blundell, C., Lillicrap, T., Kavukcuoglu, K., Wierstra, D.: Matching networks for one shot learning. arXiv preprint arXiv:1606.04080 (2016)","key":"41_CR98"},{"doi-asserted-by":"crossref","unstructured":"Wang, Y., Ma, X., Chen, Z., Luo, Y., Yi, J., Bailey, J.: Symmetric cross entropy for robust learning with noisy labels. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 322\u2013330 (2019)","key":"41_CR99","DOI":"10.1109\/ICCV.2019.00041"},{"unstructured":"Wang, Y., Kucukelbir, A., Blei, D.M.: Robust probabilistic modeling with Bayesian data reweighting. In: International Conference on Machine Learning, pp. 3646\u20133655. PMLR (2017)","key":"41_CR100"},{"doi-asserted-by":"crossref","unstructured":"Wang, Z., Hu, G., Hu, Q.: Training noise-robust deep neural networks via meta-learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4524\u20134533 (2020)","key":"41_CR101","DOI":"10.1109\/CVPR42600.2020.00458"},{"doi-asserted-by":"crossref","unstructured":"Wang, Z., Zhu, H., Dong, Z., He, X., Huang, S.L.: Less is better: unweighted data subsampling via influence function. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 6340\u20136347 (2020)","key":"41_CR102","DOI":"10.1609\/aaai.v34i04.6103"},{"unstructured":"Wei, J., Zhu, Z., Cheng, H., Liu, T., Niu, G., Liu, Y.: Learning with noisy labels revisited: a study using real-world human annotations. In: ICLR (2022)","key":"41_CR103"},{"unstructured":"Wu, P., Zheng, S., Goswami, M., Metaxas, D., Chen, C.: A topological filter for learning with label noise. arXiv preprint arXiv:2012.04835 (2020)","key":"41_CR104"},{"doi-asserted-by":"crossref","unstructured":"Wu, Y., Shu, J., Xie, Q., Zhao, Q., Meng, D.: Learning to purify noisy labels via meta soft label corrector. arXiv preprint arXiv:2008.00627 (2020)","key":"41_CR105","DOI":"10.1609\/aaai.v35i12.17244"},{"unstructured":"Xia, X., et al.: Sample selection with uncertainty of losses for learning with noisy labels. arXiv preprint arXiv:2106.00445 (2021)","key":"41_CR106"},{"unstructured":"Xia, X., et al.: Part-dependent label noise: Towards instance-dependent label noise. Adv. Neural Inf. Process. Syst. 33 (2020)","key":"41_CR107"},{"unstructured":"Xia, X., Liu, T., Wang, N., Han, B., Gong, C., Niu, G., Sugiyama, M.: Are anchor points really indispensable in label-noise learning? arXiv preprint arXiv:1906.00189 (2019)","key":"41_CR108"},{"unstructured":"Xiao, T., Xia, T., Yang, Y., Huang, C., Wang, X.: Learning from massive noisy labeled data for image classification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2691\u20132699 (2015)","key":"41_CR109"},{"unstructured":"Yang, F., Koyejo, S.: On the consistency of top-k surrogate losses. In: International Conference on Machine Learning, pp. 10727\u201310735. PMLR (2020)","key":"41_CR110"},{"doi-asserted-by":"crossref","unstructured":"Yao, J., Wu, H., Zhang, Y., Tsang, I.W., Sun, J.: Safeguarded dynamic label regression for noisy supervision. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 9103\u20139110 (2019)","key":"41_CR111","DOI":"10.1609\/aaai.v33i01.33019103"},{"unstructured":"Yao, Y., Liu, T., Gong, M., Han, B., Niu, G., Zhang, K.: Instance-dependent label-noise learning under a structural causal model. Adv. Neural Inf. Process. Syst. 34 (2021)","key":"41_CR112"},{"unstructured":"Yao, Y., et al.: Dual t: reducing estimation error for transition matrix in label-noise learning. arXiv preprint arXiv:2006.07805 (2020)","key":"41_CR113"},{"doi-asserted-by":"crossref","unstructured":"Yi, K., Wu, J.: Probabilistic end-to-end noise correction for learning with noisy labels. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7017\u20137025 (2019)","key":"41_CR114","DOI":"10.1109\/CVPR.2019.00718"},{"key":"41_CR115","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1007\/978-3-030-01246-5_5","volume-title":"Computer Vision \u2013 ECCV 2018","author":"X Yu","year":"2018","unstructured":"Yu, X., Liu, T., Gong, M., Tao, D.: Learning with biased complementary labels. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11205, pp. 69\u201385. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01246-5_5"},{"unstructured":"Zhang, C., Bengio, S., Hardt, M., Recht, B., Vinyals, O.: Understanding deep learning requires rethinking generalization. arXiv preprint arXiv:1611.03530 (2016)","key":"41_CR116"},{"unstructured":"Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412 (2017)","key":"41_CR117"},{"doi-asserted-by":"crossref","unstructured":"Zhang, X., Wu, X., Chen, F., Zhao, L., Lu, C.T.: Self-paced robust learning for leveraging clean labels in noisy data. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 6853\u20136860 (2020)","key":"41_CR118","DOI":"10.1609\/aaai.v34i04.6166"},{"unstructured":"Zhang, Y., Niu, G., Sugiyama, M.: Learning noise transition matrix from only noisy labels via total variation regularization. In: International Conference on Machine Learning, pp. 12501\u201312512. PMLR (2021)","key":"41_CR119"},{"unstructured":"Zhang, Z., Sabuncu, M.R.: Generalized cross entropy loss for training deep neural networks with noisy labels. arXiv preprint arXiv:1805.07836 (2018)","key":"41_CR120"},{"doi-asserted-by":"crossref","unstructured":"Zheng, G., Awadallah, A.H., Dumais, S.: Meta label correction for noisy label learning. In: Proceedings of the 35th AAAI Conference on Artificial Intelligence (2021)","key":"41_CR121","DOI":"10.1609\/aaai.v35i12.17319"},{"unstructured":"Zheng, S., Wu, P., Goswami, A., Goswami, M., Metaxas, D., Chen, C.: Error-bounded correction of noisy labels. In: International Conference on Machine Learning, pp. 11447\u201311457. PMLR (2020)","key":"41_CR122"},{"doi-asserted-by":"crossref","unstructured":"Zhu, Z., Liu, T., Liu, Y.: A second-order approach to learning with instance-dependent label noise. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10113\u201310123 (2021)","key":"41_CR123","DOI":"10.1109\/CVPR46437.2021.00998"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19806-9_41","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,21]],"date-time":"2022-10-21T23:19:00Z","timestamp":1666394340000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19806-9_41"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198052","9783031198069"],"references-count":123,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19806-9_41","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"20 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}