{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T19:48:24Z","timestamp":1774986504918,"version":"3.50.1"},"reference-count":69,"publisher":"Association for Computing Machinery (ACM)","issue":"3","funder":[{"name":"CCF-Baidu Open Fund","award":["CCF-Baidu202402"],"award-info":[{"award-number":["CCF-Baidu202402"]}]},{"DOI":"10.13039\/501100001809","name":"NSFC","doi-asserted-by":"crossref","award":["62472031,624B2023,61932004, 62225203, U21A20516,62427808, U2001211"],"award-info":[{"award-number":["62472031,624B2023,61932004, 62225203, U21A20516,62427808, U2001211"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100006374","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2024YFC3308200,2022YFB2702100"],"award-info":[{"award-number":["2024YFC3308200,2022YFB2702100"]}],"id":[{"id":"10.13039\/501100006374","id-type":"DOI","asserted-by":"publisher"}]},{"name":"the Liaoning Revitalization Talents Program","award":["XLYC2204005"],"award-info":[{"award-number":["XLYC2204005"]}]},{"DOI":"10.13039\/501100006374","name":"NSF","doi-asserted-by":"publisher","award":["DBI-2327954"],"award-info":[{"award-number":["DBI-2327954"]}],"id":[{"id":"10.13039\/501100006374","id-type":"DOI","asserted-by":"publisher"}]},{"name":"the BIT Research and Innovation Promoting Project","award":["2024YCXZ004"],"award-info":[{"award-number":["2024YCXZ004"]}]},{"DOI":"10.13039\/100021154","name":"Amazon Research Awards","doi-asserted-by":"crossref","id":[{"id":"10.13039\/100021154","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100003816","name":"Huawei Technologies","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100003816","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Proc. ACM Manag. Data"],"published-print":{"date-parts":[[2025,6,17]]},"abstract":"<jats:p>Using a large training dataset to train a big and powerful model -- a typical practice in modern deep learning, often suffers from two major problems: the expensive and slow training process and the error-prone labels. The existing approaches, targeting either speeding up the training by selecting a subset of representative training instances (subset selection) or eliminating the negative effect of mislabels during training (mislabel detection), do not perform well in this scenario due to overlooking one of these two problems. To fill this gap, we propose Deem, a novel data-efficient framework that selects a subset of representative training instances under label uncertainty. The key idea is to leverage the metadata produced during deep learning training, e.g., training losses and gradients, to estimate the label uncertainty and select the representative instances. In particular, we model the problem of subset selection under uncertainty as a problem of finding a subset that closely approximates the gradient of the whole training data set derived on soft labels. We show that it is an NP-hard problem with submodular property and propose a low complexity algorithm to solve this problem with an approximate ratio. Training on this small subset thus improves the training efficiency while guaranteeing the model's accuracy. Moreover, we propose an efficient strategy to dynamically refine this subset during the iterative training process. Extensive experiments on 6 datasets and 10 baselines demonstrate that Deem accelerates the training process up to 10X without sacrificing the model accuracy.<\/jats:p>","DOI":"10.1145\/3728289","type":"journal-article","created":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T21:23:29Z","timestamp":1750281809000},"page":"1-28","source":"Crossref","is-referenced-by-count":0,"title":["Two Birds with One Stone: Efficient Deep Learning over Mislabeled Data through Subset Selection"],"prefix":"10.1145","volume":"3","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-4473-4527","authenticated-orcid":false,"given":"Yuhao","family":"Deng","sequence":"first","affiliation":[{"name":"Beijing Institute of Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-5386-1330","authenticated-orcid":false,"given":"Chengliang","family":"Chai","sequence":"additional","affiliation":[{"name":"Beijing Institute of Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-7020-5404","authenticated-orcid":false,"given":"Kaisen","family":"Jin","sequence":"additional","affiliation":[{"name":"Beijing Institute of Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-4437-1556","authenticated-orcid":false,"given":"Linan","family":"Zheng","sequence":"additional","affiliation":[{"name":"University of Arizona, Tucson, AZ, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9909-8607","authenticated-orcid":false,"given":"Lei","family":"Cao","sequence":"additional","affiliation":[{"name":"University of Arizona, Tucson, USA and MIT, Cambridge, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0247-9866","authenticated-orcid":false,"given":"Ye","family":"Yuan","sequence":"additional","affiliation":[{"name":"Beijing Institute of Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0181-8379","authenticated-orcid":false,"given":"Guoren","family":"Wang","sequence":"additional","affiliation":[{"name":"Beijing Institute of Technology, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2025,6,18]]},"reference":[{"key":"e_1_2_1_1_1","unstructured":"1998. https:\/\/archive.ics.uci.edu\/ml\/datasets\/Covertype."},{"key":"e_1_2_1_2_1","unstructured":"1999. https:\/\/yann.lecun.com\/exdb\/mnist\/."},{"key":"e_1_2_1_3_1","unstructured":"2009. http:\/\/www.cs.toronto.edu\/~kriz\/cifar.html."},{"key":"e_1_2_1_4_1","unstructured":"2011. http:\/\/ufldl.stanford.edu\/housenumbers\/."},{"key":"e_1_2_1_5_1","unstructured":"2025. https:\/\/anonymous.4open.science\/r\/Deem-1030\/Deem_tech_report.pdf."},{"key":"e_1_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.48786\/EDBT.2023.43"},{"key":"e_1_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1007\/978--3-030--58517--4_9"},{"key":"e_1_2_1_8_1","volume-title":"Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9--15","volume":"321","author":"Arazo Eric","year":"2019","unstructured":"Eric Arazo, Diego Ortego, Paul Albert, Noel E. O'Connor, and Kevin McGuinness. 2019. Unsupervised Label Noise Modeling and Loss Correction. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9--15 June 2019, Long Beach, California, USA (Proceedings of Machine Learning Research, Vol. 97), Kamalika Chaudhuri and Ruslan Salakhutdinov (Eds.). PMLR, 312--321. http:\/\/proceedings.mlr.press\/v97\/arazo19a.html"},{"key":"e_1_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3219973"},{"key":"e_1_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1007\/3-540-44522-6_64"},{"key":"e_1_2_1_11_1","volume-title":"The tradeoffs of large scale learning. Advances in neural information processing systems 20","author":"Bottou L\u00e9on","year":"2007","unstructured":"L\u00e9on Bottou and Olivier Bousquet. 2007. The tradeoffs of large scale learning. Advances in neural information processing systems 20 (2007)."},{"key":"e_1_2_1_12_1","volume-title":"Random forests. Machine learning 45","author":"Breiman Leo","year":"2001","unstructured":"Leo Breiman. 2001. Random forests. Machine learning 45 (2001), 5--32."},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1613\/jair.606"},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3589302"},{"key":"e_1_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3589302"},{"key":"e_1_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599326"},{"key":"e_1_2_1_17_1","volume-title":"UAI 2010, Proceedings of the Twenty-Sixth Conference on Uncertainty in Artificial Intelligence","author":"Chen Yutian","year":"2010","unstructured":"Yutian Chen, Max Welling, and Alexander J. Smola. 2010. Super-Samples from Kernel Herding. In UAI 2010, Proceedings of the Twenty-Sixth Conference on Uncertainty in Artificial Intelligence, Catalina Island, CA, USA, July 8--11, 2010, Peter Gr\u00fcnwald and Peter Spirtes (Eds.). AUAI Press, 109--116. https:\/\/dslpitt.org\/uai\/displayArticleDetails.jsp?mmnu=1&smnu=2&article_id=2148&proceeding_id=26"},{"key":"e_1_2_1_18_1","volume-title":"8th International Conference on Learning Representations, ICLR 2020","author":"Coleman Cody","year":"2020","unstructured":"Cody Coleman, Christopher Yeh, Stephen Mussmann, Baharan Mirzasoleiman, Peter Bailis, Percy Liang, Jure Leskovec, and Matei Zaharia. 2020. Selection via Proxy: Efficient Data Selection for Deep Learning. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26--30, 2020. OpenReview.net. https:\/\/openreview.net\/forum?id=HJg2b0VYDr"},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.14778\/3648160.3648161"},{"key":"e_1_2_1_20_1","volume-title":"Adversarial Active Learning for Deep Networks: a Margin Based Approach. CoRR abs\/1802.09841","author":"Ducoffe Melanie","year":"2018","unstructured":"Melanie Ducoffe and Fr\u00e9d\u00e9ric Precioso. 2018. Adversarial Active Learning for Deep Networks: a Margin Based Approach. CoRR abs\/1802.09841 (2018). arXiv:1802.09841 http:\/\/arxiv.org\/abs\/1802.09841"},{"key":"e_1_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01551"},{"key":"e_1_2_1_22_1","volume-title":"Masking: A New Perspective of Noisy Supervision. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018","author":"Han Bo","year":"2018","unstructured":"Bo Han, Jiangchao Yao, Gang Niu, Mingyuan Zhou, Ivor W. Tsang, Ya Zhang, and Masashi Sugiyama. 2018. Masking: A New Perspective of Noisy Supervision. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, December 3--8, 2018, Montr\u00e9al, Canada, Samy Bengio, Hanna M. Wallach, Hugo Larochelle, Kristen Grauman, Nicol\u00f2 Cesa-Bianchi, and Roman Garnett (Eds.). 5841--5851. https:\/\/proceedings.neurips.cc\/paper\/2018\/hash\/aee92f16efd522b9326c25cc3237ac15-Abstract.html"},{"key":"e_1_2_1_23_1","volume-title":"Co-teaching: Robust training of deep neural networks with extremely noisy labels. In NeurIPS","author":"Han Bo","year":"2018","unstructured":"Bo Han, Quanming Yao, Xingrui Yu, Gang Niu, Miao Xu, Weihua Hu, Ivor W. Tsang, and Masashi Sugiyama. 2018. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In NeurIPS 2018. 8536--8546. https:\/\/proceedings.neurips.cc\/paper\/2018\/hash\/a19744e268754fb0148b017647355b7b-Abstract.html"},{"key":"e_1_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/1007352.1007400"},{"key":"e_1_2_1_25_1","volume-title":"NeurIPS","author":"Hendrycks Dan","year":"2018","unstructured":"Dan Hendrycks, Mantas Mazeika, Duncan Wilson, and Kevin Gimpel. 2018. Using Trusted Data to Train Deep Networks on Labels Corrupted by Severe Noise. In NeurIPS 2018. 10477--10486. https:\/\/proceedings.neurips.cc\/paper\/2018\/hash\/ad554d8c3b06d6b97ee76a2448bd7913-Abstract.html"},{"key":"e_1_2_1_26_1","volume-title":"Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsm\u00e4ssan","author":"Jiang Lu","year":"2018","unstructured":"Lu Jiang, Zhengyuan Zhou, Thomas Leung, Li-Jia Li, and Li Fei-Fei. 2018. MentorNet: Learning Data-Driven Curriculum for Very Deep Neural Networks on Corrupted Labels. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsm\u00e4ssan, Stockholm, Sweden, July 10--15, 2018 (Proceedings of Machine Learning Research, Vol. 80), Jennifer G. Dy and Andreas Krause (Eds.). PMLR, 2309--2318. http:\/\/proceedings.mlr.press\/v80\/jiang18c.html"},{"key":"e_1_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-28647-9_60"},{"key":"e_1_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.5555\/3430915.3442426"},{"key":"e_1_2_1_29_1","volume-title":"Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18--24","volume":"5474","author":"Killamsetty KrishnaTeja","year":"2021","unstructured":"KrishnaTeja Killamsetty, Durga Sivasubramanian, Ganesh Ramakrishnan, Abir De, and Rishabh K. Iyer. 2021. GRADMATCH: Gradient Matching based Data Subset Selection for Efficient Deep Model Training. In Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18--24 July 2021, Virtual Event (Proceedings of Machine Learning Research, Vol. 139), Marina Meila and Tong Zhang (Eds.). PMLR, 5464--5474. http:\/\/proceedings.mlr.press\/v139\/killamsetty21a.html"},{"key":"e_1_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1609\/AAAI.V35I9.16988"},{"key":"e_1_2_1_31_1","volume-title":"BoostClean: Automated Error Detection and Repair for Machine Learning. CoRR abs\/1711.01299","author":"Krishnan Sanjay","year":"2017","unstructured":"Sanjay Krishnan, Michael J. Franklin, Ken Goldberg, and Eugene Wu. 2017. BoostClean: Automated Error Detection and Repair for Machine Learning. CoRR abs\/1711.01299 (2017). arXiv:1711.01299 http:\/\/arxiv.org\/abs\/1711.01299"},{"key":"e_1_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.14778\/2994509.2994514"},{"key":"e_1_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.14778\/2850583.2850594"},{"key":"e_1_2_1_34_1","volume-title":"ICLR","author":"Li Junnan","year":"2020","unstructured":"Junnan Li, Richard Socher, and Steven C. H. Hoi. 2020. DivideMix: Learning with Noisy Labels as Semi-supervised Learning. In ICLR 2020. OpenReview.net. https:\/\/openreview.net\/forum?id=HJgExaVtwr"},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDE51399.2021.00009"},{"key":"e_1_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.24963\/IJCAI.2020\/356"},{"key":"e_1_2_1_37_1","volume-title":"AutoCoreset: An Automatic Practical Coreset Construction Framework. In International Conference on Machine Learning, ICML 2023","volume":"23466","author":"Maalouf Alaa","year":"2023","unstructured":"Alaa Maalouf, Murad Tukan, Vladimir Braverman, and Daniela Rus. 2023. AutoCoreset: An Automatic Practical Coreset Construction Framework. In International Conference on Machine Learning, ICML 2023, 23--29 July 2023, Honolulu, Hawaii, USA (Proceedings of Machine Learning Research, Vol. 202), Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (Eds.). PMLR, 23451--23466. https:\/\/proceedings.mlr.press\/v202\/maalouf23a.html"},{"key":"e_1_2_1_38_1","volume-title":"Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017","author":"Malach Eran","year":"2017","unstructured":"Eran Malach and Shai Shalev-Shwartz. 2017. Decoupling ''when to update'' from ''how to update''. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4--9, 2017, Long Beach, CA, USA, Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett (Eds.). 960--970. https:\/\/proceedings.neurips.cc\/paper\/2017\/hash\/58d4d1e7b1e97b258c9ed0b37e02d087-Abstract.html"},{"key":"e_1_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV45572.2020.9093342"},{"key":"e_1_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/V1\/2021.EMNLP-MAIN.51"},{"key":"e_1_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1007\/978--3--540--25966--4_29"},{"key":"e_1_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v29i1.9486"},{"key":"e_1_2_1_43_1","volume-title":"Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18","volume":"6960","author":"Mirzasoleiman Baharan","year":"2020","unstructured":"Baharan Mirzasoleiman, Jeff A. Bilmes, and Jure Leskovec. 2020. Coresets for Data-efficient Training of Machine Learning Models. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18 July 2020, Virtual Event (Proceedings of Machine Learning Research, Vol. 119). PMLR, 6950--6960. http:\/\/proceedings.mlr.press\/v119\/mirzasoleiman20a.html"},{"key":"e_1_2_1_44_1","volume-title":"Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18","volume":"6960","author":"Mirzasoleiman Baharan","year":"2020","unstructured":"Baharan Mirzasoleiman, Jeff A. Bilmes, and Jure Leskovec. 2020. Coresets for Data-efficient Training of Machine Learning Models. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18 July 2020, Virtual Event (Proceedings of Machine Learning Research, Vol. 119). PMLR, 6950--6960. http:\/\/proceedings.mlr.press\/v119\/mirzasoleiman20a.html"},{"key":"e_1_2_1_45_1","volume-title":"Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019","author":"M\u00fcller Rafael","year":"2019","unstructured":"Rafael M\u00fcller, Simon Kornblith, and Geoffrey E. Hinton. 2019. When does label smoothing help?. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8--14, 2019, Vancouver, BC, Canada, Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d'Alch\u00e9-Buc, Emily B. Fox, and Roman Garnett (Eds.). 4696--4705. https:\/\/proceedings.neurips.cc\/paper\/2019\/hash\/f1748d6b0fd9d439f71450117eba2725-Abstract.html"},{"key":"e_1_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1613\/jair.1.12125"},{"key":"e_1_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.240"},{"key":"e_1_2_1_48_1","volume-title":"Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021","author":"Paul Mansheej","year":"2021","unstructured":"Mansheej Paul, Surya Ganguli, and Gintare Karolina Dziugaite. 2021. Deep Learning on a Data Diet: Finding Important Examples Early in Training. In Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6--14, 2021, virtual, Marc'Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (Eds.). 20596--20607. https:\/\/proceedings.neurips.cc\/paper\/2021\/hash\/ac56f8fe9eea3e4a365f29f0f1957c55-Abstract.html"},{"key":"e_1_2_1_49_1","volume-title":"Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsm\u00e4ssan","author":"Ren Mengye","year":"2018","unstructured":"Mengye Ren, Wenyuan Zeng, Bin Yang, and Raquel Urtasun. 2018. Learning to Reweight Examples for Robust Deep Learning. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsm\u00e4ssan, Stockholm, Sweden, July 10--15, 2018 (Proceedings of Machine Learning Research, Vol. 80), Jennifer G. Dy and Andreas Krause (Eds.). PMLR, 4331--4340. http:\/\/proceedings.mlr.press\/v80\/ren18a.html"},{"key":"e_1_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2006.211"},{"key":"e_1_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1016\/S0167--8655(02)00225--8"},{"key":"e_1_2_1_52_1","volume-title":"6th International Conference on Learning Representations, ICLR","author":"Sener Ozan","year":"2018","unstructured":"Ozan Sener and Silvio Savarese. 2018. Active Learning for Convolutional Neural Networks: A Core-Set Approach. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. OpenReview.net. https:\/\/openreview.net\/forum?id=H1aIuk-RW"},{"key":"e_1_2_1_53_1","volume-title":"Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18","volume":"9015","author":"Sinha Samarth","year":"2020","unstructured":"Samarth Sinha, Han Zhang, Anirudh Goyal, Yoshua Bengio, Hugo Larochelle, and Augustus Odena. 2020. Small-GAN: Speeding up GAN Training using Core-Sets. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18 July 2020, Virtual Event (Proceedings of Machine Learning Research, Vol. 119). PMLR, 9005--9015. http:\/\/proceedings.mlr.press\/v119\/sinha20b.html"},{"key":"e_1_2_1_54_1","volume-title":"7th International Conference on Learning Representations, ICLR 2019","author":"Toneva Mariya","year":"2019","unstructured":"Mariya Toneva, Alessandro Sordoni, Remi Tachet des Combes, Adam Trischler, Yoshua Bengio, and Geoffrey J. Gordon. 2019. An Empirical Study of Example Forgetting during Deep Neural Network Learning. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6--9, 2019. OpenReview.net. https:\/\/openreview.net\/forum?id=BJlxm30cKm"},{"key":"e_1_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1016\/J.TCS.2021.09.008"},{"key":"e_1_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1137\/1.9781611972771.28"},{"key":"e_1_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.14778\/3561261.3561267"},{"key":"e_1_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.14778\/3561261.3561267"},{"key":"e_1_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00906"},{"key":"e_1_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00458"},{"key":"e_1_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/1553374.1553517"},{"key":"e_1_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298885"},{"key":"e_1_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01576"},{"key":"e_1_2_1_64_1","volume-title":"Learning From Biased Soft Labels. In Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023","author":"Yuan Hua","year":"2023","unstructured":"Hua Yuan, Yu Shi, Ning Xu, Xu Yang, Xin Geng, and Yong Rui. 2023. Learning From Biased Soft Labels. In Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 - 16, 2023, Alice Oh, Tristan Naumann, Amir Globerson, Kate Saenko, Moritz Hardt, and Sergey Levine (Eds.). http:\/\/papers.nips.cc\/paper_files\/paper\/2023\/hash\/bad8ddaed5feb552f9e8f2e37c0531a1-Abstract-Conference.html"},{"key":"e_1_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2567393"},{"key":"e_1_2_1_66_1","volume-title":"Simple and Automated Negative Sampling for Knowledge Graph Embedding. CoRR abs\/2010.14227","author":"Zhang Yongqi","year":"2020","unstructured":"Yongqi Zhang, Quanming Yao, and Lei Chen. 2020. Efficient, Simple and Automated Negative Sampling for Knowledge Graph Embedding. CoRR abs\/2010.14227 (2020). arXiv:2010.14227 https:\/\/arxiv.org\/abs\/2010.14227"},{"key":"e_1_2_1_67_1","volume-title":"Sabuncu","author":"Zhang Zhilu","year":"2018","unstructured":"Zhilu Zhang and Mert R. Sabuncu. 2018. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels. In NeurIPS 2018. 8792--8802. https:\/\/proceedings.neurips.cc\/paper\/2018\/hash\/f2925f97bc13ad2852a7a551802feea0-Abstract.html"},{"key":"e_1_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00931"},{"key":"e_1_2_1_69_1","volume-title":"Eliminating Class Noise in Large Datasets. In ICML 2003","author":"Zhu Xingquan","year":"2003","unstructured":"Xingquan Zhu, Xindong Wu, and Qijun Chen. 2003. Eliminating Class Noise in Large Datasets. In ICML 2003, Tom Fawcett and Nina Mishra (Eds.). AAAI Press, 920--927. http:\/\/www.aaai.org\/Library\/ICML\/2003\/icml03--119.php"}],"container-title":["Proceedings of the ACM on Management of Data"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3728289","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T18:51:49Z","timestamp":1774983109000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3728289"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,17]]},"references-count":69,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,6,17]]}},"alternative-id":["10.1145\/3728289"],"URL":"https:\/\/doi.org\/10.1145\/3728289","relation":{},"ISSN":["2836-6573"],"issn-type":[{"value":"2836-6573","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,6,17]]}}}