{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,16]],"date-time":"2026-02-16T15:37:59Z","timestamp":1771256279920,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":46,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,8,14]],"date-time":"2021-08-14T00:00:00Z","timestamp":1628899200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Institute of Information & Communications Technology Planning & Evaluation (IITP)","award":["2020-0-00862"],"award-info":[{"award-number":["2020-0-00862"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,8,14]]},"DOI":"10.1145\/3447548.3467222","type":"proceedings-article","created":{"date-parts":[[2021,8,12]],"date-time":"2021-08-12T06:12:05Z","timestamp":1628748725000},"page":"1490-1500","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":28,"title":["Robust Learning by Self-Transition for Handling Noisy Labels"],"prefix":"10.1145","author":[{"given":"Hwanjun","family":"Song","sequence":"first","affiliation":[{"name":"NAVER AI Lab, Seongnam, South Korea"}]},{"given":"Minseok","family":"Kim","sequence":"additional","affiliation":[{"name":"Korea Advanced Institute of Science and Technology, Daejeon, South Korea"}]},{"given":"Dongmin","family":"Park","sequence":"additional","affiliation":[{"name":"Korea Advanced Institute of Science and Technology, Daejeon, South Korea"}]},{"given":"Yooju","family":"Shin","sequence":"additional","affiliation":[{"name":"Korea Advanced Institute of Science and Technology, Daejeon, South Korea"}]},{"given":"Jae-Gil","family":"Lee","sequence":"additional","affiliation":[{"name":"Korea Advanced Institute of Science and Technology, Daejeon, South Korea"}]}],"member":"320","published-online":{"date-parts":[[2021,8,14]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Eric Arazo Diego Ortego Paul Albert Noel O'Connor and Kevin Mcguinness. 2019. Unsupervised label noise modeling and loss correction. In ICML. 312--321.  Eric Arazo Diego Ortego Paul Albert Noel O'Connor and Kevin Mcguinness. 2019. Unsupervised label noise modeling and loss correction. In ICML. 312--321."},{"key":"e_1_3_2_1_2_1","unstructured":"Devansh Arpit Stanislaw Jastrzebski Nicolas Ballas David Krueger Emmanuel Bengio Maxinder S Kanwal Tegan Maharaj Asja Fischer Aaron Courville Yoshua Bengio etal 2017. A closer look at memorization in deep networks. In ICML. 233--242.  Devansh Arpit Stanislaw Jastrzebski Nicolas Ballas David Krueger Emmanuel Bengio Maxinder S Kanwal Tegan Maharaj Asja Fischer Aaron Courville Yoshua Bengio et al. 2017. A closer look at memorization in deep networks. In ICML. 233--242."},{"key":"e_1_3_2_1_3_1","unstructured":"David Berthelot Nicholas Carlini Ian Goodfellow Nicolas Papernot Avital Oliver and Colin A Raffel. 2019. MixMatch: A holistic approach to semi-supervised learning. In NeurIPS. 5049--5059.  David Berthelot Nicholas Carlini Ian Goodfellow Nicolas Papernot Avital Oliver and Colin A Raffel. 2019. MixMatch: A holistic approach to semi-supervised learning. In NeurIPS. 5049--5059."},{"key":"e_1_3_2_1_4_1","volume-title":"Active Bias: Training more accurate neural networks by emphasizing high variance samples. In NeurIPS. 1002--1012.","author":"Chang Haw-Shiuan","year":"2017","unstructured":"Haw-Shiuan Chang , Erik Learned-Miller , and Andrew McCallum . 2017 . Active Bias: Training more accurate neural networks by emphasizing high variance samples. In NeurIPS. 1002--1012. Haw-Shiuan Chang, Erik Learned-Miller, and Andrew McCallum. 2017. Active Bias: Training more accurate neural networks by emphasizing high variance samples. In NeurIPS. 1002--1012."},{"key":"e_1_3_2_1_5_1","volume-title":"Guangyong Chen, and Shengyu Zhang.","author":"Chen Pengfei","year":"2019","unstructured":"Pengfei Chen , Ben Ben Liao , Guangyong Chen, and Shengyu Zhang. 2019 . Understanding and utilizing deep neural networks trained with noisy labels. In ICML. 1062--1070. Pengfei Chen, Ben Ben Liao, Guangyong Chen, and Shengyu Zhang. 2019. Understanding and utilizing deep neural networks trained with noisy labels. In ICML. 1062--1070."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"Jia Deng Wei Dong Richard Socher Li-Jia Li Kai Li and Li Fei-Fei. 2009. ImageNet: A large-scale hierarchical image database. In CVPR. 248--255.  Jia Deng Wei Dong Richard Socher Li-Jia Li Kai Li and Li Fei-Fei. 2009. ImageNet: A large-scale hierarchical image database. In CVPR. 248--255.","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2013.2292894"},{"key":"e_1_3_2_1_8_1","volume-title":"SIGUA: Forgetting may make learning with noisy labels more robust. In ICML. 4006--4016.","author":"Han Bo","year":"2020","unstructured":"Bo Han , Gang Niu , Xingrui Yu , Quanming Yao , Miao Xu , Ivor Tsang , and Masashi Sugiyama . 2020 . SIGUA: Forgetting may make learning with noisy labels more robust. In ICML. 4006--4016. Bo Han, Gang Niu, Xingrui Yu, Quanming Yao, Miao Xu, Ivor Tsang, and Masashi Sugiyama. 2020. SIGUA: Forgetting may make learning with noisy labels more robust. In ICML. 4006--4016."},{"key":"e_1_3_2_1_9_1","volume-title":"Co-teaching: Robust training of deep neural networks with extremely noisy labels. In NeurIPS. 8536--8546.","author":"Han Bo","year":"2018","unstructured":"Bo Han , Quanming Yao , Xingrui Yu , Gang Niu , Miao Xu , Weihua Hu , Ivor Tsang , and Masashi Sugiyama . 2018 . Co-teaching: Robust training of deep neural networks with extremely noisy labels. In NeurIPS. 8536--8546. Bo Han, Quanming Yao, Xingrui Yu, Gang Niu, Miao Xu, Weihua Hu, Ivor Tsang, and Masashi Sugiyama. 2018. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In NeurIPS. 8536--8546."},{"key":"e_1_3_2_1_10_1","volume-title":"Distilling the knowledge in a neural network. CoRR","author":"Hinton Geoffrey","year":"2015","unstructured":"Geoffrey Hinton , Oriol Vinyals , and Jeff Dean . 2015. Distilling the knowledge in a neural network. CoRR ( 2015 ). Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015. Distilling the knowledge in a neural network. CoRR (2015)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"crossref","unstructured":"Jinchi Huang Lie Qu Rongfei Jia and Binqiang Zhao. 2019. O2U-Net: A simple noisy label detection approach for deep neural networks. In ICCV. 3326--3334.  Jinchi Huang Lie Qu Rongfei Jia and Binqiang Zhao. 2019. O2U-Net: A simple noisy label detection approach for deep neural networks. In ICCV. 3326--3334.","DOI":"10.1109\/ICCV.2019.00342"},{"key":"e_1_3_2_1_12_1","unstructured":"Lu Jiang Zhengyuan Zhou Thomas Leung Li-Jia Li and Li Fei-Fei. 2018. MentorNet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In ICML. 2309--2318.  Lu Jiang Zhengyuan Zhou Thomas Leung Li-Jia Li and Li Fei-Fei. 2018. MentorNet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In ICML. 2309--2318."},{"key":"e_1_3_2_1_13_1","unstructured":"Mohammad Mahdi Kamani Sadegh Farhang Mehrdad Mahdavi and James Z Wang. 2020. Targeted data-driven regularization for out-of-distribution generalization. In KDD. 882--891.  Mohammad Mahdi Kamani Sadegh Farhang Mehrdad Mahdavi and James Z Wang. 2020. Targeted data-driven regularization for out-of-distribution generalization. In KDD. 882--891."},{"key":"e_1_3_2_1_14_1","unstructured":"Alex Krizhevsky Vinod Nair and Geoffrey Hinton. 2014. CIFAR-10 and CIFAR-100 datasets. https:\/\/www.cs.toronto.edu\/ kriz\/cifar.html.  Alex Krizhevsky Vinod Nair and Geoffrey Hinton. 2014. CIFAR-10 and CIFAR-100 datasets. https:\/\/www.cs.toronto.edu\/ kriz\/cifar.html."},{"key":"e_1_3_2_1_15_1","unstructured":"Alex Krizhevsky Ilya Sutskever and Geoffrey E Hinton. 2012. ImageNet classification with deep convolutional neural networks. In NeurIPS. 1097--1105.  Alex Krizhevsky Ilya Sutskever and Geoffrey E Hinton. 2012. ImageNet classification with deep convolutional neural networks. In NeurIPS. 1097--1105."},{"key":"e_1_3_2_1_16_1","unstructured":"Samuli Laine and Tim Aila. 2017. Temporal ensembling for semi-supervised learning. In ICLR.  Samuli Laine and Tim Aila. 2017. Temporal ensembling for semi-supervised learning. In ICLR."},{"key":"e_1_3_2_1_17_1","unstructured":"Kuang-Huei Lee Xiaodong He Lei Zhang and Linjun Yang. 2018. CleanNet: Transfer learning for scalable image classifier training with label noise. In CVPR. 5447--5456.  Kuang-Huei Lee Xiaodong He Lei Zhang and Linjun Yang. 2018. CleanNet: Transfer learning for scalable image classifier training with label noise. In CVPR. 5447--5456."},{"key":"e_1_3_2_1_18_1","unstructured":"Junnan Li Richard Socher and Steven CH Hoi. 2020 b. DivideMix: Learning with noisy labels as semi-supervised learning. In ICLR.  Junnan Li Richard Socher and Steven CH Hoi. 2020 b. DivideMix: Learning with noisy labels as semi-supervised learning. In ICLR."},{"key":"e_1_3_2_1_19_1","unstructured":"Mingchen Li Mahdi Soltanolkotabi and Samet Oymak. 2020 c. Gradient descent with early stopping is provably robust to label noise for overparameterized neural networks. In AISTATS. 4313--4324.  Mingchen Li Mahdi Soltanolkotabi and Samet Oymak. 2020 c. Gradient descent with early stopping is provably robust to label noise for overparameterized neural networks. In AISTATS. 4313--4324."},{"key":"e_1_3_2_1_20_1","volume-title":"2020 a. Product image recognition with guidance learning and noisy supervision. Computer Vision and Image Understanding","author":"Li Qing","year":"2020","unstructured":"Qing Li , Xiaojiang Peng , Liangliang Cao , Wenbin Du , Hao Xing , Yu Qiao , and Qiang Peng . 2020 a. Product image recognition with guidance learning and noisy supervision. Computer Vision and Image Understanding ( 2020 ), 102963. Qing Li, Xiaojiang Peng, Liangliang Cao, Wenbin Du, Hao Xing, Yu Qiao, and Qiang Peng. 2020 a. Product image recognition with guidance learning and noisy supervision. Computer Vision and Image Understanding (2020), 102963."},{"key":"e_1_3_2_1_21_1","volume-title":"Webvision database: Visual learning and understanding from web data. arXiv preprint arXiv:1708.02862","author":"Li Wen","year":"2017","unstructured":"Wen Li , Limin Wang , Wei Li , Eirikur Agustsson , and Luc Van Gool . 2017. Webvision database: Visual learning and understanding from web data. arXiv preprint arXiv:1708.02862 ( 2017 ). Wen Li, Limin Wang, Wei Li, Eirikur Agustsson, and Luc Van Gool. 2017. Webvision database: Visual learning and understanding from web data. arXiv preprint arXiv:1708.02862 (2017)."},{"key":"e_1_3_2_1_22_1","unstructured":"Sheng Liu Jonathan Niles-Weed Narges Razavian and Carlos Fernandez-Granda. 2020. Early-learning regularization prevents memorization of noisy labels. In NeurIPS.  Sheng Liu Jonathan Niles-Weed Narges Razavian and Carlos Fernandez-Granda. 2020. Early-learning regularization prevents memorization of noisy labels. In NeurIPS."},{"key":"e_1_3_2_1_23_1","unstructured":"Xingjun Ma Yisen Wang Michael E Houle Shuo Zhou Sarah M Erfani Shu-Tao Xia Sudanthi Wijewickrema and James Bailey. 2018. Dimensionality-driven learning with noisy labels. In ICML. 3361--3370.  Xingjun Ma Yisen Wang Michael E Houle Shuo Zhou Sarah M Erfani Shu-Tao Xia Sudanthi Wijewickrema and James Bailey. 2018. Dimensionality-driven learning with noisy labels. In ICML. 3361--3370."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2011.63"},{"key":"e_1_3_2_1_25_1","unstructured":"Eran Malach and Shai Shalev-Shwartz. 2017. Decoupling \"when to update\" from \"how to update\". In NeurIPS. 960--970.  Eran Malach and Shai Shalev-Shwartz. 2017. Decoupling \"when to update\" from \"how to update\". In NeurIPS. 960--970."},{"key":"e_1_3_2_1_26_1","volume-title":"Richard Nock, and Lizhen Qu.","author":"Patrini Giorgio","year":"2017","unstructured":"Giorgio Patrini , Alessandro Rozza , Aditya Krishna Menon , Richard Nock, and Lizhen Qu. 2017 . Making deep neural networks robust to label noise: A loss correction approach. In CVPR. 2233--2241. Giorgio Patrini, Alessandro Rozza, Aditya Krishna Menon, Richard Nock, and Lizhen Qu. 2017. Making deep neural networks robust to label noise: A loss correction approach. In CVPR. 2233--2241."},{"key":"e_1_3_2_1_27_1","volume-title":"Weinberger","author":"Pleiss Geoff","year":"2020","unstructured":"Geoff Pleiss , Tianyi Zhang , Ethan R. Elenberg , and Kilian Q . Weinberger . 2020 . Detecting noisy training data with loss curves. https:\/\/openreview.net\/forum?id=HyenUkrtDB Geoff Pleiss, Tianyi Zhang, Ethan R. Elenberg, and Kilian Q. Weinberger. 2020. Detecting noisy training data with loss curves. https:\/\/openreview.net\/forum?id=HyenUkrtDB"},{"key":"e_1_3_2_1_28_1","unstructured":"Scott Reed Honglak Lee Dragomir Anguelov Christian Szegedy Dumitru Erhan and Andrew Rabinovich. 2015. Training deep neural networks on noisy labels with bootstrapping. In ICLR.  Scott Reed Honglak Lee Dragomir Anguelov Christian Szegedy Dumitru Erhan and Andrew Rabinovich. 2015. Training deep neural networks on noisy labels with bootstrapping. In ICLR."},{"key":"e_1_3_2_1_29_1","unstructured":"Mengye Ren Wenyuan Zeng Bin Yang and Raquel Urtasun. 2018. Learning to reweight examples for robust deep learning. In ICML. 4334--4343.  Mengye Ren Wenyuan Zeng Bin Yang and Raquel Urtasun. 2018. Learning to reweight examples for robust deep learning. In ICML. 4334--4343."},{"key":"e_1_3_2_1_30_1","unstructured":"Ohad Shamir and Tong Zhang. 2013. Stochastic gradient descent for non-smooth optimization: Convergence results and optimal averaging schemes. In ICML. 71--79.  Ohad Shamir and Tong Zhang. 2013. Stochastic gradient descent for non-smooth optimization: Convergence results and optimal averaging schemes. In ICML. 71--79."},{"key":"e_1_3_2_1_31_1","unstructured":"Yanyao Shen and Sujay Sanghavi. 2019. Learning with bad training data via iterative trimmed loss minimization. In ICML. 5739--5748.  Yanyao Shen and Sujay Sanghavi. 2019. Learning with bad training data via iterative trimmed loss minimization. In ICML. 5739--5748."},{"key":"e_1_3_2_1_32_1","volume-title":"Meta-weight-net: Learning an explicit mapping for sample weighting. In NeurIPS. 1917--1928.","author":"Shu Jun","year":"2019","unstructured":"Jun Shu , Qi Xie , Lixuan Yi , Qian Zhao , Sanping Zhou , Zongben Xu , and Deyu Meng . 2019 . Meta-weight-net: Learning an explicit mapping for sample weighting. In NeurIPS. 1917--1928. Jun Shu, Qi Xie, Lixuan Yi, Qian Zhao, Sanping Zhou, Zongben Xu, and Deyu Meng. 2019. Meta-weight-net: Learning an explicit mapping for sample weighting. In NeurIPS. 1917--1928."},{"key":"e_1_3_2_1_33_1","volume-title":"SELFIE: Refurbishing unclean samples for robust deep learning. In ICML. 5907--5915.","author":"Song Hwanjun","year":"2019","unstructured":"Hwanjun Song , Minseok Kim , and Jae-Gil Lee . 2019 . SELFIE: Refurbishing unclean samples for robust deep learning. In ICML. 5907--5915. Hwanjun Song, Minseok Kim, and Jae-Gil Lee. 2019. SELFIE: Refurbishing unclean samples for robust deep learning. In ICML. 5907--5915."},{"key":"e_1_3_2_1_34_1","unstructured":"Hwanjun Song Minseok Kim Dongmin Park and Jae-Gil Lee. 2020 a. How does early stopping help generalization against label noise?. In ICMLW.  Hwanjun Song Minseok Kim Dongmin Park and Jae-Gil Lee. 2020 a. How does early stopping help generalization against label noise?. In ICMLW."},{"key":"e_1_3_2_1_35_1","volume-title":"2020 b. Learning from noisy labels with deep neural networks: A survey. arXiv preprint arXiv:2007.08199","author":"Song Hwanjun","year":"2020","unstructured":"Hwanjun Song , Minseok Kim , Dongmin Park , and Jae-Gil Lee . 2020 b. Learning from noisy labels with deep neural networks: A survey. arXiv preprint arXiv:2007.08199 ( 2020 ). Hwanjun Song, Minseok Kim, Dongmin Park, and Jae-Gil Lee. 2020 b. Learning from noisy labels with deep neural networks: A survey. arXiv preprint arXiv:2007.08199 (2020)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"crossref","unstructured":"Christian Szegedy Sergey Ioffe Vincent Vanhoucke and Alexander A Alemi. 2017. Inception-v4 inception-resnet and the impact of residual connections on learning. In AAAI.  Christian Szegedy Sergey Ioffe Vincent Vanhoucke and Alexander A Alemi. 2017. Inception-v4 inception-resnet and the impact of residual connections on learning. In AAAI.","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"e_1_3_2_1_37_1","volume-title":"MSURU: Large scale e-commerce image classification with weakly supervised search data. In KDD. 2518--2526.","author":"Tang Yina","year":"2019","unstructured":"Yina Tang , Fedor Borisyuk , Siddarth Malreddy , Yixuan Li , Yiqun Liu , and Sergey Kirshner . 2019 . MSURU: Large scale e-commerce image classification with weakly supervised search data. In KDD. 2518--2526. Yina Tang, Fedor Borisyuk, Siddarth Malreddy, Yixuan Li, Yiqun Liu, and Sergey Kirshner. 2019. MSURU: Large scale e-commerce image classification with weakly supervised search data. In KDD. 2518--2526."},{"key":"e_1_3_2_1_38_1","unstructured":"Antti Tarvainen and Harri Valpola. 2017. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In NeurIPS. 1195--1204.  Antti Tarvainen and Harri Valpola. 2017. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In NeurIPS. 1195--1204."},{"key":"e_1_3_2_1_39_1","volume-title":"Adam Trischler, Yoshua Bengio, and Geoffrey J Gordon.","author":"Toneva Mariya","year":"2019","unstructured":"Mariya Toneva , Alessandro Sordoni , Remi Tachet des Combes , Adam Trischler, Yoshua Bengio, and Geoffrey J Gordon. 2019 . An empirical study of example forgetting during deep neural network learning. In ICLR. Mariya Toneva, Alessandro Sordoni, Remi Tachet des Combes, Adam Trischler, Yoshua Bengio, and Geoffrey J Gordon. 2019. An empirical study of example forgetting during deep neural network learning. In ICLR."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"crossref","unstructured":"Yisen Wang Weiyang Liu Xingjun Ma James Bailey Hongyuan Zha Le Song and Shu-Tao Xia. 2018. Iterative learning with open-set noisy labels. In CVPR. 8688--8696.  Yisen Wang Weiyang Liu Xingjun Ma James Bailey Hongyuan Zha Le Song and Shu-Tao Xia. 2018. Iterative learning with open-set noisy labels. In CVPR. 8688--8696.","DOI":"10.1109\/CVPR.2018.00906"},{"key":"e_1_3_2_1_41_1","unstructured":"Han Xiao Huang Xiao and Claudia Eckert. 2012. Adversarial label flips attack on support vector machines. In ECAI. 870--875.  Han Xiao Huang Xiao and Claudia Eckert. 2012. Adversarial label flips attack on support vector machines. In ECAI. 870--875."},{"key":"e_1_3_2_1_42_1","unstructured":"Xingrui Yu Bo Han Jiangchao Yao Gang Niu Ivor Tsang and Masashi Sugiyama. 2019. How does disagreement help generalization against label corruption?. In ICML. 7164--7173.  Xingrui Yu Bo Han Jiangchao Yao Gang Niu Ivor Tsang and Masashi Sugiyama. 2019. How does disagreement help generalization against label corruption?. In ICML. 7164--7173."},{"key":"e_1_3_2_1_43_1","unstructured":"Chiyuan Zhang Samy Bengio Moritz Hardt Michael C Mozer and Yoram Singer. 2020 a. Identity Crisis: Memorization and generalization under extreme overparameterization. In ICLR.  Chiyuan Zhang Samy Bengio Moritz Hardt Michael C Mozer and Yoram Singer. 2020 a. Identity Crisis: Memorization and generalization under extreme overparameterization. In ICLR."},{"key":"e_1_3_2_1_44_1","unstructured":"Chiyuan Zhang Samy Bengio Moritz Hardt Benjamin Recht and Oriol Vinyals. 2017. Understanding deep learning requires rethinking generalization. In ICLR.  Chiyuan Zhang Samy Bengio Moritz Hardt Benjamin Recht and Oriol Vinyals. 2017. Understanding deep learning requires rethinking generalization. In ICLR."},{"key":"e_1_3_2_1_45_1","unstructured":"Han Zhang Zizhao Zhang Augustus Odena and Honglak Lee. 2020 b. Consistency regularization for generative adversarial networks. In ICLR.  Han Zhang Zizhao Zhang Augustus Odena and Honglak Lee. 2020 b. Consistency regularization for generative adversarial networks. In ICLR."},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"crossref","unstructured":"Bohan Zhuang Lingqiao Liu Yao Li Chunhua Shen and Ian Reid. 2017. Attend in groups: A weakly-supervised deep learning framework for learning from web data. In CVPR. 1878--1887.  Bohan Zhuang Lingqiao Liu Yao Li Chunhua Shen and Ian Reid. 2017. Attend in groups: A weakly-supervised deep learning framework for learning from web data. In CVPR. 1878--1887.","DOI":"10.1109\/CVPR.2017.311"}],"event":{"name":"KDD '21: The 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Virtual Event Singapore","acronym":"KDD '21","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &amp; Data Mining"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3447548.3467222","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3447548.3467222","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:18:28Z","timestamp":1750191508000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3447548.3467222"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,8,14]]},"references-count":46,"alternative-id":["10.1145\/3447548.3467222","10.1145\/3447548"],"URL":"https:\/\/doi.org\/10.1145\/3447548.3467222","relation":{},"subject":[],"published":{"date-parts":[[2021,8,14]]},"assertion":[{"value":"2021-08-14","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}