{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,6]],"date-time":"2025-08-06T12:01:01Z","timestamp":1754481661361,"version":"3.40.3"},"publisher-location":"Cham","reference-count":19,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031439865"},{"type":"electronic","value":"9783031439872"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-43987-2_8","type":"book-chapter","created":{"date-parts":[[2023,9,30]],"date-time":"2023-09-30T23:07:48Z","timestamp":1696115268000},"page":"75-84","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Gradient and\u00a0Feature Conformity-Steered Medical Image Classification with\u00a0Noisy Labels"],"prefix":"10.1007","author":[{"given":"Xiaohan","family":"Xing","sequence":"first","affiliation":[]},{"given":"Zhen","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Zhifan","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Yixuan","family":"Yuan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,10,1]]},"reference":[{"issue":"3","key":"8_CR1","doi-asserted-by":"publisher","first-page":"107","DOI":"10.1145\/3446776","volume":"64","author":"C Zhang","year":"2021","unstructured":"Zhang, C., Bengio, S., Hardt, M., Recht, B., Vinyals, O.: Understanding deep learning (still) requires rethinking generalization. Commun. ACM 64(3), 107\u2013115 (2021)","journal-title":"Commun. ACM"},{"key":"8_CR2","first-page":"20331","volume":"33","author":"S Liu","year":"2020","unstructured":"Liu, S., Niles-Weed, J., Razavian, N., Fernandez-Granda, C.: Early-learning regularization prevents memorization of noisy labels. NeurIPS 33, 20331\u201320342 (2020)","journal-title":"NeurIPS"},{"key":"8_CR3","unstructured":"Li, J., Socher, R., Hoi, S.C.H.: DivideMix: learning with noisy labels as semi-supervised learning. arXiv preprint arXiv:2002.07394 (2020)"},{"key":"8_CR4","unstructured":"Han, B., et al.: Co-teaching: robust training of deep neural networks with extremely noisy labels. In: NeurIPS, vol. 31 (2018)"},{"key":"8_CR5","doi-asserted-by":"crossref","unstructured":"Bai, Y., Liu, T.: Me-momentum: extracting hard confident examples from noisily labeled data. In: ICCV, pp. 9312\u20139321 (2021)","DOI":"10.1109\/ICCV48922.2021.00918"},{"key":"8_CR6","doi-asserted-by":"crossref","unstructured":"Li, J., Li, G., Liu, F., Yu, Y.: Neighborhood collective estimation for noisy label identification and correction. arXiv preprint arXiv:2208.03207 (2022)","DOI":"10.1007\/978-3-031-20053-3_8"},{"key":"8_CR7","doi-asserted-by":"publisher","unstructured":"Zhao, G., Li, G., Qin, Y., Liu, F., Yu, Y.: Centrality and consistency: two-stage clean samples identification for learning with instance-dependent noisy labels. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision - ECCV 2022. ECCV 2022. LNCS, vol. 13685, pp 21\u201337. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19806-9_2","DOI":"10.1007\/978-3-031-19806-9_2"},{"key":"8_CR8","doi-asserted-by":"crossref","unstructured":"Tang, H., Jia, K.: Towards discovering the effectiveness of moderately confident samples for semi-supervised learning. In: CVPR, pp. 14658\u201314667 (2022)","DOI":"10.1109\/CVPR52688.2022.01425"},{"key":"8_CR9","doi-asserted-by":"crossref","unstructured":"Iscen, A., Valmadre, J., Arnab, A., Schmid, C.: Learning with neighbor consistency for noisy labels. In: CVPR, pp. 4672\u20134681 (2022)","DOI":"10.1109\/CVPR52688.2022.00463"},{"issue":"4","key":"8_CR10","doi-asserted-by":"publisher","first-page":"881","DOI":"10.1109\/TMI.2021.3125459","volume":"41","author":"C Zhu","year":"2021","unstructured":"Zhu, C., Chen, W., Peng, T., Wang, Y., Jin, M.: Hard sample aware noise robust learning for histopathology image classification. IEEE Trans. Med. Imaging 41(4), 881\u2013894 (2021)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"8_CR11","unstructured":"Berthelot, D., Carlini, N., Goodfellow, I., Papernot, N., Oliver, A., Raffel, C.A.:. MixMatch: a holistic approach to semi-supervised learning. In: NeurIPS, vol. 32 (2019)"},{"key":"8_CR12","unstructured":"Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: beyond empirical risk minimization. arXiv preprint arXiv:1710.09412 (2017)"},{"key":"8_CR13","unstructured":"Jiang, L., Zhou, Z., Leung, T., Li, L.J., Fei-Fei, L.: MentorNet: learning data-driven curriculum for very deep neural networks on corrupted labels. In: ICML, pp. 2304\u20132313. PMLR (2018)"},{"key":"8_CR14","unstructured":"Yu, X., Han, B., Yao, J., Niu, G., Tsang, I., Sugiyama, M.: How does disagreement help generalization against label corruption? In: ICML, pp. 7164\u20137173. PMLR (2019)"},{"key":"8_CR15","unstructured":"Arpit, D., et al.: A closer look at memorization in deep networks. In: ICML, pp. 233\u2013242. PMLR (2017)"},{"key":"8_CR16","doi-asserted-by":"crossref","unstructured":"Dray, X., et al.: Cad-cap: UNE base de donn\u00e9es fran\u00e7aise \u00e0 vocation internationale, pour le d\u00e9veloppement et la validation d\u2019outils de diagnostic assist\u00e9 par ordinateur en vid\u00e9ocapsule endoscopique du gr\u00eale. Endoscopy 50(03), 000441 (2018)","DOI":"10.1055\/s-0038-1623358"},{"key":"8_CR17","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"8_CR18","unstructured":"Paszke, A., et al.: PyTorch: an imperative style, high-performance deep learning library. In: NeurIPS, vol. 32 (2019)"},{"key":"8_CR19","doi-asserted-by":"crossref","unstructured":"Wei, Q., Sun, H., Lu, X., Yin, Y.: Self-filtering: a noise-aware sample selection for label noise with confidence penalization. arXiv preprint arXiv:2208.11351 (2022)","DOI":"10.1007\/978-3-031-20056-4_30"}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2023"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-43987-2_8","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,11]],"date-time":"2024-03-11T15:28:27Z","timestamp":1710170907000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-43987-2_8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031439865","9783031439872"],"references-count":19,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-43987-2_8","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"1 October 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vancouver, BC","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Canada","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/conferences.miccai.org\/2023\/en\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2250","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"730","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"32% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}