{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:14:36Z","timestamp":1750220076409,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":78,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,4,30]],"date-time":"2023-04-30T00:00:00Z","timestamp":1682812800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,4,30]]},"DOI":"10.1145\/3543507.3583211","type":"proceedings-article","created":{"date-parts":[[2023,4,26]],"date-time":"2023-04-26T23:30:25Z","timestamp":1682551825000},"page":"2011-2021","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Curriculum Graph Poisoning"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7415-6839","authenticated-orcid":false,"given":"Hanwen","family":"Liu","sequence":"first","affiliation":[{"name":"Peking University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8543-3953","authenticated-orcid":false,"given":"Peilin","family":"Zhao","sequence":"additional","affiliation":[{"name":"Tencent AI Lab, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8487-9045","authenticated-orcid":false,"given":"Tingyang","family":"Xu","sequence":"additional","affiliation":[{"name":"Tencent AI Lab, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2368-4084","authenticated-orcid":false,"given":"Yatao","family":"Bian","sequence":"additional","affiliation":[{"name":"Tencent AI Lab, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9548-1227","authenticated-orcid":false,"given":"Junzhou","family":"Huang","sequence":"additional","affiliation":[{"name":"University of Texas at Arlington, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2524-6800","authenticated-orcid":false,"given":"Yuesheng","family":"Zhu","sequence":"additional","affiliation":[{"name":"Peking University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7815-3750","authenticated-orcid":false,"given":"Yadong","family":"Mu","sequence":"additional","affiliation":[{"name":"Peking University, China"}]}],"member":"320","published-online":{"date-parts":[[2023,4,30]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"[1] Yingbin Bai Erkun Yang Bo Han Yanhua Yang Jiatong Li Yinian Mao Gang Niu and Tongliang Liu. 2021. Understanding and Improving Early Stopping for Learning with Noisy Labels. In NeurIPS. 24392\u201324403."},{"key":"e_1_3_2_1_2_1","unstructured":"[2] Aseem Baranwal Kimon Fountoulakis and Aukosh Jagannath. 2021. Graph Convolution for Semi-Supervised Classification: Improved Linear Separability and Out-of-Distribution Generalization. In ICML(Proceedings of Machine Learning Research Vol.\u00a0139). PMLR 684\u2013693."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/1553374.1553380"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1137\/0217014"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2018.07.023"},{"key":"e_1_3_2_1_6_1","unstructured":"[6] Aleksandar Bojchevski and Stephan G\u00fcnnemann. 2018. Deep Gaussian Embedding of Graphs: Unsupervised Inductive Learning via Ranking. In ICLR (Poster). OpenReview.net."},{"key":"e_1_3_2_1_7_1","unstructured":"[7] Aleksandar Bojchevski and Stephan G\u00fcnnemann. 2019. Adversarial Attacks on Node Embeddings via Graph Poisoning. In ICML(Proceedings of Machine Learning Research Vol.\u00a097). PMLR 695\u2013704."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"crossref","unstructured":"[8] Aleksandar Bojchevski Johannes Klicpera Bryan Perozzi Amol Kapoor Martin Blais Benedek R\u00f3zemberczki Michal Lukasik and Stephan G\u00fcnnemann. 2020. Scaling Graph Neural Networks with Approximate PageRank. In KDD. ACM 2464\u20132473.","DOI":"10.1145\/3394486.3403296"},{"key":"e_1_3_2_1_9_1","volume-title":"CoRR abs\/1606.01540","author":"Brockman Greg","year":"2016","unstructured":"[9] Greg Brockman, Vicki Cheung, Ludwig Pettersson, Jonas Schneider, John Schulman, Jie Tang, and Wojciech Zaremba. 2016. OpenAI Gym. CoRR abs\/1606.01540 (2016)."},{"key":"e_1_3_2_1_10_1","unstructured":"[10] Heng Chang Yu Rong Tingyang Xu Yatao Bian Shiji Zhou Xin Wang Junzhou Huang and Wenwu Zhu. 2021. Not All Low-Pass Filters are Robust in Graph Convolutional Networks. In NeurIPS. 25058\u201325071."},{"key":"e_1_3_2_1_11_1","unstructured":"[11] Hong Chen Yudong Chen Xin Wang Ruobing Xie Rui Wang Feng Xia and Wenwu Zhu. 2021. Curriculum Disentangled Recommendation with Noisy Multi-feedback. In NeurIPS. 26924\u201326936."},{"key":"e_1_3_2_1_12_1","unstructured":"[12] Jiefeng Chen Frederick Liu Besim Avci Xi Wu Yingyu Liang and Somesh Jha. 2021. Detecting Errors and Estimating Accuracy on Unlabeled Data with Self-training Ensembles. In NeurIPS. 14980\u201314992."},{"key":"e_1_3_2_1_13_1","unstructured":"[13] Jie Chen Tengfei Ma and Cao Xiao. 2018. FastGCN: Fast Learning with Graph Convolutional Networks via Importance Sampling. In ICLR (Poster). OpenReview.net."},{"key":"e_1_3_2_1_14_1","unstructured":"[14] Ming Chen Zhewei Wei Zengfeng Huang Bolin Ding and Yaliang Li. 2020. Simple and Deep Graph Convolutional Networks. In ICML(Proceedings of Machine Learning Research Vol.\u00a0119). PMLR 1725\u20131735."},{"key":"e_1_3_2_1_15_1","unstructured":"[15] Yongqiang Chen Han Yang Yonggang Zhang Kaili Ma Tongliang Liu Bo Han and James Cheng. 2022. Understanding and Improving Graph Injection Attack by Promoting Unnoticeability. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_16_1","unstructured":"[16] Hanjun Dai Hui Li Tian Tian Xin Huang Lin Wang Jun Zhu and Le Song. 2018. Adversarial Attack on Graph Structured Data. In ICML(Proceedings of Machine Learning Research Vol.\u00a080). PMLR 1123\u20131132."},{"key":"e_1_3_2_1_17_1","unstructured":"[17] Minghong Fang Neil\u00a0Zhenqiang Gong and Jia Liu. 2020. Influence Function based Data Poisoning Attacks to Top-N Recommender Systems. In WWW. ACM \/ IW3C2 3019\u20133025."},{"key":"e_1_3_2_1_18_1","unstructured":"[18] Minghong Fang Minghao Sun Qi Li Neil\u00a0Zhenqiang Gong Jin Tian and Jia Liu. 2021. Data Poisoning Attacks and Defenses to Crowdsourcing Systems. In WWW. ACM \/ IW3C2 969\u2013980."},{"key":"e_1_3_2_1_19_1","unstructured":"[19] Chelsea Finn Pieter Abbeel and Sergey Levine. 2017. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks. In ICML(Proceedings of Machine Learning Research Vol.\u00a070). PMLR 1126\u20131135."},{"key":"e_1_3_2_1_20_1","volume-title":"Witches\u2019 Brew: Industrial Scale Data Poisoning via Gradient Matching. In ICLR. OpenReview.net.","author":"Geiping Jonas","year":"2021","unstructured":"[20] Jonas Geiping, Liam\u00a0H. Fowl, W.\u00a0Ronny Huang, Wojciech Czaja, Gavin Taylor, Michael Moeller, and Tom Goldstein. 2021. Witches\u2019 Brew: Industrial Scale Data Poisoning via Gradient Matching. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_21_1","volume-title":"Robustness of Graph Neural Networks at Scale. CoRR abs\/2110.14038","author":"Geisler Simon","year":"2021","unstructured":"[21] Simon Geisler, Tobias Schmidt, Hakan Sirin, Daniel Z\u00fcgner, Aleksandar Bojchevski, and Stephan G\u00fcnnemann. 2021. Robustness of Graph Neural Networks at Scale. CoRR abs\/2110.14038 (2021)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"crossref","unstructured":"[22] C.\u00a0Lee Giles Kurt\u00a0D. Bollacker and Steve Lawrence. 1998. CiteSeer: An Automatic Citation Indexing System. In ACM DL. ACM 89\u201398.","DOI":"10.1145\/276675.276685"},{"key":"e_1_3_2_1_23_1","unstructured":"[23] Ian\u00a0J. Goodfellow Jonathon Shlens and Christian Szegedy. 2015. Explaining and Harnessing Adversarial Examples. In ICLR (Poster)."},{"key":"e_1_3_2_1_24_1","unstructured":"[24] Guy Hacohen and Daphna Weinshall. 2019. On The Power of Curriculum Learning in Training Deep Networks. In ICML(Proceedings of Machine Learning Research Vol.\u00a097). PMLR 2535\u20132544."},{"key":"e_1_3_2_1_25_1","unstructured":"[25] William\u00a0L. Hamilton Zhitao Ying and Jure Leskovec. 2017. Inductive Representation Learning on Large Graphs. In NIPS. 1024\u20131034."},{"key":"e_1_3_2_1_26_1","volume-title":"Rainbow: Combining Improvements in Deep Reinforcement Learning","author":"Hessel Matteo","year":"2018","unstructured":"[26] Matteo Hessel, Joseph Modayil, Hado van Hasselt, Tom Schaul, Georg Ostrovski, Will Dabney, Dan Horgan, Bilal Piot, Mohammad\u00a0Gheshlaghi Azar, and David Silver. 2018. Rainbow: Combining Improvements in Deep Reinforcement Learning. In AAAI. AAAI Press, 3215\u20133222."},{"key":"e_1_3_2_1_27_1","unstructured":"[27] Weihua Hu Matthias Fey Marinka Zitnik Yuxiao Dong Hongyu Ren Bowen Liu Michele Catasta and Jure Leskovec. 2020. Open Graph Benchmark: Datasets for Machine Learning on Graphs. In NeurIPS."},{"key":"e_1_3_2_1_28_1","unstructured":"[28] W.\u00a0Ronny Huang Jonas Geiping Liam Fowl Gavin Taylor and Tom Goldstein. 2020. MetaPoison: Practical General-purpose Clean-label Data Poisoning. In NeurIPS."},{"key":"e_1_3_2_1_29_1","volume-title":"Adversarial Attacks and Defenses on Graphs: A Review and Empirical Study. CoRR abs\/2003.00653","author":"Jin Wei","year":"2020","unstructured":"[29] Wei Jin, Yaxin Li, Han Xu, Yiqi Wang, and Jiliang Tang. 2020. Adversarial Attacks and Defenses on Graphs: A Review and Empirical Study. CoRR abs\/2003.00653 (2020)."},{"key":"e_1_3_2_1_30_1","volume-title":"Kipf and Max Welling","author":"N.","year":"2017","unstructured":"[30] Thomas\u00a0N. Kipf and Max Welling. 2017. Semi-Supervised Classification with Graph Convolutional Networks. In ICLR (Poster). OpenReview.net."},{"volume-title":"Deeper Insights Into Graph Convolutional Networks for Semi-Supervised Learning","author":"Li Qimai","key":"e_1_3_2_1_31_1","unstructured":"[31] Qimai Li, Zhichao Han, and Xiao-Ming Wu. 2018. Deeper Insights Into Graph Convolutional Networks for Semi-Supervised Learning. In AAAI. AAAI Press, 3538\u20133545."},{"volume-title":"Exploratory Adversarial Attacks on Graph Neural Networks","author":"Lin Xixun","key":"e_1_3_2_1_32_1","unstructured":"[32] Xixun Lin, Chuan Zhou, Hong Yang, Jia Wu, Haibo Wang, Yanan Cao, and Bin Wang. 2020. Exploratory Adversarial Attacks on Graph Neural Networks. In ICDM. IEEE, 1136\u20131141."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"crossref","unstructured":"[33] Can Liu Li Sun Xiang Ao Jinghua Feng Qing He and Hao Yang. 2021. Intention-aware Heterogeneous Graph Attention Networks for Fraud Transactions Detection. In KDD. ACM 3280\u20133288.","DOI":"10.1145\/3447548.3467142"},{"key":"e_1_3_2_1_34_1","unstructured":"[34] Xuanqing Liu Si Si Jerry Zhu Yang Li and Cho-Jui Hsieh. 2019. A Unified Framework for Data Poisoning Attack to Graph-based Semi-supervised Learning. In NeurIPS. 9777\u20139787."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"crossref","unstructured":"[35] Zemin Liu Qiheng Mao Chenghao Liu Yuan Fang and Jianling Sun. 2022. On Size-Oriented Long-Tailed Graph Classification of Graph Neural Networks. In WWW. ACM 1506\u20131516.","DOI":"10.1145\/3485447.3512197"},{"key":"e_1_3_2_1_36_1","unstructured":"[36] Roi Livni Shai Shalev-Shwartz and Ohad Shamir. 2014. On the Computational Efficiency of Training Neural Networks. In NIPS. 855\u2013863."},{"volume-title":"Data Classification: Algorithms and Applications","author":"London Ben","key":"e_1_3_2_1_37_1","unstructured":"[37] Ben London and Lise Getoor. 2014. Collective Classification of Network Data. In Data Classification: Algorithms and Applications. CRC Press, 399\u2013416."},{"key":"e_1_3_2_1_38_1","unstructured":"[38] Jiaqi Ma Shuangrui Ding and Qiaozhu Mei. 2020. Towards More Practical Adversarial Attacks on Graph Neural Networks. In NeurIPS."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"crossref","unstructured":"[39] Yao Ma Suhang Wang Tyler Derr Lingfei Wu and Jiliang Tang. 2021. Graph Adversarial Attack via Rewiring. In KDD. ACM 1161\u20131169.","DOI":"10.1145\/3447548.3467416"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1023\/A:1009953814988"},{"volume-title":"PyTorch: An Imperative Style","author":"Paszke Adam","key":"e_1_3_2_1_41_1","unstructured":"[41] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas K\u00f6pf, Edward\u00a0Z. Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In NeurIPS. 8024\u20138035."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.5555\/2627435.2627446"},{"key":"e_1_3_2_1_43_1","volume-title":"Proximal Policy Optimization Algorithms. CoRR abs\/1707.06347","author":"Schulman John","year":"2017","unstructured":"[43] John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. 2017. Proximal Policy Optimization Algorithms. CoRR abs\/1707.06347 (2017)."},{"key":"e_1_3_2_1_44_1","unstructured":"[44] Avi Schwarzschild Micah Goldblum Arjun Gupta John\u00a0P. Dickerson and Tom Goldstein. 2021. Just How Toxic is Data Poisoning? A Unified Benchmark for Backdoor and Data Poisoning Attacks. In ICML(Proceedings of Machine Learning Research Vol.\u00a0139). PMLR 9389\u20139398."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1609\/aimag.v29i3.2157"},{"key":"e_1_3_2_1_46_1","unstructured":"[46] Ali Shafahi W.\u00a0Ronny Huang Mahyar Najibi Octavian Suciu Christoph Studer Tudor Dumitras and Tom Goldstein. 2018. Poison Frogs! Targeted Clean-Label Poisoning Attacks on Neural Networks. In NeurIPS. 6106\u20136116."},{"key":"e_1_3_2_1_47_1","volume-title":"Pitfalls of Graph Neural Network Evaluation. CoRR abs\/1811.05868","author":"Shchur Oleksandr","year":"2018","unstructured":"[47] Oleksandr Shchur, Maximilian Mumme, Aleksandar Bojchevski, and Stephan G\u00fcnnemann. 2018. Pitfalls of Graph Neural Network Evaluation. CoRR abs\/1811.05868 (2018)."},{"key":"e_1_3_2_1_48_1","unstructured":"[48] Samarth Sinha Animesh Garg and Hugo Larochelle. 2020. Curriculum By Smoothing. In NeurIPS."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"crossref","unstructured":"[49] Yiwei Sun Suhang Wang Xianfeng Tang Tsung-Yu Hsieh and Vasant\u00a0G. Honavar. 2020. Adversarial Attacks on Graph Neural Networks via Node Injections: A Hierarchical Reinforcement Learning Approach. In WWW. ACM \/ IW3C2 673\u2013683.","DOI":"10.1145\/3366423.3380149"},{"key":"e_1_3_2_1_50_1","unstructured":"[50] Fnu Suya Saeed Mahloujifar Anshuman Suri David Evans and Yuan Tian. 2021. Model-Targeted Poisoning Attacks with Provable Convergence. In ICML(Proceedings of Machine Learning Research Vol.\u00a0139). PMLR 10000\u201310010."},{"key":"e_1_3_2_1_51_1","unstructured":"[51] Kai\u00a0Sheng Tai Peter Bailis and Gregory Valiant. 2021. Sinkhorn Label Allocation: Semi-Supervised Classification via Annealed Self-Training. In ICML(Proceedings of Machine Learning Research Vol.\u00a0139). PMLR 10065\u201310075."},{"key":"e_1_3_2_1_52_1","unstructured":"[52] Petar Velickovic Guillem Cucurull Arantxa Casanova Adriana Romero Pietro Li\u00f2 and Yoshua Bengio. 2018. Graph Attention Networks. In ICLR (Poster). OpenReview.net."},{"key":"e_1_3_2_1_53_1","volume-title":"CAPTCHA: Using Hard AI Problems for Security. In EUROCRYPT(Lecture Notes in Computer Science, Vol.\u00a02656)","author":"von Ahn Luis","year":"2003","unstructured":"[53] Luis von Ahn, Manuel Blum, Nicholas\u00a0J. Hopper, and John Langford. 2003. CAPTCHA: Using Hard AI Problems for Security. In EUROCRYPT(Lecture Notes in Computer Science, Vol.\u00a02656). Springer, 294\u2013311."},{"key":"e_1_3_2_1_54_1","volume-title":"Adversarial Attacks on Graph Classification via Bayesian Optimisation. CoRR abs\/2111.02842","author":"Wan Xingchen","year":"2021","unstructured":"[54] Xingchen Wan, Henry Kenlay, Binxin Ru, Arno Blaas, Michael\u00a0A. Osborne, and Xiaowen Dong. 2021. Adversarial Attacks on Graph Classification via Bayesian Optimisation. CoRR abs\/2111.02842 (2021)."},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10618-020-00696-7"},{"key":"e_1_3_2_1_56_1","unstructured":"[56] Yunjuan Wang Poorya Mianjy and Raman Arora. 2021. Robust Learning for Data Poisoning Attacks. In ICML(Proceedings of Machine Learning Research Vol.\u00a0139). PMLR 10859\u201310869."},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41562-017-0290-3"},{"key":"e_1_3_2_1_58_1","unstructured":"[58] Colin Wei Kendrick Shen Yining Chen and Tengyu Ma. 2021. Theoretical Analysis of Self-Training with Deep Networks on Unlabeled Data. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_59_1","unstructured":"[59] Qitian Wu Hengrui Zhang Junchi Yan and David Wipf. 2022. Handling Distribution Shifts on Graphs: An Invariance Perspective. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_60_1","volume-title":"Graph Backdoor. In USENIX Security Symposium. USENIX Association, 1523\u20131540","author":"Xi Zhaohan","year":"2021","unstructured":"[60] Zhaohan Xi, Ren Pang, Shouling Ji, and Ting Wang. 2021. Graph Backdoor. In USENIX Security Symposium. USENIX Association, 1523\u20131540."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"crossref","unstructured":"[61] Chang Xu Jun Wang Yuqing Tang Francisco Guzm\u00e1n Benjamin I.\u00a0P. Rubinstein and Trevor Cohn. 2021. A Targeted Attack on Black-Box Neural Machine Translation with Parallel Data Poisoning. In WWW. ACM \/ IW3C2 3638\u20133650.","DOI":"10.1145\/3442381.3450034"},{"key":"e_1_3_2_1_62_1","unstructured":"[62] Keyulu Xu Weihua Hu Jure Leskovec and Stefanie Jegelka. 2019. How Powerful are Graph Neural Networks?. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"crossref","unstructured":"[63] Zhe Xu Boxin Du and Hanghang Tong. 2022. Graph Sanitation with Application to Node Classification. In WWW. ACM 1136\u20131147.","DOI":"10.1145\/3485447.3512180"},{"key":"e_1_3_2_1_64_1","unstructured":"[64] Zhitao Ying Jiaxuan You Christopher Morris Xiang Ren William\u00a0L. Hamilton and Jure Leskovec. 2018. Hierarchical Graph Representation Learning with Differentiable Pooling. In NeurIPS. 4805\u20134815."},{"key":"e_1_3_2_1_65_1","unstructured":"[65] Jaemin Yoo Sooyeon Shim and U Kang. 2022. Model-Agnostic Augmentation for Accurate Graph Classification. In WWW. ACM 1281\u20131291."},{"key":"e_1_3_2_1_66_1","unstructured":"[66] Bowen Zhang Yidong Wang Wenxin Hou Hao Wu Jindong Wang Manabu Okumura and Takahiro Shinozaki. 2021. FlexMatch: Boosting Semi-Supervised Learning with Curriculum Pseudo Labeling. In NeurIPS. 18408\u201318419."},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"crossref","unstructured":"[67] Hengtong Zhang Yaliang Li Bolin Ding and Jing Gao. 2020. Practical Data Poisoning Attack against Next-Item Recommendation. In WWW. ACM \/ IW3C2 2458\u20132464.","DOI":"10.1145\/3366423.3379992"},{"volume-title":"Adversarial Label-Flipping Attack and Defense for Graph Neural Networks","author":"Zhang Mengmei","key":"e_1_3_2_1_68_1","unstructured":"[68] Mengmei Zhang, Linmei Hu, Chuan Shi, and Xiao Wang. 2020. Adversarial Label-Flipping Attack and Defense for Graph Neural Networks. In ICDM. IEEE, 791\u2013800."},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"crossref","unstructured":"[69] Sixiao Zhang Hongxu Chen Xiangguo Sun Yicong Li and Guandong Xu. 2022. Unsupervised Graph Poisoning Attack via Contrastive Loss Back-propagation. In WWW. ACM 1322\u20131330.","DOI":"10.1145\/3485447.3512179"},{"key":"e_1_3_2_1_70_1","unstructured":"[70] Zijie Zhang Zeru Zhang Yang Zhou Yelong Shen Ruoming Jin and Dejing Dou. 2020. Adversarial Attacks on Deep Graph Matching. In NeurIPS."},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"crossref","unstructured":"[71] Tianxiang Zhao Xiang Zhang and Suhang Wang. 2022. Exploring Edge Disentanglement for Node Classification. In WWW. ACM 1028\u20131036.","DOI":"10.1145\/3485447.3511929"},{"key":"e_1_3_2_1_72_1","unstructured":"[72] Tianyi Zhou Shengjie Wang and Jeff\u00a0A. Bilmes. 2020. Curriculum Learning by Dynamic Instance Hardness. In NeurIPS."},{"key":"e_1_3_2_1_73_1","unstructured":"[73] Dingyuan Zhu Ziwei Zhang Peng Cui and Wenwu Zhu. 2019. Robust Graph Convolutional Networks Against Adversarial Attacks. In KDD. ACM 1399\u20131407."},{"key":"e_1_3_2_1_74_1","unstructured":"[74] Jiong Zhu Yujun Yan Lingxiao Zhao Mark Heimann Leman Akoglu and Danai Koutra. 2020. Beyond Homophily in Graph Neural Networks: Current Limitations and Effective Designs. In NeurIPS."},{"key":"e_1_3_2_1_75_1","volume-title":"TDGIA: Effective Injection Attacks on Graph Neural Networks. In KDD. ACM, 2461\u20132471.","author":"Zou Xu","year":"2021","unstructured":"[75] Xu Zou, Qinkai Zheng, Yuxiao Dong, Xinyu Guan, Evgeny Kharlamov, Jialiang Lu, and Jie Tang. 2021. TDGIA: Effective Injection Attacks on Graph Neural Networks. In KDD. ACM, 2461\u20132471."},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"crossref","unstructured":"[76] Daniel Z\u00fcgner Amir Akbarnejad and Stephan G\u00fcnnemann. 2018. Adversarial Attacks on Neural Networks for Graph Data. In KDD. ACM 2847\u20132856.","DOI":"10.1145\/3219819.3220078"},{"key":"e_1_3_2_1_77_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394520"},{"key":"e_1_3_2_1_78_1","doi-asserted-by":"crossref","unstructured":"[78] Daniel Z\u00fcgner and Stephan G\u00fcnnemann. 2019. Adversarial Attacks on Graph Neural Networks via Meta Learning. In ICLR (Poster). OpenReview.net.","DOI":"10.24963\/ijcai.2019\/872"}],"event":{"name":"WWW '23: The ACM Web Conference 2023","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Austin TX USA","acronym":"WWW '23"},"container-title":["Proceedings of the ACM Web Conference 2023"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3543507.3583211","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3543507.3583211","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T18:09:18Z","timestamp":1750183758000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3543507.3583211"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,30]]},"references-count":78,"alternative-id":["10.1145\/3543507.3583211","10.1145\/3543507"],"URL":"https:\/\/doi.org\/10.1145\/3543507.3583211","relation":{},"subject":[],"published":{"date-parts":[[2023,4,30]]},"assertion":[{"value":"2023-04-30","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}