{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,8]],"date-time":"2025-10-08T00:31:20Z","timestamp":1759883480029,"version":"build-2065373602"},"publisher-location":"New York, NY, USA","reference-count":55,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,5,8]],"date-time":"2025-05-08T00:00:00Z","timestamp":1746662400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"NSF","award":["IIS-2334193, IIS-2340346, CNS-2426514, and CMMI-2146076"],"award-info":[{"award-number":["IIS-2334193, IIS-2340346, CNS-2426514, and CMMI-2146076"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,5,8]]},"DOI":"10.1145\/3701716.3715858","type":"proceedings-article","created":{"date-parts":[[2025,5,23]],"date-time":"2025-05-23T16:06:11Z","timestamp":1748016371000},"page":"77-80","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Resource-Efficient Learning for the Web"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8349-7926","authenticated-orcid":false,"given":"Chuxu","family":"Zhang","sequence":"first","affiliation":[{"name":"University of Connecticut, Storrs, CT, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6684-6752","authenticated-orcid":false,"given":"Kaize","family":"Ding","sequence":"additional","affiliation":[{"name":"Northwestern University, Evanston, IL, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1878-817X","authenticated-orcid":false,"given":"Jundong","family":"Li","sequence":"additional","affiliation":[{"name":"University of Virginia, Charlottesville, VA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1456-9658","authenticated-orcid":false,"given":"Dongkuan","family":"Xu","sequence":"additional","affiliation":[{"name":"North Carolina State University, Raleigh, NC, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7485-6213","authenticated-orcid":false,"given":"Haoyu","family":"Wang","sequence":"additional","affiliation":[{"name":"University at Albany, Albany, NY, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-7943-8328","authenticated-orcid":false,"given":"Derek Zhiyuan","family":"Cheng","sequence":"additional","affiliation":[{"name":"Google DeepMind, Mountain View, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3264-7904","authenticated-orcid":false,"given":"Huan","family":"Liu","sequence":"additional","affiliation":[{"name":"Arizona State University, Tempe, AZ, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,5,23]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"A survey on data selection for language models. arXiv preprint arXiv:2402.16827","author":"Albalak A.","year":"2024","unstructured":"Albalak, A., Elazar, Y., Xie, S. M., Longpre, S., Lambert, N., Wang, X., Muennighoff, N., Hou, B., Pan, L., Jeong, H., et al. A survey on data selection for language models. arXiv preprint arXiv:2402.16827 (2024)."},{"key":"e_1_3_2_1_2_1","volume-title":"ICML","author":"Chen T.","year":"2021","unstructured":"Chen, T., Sui, Y., Chen, X., Zhang, A., and Wang, Z. A unified lottery ticket hypothesis for graph neural networks. In ICML (2021)."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1137\/1.9781611975673.67"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i6.20605"},{"key":"e_1_3_2_1_5_1","volume-title":"Weakly-supervised graph meta-learning for few-shot node classification. arXiv preprint arXiv:2106.06873","author":"Ding K.","year":"2021","unstructured":"Ding, K., Wang, J., Li, J., Caverlee, J., and Liu, H. Weakly-supervised graph meta-learning for few-shot node classification. arXiv preprint arXiv:2106.06873 (2021)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3411922"},{"key":"e_1_3_2_1_7_1","volume-title":"KDD","author":"Ding K.","year":"2022","unstructured":"Ding, K., Zhang, C., Tang, J., Chawla, N., and Liu, H. Toward graph minimallysupervised learning. In KDD (2022)."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442381.3449922"},{"key":"e_1_3_2_1_9_1","volume-title":"NeurIPS","author":"Fang T.","year":"2023","unstructured":"Fang, T., Zhang, Y., Yang, Y., Wang, C., and Chen, L. Universal prompt tuning for graph neural networks. In NeurIPS (2023)."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i6.25944"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442381.3450112"},{"key":"e_1_3_2_1_12_1","volume-title":"Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149","author":"Han S.","year":"2015","unstructured":"Han, S., Mao, H., and Dally, W. J. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149 (2015)."},{"key":"e_1_3_2_1_13_1","volume-title":"Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531","author":"Hinton G.","year":"2015","unstructured":"Hinton, G. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)."},{"key":"e_1_3_2_1_14_1","unstructured":"Huang S. Xu D. Yen I. E. Wang Y. Chang S.-E. Li B. Chen S. Xie M. Rajasekaran S. Liu H. et al. Sparse progressive distillation: Resolving overfitting under pretrain-and-finetune paradigm."},{"key":"e_1_3_2_1_15_1","volume-title":"ICLR","author":"Ju M.","year":"2023","unstructured":"Ju, M., Zhao, T., Wen, Q., Yu, W., Shah, N., Ye, Y., and Zhang, C. Multi-task self-supervised graph neural networks enable stronger task generalization. In ICLR (2023)."},{"key":"e_1_3_2_1_16_1","volume-title":"Variational graph auto-encoders. arXiv preprint arXiv:1611.07308","author":"Kipf T. N.","year":"2016","unstructured":"Kipf, T. N., and Welling, M. Variational graph auto-encoders. arXiv preprint arXiv:1611.07308 (2016)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.79"},{"key":"e_1_3_2_1_18_1","volume-title":"TMLR","author":"Li J.","year":"2025","unstructured":"Li, J., Li, J., and Zhang, C. Instance-aware graph prompt learning. In TMLR (2025)."},{"key":"e_1_3_2_1_19_1","volume-title":"Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. CSUR","author":"Liu P.","year":"2023","unstructured":"Liu, P., Yuan,W., Fu, J., Jiang, Z., Hayashi, H., and Neubig, G. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. CSUR (2023)."},{"key":"e_1_3_2_1_20_1","volume-title":"Selfsupervised learning: Generative or contrastive. TKDE","author":"Liu X.","year":"2021","unstructured":"Liu, X., Zhang, F., Hou, Z., Mian, L., Wang, Z., Zhang, J., and Tang, J. Selfsupervised learning: Generative or contrastive. TKDE (2021)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i5.16551"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403168"},{"key":"e_1_3_2_1_23_1","volume-title":"Exploring the limits of transfer learning with a unified text-to-text transformer. JMLR","author":"Raffel C.","year":"2020","unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., and Liu, P. J. Exploring the limits of transfer learning with a unified text-to-text transformer. JMLR (2020)."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3640457.3688052"},{"key":"e_1_3_2_1_25_1","volume-title":"How to train data-efficient llms. arXiv preprint arXiv:2402.09668","author":"Sachdeva N.","year":"2024","unstructured":"Sachdeva, N., Coleman, B., Kang, W.-C., Ni, J., Hong, L., Chi, E. H., Caverlee, J., McAuley, J., and Cheng, D. Z. How to train data-efficient llms. arXiv preprint arXiv:2402.09668 (2024)."},{"key":"e_1_3_2_1_26_1","volume-title":"Farzi data: Autoregressive data distillation. arXiv preprint arXiv:2310.09983","author":"Sachdeva N.","year":"2023","unstructured":"Sachdeva, N., He, Z., Kang, W.-C., Ni, J., Cheng, D. Z., and McAuley, J. Farzi data: Autoregressive data distillation. arXiv preprint arXiv:2310.09983 (2023)."},{"key":"e_1_3_2_1_27_1","volume-title":"Data distillation: A survey. arXiv preprint arXiv:2301.04272","author":"Sachdeva N.","year":"2023","unstructured":"Sachdeva, N., and McAuley, J. Data distillation: A survey. arXiv preprint arXiv:2301.04272 (2023)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599256"},{"key":"e_1_3_2_1_29_1","volume-title":"Knowledge distillation on graphs: A survey. arXiv preprint arXiv:2302.00219","author":"Tian Y.","year":"2023","unstructured":"Tian, Y., Pei, S., Zhang, X., Zhang, C., and Chawla, N. V. Knowledge distillation on graphs: A survey. arXiv preprint arXiv:2302.00219 (2023)."},{"key":"e_1_3_2_1_30_1","volume-title":"KDD","author":"Wang H.","year":"2023","unstructured":"Wang, H., Li, R., Jiang, H., Wang, Z., Tang, X., Bi, B., Cheng, M., Yin, B., Wang, Y., Zhao, T., et al. Lighttoken: A task and model-agnostic lightweight token embedding framework for pre-trained language models. In KDD (2023)."},{"key":"e_1_3_2_1_31_1","volume-title":"Roselora: Row and column-wise sparse low-rank adaptation of pre-trained language model for knowledge editing and fine-tuning. arXiv preprint arXiv:2406.10777","author":"Wang H.","year":"2024","unstructured":"Wang, H., Liu, T., Zhao, T., and Gao, J. Roselora: Row and column-wise sparse low-rank adaptation of pre-trained language model for knowledge editing and fine-tuning. arXiv preprint arXiv:2406.10777 (2024)."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.283"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3411923"},{"key":"e_1_3_2_1_34_1","volume-title":"NeurIPS","author":"Wang S.","year":"2022","unstructured":"Wang, S., Chen, C., and Li, J. Graph few-shot learning with task-specific structures. In NeurIPS (2022)."},{"key":"e_1_3_2_1_35_1","volume-title":"Generalizing from a few examples: A survey on few-shot learning. CSUR","author":"Wang Y.","year":"2020","unstructured":"Wang, Y., Yao, Q., Kwok, J. T., and Ni, L. M. Generalizing from a few examples: A survey on few-shot learning. CSUR (2020)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3701551.3703550"},{"key":"e_1_3_2_1_37_1","volume-title":"ICML","author":"Wen Q.","year":"2024","unstructured":"Wen, Q., Ju, M., Ouyang, Z., Zhang, C., and Ye, Y. From coarse to fine: Enable comprehensive graph self-supervised learning with multi-granular semantic ensemble. In ICML (2024)."},{"key":"e_1_3_2_1_38_1","volume-title":"NeurIPS","author":"Wu L.","year":"2022","unstructured":"Wu, L., Lin, H., Huang, Y., and Li, S. Z. Knowledge distillation improves graph structure augmentation for graph neural networks. In NeurIPS (2022)."},{"key":"e_1_3_2_1_39_1","volume-title":"A comprehensive survey on graph neural networks. TNNLS","author":"Wu Z.","year":"2020","unstructured":"Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C., and Philip, S. Y. A comprehensive survey on graph neural networks. TNNLS (2020)."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.188"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.6142"},{"key":"e_1_3_2_1_42_1","volume-title":"NeurIPS","author":"You Y.","year":"2020","unstructured":"You, Y., Chen, T., Sui, Y., Chen, T., Wang, Z., and Shen, Y. Graph contrastive learning with augmentations. In NeurIPS (2020)."},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2022\/789"},{"key":"e_1_3_2_1_44_1","volume-title":"ICML","author":"Zhang C.","year":"2023","unstructured":"Zhang, C., Huang, C., Tian, Y., Wen, Q., Ouyang, Z., Li, Y., Ye, Y., and Zhang, C. When sparsity meets contrastive models: less graph data can bring better class-balanced representations. In ICML (2023)."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3470813"},{"key":"e_1_3_2_1_46_1","volume-title":"Mind the gap: Mitigating the distribution gap in graph few-shot learning. TMLR","author":"Zhang C.","year":"2023","unstructured":"Zhang, C., Liu, H., Li, J., Ye, Y., and Zhang, C. Mind the gap: Mitigating the distribution gap in graph few-shot learning. TMLR (2023)."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330961"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671487"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i03.5698"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.51"},{"key":"e_1_3_2_1_51_1","volume-title":"ICLR","author":"Zhang S.","year":"2022","unstructured":"Zhang, S., Liu, Y., Sun, Y., and Shah, N. Graph-less neural networks: Teaching old mlps new tricks via distillation. In ICLR (2022)."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1137\/1.9781611977172.75"},{"key":"e_1_3_2_1_53_1","volume-title":"A survey of large language models. arXiv preprint arXiv:2303.18223","author":"Zhao W. X.","year":"2023","unstructured":"Zhao, W. X., Zhou, K., Li, J., Tang, T., Wang, X., Hou, Y., Min, Y., Zhang, B., Zhang, J., Dong, Z., et al. A survey of large language models. arXiv preprint arXiv:2303.18223 (2023)."},{"key":"e_1_3_2_1_54_1","volume-title":"A brief introduction to weakly supervised learning. National Science Review","author":"Zhou Z.-H.","year":"2018","unstructured":"Zhou, Z.-H. A brief introduction to weakly supervised learning. National Science Review (2018)."},{"key":"e_1_3_2_1_55_1","volume-title":"A survey on model compression for large language models. arXiv preprint arXiv:2308.07633","author":"Zhu X.","year":"2023","unstructured":"Zhu, X., Li, J., Liu, Y., Ma, C., and Wang, W. A survey on model compression for large language models. arXiv preprint arXiv:2308.07633 (2023)."}],"event":{"name":"WWW '25: The ACM Web Conference 2025","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Sydney NSW Australia","acronym":"WWW '25"},"container-title":["Companion Proceedings of the ACM on Web Conference 2025"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3701716.3715858","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3701716.3715858","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,7]],"date-time":"2025-10-07T17:51:11Z","timestamp":1759859471000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3701716.3715858"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,8]]},"references-count":55,"alternative-id":["10.1145\/3701716.3715858","10.1145\/3701716"],"URL":"https:\/\/doi.org\/10.1145\/3701716.3715858","relation":{},"subject":[],"published":{"date-parts":[[2025,5,8]]},"assertion":[{"value":"2025-05-23","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}