{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T00:19:45Z","timestamp":1759969185936,"version":"build-2065373602"},"publisher-location":"New York, NY, USA","reference-count":43,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,5,8]],"date-time":"2025-05-08T00:00:00Z","timestamp":1746662400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"NSF","award":["IIS-2334193, IIS-2340346, CNS-2426514, and CMMI-2146076"],"award-info":[{"award-number":["IIS-2334193, IIS-2340346, CNS-2426514, and CMMI-2146076"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,5,8]]},"DOI":"10.1145\/3701716.3717558","type":"proceedings-article","created":{"date-parts":[[2025,5,23]],"date-time":"2025-05-23T16:12:56Z","timestamp":1748016776000},"page":"2468-2471","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["RelWeb 2025: The International Workshop on Resource-Efficient Learning for the Web"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8349-7926","authenticated-orcid":false,"given":"Chuxu","family":"Zhang","sequence":"first","affiliation":[{"name":"University of Connecticut, Storrs, CT, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6684-6752","authenticated-orcid":false,"given":"Kaize","family":"Ding","sequence":"additional","affiliation":[{"name":"Northwestern University, Evanston, IL, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1878-817X","authenticated-orcid":false,"given":"Jundong","family":"Li","sequence":"additional","affiliation":[{"name":"University of Virginia, Charlottesville, VA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1456-9658","authenticated-orcid":false,"given":"Dongkuan","family":"Xu","sequence":"additional","affiliation":[{"name":"North Carolina State University, Raleigh, NC, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7485-6213","authenticated-orcid":false,"given":"Haoyu","family":"Wang","sequence":"additional","affiliation":[{"name":"University at Albany, Albany, NY, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-7943-8328","authenticated-orcid":false,"given":"Derek Zhiyuan","family":"Cheng","sequence":"additional","affiliation":[{"name":"Google DeepMind, Mountain View, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3264-7904","authenticated-orcid":false,"given":"Huan","family":"Liu","sequence":"additional","affiliation":[{"name":"Arizona State University, Tempe, AZ, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,5,23]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"A survey on data selection for language models. arXiv preprint arXiv:2402.16827","author":"Albalak A.","year":"2024","unstructured":"Albalak, A., Elazar, Y., Xie, S. M., Longpre, S., Lambert, N., Wang, X., Muennighoff, N., Hou, B., Pan, L., Jeong, H., et al. A survey on data selection for language models. arXiv preprint arXiv:2402.16827 (2024)."},{"key":"e_1_3_2_1_2_1","volume-title":"ICML","author":"Chen T.","year":"2020","unstructured":"Chen, T., Kornblith, S., Norouzi, M., and Hinton, G. A simple framework for contrastive learning of visual representations. In ICML (2020)."},{"key":"e_1_3_2_1_3_1","volume-title":"Big self-supervised models are strong semi-supervised learners. NeurIPS","author":"Chen T.","year":"2020","unstructured":"Chen, T., Kornblith, S., Swersky, K., Norouzi, M., and Hinton, G. E. Big self-supervised models are strong semi-supervised learners. NeurIPS (2020)."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3411922"},{"key":"e_1_3_2_1_5_1","volume-title":"NeurIPS","author":"Fang T.","year":"2023","unstructured":"Fang, T., Zhang, Y., Yang, Y., Wang, C., and Chen, L. Universal prompt tuning for graph neural networks. In NeurIPS (2023)."},{"key":"e_1_3_2_1_6_1","volume-title":"ICML","author":"Finn C.","year":"2017","unstructured":"Finn, C., Abbeel, P., and Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. In ICML (2017)."},{"key":"e_1_3_2_1_7_1","volume-title":"Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149","author":"Han S.","year":"2015","unstructured":"Han, S., Mao, H., and Dally, W. J. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149 (2015)."},{"key":"e_1_3_2_1_8_1","volume-title":"Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531","author":"Hinton G.","year":"2015","unstructured":"Hinton, G. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)."},{"key":"e_1_3_2_1_9_1","unstructured":"Huang S. Xu D. Yen I. E. Wang Y. Chang S.-E. Li B. Chen S. Xie M. Rajasekaran S. Liu H. et al. Sparse progressive distillation: Resolving overfitting under pretrain-and-finetune paradigm."},{"key":"e_1_3_2_1_10_1","volume-title":"ICLR","author":"Ju M.","year":"2023","unstructured":"Ju, M., Zhao, T., Wen, Q., Yu, W., Shah, N., Ye, Y., and Zhang, C. Multi-task self-supervised graph neural networks enable stronger task generalization. In ICLR (2023)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.79"},{"key":"e_1_3_2_1_12_1","volume-title":"Instance-aware graph prompt learning. arXiv","author":"Li J.","year":"2024","unstructured":"Li, J., Li, J., and Zhang, C. Instance-aware graph prompt learning. arXiv (2024)."},{"key":"e_1_3_2_1_13_1","volume-title":"Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. CSUR","author":"Liu P.","year":"2023","unstructured":"Liu, P., Yuan,W., Fu, J., Jiang, Z., Hayashi, H., and Neubig, G. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. CSUR (2023)."},{"key":"e_1_3_2_1_14_1","volume-title":"Selfsupervised learning: Generative or contrastive. TKDE","author":"Liu X.","year":"2021","unstructured":"Liu, X., Zhang, F., Hou, Z., Mian, L., Wang, Z., Zhang, J., and Tang, J. Selfsupervised learning: Generative or contrastive. TKDE (2021)."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403168"},{"key":"e_1_3_2_1_16_1","volume-title":"Exploring the limits of transfer learning with a unified text-to-text transformer. JMLR","author":"Raffel C.","year":"2020","unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., and Liu, P. J. Exploring the limits of transfer learning with a unified text-to-text transformer. JMLR (2020)."},{"key":"e_1_3_2_1_17_1","volume-title":"How to train data-efficient llms. arXiv preprint arXiv:2402.09668","author":"Sachdeva N.","year":"2024","unstructured":"Sachdeva, N., Coleman, B., Kang, W.-C., Ni, J., Hong, L., Chi, E. H., Caverlee, J., McAuley, J., and Cheng, D. Z. How to train data-efficient llms. arXiv preprint arXiv:2402.09668 (2024)."},{"key":"e_1_3_2_1_18_1","volume-title":"Data distillation: A survey. arXiv preprint arXiv:2301.04272","author":"Sachdeva N.","year":"2023","unstructured":"Sachdeva, N., and McAuley, J. Data distillation: A survey. arXiv preprint arXiv:2301.04272 (2023)."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISVLSI.2017.124"},{"key":"e_1_3_2_1_20_1","volume-title":"A survey on deep learning hardware accelerators for heterogeneous hpc platforms. arXiv preprint arXiv:2306.15552","author":"Silvano C.","year":"2023","unstructured":"Silvano, C., Ielmini, D., Ferrandi, F., Fiorin, L., Curzel, S., Benini, L., Conti, F., Garofalo, A., Zambelli, C., Calore, E., et al. A survey on deep learning hardware accelerators for heterogeneous hpc platforms. arXiv preprint arXiv:2306.15552 (2023)."},{"key":"e_1_3_2_1_21_1","volume-title":"Knowledge distillation on graphs: A survey. arXiv preprint arXiv:2302.00219","author":"Tian Y.","year":"2023","unstructured":"Tian, Y., Pei, S., Zhang, X., Zhang, C., and Chawla, N. V. Knowledge distillation on graphs: A survey. arXiv preprint arXiv:2302.00219 (2023)."},{"key":"e_1_3_2_1_22_1","volume-title":"KDD","author":"Wang H.","year":"2023","unstructured":"Wang, H., Li, R., Jiang, H., Wang, Z., Tang, X., Bi, B., Cheng, M., Yin, B., Wang, Y., Zhao, T., et al. Lighttoken: A task and model-agnostic lightweight token embedding framework for pre-trained language models. In KDD (2023)."},{"key":"e_1_3_2_1_23_1","volume-title":"Roselora: Row and column-wise sparse low-rank adaptation of pre-trained language model for knowledge editing and fine-tuning. arXiv preprint arXiv:2406.10777","author":"Wang H.","year":"2024","unstructured":"Wang, H., Liu, T., Zhao, T., and Gao, J. Roselora: Row and column-wise sparse low-rank adaptation of pre-trained language model for knowledge editing and fine-tuning. arXiv preprint arXiv:2406.10777 (2024)."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.283"},{"key":"e_1_3_2_1_25_1","volume-title":"NeurIPS","author":"Wang S.","year":"2022","unstructured":"Wang, S., Chen, C., and Li, J. Graph few-shot learning with task-specific structures. In NeurIPS (2022)."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3701551.3703550"},{"key":"e_1_3_2_1_27_1","volume-title":"NeurIPS","author":"Wang Z.","year":"2024","unstructured":"Wang, Z., Zhang, Z., Chawla, N. V., Zhang, C., and Ye, Y. Gft: Graph foundation model with transferable tree vocabulary. In NeurIPS (2024)."},{"key":"e_1_3_2_1_28_1","volume-title":"Learning crosstask generalities across graphs via task-trees. arXiv preprint arXiv:2412.16441","author":"Wang Z.","year":"2024","unstructured":"Wang, Z., Zhang, Z., Ma, T., Chawla, N. V., Zhang, C., and Ye, Y. Learning crosstask generalities across graphs via task-trees. arXiv preprint arXiv:2412.16441 (2024)."},{"key":"e_1_3_2_1_29_1","volume-title":"ICML","author":"Wen Q.","year":"2024","unstructured":"Wen, Q., Ju, M., Ouyang, Z., Zhang, C., and Ye, Y. From coarse to fine: Enable comprehensive graph self-supervised learning with multi-granular semantic ensemble. In ICML (2024)."},{"key":"e_1_3_2_1_30_1","volume-title":"A comprehensive survey on graph neural networks. TNNLS","author":"Wu Z.","year":"2020","unstructured":"Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C., and Philip, S. Y. A comprehensive survey on graph neural networks. TNNLS (2020)."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.188"},{"key":"e_1_3_2_1_32_1","volume-title":"AAAI","author":"Yu L.","year":"2022","unstructured":"Yu, L., Pei, S., Ding, L., Zhou, J., Li, L., Zhang, C., and Zhang, X. Sail: Selfaugmented graph contrastive learning. In AAAI (2022)."},{"key":"e_1_3_2_1_33_1","volume-title":"ICML","author":"Zhang C.","year":"2023","unstructured":"Zhang, C., Huang, C., Tian, Y., Wen, Q., Ouyang, Z., Li, Y., Ye, Y., and Zhang, C. When sparsity meets contrastive models: less graph data can bring better class-balanced representations. In ICML (2023)."},{"key":"e_1_3_2_1_34_1","volume-title":"Mind the gap: Mitigating the distribution gap in graph few-shot learning. TMLR","author":"Zhang C.","year":"2023","unstructured":"Zhang, C., Liu, H., Li, J., Ye, Y., and Zhang, C. Mind the gap: Mitigating the distribution gap in graph few-shot learning. TMLR (2023)."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671487"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599228"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i03.5698"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.51"},{"key":"e_1_3_2_1_39_1","volume-title":"ICLR","author":"Zhang S.","year":"2022","unstructured":"Zhang, S., Liu, Y., Sun, Y., and Shah, N. Graph-less neural networks: Teaching old mlps new tricks via distillation. In ICLR (2022)."},{"key":"e_1_3_2_1_40_1","unstructured":"Zhang X. Jiang W. Shi Y. and Hu J. When neural architecture search meets hardware implementation: from hardware awareness to co-design. In ISVLSI."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1137\/1.9781611977172.75"},{"key":"e_1_3_2_1_42_1","volume-title":"A survey of large language models. arXiv preprint arXiv:2303.18223","author":"Zhao W. X.","year":"2023","unstructured":"Zhao, W. X., Zhou, K., Li, J., Tang, T., Wang, X., Hou, Y., Min, Y., Zhang, B., Zhang, J., Dong, Z., et al. A survey of large language models. arXiv preprint arXiv:2303.18223 (2023)."},{"key":"e_1_3_2_1_43_1","volume-title":"A brief introduction to weakly supervised learning. National science review","author":"Zhou Z.-H.","year":"2018","unstructured":"Zhou, Z.-H. A brief introduction to weakly supervised learning. National science review (2018)."}],"event":{"name":"WWW '25: The ACM Web Conference 2025","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Sydney NSW Australia","acronym":"WWW '25"},"container-title":["Companion Proceedings of the ACM on Web Conference 2025"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3701716.3717558","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3701716.3717558","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,8]],"date-time":"2025-10-08T03:04:14Z","timestamp":1759892654000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3701716.3717558"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,8]]},"references-count":43,"alternative-id":["10.1145\/3701716.3717558","10.1145\/3701716"],"URL":"https:\/\/doi.org\/10.1145\/3701716.3717558","relation":{},"subject":[],"published":{"date-parts":[[2025,5,8]]},"assertion":[{"value":"2025-05-23","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}