{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,25]],"date-time":"2026-02-25T18:07:56Z","timestamp":1772042876162,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":77,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T00:00:00Z","timestamp":1745280000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"the National Natural Science Foundation of China","award":["62441239,U23A20319,62172056,62472394,62192784,U22B2038"],"award-info":[{"award-number":["62441239,U23A20319,62172056,62472394,62192784,U22B2038"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,4,28]]},"DOI":"10.1145\/3696410.3714613","type":"proceedings-article","created":{"date-parts":[[2025,5,5]],"date-time":"2025-05-05T16:42:02Z","timestamp":1746463322000},"page":"3994-4003","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["TD3: Tucker Decomposition Based Dataset Distillation Method for Sequential Recommendation"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-1039-9735","authenticated-orcid":false,"given":"Jiaqing","family":"Zhang","sequence":"first","affiliation":[{"name":"State Key Laboratory of Cognitive Intelligence, University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-0853-1089","authenticated-orcid":false,"given":"Mingjia","family":"Yin","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Cognitive Intelligence, University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9921-2078","authenticated-orcid":false,"given":"Hao","family":"Wang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Cognitive Intelligence, University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2662-3444","authenticated-orcid":false,"given":"Yawen","family":"Li","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1513-7814","authenticated-orcid":false,"given":"Yuyang","family":"Ye","sequence":"additional","affiliation":[{"name":"Rutgers Business School, Newark, NJ, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-3180-0668","authenticated-orcid":false,"given":"Xingyu","family":"Lou","sequence":"additional","affiliation":[{"name":"Sun Yat-sen University, Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9402-3806","authenticated-orcid":false,"given":"Junping","family":"Du","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4835-4102","authenticated-orcid":false,"given":"Enhong","family":"Chen","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Cognitive Intelligence, University of Science and Technology of China, Hefei, China"}]}],"member":"320","published-online":{"date-parts":[[2025,4,22]]},"reference":[{"key":"e_1_3_2_1_1_1","first-page":"4750","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Cazenavette George","year":"2022","unstructured":"George Cazenavette, Tongzhou Wang, Antonio Torralba, Alexei A Efros, and Jun-Yan Zhu. Dataset distillation by matching training trajectories. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pages 4750--4759, 2022."},{"key":"e_1_3_2_1_2_1","first-page":"581","volume-title":"Bilevel optimization: theory, algorithms, applications and a bibliography. Bilevel optimization: advances and next challenges","author":"Dempe Stephan","year":"2020","unstructured":"Stephan Dempe. Bilevel optimization: theory, algorithms, applications and a bibliography. Bilevel optimization: advances and next challenges, pages 581--672, 2020."},{"key":"e_1_3_2_1_3_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018."},{"key":"e_1_3_2_1_4_1","volume-title":"Adam: A method for stochastic optimization. (No Title)","author":"Diederik P Kingma","year":"2014","unstructured":"P Kingma Diederik. Adam: A method for stochastic optimization. (No Title), 2014."},{"key":"e_1_3_2_1_5_1","first-page":"80644","article-title":"Fair graph distillation","volume":"36","author":"Feng Qizhang","year":"2023","unstructured":"Qizhang Feng, Zhimeng Stephen Jiang, Ruiquan Li, Yicheng Wang, Na Zou, Jiang Bian, and Xia Hu. Fair graph distillation. Advances in Neural Information Processing Systems, 36:80644--80660, 2023.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_6_1","volume-title":"The Twelfth International Conference on Learning Representations","author":"Feng Yunzhen","year":"2023","unstructured":"Yunzhen Feng, Shanmukha Ramakrishna Vedantam, and Julia Kempe. Embarrassingly simple dataset distillation. In The Twelfth International Conference on Learning Representations, 2023."},{"key":"e_1_3_2_1_7_1","volume-title":"Artem Molchanov, Franziska Meier, Douwe Kiela, Kyunghyun Cho, and Soumith Chintala. Generalized inner loop meta-learning. arXiv preprint arXiv:1910.01727","author":"Grefenstette Edward","year":"2019","unstructured":"Edward Grefenstette, Brandon Amos, Denis Yarats, Phu Mon Htut, Artem Molchanov, Franziska Meier, Douwe Kiela, Kyunghyun Cho, and Soumith Chintala. Generalized inner loop meta-learning. arXiv preprint arXiv:1910.01727, 2019."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.01495"},{"key":"e_1_3_2_1_9_1","volume-title":"Summarizing stream data for memory-restricted online continual learning. arXiv preprint arXiv:2305.16645, 2","author":"Gu Jianyang","year":"2023","unstructured":"Jianyang Gu, Kai Wang, Wei Jiang, and Yang You. Summarizing stream data for memory-restricted online continual learning. arXiv preprint arXiv:2305.16645, 2, 2023."},{"key":"e_1_3_2_1_10_1","volume-title":"Zhongzhou Liu, Kai Cheng, Qiushi Pan, Yi Quan Lee, Wanqi Xue, Tingjia Shen, et al. Scaling new frontiers: Insights into large recommendation models. arXiv preprint arXiv:2412.00714","author":"Guo Wei","year":"2024","unstructured":"Wei Guo, Hao Wang, Luankang Zhang, Jin Yao Chin, Zhongzhou Liu, Kai Cheng, Qiushi Pan, Yi Quan Lee, Wanqi Xue, Tingjia Shen, et al. Scaling new frontiers: Insights into large recommendation models. arXiv preprint arXiv:2412.00714, 2024."},{"key":"e_1_3_2_1_11_1","volume-title":"Mirage: Model-agnostic graph distillation for graph classification. arXiv preprint arXiv:2310.09486","author":"Gupta Mridul","year":"2023","unstructured":"Mridul Gupta, Sahil Manchanda, Sayan Ranu, and Hariprasad Kodamana. Mirage: Model-agnostic graph distillation for graph classification. arXiv preprint arXiv:2310.09486, 2023."},{"key":"e_1_3_2_1_12_1","volume-title":"End4rec: Efficient noise-decoupling for multi-behavior sequential recommendation. arXiv preprint arXiv:2403.17603","author":"Han Yongqiang","year":"2024","unstructured":"Yongqiang Han, Hao Wang, Kefan Wang, Likang Wu, Zhi Li, Wei Guo, Yong Liu, Defu Lian, and Enhong Chen. End4rec: Efficient noise-decoupling for multi-behavior sequential recommendation. arXiv preprint arXiv:2403.17603, 2024."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3604915.3610639"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2016.0030"},{"key":"e_1_3_2_1_15_1","volume-title":"Session-based recommendations with recurrent neural networks. arXiv preprint arXiv:1511.06939","author":"Hidasi Bal\u00e1zs","year":"2015","unstructured":"Bal\u00e1zs Hidasi, Alexandros Karatzoglou, Linas Baltrunas, and Domonkos Tikk. Session-based recommendations with recurrent neural networks. arXiv preprint arXiv:1511.06939, 2015."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3589335.3648307"},{"key":"e_1_3_2_1_17_1","volume-title":"Overcoming data and model heterogeneities in decentralized federated learning via synthetic anchors. arXiv preprint arXiv:2405.11525","author":"Huang Chun-Yin","year":"2024","unstructured":"Chun-Yin Huang, Kartik Srinivas, Xin Zhang, and Xiaoxiao Li. Overcoming data and model heterogeneities in decentralized federated learning via synthetic anchors. arXiv preprint arXiv:2405.11525, 2024."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2018.00035"},{"key":"e_1_3_2_1_19_1","volume-title":"A survey on data-centric recommender systems. arXiv preprint arXiv:2401.17878","author":"Lai Riwei","year":"2024","unstructured":"Riwei Lai, Li Chen, Rui Chen, and Chi Zhang. A survey on data-centric recommender systems. arXiv preprint arXiv:2401.17878, 2024."},{"key":"e_1_3_2_1_20_1","article-title":"A comprehensive survey of dataset distillation","author":"Lei Shiye","year":"2023","unstructured":"Shiye Lei and Dacheng Tao. A comprehensive survey of dataset distillation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"e_1_3_2_1_21_1","volume-title":"Visualizing the loss landscape of neural nets. Advances in neural information processing systems, 31","author":"Li Hao","year":"2018","unstructured":"Hao Li, Zheng Xu, Gavin Taylor, Christoph Studer, and Tom Goldstein. Visualizing the loss landscape of neural nets. Advances in neural information processing systems, 31, 2018."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599519"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3132847.3132926"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3132847.3132926"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.02522"},{"key":"e_1_3_2_1_26_1","volume-title":"Data distillation for text classification. arXiv preprint arXiv:2104.08448","author":"Li Yongqi","year":"2021","unstructured":"Yongqi Li and Wenjie Li. Data distillation for text classification. arXiv preprint arXiv:2104.08448, 2021."},{"key":"e_1_3_2_1_27_1","volume-title":"Meta knowledge condensation for federated learning. arXiv preprint arXiv:2209.14851","author":"Liu Ping","year":"2022","unstructured":"Ping Liu, Xin Yu, and Joey Tianyi Zhou. Meta knowledge condensation for federated learning. arXiv preprint arXiv:2209.14851, 2022."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3583780.3615134"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3463036"},{"key":"e_1_3_2_1_30_1","volume-title":"Can pre-trained models assist in dataset distillation? arXiv preprint arXiv:2310.03295","author":"Lu Yao","year":"2023","unstructured":"Yao Lu, Xuguang Chen, Yuchen Zhang, Jianyang Gu, Tianle Zhang, Yifan Zhang, Xiaoniu Yang, Qi Xuan, Kai Wang, and Yang You. Can pre-trained models assist in dataset distillation? arXiv preprint arXiv:2310.03295, 2023."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-short.12"},{"key":"e_1_3_2_1_32_1","volume-title":"Dilm: Distilling dataset into language model for text-level dataset distillation. arXiv preprint arXiv:2404.00264","author":"Maekawa Aru","year":"2024","unstructured":"Aru Maekawa, Satoshi Kosugi, Kotaro Funakoshi, and Manabu Okumura. Dilm: Distilling dataset into language model for text-level dataset distillation. arXiv preprint arXiv:2404.00264, 2024."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW50498.2020.00134"},{"key":"e_1_3_2_1_34_1","first-page":"4556","volume-title":"International Conference on Machine Learning","author":"Metz Luke","year":"2019","unstructured":"Luke Metz, Niru Maheswaranathan, Jeremy Nixon, Daniel Freeman, and Jascha Sohl-Dickstein. Understanding and correcting pathologies in the training of learned optimizers. In International Conference on Machine Learning, pages 4556--4565. PMLR, 2019."},{"key":"e_1_3_2_1_35_1","volume-title":"Landing ai. Landing AI. Available online: https:\/\/landing.ai\/(accessed on","author":"Ng Andrew","year":"2023","unstructured":"Andrew Ng. Landing ai. Landing AI. Available online: https:\/\/landing.ai\/(accessed on 8 February 2023), 2023."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICNN.1994.374611"},{"key":"e_1_3_2_1_37_1","volume-title":"Sequence-aware recommender systems. ACM computing surveys (CSUR), 51(4):1--36","author":"Quadrana Massimo","year":"2018","unstructured":"Massimo Quadrana, Paolo Cremonesi, and Dietmar Jannach. Sequence-aware recommender systems. ACM computing surveys (CSUR), 51(4):1--36, 2018."},{"key":"e_1_3_2_1_38_1","volume-title":"Scale ai. Snorkel AI. Available online: https:\/\/snorkel.ai (accessed on","author":"Ratner Alexander","year":"2023","unstructured":"Alexander Ratner. Scale ai. Snorkel AI. Available online: https:\/\/snorkel.ai (accessed on 8 February 2023), 2023."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/1772690.1772773"},{"key":"e_1_3_2_1_40_1","first-page":"31292","article-title":"Infinite recommendation networks: A data-centric approach","volume":"35","author":"Sachdeva Noveen","year":"2022","unstructured":"Noveen Sachdeva, Mehak Dhaliwal, Carole-Jean Wu, and Julian McAuley. Infinite recommendation networks: A data-centric approach. Advances in Neural Information Processing Systems, 35:31292--31305, 2022.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_41_1","volume-title":"Derek Zhiyuan Cheng, and Julian McAuley. Farzi data: Autoregressive data distillation. arXiv preprint arXiv:2310.09983","author":"Sachdeva Noveen","year":"2023","unstructured":"Noveen Sachdeva, Zexue He, Wang-Cheng Kang, Jianmo Ni, Derek Zhiyuan Cheng, and Julian McAuley. Farzi data: Autoregressive data distillation. arXiv preprint arXiv:2310.09983, 2023."},{"key":"e_1_3_2_1_42_1","volume-title":"Data distillation: A survey. arXiv preprint arXiv:2301.04272","author":"Sachdeva Noveen","year":"2023","unstructured":"Noveen Sachdeva and Julian McAuley. Data distillation: A survey. arXiv preprint arXiv:2301.04272, 2023."},{"key":"e_1_3_2_1_43_1","unstructured":"Tingjia Shen Hao Wang Chuhan Wu Jin Yao Chin Wei Guo Yong Liu Huifeng Guo Defu Lian Ruiming Tang and Enhong Chen. Predictive models in sequential recommendations: Bridging performance laws with data quality insights. arXiv preprint arXiv:2412.00430 2024."},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357895"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3159652.3159656"},{"key":"e_1_3_2_1_46_1","volume-title":"The extension of factor analysis to three-dimensional matrices. Contributions to mathematical psychology, 110119:110--182","author":"Tucker Ledyard R","year":"1964","unstructured":"Ledyard R Tucker et al. The extension of factor analysis to three-dimensional matrices. Contributions to mathematical psychology, 110119:110--182, 1964."},{"key":"e_1_3_2_1_47_1","first-page":"10553","volume-title":"International Conference on Machine Learning","author":"Vicol Paul","year":"2021","unstructured":"Paul Vicol, Luke Metz, and Jascha Sohl-Dickstein. Unbiased gradient estimation in unrolled computation graphs with persistent evolution strategies. In International Conference on Machine Learning, pages 10553--10563. PMLR, 2021."},{"key":"e_1_3_2_1_48_1","volume-title":"Scale ai. Scale AI. Available online: https:\/\/scale.com\/(accessed on","author":"Wang Alexandr","year":"2023","unstructured":"Alexandr Wang. Scale ai. Scale AI. Available online: https:\/\/scale.com\/(accessed on 8 February 2023), 2023."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/3604915.3608769"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3477495.3532685"},{"key":"e_1_3_2_1_51_1","volume-title":"Dataset distillation. arXiv preprint arXiv:1811.10959","author":"Wang Tongzhou","year":"2018","unstructured":"Tongzhou Wang, Jun-Yan Zhu, Antonio Torralba, and Alexei A Efros. Dataset distillation. arXiv preprint arXiv:1811.10959, 2018."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.02478"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401142"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1990.2.4.490"},{"key":"e_1_3_2_1_55_1","volume-title":"Dataset condensation for recommendation. arXiv preprint arXiv:2310.01038","author":"Wu Jiahao","year":"2023","unstructured":"Jiahao Wu, Wenqi Fan, Shengcai Liu, Qijiong Liu, Rui He, Qing Li, and Ke Tang. Dataset condensation for recommendation. arXiv preprint arXiv:2310.01038, 2023."},{"key":"e_1_3_2_1_56_1","volume-title":"Leveraging large language models (llms) to empower training-free dataset condensation for content-based recommendation. arXiv preprint arXiv:2310.09874","author":"Wu Jiahao","year":"2023","unstructured":"Jiahao Wu, Qijiong Liu, Hengchang Hu, Wenqi Fan, Shengcai Liu, Qing Li, Xiao-Ming Wu, and Ke Tang. Leveraging large language models (llms) to empower training-free dataset condensation for content-based recommendation. arXiv preprint arXiv:2310.09874, 2023."},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.3301346"},{"key":"e_1_3_2_1_58_1","volume-title":"Understanding short-horizon bias in stochastic meta-optimization. arXiv preprint arXiv:1803.02021","author":"Wu Yuhuai","year":"2018","unstructured":"Yuhuai Wu, Mengye Ren, Renjie Liao, and Roger Grosse. Understanding short-horizon bias in stochastic meta-optimization. arXiv preprint arXiv:1803.02021, 2018."},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-acl.458"},{"key":"e_1_3_2_1_60_1","volume-title":"Breaking determinism: Fuzzy modeling of sequential recommendation using discrete state space diffusion model. arXiv preprint arXiv:2410.23994","author":"Xie Wenjia","year":"2024","unstructured":"Wenjia Xie, Hao Wang, Luankang Zhang, Rui Zhou, Defu Lian, and Enhong Chen. Breaking determinism: Fuzzy modeling of sequential recommendation using discrete state space diffusion model. arXiv preprint arXiv:2410.23994, 2024."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01566"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2019\/547"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3358113"},{"key":"e_1_3_2_1_64_1","first-page":"36","article-title":"Does graph distillation see like vision dataset counterpart?","author":"Yang Beining","year":"2024","unstructured":"Beining Yang, Kai Wang, Qingyun Sun, Cheng Ji, Xingcheng Fu, Hao Tang, Yang You, and Jianxin Li. Does graph distillation see like vision dataset counterpart? Advances in Neural Information Processing Systems, 36, 2024.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_65_1","first-page":"36","article-title":"An efficient dataset condensation plugin and its application to continual learning","author":"Yang Enneng","year":"2023","unstructured":"Enneng Yang, Li Shen, Zhenyi Wang, Tongliang Liu, and Guibing Guo. An efficient dataset condensation plugin and its application to continual learning. Advances in Neural Information Processing Systems, 36, 2023.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1145\/3626772.3657762"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671841"},{"key":"e_1_3_2_1_68_1","article-title":"A comprehensive review","author":"Yu Ruonan","year":"2023","unstructured":"Ruonan Yu, Songhua Liu, and Xinchao Wang. Dataset distillation: A comprehensive review. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1145\/3511808.3557348"},{"key":"e_1_3_2_1_70_1","volume-title":"Learning system dynamics without forgetting. arXiv preprint arXiv:2407.00717","author":"Zhang Xikun","year":"2024","unstructured":"Xikun Zhang, Dongjin Song, Yushan Jiang, Yixin Chen, and Dacheng Tao. Learning system dynamics without forgetting. arXiv preprint arXiv:2407.00717, 2024."},{"key":"e_1_3_2_1_71_1","volume-title":"Navigating complexity: Toward lossless graph condensation via expanding window matching. arXiv preprint arXiv:2402.05011","author":"Zhang Yuchen","year":"2024","unstructured":"Yuchen Zhang, Tianle Zhang, Kai Wang, Ziyao Guo, Yuxuan Liang, Xavier Bresson, Wei Jin, and Yang You. Navigating complexity: Toward lossless graph condensation via expanding window matching. arXiv preprint arXiv:2402.05011, 2024."},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV56688.2023.00645"},{"key":"e_1_3_2_1_73_1","volume-title":"Konda Reddy Mopuri, and Hakan Bilen. Dataset condensation with gradient matching. arXiv preprint arXiv:2006.05929","author":"Zhao Bo","year":"2020","unstructured":"Bo Zhao, Konda Reddy Mopuri, and Hakan Bilen. Dataset condensation with gradient matching. arXiv preprint arXiv:2006.05929, 2020."},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1145\/2661829.2661998"},{"key":"e_1_3_2_1_75_1","first-page":"4722","volume-title":"CIKM","author":"Zhao Wayne Xin","year":"2022","unstructured":"Wayne Xin Zhao, Yupeng Hou, Xingyu Pan, Chen Yang, Zeyu Zhang, Zihan Lin, Jingsen Zhang, Shuqing Bian, Jiakai Tang, Wenqi Sun, Yushuo Chen, Lanling Xu, Gaowei Zhang, Zhen Tian, Changxin Tian, Shanlei Mu, Xinyan Fan, Xu Chen, and Ji-Rong Wen. Recbole 2.0: Towards a more up-to-date recommendation library. In CIKM, pages 4722--4726. ACM, 2022."},{"key":"e_1_3_2_1_76_1","article-title":"Enhanced self-attention mechanism for long and short term sequential recommendation models","author":"Zheng Xiaoyao","year":"2024","unstructured":"Xiaoyao Zheng, Xingwang Li, Zhenghua Chen, Liping Sun, Qingying Yu, Liangmin Guo, and Yonglong Luo. Enhanced self-attention mechanism for long and short term sequential recommendation models. IEEE Transactions on Emerging Topics in Computational Intelligence, 2024.","journal-title":"IEEE Transactions on Emerging Topics in Computational Intelligence"},{"key":"e_1_3_2_1_77_1","doi-asserted-by":"publisher","DOI":"10.1145\/3485447.3512111"}],"event":{"name":"WWW '25: The ACM Web Conference 2025","location":"Sydney NSW Australia","acronym":"WWW '25","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Proceedings of the ACM on Web Conference 2025"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696410.3714613","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3696410.3714613","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:34Z","timestamp":1750295914000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696410.3714613"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,22]]},"references-count":77,"alternative-id":["10.1145\/3696410.3714613","10.1145\/3696410"],"URL":"https:\/\/doi.org\/10.1145\/3696410.3714613","relation":{},"subject":[],"published":{"date-parts":[[2025,4,22]]},"assertion":[{"value":"2025-04-22","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}