{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,24]],"date-time":"2025-09-24T00:14:58Z","timestamp":1758672898632,"version":"3.44.0"},"publisher-location":"California","reference-count":0,"publisher":"International Joint Conferences on Artificial Intelligence Organization","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,9]]},"abstract":"<jats:p>Self-supervised learning (SSL) has garnered increasing attention in electrocardiogram (ECG) analysis for its effectiveness in resource-limited settings. Existing state-of-the-art SSL methods rely on time-frequency detail reconstruction, but due to the inherent redundancy of ECG signals and individual variability, these approaches often yield suboptimal performance. In contrast, discrete label prediction becomes a superior pre-training objective by encouraging models to efficiently abstract ECG high-level semantics. However, the continuity and significant variability of ECG signals pose a challenge in generating semantically discrete labels. To address this issue, we propose an ECG pretraining framework with a self-distillation semantic tokenizer (ECG2TOK), which maps continuous ECG signals into discrete labels for self-supervised training. Specifically, the tokenizer extracts semantically aware embeddings of ECG by self-distillation and performs online clustering to generate semantically rich discrete labels. Subsequently, the SSL model is trained in conjunction with masking strategies and discrete label prediction to facilitate the abstraction of high-level semantic representations. We evaluate ECG2TOK in six downstream tasks, demonstrating that ECG2TOK efficiently achieves state-of-the-art performance and up to a 30.73% AUC increase in low-resource scenarios. Moreover, visualization experiments demonstrate that the discrete labels generated by ECG2TOK exhibit consistent semantics closely associated with clinical features. Our code is available on https:\/\/github.com\/YXYanova\/ECG2TOK.<\/jats:p>","DOI":"10.24963\/ijcai.2025\/1110","type":"proceedings-article","created":{"date-parts":[[2025,9,19]],"date-time":"2025-09-19T08:10:40Z","timestamp":1758269440000},"page":"9990-9998","source":"Crossref","is-referenced-by-count":0,"title":["ECG2TOK: ECG Pre-Training with Self-Distillation Semantic Tokenizers"],"prefix":"10.24963","author":[{"given":"Xiaoyan","family":"Yuan","sequence":"first","affiliation":[{"name":"Shenzhen MSU-BIT University"},{"name":"Beijing Institute of Technology"}]},{"given":"Wei","family":"Wang","sequence":"additional","affiliation":[{"name":"Shenzhen MSU-BIT University"},{"name":"Beijing Institute of Technology"}]},{"given":"Han","family":"Liu","sequence":"additional","affiliation":[{"name":"Dalian University of Technology"}]},{"given":"Jian","family":"Chen","sequence":"additional","affiliation":[{"name":"Shenzhen MSU-BIT University"}]},{"given":"Xiping","family":"Hu","sequence":"additional","affiliation":[{"name":"Shenzhen MSU-BIT University"},{"name":"Beijing Institute of Technology"}]}],"member":"10584","event":{"number":"34","sponsor":["International Joint Conferences on Artificial Intelligence Organization (IJCAI)"],"acronym":"IJCAI-2025","name":"Thirty-Fourth International Joint Conference on Artificial Intelligence {IJCAI-25}","start":{"date-parts":[[2025,8,16]]},"theme":"Artificial Intelligence","location":"Montreal, Canada","end":{"date-parts":[[2025,8,22]]}},"container-title":["Proceedings of the Thirty-Fourth International Joint Conference on Artificial Intelligence"],"original-title":[],"deposited":{"date-parts":[[2025,9,23]],"date-time":"2025-09-23T11:36:09Z","timestamp":1758627369000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.ijcai.org\/proceedings\/2025\/1110"}},"subtitle":[],"proceedings-subject":"Artificial Intelligence Research Articles","short-title":[],"issued":{"date-parts":[[2025,9]]},"references-count":0,"URL":"https:\/\/doi.org\/10.24963\/ijcai.2025\/1110","relation":{},"subject":[],"published":{"date-parts":[[2025,9]]}}}