{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T00:52:18Z","timestamp":1769129538353,"version":"3.49.0"},"reference-count":18,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,6,8]],"date-time":"2025-06-08T00:00:00Z","timestamp":1749340800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,6,8]],"date-time":"2025-06-08T00:00:00Z","timestamp":1749340800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,6,8]]},"DOI":"10.1109\/icc52391.2025.11160774","type":"proceedings-article","created":{"date-parts":[[2025,9,26]],"date-time":"2025-09-26T17:34:55Z","timestamp":1758908095000},"page":"1019-1024","source":"Crossref","is-referenced-by-count":0,"title":["Towards Energy-Efficient Edge Inference in Radio Cpns: a Mixture-of-Depths Transformer Based Tri-Parallel Distributed Approach"],"prefix":"10.1109","author":[{"given":"Liu","family":"Gao","sequence":"first","affiliation":[{"name":"Beijing University of Posts and Telecommunications,State Key Laboratory of Networking and Switching Technology,Beijing,China"}]},{"given":"Dixiang","family":"Gao","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications,State Key Laboratory of Networking and Switching Technology,Beijing,China"}]},{"given":"Nian","family":"Xia","sequence":"additional","affiliation":[{"name":"Nanjing Normal University,School of Computer and Electronic Information\/School of Artificial Intelligence,Nanjing,China"}]},{"given":"Mugen","family":"Peng","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications,State Key Laboratory of Networking and Switching Technology,Beijing,China"}]},{"given":"Dong","family":"Wang","sequence":"additional","affiliation":[{"name":"China Telecom Research Institute,Beijing,China"}]},{"given":"Xiqing","family":"Liu","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications,State Key Laboratory of Networking and Switching Technology,Beijing,China"}]}],"member":"263","reference":[{"key":"ref1","volume-title":"Rec.Y.2501: Next Generation Networks - Computing power networks","year":"2021"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.23919\/JCC.ja.2021-0776"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/MNET.2024.3422309"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1145\/3603269.3610856"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/3437801.3441578"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1145\/3641512.3686358"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM52122.2024.10621342"},{"key":"ref8","author":"Brakel","year":"2024","journal-title":"Model Parallelism on Distributed Infrastructure: A Literature Review from Theory to LLM CaseStudies"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/uic-atc.2017.8397638"},{"key":"ref10","author":"Karakus","year":"2021","journal-title":"Amazon sagemaker model parallelism: A general and flexible framework for large model training"},{"key":"ref11","author":"Raposo","year":"2024","journal-title":"Mixture-of-Depths: Dynamically Allocating Compute in Transformer-Based Language Models"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TC.2024.3354033"},{"key":"ref13","author":"Sanh","year":"2020","journal-title":"Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter"},{"key":"ref14","author":"Devlin","year":"2019","journal-title":"Bert: Pre-training of deep bidirectional transformers for language understanding"},{"key":"ref15","author":"Radford","year":"2019","journal-title":"Language models are unsupervised multitask learners"},{"key":"ref16","author":"Zhang","year":"2022","journal-title":"Opt: Open pre-trained transformer language models"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W18-5446"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2950443"}],"event":{"name":"ICC 2025 - IEEE International Conference on Communications","location":"Montreal, QC, Canada","start":{"date-parts":[[2025,6,8]]},"end":{"date-parts":[[2025,6,12]]}},"container-title":["ICC 2025 - IEEE International Conference on Communications"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11160703\/11160705\/11160774.pdf?arnumber=11160774","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,30]],"date-time":"2025-09-30T14:36:05Z","timestamp":1759242965000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11160774\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,8]]},"references-count":18,"URL":"https:\/\/doi.org\/10.1109\/icc52391.2025.11160774","relation":{},"subject":[],"published":{"date-parts":[[2025,6,8]]}}}