{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T13:57:25Z","timestamp":1762005445094,"version":"3.33.0"},"reference-count":7,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2023,3,31]],"date-time":"2023-03-31T00:00:00Z","timestamp":1680220800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,3,31]],"date-time":"2023-03-31T00:00:00Z","timestamp":1680220800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Front. Comput. Sci."],"published-print":{"date-parts":[[2023,12]]},"DOI":"10.1007\/s11704-023-2230-x","type":"journal-article","created":{"date-parts":[[2023,3,31]],"date-time":"2023-03-31T16:02:54Z","timestamp":1680278574000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Fine-grained sequence-to-sequence lip reading based on self-attention and self-distillation"],"prefix":"10.1007","volume":"17","author":[{"given":"Junxiao","family":"Xue","sequence":"first","affiliation":[]},{"given":"Shibo","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Huawei","family":"Song","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Shi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,3,31]]},"reference":[{"key":"2230_CR1","doi-asserted-by":"crossref","unstructured":"Xiao J, Yang S, Zhang Y, Shan S, Chen X. Deformation flow based two-stream network for lip reading. In: Proceedings of the 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020). 2020: 364\u2013370","DOI":"10.1109\/FG47880.2020.00132"},{"key":"2230_CR2","unstructured":"Assael Y M, Shillingford B, Whiteson S, De Freitas N. LipNet: End-to-end sentence-level lipreading. 2017, arXiv preprint arXiv: 1611, 0159: 9"},{"key":"2230_CR3","doi-asserted-by":"crossref","unstructured":"Chung J S, Senior A, Vinyals O, et al. Lip reading sentences in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017: 3444\u20133453","DOI":"10.1109\/CVPR.2017.367"},{"key":"2230_CR4","doi-asserted-by":"crossref","unstructured":"Xu K, Li D, Cassimatis N, Wang X. LCANet: End-to-end lipreading with cascaded attention-CTC. In: Proceedings of the 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018). 2018: 548\u2013555","DOI":"10.1109\/FG.2018.00088"},{"key":"2230_CR5","doi-asserted-by":"crossref","unstructured":"Zhang Y, Yang S, Xiao J, et al. Can we read speech beyond the lips? rethinking roi selection for deep visual speech recognition In: Proceedings of the 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020). 2020: 356\u2013363","DOI":"10.1109\/FG47880.2020.00134"},{"key":"2230_CR6","doi-asserted-by":"crossref","unstructured":"Luo M, Yang S, Shan S, Chen X. Pseudo-convolutional policy gradient for sequence-to-sequence lip-reading. In: Proceedings of the 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020). 2020: 273\u2013280","DOI":"10.1109\/FG47880.2020.00010"},{"key":"2230_CR7","doi-asserted-by":"crossref","unstructured":"Zhang X, Cheng F, Wang S. Spatio-temporal fusion based convolutional sequence learning for lip reading. In: Proceedings of 2019 IEEE\/CVF International Conference on Computer Vision. 2019: 713\u2013722","DOI":"10.1109\/ICCV.2019.00080"}],"container-title":["Frontiers of Computer Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11704-023-2230-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11704-023-2230-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11704-023-2230-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,18]],"date-time":"2025-01-18T20:42:36Z","timestamp":1737232956000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11704-023-2230-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,3,31]]},"references-count":7,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2023,12]]}},"alternative-id":["2230"],"URL":"https:\/\/doi.org\/10.1007\/s11704-023-2230-x","relation":{},"ISSN":["2095-2228","2095-2236"],"issn-type":[{"type":"print","value":"2095-2228"},{"type":"electronic","value":"2095-2236"}],"subject":[],"published":{"date-parts":[[2023,3,31]]},"assertion":[{"value":"19 April 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 February 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 March 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"176344"}}