{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,18]],"date-time":"2026-02-18T23:46:46Z","timestamp":1771458406950,"version":"3.50.1"},"publisher-location":"Singapore","reference-count":19,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819981403","type":"print"},{"value":"9789819981410","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,11,26]],"date-time":"2023-11-26T00:00:00Z","timestamp":1700956800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,11,26]],"date-time":"2023-11-26T00:00:00Z","timestamp":1700956800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8141-0_26","type":"book-chapter","created":{"date-parts":[[2023,11,25]],"date-time":"2023-11-25T09:02:16Z","timestamp":1700902936000},"page":"347-356","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Lip Reading Using Temporal Adaptive Module"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-1898-3756","authenticated-orcid":false,"given":"Jian","family":"Huang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6523-9731","authenticated-orcid":false,"given":"Lianwei","family":"Teng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9689-3760","authenticated-orcid":false,"given":"Yewei","family":"Xiao","sequence":"additional","affiliation":[]},{"given":"Aosu","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Xuanming","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,11,26]]},"reference":[{"key":"26_CR1","doi-asserted-by":"crossref","unstructured":"Sun, K., Yu, C., Shi, W., Liu, L., Shi, Y.: Lip-interact: improving mobile device interaction with silent speech commands. In: Proceedings of the 31st Annual ACM Symposium on User Interface Software and Technology, pp. 581\u2013593 (2018)","DOI":"10.1145\/3242587.3242599"},{"key":"26_CR2","doi-asserted-by":"crossref","unstructured":"Jha, A., Namboodiri, V.P., Jawahar, C.V.: Word spotting in silent lip videos. In: 2018 IEEE Winter Conference on Applications of Computer Vision (WACV). IEEE (2018)","DOI":"10.1109\/WACV.2018.00023"},{"key":"26_CR3","unstructured":"Afouras, T., et al.: Deep audio-visual speech recognition. IEEE Trans. Pattern Anal. Mach. Intell. (2018)"},{"key":"26_CR4","doi-asserted-by":"publisher","unstructured":"Rufai, S.Z., Selwal, A., Sharma, D.: On analysis of face liveness detection mechanisms via deep learning models. In: International Conference on Sustainable Computing and Data Communication Systems (ICSCDS), vol. 2022, pp. 59\u201364 (2022). https:\/\/doi.org\/10.1109\/ICSCDS53736.2022.9760922","DOI":"10.1109\/ICSCDS53736.2022.9760922"},{"key":"26_CR5","doi-asserted-by":"crossref","unstructured":"Noda, K., Yamaguchi, Y., Nakadai, K., Okuno, H.G., Ogata, T.: Lipreading using convolutional neural network. In: Fifteenth Annual Conference of the International Speech Communication Association (2014)","DOI":"10.21437\/Interspeech.2014-293"},{"key":"26_CR6","doi-asserted-by":"crossref","unstructured":"Stafylakis, T., Tzimiropoulos, G.J.A.P.A.: Combining residual networks with LSTMs for lipreading (2017)","DOI":"10.21437\/Interspeech.2017-85"},{"key":"26_CR7","doi-asserted-by":"crossref","unstructured":"Martinez, B., Ma, P., Petridis, S., Pantic, M.: Lipreading using temporal convolutional networks. In: ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 6319\u20136323. IEEE (2020)","DOI":"10.1109\/ICASSP40776.2020.9053841"},{"key":"26_CR8","doi-asserted-by":"crossref","unstructured":"Ma, P., Wang, Y., Shen, J., Petridis, S., Pantic, M.: Lip-reading with densely connected temporal convolutional networks. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 2857\u20132866 (2021)","DOI":"10.1109\/WACV48630.2021.00290"},{"key":"26_CR9","doi-asserted-by":"crossref","unstructured":"Xiao, J., Yang, S., Zhang, Y., Shan, S., Chen, X.: Deformation flow based two-stream network for lip reading. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), pp. 364\u2013370. IEEE (2020)","DOI":"10.1109\/FG47880.2020.00132"},{"key":"26_CR10","doi-asserted-by":"crossref","unstructured":"Hao, M., et al.: How to use time information effectively? Combining with time shift module for lipreading. In: ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE (2021)","DOI":"10.1109\/ICASSP39728.2021.9414659"},{"key":"26_CR11","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: TAM: temporal adaptive module for video recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2021)","DOI":"10.1109\/ICCV48922.2021.01345"},{"key":"26_CR12","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"87","DOI":"10.1007\/978-3-319-54184-6_6","volume-title":"Computer Vision \u2013 ACCV 2016","author":"JS Chung","year":"2017","unstructured":"Chung, J.S., Zisserman, A.: Lip reading in the wild. In: Lai, S.-H., Lepetit, V., Nishino, K., Sato, Y. (eds.) ACCV 2016. LNCS, vol. 10112, pp. 87\u2013103. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-54184-6_6"},{"key":"26_CR13","unstructured":"Mediapipe. https:\/\/mediapipe.dev\/"},{"key":"26_CR14","doi-asserted-by":"publisher","unstructured":"Ma, P., Wang, Y., Petridis, S., Shen, J., Pantic, M.: Training strategies for improved lip-reading. In: ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 8472\u20138476 (2022). https:\/\/doi.org\/10.1109\/ICASSP43922.2022.9746706","DOI":"10.1109\/ICASSP43922.2022.9746706"},{"key":"26_CR15","doi-asserted-by":"crossref","unstructured":"Miao, Z., Liu, H., Yang, B.: Part-based lipreading for audio-visual speech recognition. In: 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC). IEEE (2020)","DOI":"10.1109\/SMC42975.2020.9283044"},{"key":"26_CR16","doi-asserted-by":"publisher","unstructured":"Tian, W., Zhang, H., Peng, C., Zhao, Z.-Q.: lipreading model based on whole-part collaborative learning. In: ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2425\u20132429 (2022). https:\/\/doi.org\/10.1109\/ICASSP43922.2022.9747052","DOI":"10.1109\/ICASSP43922.2022.9747052"},{"key":"26_CR17","doi-asserted-by":"publisher","unstructured":"Ma, P., Martinez, B., Petridis, S., Pantic, M.: Towards practical lipreading with distilled and efficient models. In: ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 7608-7612 (2021). https:\/\/doi.org\/10.1109\/ICASSP39728.2021.9415063","DOI":"10.1109\/ICASSP39728.2021.9415063"},{"key":"26_CR18","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"26_CR19","doi-asserted-by":"publisher","unstructured":"Zhao, X., Yang, S., Shan, S., Chen, X.: Mutual information maximization for effective lip reading. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), pp. 420-427 (2020). https:\/\/doi.org\/10.1109\/FG47880.2020.00133","DOI":"10.1109\/FG47880.2020.00133"}],"container-title":["Communications in Computer and Information Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8141-0_26","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T15:41:02Z","timestamp":1710344462000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8141-0_26"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,26]]},"ISBN":["9789819981403","9789819981410"],"references-count":19,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8141-0_26","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,11,26]]},"assertion":[{"value":"26 November 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Changsha","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 November 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 November 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/iconip2023.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1274","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"650","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"51% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4.14","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.46","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}