{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T00:16:42Z","timestamp":1742948202035,"version":"3.40.3"},"publisher-location":"Cham","reference-count":19,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031202322"},{"type":"electronic","value":"9783031202339"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-20233-9_29","type":"book-chapter","created":{"date-parts":[[2022,11,3]],"date-time":"2022-11-03T00:02:48Z","timestamp":1667433768000},"page":"287-296","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Adaptive Enhanced Micro-expression Spotting Network Based on\u00a0Multi-stage Features Extraction"],"prefix":"10.1007","author":[{"given":"Zhihua","family":"Xie","sequence":"first","affiliation":[]},{"given":"Sijia","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Xiaoyu","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jiawei","family":"Fan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,3]]},"reference":[{"key":"29_CR1","doi-asserted-by":"crossref","unstructured":"Porter, S., Brinke, L.: Reading between the lies: identifying concealed and falsified emotions in universal facial expressions. Psychol. Sci. 19(5), 508\u2013514 (2008)","DOI":"10.1111\/j.1467-9280.2008.02116.x"},{"key":"29_CR2","doi-asserted-by":"crossref","unstructured":"Ekman, P., Friesen, W. V.: Nonverbal leakage and clues to deception. Psychiatry 32(1), 88\u2013106 (1969)","DOI":"10.1080\/00332747.1969.11023575"},{"key":"29_CR3","doi-asserted-by":"crossref","unstructured":"Porter, S., Brinke, L.: Reading between the lies identifying concealed and falsified emotions in universal facial expressions. Psychol. Sci. 19(5), 508\u2013514 (2008)","DOI":"10.1111\/j.1467-9280.2008.02116.x"},{"key":"29_CR4","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Chen, T. H., Liu, G., Fu, X.: SMEconvnet: a convolutional neural network for spotting spontaneous facial micro-expression from long videos. IEEE Access 6(71), 143\u2013171 (2018)","DOI":"10.1109\/ACCESS.2018.2879485"},{"key":"29_CR5","unstructured":"Antti, M., Guoying, Z., Matti, P.: Spotting rapid facial movements from videos using appearance-based feature difference analysis. In: International Conference on Pattern Recognition, pp. 1722\u20131727 (2014)"},{"key":"29_CR6","unstructured":"Adrian, D.K., Moi, Y.H., Cliff, L.: Micro-facial movement detection using individualised baselines and histogram based descriptors. In: International Conference on Systems, Man, and Cybernetics, pp. 1864\u20131869 (2015)"},{"key":"29_CR7","doi-asserted-by":"crossref","unstructured":"Adrian, D., Walied, M., Cliff, L., Choon, N.C., Moi, Y.H.: Objective micro-facial movement detection using FACS-based regions and baseline evaluation. In: International Conference on Automatic Face and Gesture Recognition (FG), pp. 642\u2013649 (2018)","DOI":"10.1109\/FG.2018.00101"},{"key":"29_CR8","doi-asserted-by":"crossref","unstructured":"Devangini, P., Guoying, Z., Matti, P.: Spatiotemporal integration of optical flow vectors for micro-expression detection. In: International Conference on Advanced Concepts for Intelligent Vision Systems, pp. 369\u2013380 (2015)","DOI":"10.1007\/978-3-319-25903-1_32"},{"key":"29_CR9","doi-asserted-by":"crossref","unstructured":"Thuong, T.K., Xiaopeng, H., Guoying, Z.: Sliding window based micro-expression spotting: a benchmark. In: International Conference on Advanced Concepts for Intelligent Vision Systems, pp. 542\u2013553 (2017)","DOI":"10.1007\/978-3-319-70353-4_46"},{"key":"29_CR10","doi-asserted-by":"crossref","unstructured":"Sujing, W., Shuhuang, W., Xingsheng, Q., Jingxiu, L., Xiaolan, F.: A main directional maximal difference analysis for spotting facial movements from long-term videos. Neurocomputing 382\u2013389 (2017)","DOI":"10.1016\/j.neucom.2016.12.034"},{"key":"29_CR11","unstructured":"Genbing, L., See, J., Laikuan, W.: Shallow optical flow three-stream CNN for macro-and micro-expression spotting from long videos. In: ICIP, pp. 2643\u20132647 (2021)"},{"key":"29_CR12","unstructured":"Wangwang, Y., Jingwen, J., Yongjie, L.: LSSNET: a two-stream convolutional neural network for spotting macro-and micro-expression in long videos. In: ACM Conference on Multimedia, pp. 4745\u20134749 (2021)"},{"key":"29_CR13","unstructured":"Xiaolong, W., Girshick, R., Gupta, A., Kaiming, H.: Non-local neural networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 7794\u20137803 (2018)"},{"key":"29_CR14","doi-asserted-by":"crossref","unstructured":"Li, J., Soladie, C., Seguier, R., Wang, S.-J., Yap, M.H.: Spotting micro-expressions on long videos sequences. In: IEEE International Conference on Automatic Face and Gesture Recognition, pp. 1\u20135 (2019)","DOI":"10.1109\/FG.2019.8756626"},{"key":"29_CR15","doi-asserted-by":"crossref","unstructured":"Qu, F., Wang, S.J., Yan, W. J., Li, H., Wu, S., Fu, X.: CAS(ME) 2: a database for spontaneous macro-expression and micro-expression spotting and recognition. IEEE Trans. Affect. Comput. 9(4), 424\u2013436 (2017)","DOI":"10.1109\/TAFFC.2017.2654440"},{"key":"29_CR16","doi-asserted-by":"crossref","unstructured":"Yap, C., Kendrick, C., Yap, M.: SAMM long videos: a spontaneous facial micro-and macro-expressions dataset. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition, pp. 194\u2013199 (2020)","DOI":"10.1109\/FG47880.2020.00029"},{"key":"29_CR17","doi-asserted-by":"crossref","unstructured":"Yap, C., Yap, M., Davison, A., Cunningham, R.: 3D-CNN for facial micro- and macro-expression spotting on long video sequences using temporal oriented reference frame. arXiv:2105.06340 (2021)","DOI":"10.1145\/3503161.3551570"},{"key":"29_CR18","doi-asserted-by":"crossref","unstructured":"Wang, S., He, Y., Li, J., Fu, X.: MESNet: a convolutional neural network for spotting multi-scale micro-expression intervals in long videos. IEEE Trans. Image Process. 3956\u20133969 (2021)","DOI":"10.1109\/TIP.2021.3064258"},{"key":"29_CR19","doi-asserted-by":"crossref","unstructured":"Zhang, L., et al.: Spatio-temporal fusion for macro-and micro-expression spotting in long video sequences. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition, pp. 245\u2013252 (2022)","DOI":"10.1109\/FG47880.2020.00037"}],"container-title":["Lecture Notes in Computer Science","Biometric Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-20233-9_29","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,7]],"date-time":"2024-10-07T03:59:57Z","timestamp":1728273597000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-20233-9_29"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031202322","9783031202339"],"references-count":19,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-20233-9_29","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"3 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CCBR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Biometric Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Beijing","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccbr2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ccbr99.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"115","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"70","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"61% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}