{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T04:25:00Z","timestamp":1743049500814,"version":"3.40.3"},"publisher-location":"Cham","reference-count":37,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030873578"},{"type":"electronic","value":"9783030873585"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-87358-5_22","type":"book-chapter","created":{"date-parts":[[2021,9,29]],"date-time":"2021-09-29T23:54:11Z","timestamp":1632959651000},"page":"275-287","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Binary Convolutional Neural Networks for Facial Action Unit Detection"],"prefix":"10.1007","author":[{"given":"Mengyi","family":"Liu","sequence":"first","affiliation":[]},{"given":"Fei","family":"Wan","sequence":"additional","affiliation":[]},{"given":"Ruicong","family":"Zhi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,9,30]]},"reference":[{"key":"22_CR1","volume-title":"What the Face Reveals: Basic and Applied Studies of Spontaneous Expression using the Facial Action Coding System (FACS)","author":"P Ekman","year":"1997","unstructured":"Ekman, P., Rosenberg, E.L.: What the Face Reveals: Basic and Applied Studies of Spontaneous Expression using the Facial Action Coding System (FACS). Oxford University Press, USA (1997)"},{"issue":"5","key":"22_CR2","doi-asserted-by":"publisher","first-page":"1067","DOI":"10.1007\/s00371-019-01707-5","volume":"36","author":"R Zhi","year":"2019","unstructured":"Zhi, R., Liu, M., Zhang, D.: A comprehensive survey on automatic facial action unit analysis. Vis. Comput. 36(5), 1067\u20131093 (2019). https:\/\/doi.org\/10.1007\/s00371-019-01707-5","journal-title":"Vis. Comput."},{"issue":"9","key":"22_CR3","first-page":"1","volume":"13","author":"B Martinez","year":"2017","unstructured":"Martinez, B., Valstar, M.F., Jiang, B., Pantic, M.: Automatic analysis of facial actions: a survey. IEEE Trans. Affect. Comput. 13(9), 1\u201322 (2017)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"22_CR4","doi-asserted-by":"crossref","unstructured":"Bazzo, J.J, Lamar, M.V.: Recognizing facial actions using Gabor wavelets with neutral face average difference. In: Sixth IEEE International Conference on Automatic Face and Gesture Recognition, Seoul, Korea: IEEE, pp. 505\u2013510 (2004)","DOI":"10.14209\/sbrt.2004.68"},{"key":"22_CR5","unstructured":"Tian, Y., Kanade, T., Cohn, J.F.: Evaluation of Gabor-wavelet-based facial action unit recognition in image sequences of increasing complexity. In: Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition. Washington, DC, USA: IEEE, pp. 229\u2013234 (2002)"},{"key":"22_CR6","doi-asserted-by":"crossref","unstructured":"Zhao, S., Gao, Y., Zhang, B.: Sobel-lbp. In: 2008 15th IEEE International Conference on Image Processing. San Diego, USA: IEEE, pp. 2144\u20132147 (2008)","DOI":"10.1109\/ICIP.2008.4712212"},{"key":"22_CR7","doi-asserted-by":"crossref","unstructured":"Senechal, T., Rapp, V., Salam, H., Seguier, R., Bailly, K., Prevost, L.: Facial action recognition combining heterogeneous features via multikernel learning. IEEE Trans. Syst. Man Cybern. Part B (Cybernetics) 42(4), 993\u20131005 (2012)","DOI":"10.1109\/TSMCB.2012.2193567"},{"key":"22_CR8","doi-asserted-by":"crossref","unstructured":"Zhang, S., Yao, H., Liu, S.: Dynamic background modeling and subtraction using spatio-temporal local binary patterns. In: 2008 15th IEEE International Conference on Image Processing. San Diego, USA: IEEE, pp. 1556\u20131559 (2008)","DOI":"10.1109\/ICIP.2008.4712065"},{"issue":"6","key":"22_CR9","doi-asserted-by":"publisher","first-page":"765","DOI":"10.1109\/TSMCC.2011.2118750","volume":"41","author":"D Huang","year":"2011","unstructured":"Huang, D., Shan, C., Ardabilian, M., Wang, Y., Chen, L.: Local binary patterns and its application to facial image analysis: a survey. IEEE Trans. Syst. Man. Cybern. Part C 41(6), 765\u2013781 (2011)","journal-title":"IEEE Trans. Syst. Man. Cybern. Part C"},{"issue":"2","key":"22_CR10","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1109\/34.908962","volume":"23","author":"YI Tian","year":"2001","unstructured":"Tian, Y.I., Kanade, T., Cohn, J.F.: Recognizing action units for facial expression analysis. IEEE Trans. Pattern Anal. Mach. Intell. 23(2), 97\u2013115 (2001)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3","key":"22_CR11","doi-asserted-by":"publisher","first-page":"1449","DOI":"10.1109\/TSMCB.2004.825931","volume":"34","author":"M Pantic","year":"2004","unstructured":"Pantic, M., Rothkrantz, L.J.M.: Facial action recognition for facial expression analysis from static face images. IEEE Tran. Syst. Man Cybern. Part B: Cybern. 34(3), 1449\u20131461 (2004)","journal-title":"IEEE Tran. Syst. Man Cybern. Part B: Cybern."},{"key":"22_CR12","doi-asserted-by":"crossref","unstructured":"Nicolle, J., Bailly, K., Chetouani, M.: Facial action unit intensity prediction via hard multi-task metric learning for kernel regression. In: 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG). Ljubljana, SloveniaI: IEEE, pp. 1\u20136 (2015)","DOI":"10.1109\/FG.2015.7284868"},{"key":"22_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.imavis.2016.03.004","volume":"52","author":"J Nicolle","year":"2016","unstructured":"Nicolle, J., Bailly, K., Chetouani, M.: Real-time facial action unit intensity prediction with regularized metric learning. Image Vis. Comput. 52, 1\u201314 (2016)","journal-title":"Image Vis. Comput."},{"key":"22_CR14","doi-asserted-by":"crossref","unstructured":"Ming, Z., Bugeau, A., Rouas, J.L., Shochi, T.: Facial action units intensity estimation by the fusion of features with multi-kernel support vector machine. In: 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG). Ljubljana, SloveniaI: IEEE, pp. 1\u20136 (2015)","DOI":"10.1109\/FG.2015.7284870"},{"key":"22_CR15","doi-asserted-by":"publisher","unstructured":"Chang, W.Y., Chen, C.S., Hung, Y.P.: Analyzing facial expression by fusing manifolds. Asian Conference on Computer Vision, pp. 621\u2013630. Springer, Heidelberg (2007). https:\/\/doi.org\/10.1007\/978-3-540-76390-1_61","DOI":"10.1007\/978-3-540-76390-1_61"},{"key":"22_CR16","doi-asserted-by":"crossref","unstructured":"Zhao, K., Chu, W., Zhang, H.: Deep region and multi-label learning for facial action unit detection. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition, pp. 3391\u20133399 (2016)","DOI":"10.1109\/CVPR.2016.369"},{"issue":"11","key":"22_CR17","doi-asserted-by":"publisher","first-page":"2583","DOI":"10.1109\/TPAMI.2018.2791608","volume":"40","author":"W Li","year":"2018","unstructured":"Li, W., Abtahi, F., Zhu, Z., Yin, L.: EAC-Net: deep nets with enhancing and cropping for facial action unit detection. IEEE Trans. Pattern Anal. Mach. Intell. 40(11), 2583\u20132596 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"22_CR18","unstructured":"Simonyan K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"22_CR19","doi-asserted-by":"crossref","unstructured":"Mei C., Jiang F., Shen R., Hu, Q.: Region and temporal dependency fusion for multi-label action unit detection. In: 2018 24th International Conference on Pattern Recognition (ICPR), pp. 848\u2013853 (2018)","DOI":"10.1109\/ICPR.2018.8545069"},{"key":"22_CR20","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"725","DOI":"10.1007\/978-3-030-01261-8_43","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Z Shao","year":"2018","unstructured":"Shao, Z., Liu, Z., Cai, J., Ma, L.: Deep adaptive attention for joint facial action unit detection and face alignment. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11217, pp. 725\u2013740. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01261-8_43"},{"key":"22_CR21","doi-asserted-by":"crossref","unstructured":"Hu, Q., Jiang, F., Mei, C., Shen, R.: CCT: a cross-concat and temporal neural network for multi-label action unit detection. In: 2018 IEEE International Conference on Multimedia and Expo (ICME). IEEE (2018)","DOI":"10.1109\/ICME.2018.8486516"},{"key":"22_CR22","doi-asserted-by":"crossref","unstructured":"Jaiswal, S., Valstar, M.: Deep learning the dynamic appearance and shape of facial action units. In: IEEE Winter Conference on Applications of Computer Vision(WACV), pp. 1\u20138 (2016)","DOI":"10.1109\/WACV.2016.7477625"},{"key":"22_CR23","doi-asserted-by":"crossref","unstructured":"Li, W., Abtahi, F., Zhu, Z.: Action unit detection with region adaptation, multi-labeling learning, and optimal temporal fusing. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6766\u20136775 (2017)","DOI":"10.1109\/CVPR.2017.716"},{"key":"22_CR24","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"22_CR25","doi-asserted-by":"crossref","unstructured":"Girshick, R.: Fast R-CNN. In: IEEE International Conference on Computer Vision, 1440\u20131448 (2015)","DOI":"10.1109\/ICCV.2015.169"},{"key":"22_CR26","doi-asserted-by":"crossref","unstructured":"Li, W., Abtahi, F., Zhu, Z.: A deep feature based multi-kernel learning approach for video emotion recognition. In: ACM International Conference on Multimodal Interaction, pp. 483\u2013490 (2015)","DOI":"10.1145\/2818346.2830583"},{"key":"22_CR27","doi-asserted-by":"crossref","unstructured":"Shen, F., Xu, Y., Liu, L., Yang, Y., Huang, Z., Shen, H.: unsupervised deep hashing with similarity-adaptive and discrete optimization. IEEE Trans. Pattern Anal. Mach. Intell. (99),1 (2018).","DOI":"10.1109\/TPAMI.2018.2789887"},{"key":"22_CR28","unstructured":"Molchanov, P., Tyree, S., Karras, T., Aila, T., Kautz, J.: Pruning convolutional neural networks for resource efficient inference. In: International Conference on Learning Representations (2016)."},{"key":"22_CR29","doi-asserted-by":"crossref","unstructured":"Liu, Z., Li, J., Shen Z., Huang, G., Yan, S., Zhang, C.: Learning efficient convolutional networks through network slimming. In: International Conference on Computer Vision(ICCV) (2017).=","DOI":"10.1109\/ICCV.2017.298"},{"key":"22_CR30","unstructured":"Zhuang, Z., et al.: Discrimination-aware channel pruning for deep neural networks. In: 32nd Conference on Neural Information Processing Systems (NeurlPS) (2018)"},{"key":"22_CR31","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"525","DOI":"10.1007\/978-3-319-46493-0_32","volume-title":"Computer Vision \u2013 ECCV 2016","author":"M Rastegari","year":"2016","unstructured":"Rastegari, M., Ordonez, V., Redmon, J., Farhadi, A.: XNOR-Net: imagenet classification using binary convolutional neural networks. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9908, pp. 525\u2013542. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46493-0_32"},{"key":"22_CR32","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"747","DOI":"10.1007\/978-3-030-01267-0_44","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Z Liu","year":"2018","unstructured":"Liu, Z., Wu, B., Luo, W., Yang, X., Liu, W., Cheng, K.-T.: Bi-Real Net: enhancing the performance of 1-bit cnns with improved representational capability and advanced training algorithm. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11219, pp. 747\u2013763. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01267-0_44"},{"key":"22_CR33","unstructured":"Courbariaux, M., Hubara, I., Soudry, D., El-Yaniv, R., Bengio, Y.: BinaryNet: training deep neural networks with weights and activations constrained to +1 or \u22121. arXiv: 1602.02830v3 (2016)"},{"key":"22_CR34","doi-asserted-by":"crossref","unstructured":"Li, Z., Ni, B., Zhang, W., Yang, X., Wen, G.: Performance guaranteed network acceleration via high-order residual quantization. In: 2017 IEEE International Conference on Computer Vision (ICCV). IEEE Computer Society (2017)","DOI":"10.1109\/ICCV.2017.282"},{"issue":"10","key":"22_CR35","doi-asserted-by":"publisher","first-page":"692","DOI":"10.1016\/j.imavis.2014.06.002","volume":"32","author":"X Zhang","year":"2014","unstructured":"Zhang, X., et al.: BP4D-Spontaneous: a high-resolution spontaneous 3D dynamic facial expression database. Image Vis. Comput. 32(10), 692\u2013706 (2014)","journal-title":"Image Vis. Comput."},{"key":"22_CR36","doi-asserted-by":"crossref","unstructured":"Zhao, K., Chu, W.-S., Torre, F., Cohn, J.F., Zhang, H.: Joint patch and multi-label learning for facial action unit detection. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2207\u20132216 (2015)","DOI":"10.1109\/CVPR.2015.7298833"},{"key":"22_CR37","doi-asserted-by":"crossref","unstructured":"Zeng, J., Chu, W.-S., Torre, F., Cohn, J.F., Xiong, Z.: Confidence preserving machine for facial action unit detection. In: IEEE International Conference on Computer Vision, pp. 3622\u20133630 (2015)","DOI":"10.1109\/ICCV.2015.413"}],"container-title":["Lecture Notes in Computer Science","Image and Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-87358-5_22","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,9]],"date-time":"2024-09-09T01:28:19Z","timestamp":1725845299000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-87358-5_22"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030873578","9783030873585"],"references-count":37,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-87358-5_22","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"30 September 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIG","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Image and Graphics","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Haikou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 August 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 August 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icig2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/icig2021.csig.org.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"421","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"198","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"47% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Conference was postponed due to the COVID19 pandemic.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}