{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T03:13:47Z","timestamp":1742958827240,"version":"3.40.3"},"publisher-location":"Cham","reference-count":27,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030922375"},{"type":"electronic","value":"9783030922382"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-92238-2_8","type":"book-chapter","created":{"date-parts":[[2021,12,4]],"date-time":"2021-12-04T22:02:35Z","timestamp":1638655355000},"page":"88-99","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["AUPro: Multi-label Facial Action Unit Proposal Generation for\u00a0Sequence-Level Analysis"],"prefix":"10.1007","author":[{"given":"Yingjie","family":"Chen","sequence":"first","affiliation":[]},{"given":"Jiarui","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Diqi","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Tao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yizhou","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yun","family":"Liang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,12,5]]},"reference":[{"key":"8_CR1","doi-asserted-by":"crossref","unstructured":"Bodla, N., Singh, B., Chellappa, R., Davis, L.S.: Soft-NMS - improving object detection with one line of code. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), October 2017 (2017)","DOI":"10.1109\/ICCV.2017.593"},{"issue":"4","key":"8_CR2","doi-asserted-by":"publisher","first-page":"7619","DOI":"10.1109\/LRA.2021.3098944","volume":"6","author":"Y Chen","year":"2021","unstructured":"Chen, Y., Wu, H., Wang, T., Wang, Y., Liang, Y.: Cross-modal representation learning for lightweight and accurate facial action unit detection. IEEE Robot. Autom. Lett. 6(4), 7619\u20137626 (2021)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"8_CR3","doi-asserted-by":"crossref","unstructured":"Cohn, J.F., Schmidt, K.: The timing of facial motion in posed and spontaneous smiles. In: Active Media Technology, pp. 57\u201369. World Scientific (2003)","DOI":"10.1142\/9789812704313_0005"},{"key":"8_CR4","doi-asserted-by":"crossref","unstructured":"Ekman, P., Friesen, W.: Facial action coding system: a technique for the measurement of facial movement (1978)","DOI":"10.1037\/t27734-000"},{"key":"8_CR5","doi-asserted-by":"crossref","unstructured":"Gao, J., Yang, Z., Chen, K., Sun, C., Nevatia, R.: Turn TAP: temporal unit regression network for temporal action proposals. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), October 2017 (2017)","DOI":"10.1109\/ICCV.2017.392"},{"key":"8_CR6","unstructured":"Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: Teh, Y.W., Titterington, M. (eds.) Proceedings of the 13th International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, Chia Laguna Resort, Sardinia, Italy, 13\u201315 May 2010, vol. 9, pp. 249\u2013256. PMLR (2010)"},{"key":"8_CR7","doi-asserted-by":"crossref","unstructured":"He, J., Li, D., Yang, B., Cao, S., Sun, B., Yu, L.: Multi view facial action unit detection based on CNN and BLSTM-RNN. In: International Conference on Automatic Face and Gesture Recognition (2017)","DOI":"10.1109\/FG.2017.108"},{"key":"8_CR8","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"8_CR9","doi-asserted-by":"crossref","unstructured":"Li, G., Zhu, X., Zeng, Y., Wang, Q., Lin, L.: Semantic relationships guided representation learning for facial action unit recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 8594\u20138601 (2019)","DOI":"10.1609\/aaai.v33i01.33018594"},{"key":"8_CR10","doi-asserted-by":"crossref","unstructured":"Li, W., Abtahi, F., Zhu, Z.: Action unit detection with region adaptation, multi-labeling learning and optimal temporal fusing. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2017)","DOI":"10.1109\/CVPR.2017.716"},{"key":"8_CR11","doi-asserted-by":"crossref","unstructured":"Lin, C., et al.: Fast learning of temporal action proposal via dense boundary generator. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 11499\u201311506 (2020)","DOI":"10.1609\/aaai.v34i07.6815"},{"key":"8_CR12","doi-asserted-by":"crossref","unstructured":"Lin, T., Liu, X., Li, X., Ding, E., Wen, S.: BMN: boundary-matching network for temporal action proposal generation (2019)","DOI":"10.1109\/ICCV.2019.00399"},{"key":"8_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-01225-0_1","volume-title":"Computer Vision \u2013 ECCV 2018","author":"T Lin","year":"2018","unstructured":"Lin, T., Zhao, X., Su, H., Wang, C., Yang, M.: BSN: boundary sensitive network for temporal action proposal generation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11208, pp. 3\u201321. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01225-0_1"},{"key":"8_CR14","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization (2019)"},{"key":"8_CR15","doi-asserted-by":"publisher","first-page":"151","DOI":"10.1109\/T-AFFC.2013.4","volume":"4","author":"SM Mavadati","year":"2013","unstructured":"Mavadati, S.M., Mahoor, M.H., Bartlett, K., Trinh, P., Cohn, J.F.: DISFA: a spontaneous facial action intensity database. IEEE Trans. Affect. Comput. 4, 151\u2013160 (2013)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"8_CR16","unstructured":"Niu, X., Han, H., Shan, S., Chen, X.: Multi-label co-regularization for semi-supervised facial action unit recognition. arXiv preprint arXiv:1910.11012 (2019)"},{"issue":"1","key":"8_CR17","doi-asserted-by":"publisher","first-page":"37","DOI":"10.1007\/s10919-005-0003-x","volume":"30","author":"KL Schmidt","year":"2006","unstructured":"Schmidt, K.L., Ambadar, Z., Cohn, J.F., Reed, L.I.: Movement differences between deliberate and spontaneous facial expressions: Zygomaticus major action in smiling. J. Nonverbal Behav. 30(1), 37\u201352 (2006)","journal-title":"J. Nonverbal Behav."},{"key":"8_CR18","doi-asserted-by":"crossref","unstructured":"Senechal, T., Rapp, V., Salam, H., Seguier, R., Bailly, K., Prevost, L.: Combining AAM coefficients with LGBP histograms in the multi-kernel SVM framework to detect facial action units. In 2011 IEEE International Conference on Automatic Face Gesture Recognition (FG), pp. 860\u2013865 (2011)","DOI":"10.1109\/FG.2011.5771363"},{"key":"8_CR19","doi-asserted-by":"crossref","unstructured":"Shou, Z., Wang, D., Chang, S.-F.: Temporal action localization in untrimmed videos via multi-stage CNNs. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016 (2016)","DOI":"10.1109\/CVPR.2016.119"},{"issue":"10","key":"8_CR20","doi-asserted-by":"publisher","first-page":"1683","DOI":"10.1109\/TPAMI.2007.1094","volume":"29","author":"Y Tong","year":"2007","unstructured":"Tong, Y., Liao, W., Ji, Q.: Facial action unit recognition by exploiting their dynamic and semantic relationships. IEEE Trans. Pattern Anal. Mach. Intell. 29(10), 1683\u20131699 (2007)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3","key":"8_CR21","doi-asserted-by":"publisher","first-page":"297","DOI":"10.1109\/TAFFC.2017.2728534","volume":"10","author":"R Walecki","year":"2017","unstructured":"Walecki, R., Rudovic, O., Pavlovic, V., Pantic, M.: Copula ordinal regression framework for joint estimation of facial action unit intensity. IEEE Trans. Affect. Comput. 10(3), 297\u2013312 (2017)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"8_CR22","doi-asserted-by":"crossref","unstructured":"Wang, C., Wang, S.: Personalized multiple facial action unit recognition through generative adversarial recognition network. In: Proceedings of the 26th ACM international conference on Multimedia, pp. 302\u2013310 (2018)","DOI":"10.1145\/3240508.3240613"},{"issue":"10","key":"8_CR23","doi-asserted-by":"publisher","first-page":"692","DOI":"10.1016\/j.imavis.2014.06.002","volume":"32","author":"X Zhang","year":"2014","unstructured":"Zhang, X., et al.: BP4D-Spontaneous: a high-resolution spontaneous 3D dynamic facial expression database. Image Vis. Comput. 32(10), 692\u2013706 (2014)","journal-title":"Image Vis. Comput."},{"key":"8_CR24","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Jiang, H., Wu, B., Fan, Y., Ji, Q.: Context-aware feature and label fusion for facial action unit intensity estimation with partially labeled data. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 733\u2013742 (2019)","DOI":"10.1109\/ICCV.2019.00082"},{"key":"8_CR25","doi-asserted-by":"crossref","unstructured":"Zhao, K., Chu, W., Zhang, H.: Deep region and multi-label learning for facial action unit detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)","DOI":"10.1109\/CVPR.2016.369"},{"key":"8_CR26","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Xiong, Y., Wang, L., Wu, Z., Tang, X., Lin, D.: Temporal action detection with structured segment networks. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), October 2017 (2017)","DOI":"10.1109\/ICCV.2017.317"},{"issue":"2","key":"8_CR27","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1109\/T-AFFC.2011.10","volume":"2","author":"Y Zhu","year":"2011","unstructured":"Zhu, Y., De la Torre, F., Cohn, J.F., Zhang, Y.J.: Dynamic cascades with bidirectional bootstrapping for action unit detection in spontaneous facial behavior. IEEE Trans. Affect. Comput. 2(2), 79\u201391 (2011)","journal-title":"IEEE Trans. Affect. Comput."}],"container-title":["Lecture Notes in Computer Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-92238-2_8","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T18:50:00Z","timestamp":1710355800000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-92238-2_8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030922375","9783030922382"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-92238-2_8","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"5 December 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Sanur, Bali","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Indonesia","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 December 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 December 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iconip2021.apnns.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1093","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"226","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"177","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"21% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.57","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"6","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Due to the COVID-19 pandemic the conference was held online.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}