{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T07:28:22Z","timestamp":1740122902257,"version":"3.37.3"},"reference-count":35,"publisher":"Springer Science and Business Media LLC","issue":"17","license":[{"start":{"date-parts":[[2023,3,2]],"date-time":"2023-03-02T00:00:00Z","timestamp":1677715200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,3,2]],"date-time":"2023-03-02T00:00:00Z","timestamp":1677715200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2023,7]]},"DOI":"10.1007\/s11042-023-14527-6","type":"journal-article","created":{"date-parts":[[2023,3,2]],"date-time":"2023-03-02T03:03:05Z","timestamp":1677726185000},"page":"25733-25746","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Deep-block network for AU recognition and expression migration"],"prefix":"10.1007","volume":"82","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8062-2982","authenticated-orcid":false,"given":"Minghua","family":"Zhao","sequence":"first","affiliation":[]},{"given":"Yuxing","family":"Zhi","sequence":"additional","affiliation":[]},{"given":"Fei","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Junhuai","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jing","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Shuangli","family":"Du","sequence":"additional","affiliation":[]},{"given":"Zhenghao","family":"Shi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,3,2]]},"reference":[{"key":"14527_CR1","volume-title":"AUMPNet: simultaneous action units detection and intensity estimation on multipose facial images using a single convolutional neural network[C]","author":"JC Batista","year":"2017","unstructured":"Batista JC, Albiero V, Bellon ORP et al (2017) AUMPNet: simultaneous action units detection and intensity estimation on multipose facial images using a single convolutional neural network[C]. IEEE International Conference on Automatic Face & Gesture Recognition"},{"key":"14527_CR2","doi-asserted-by":"crossref","unstructured":"Chen D, Hua G, Wen F et al (2016) Supervised transformer network for efficient face detection[C]. European Conference on Computer Vision:122\u2013138","DOI":"10.1007\/978-3-319-46454-1_8"},{"issue":"2","key":"14527_CR3","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1007\/s00779-004-0267-x","volume":"8","author":"AD Cheok","year":"2004","unstructured":"Cheok AD, Goh KH, Liu W et al (2004) Human Pacman: a Mobile, wide-area entertainment system based on physical, social, and ubiquitous computing[J]. Pers Ubiquit Comput 8(2):71\u201381","journal-title":"Pers Ubiquit Comput"},{"key":"14527_CR4","doi-asserted-by":"crossref","unstructured":"Cootes TF, Taylor CJ (1993) Active shape model search using local Grey-level models: a quantitative evaluation[C]. British Machine Vision Conference","DOI":"10.5244\/C.7.64"},{"key":"14527_CR5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2005.177","volume-title":"Histograms of oriented gradients for human detection[C]","author":"N Dalal","year":"2005","unstructured":"Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection[C]. IEEE Computer Society Conference on Computer Vision & Pattern Recognition"},{"key":"14527_CR6","doi-asserted-by":"crossref","unstructured":"Deng J, Guo J, Ververas E et al (2020) Retinaface: single-shot multi-level face localisation in the wild[C]. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition:5203\u20135212","DOI":"10.1109\/CVPR42600.2020.00525"},{"issue":"2","key":"14527_CR7","first-page":"126","volume":"47","author":"P Ekman","year":"1978","unstructured":"Ekman P, Friesen WV (1978) Facial action coding system (FACS): a technique for the measurement of facial actions[J]. Rivista Di Psichiatria 47(2):126\u2013138","journal-title":"Rivista Di Psichiatria"},{"key":"14527_CR8","first-page":"609","volume-title":"A multi-label convolutional neural network approach to cross-domain action unit detection[C]","author":"S Ghosh","year":"2015","unstructured":"Ghosh S, Laksana E, Scherer S et al (2015) A multi-label convolutional neural network approach to cross-domain action unit detection[C]. International Conference on Affective Computing and Intelligent Interaction (ACII), pp 609\u2013615"},{"key":"14527_CR9","first-page":"1","volume-title":"Deep learning based FACS Action Unit occurrence and intensity estimation[C]","author":"A Gudi","year":"2015","unstructured":"Gudi A, Tasli HE, Den Uyl TM et al (2015) Deep learning based FACS Action Unit occurrence and intensity estimation[C]. IEEE International Conference & Workshops on Automatic Face & Gesture Recognition, pp 1\u20135"},{"key":"14527_CR10","first-page":"2584","volume-title":"Reliable Crowdsourcing and Deep Locality-Preserving Learning for Expression Recognition in the Wild[C]","author":"S Li","year":"2017","unstructured":"Li S, Deng W, Du J (2017) Reliable Crowdsourcing and Deep Locality-Preserving Learning for Expression Recognition in the Wild[C]. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 2584\u20132593"},{"key":"14527_CR11","first-page":"522","volume-title":"Adaptive Deep Metric Learning for Identity-Aware Facial Expression Recognition[C]","author":"X Liu","year":"2017","unstructured":"Liu X, Kumar BVKV, You J et al (2017) Adaptive Deep Metric Learning for Identity-Aware Facial Expression Recognition[C]. IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp 522\u2013531"},{"issue":"2","key":"14527_CR12","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1023\/B:VISI.0000029664.99615.94","volume":"60","author":"DG Lowe","year":"2004","unstructured":"Lowe DG (2004) Distinctive image features from scale-invariant Keypoints[J]. Int J Comput Vis 60(2):91\u2013110","journal-title":"Int J Comput Vis"},{"key":"14527_CR13","volume-title":"The extended Cohn-Kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression[C]","author":"P Lucey","year":"2010","unstructured":"Lucey P, Cohn JF, Kanade T et al (2010) The extended Cohn-Kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression[C]. Computer Vision & Pattern Recognition Workshops"},{"issue":"S2","key":"14527_CR14","doi-asserted-by":"publisher","first-page":"507","DOI":"10.1007\/s10339-011-0419-7","volume":"13","author":"L Maaten","year":"2012","unstructured":"Maaten L, Hendriks E (2012) Action unit classification using active appearance models and conditional random fields[J]. Cogn Process 13(S2):507\u2013518","journal-title":"Cogn Process"},{"key":"14527_CR15","doi-asserted-by":"crossref","unstructured":"Parr LA, Waller BM, Burrows AM et al (2010) Brief communication: MaqFACS: a muscle-based facial movement coding system for the rhesus macaque[J]. Am J Phys Anthropol 143(4)","DOI":"10.1002\/ajpa.21401"},{"key":"14527_CR16","first-page":"2188","volume-title":"Weakly Supervised Facial Action Unit Recognition Through Adversarial Training[C]","author":"G Peng","year":"2018","unstructured":"Peng G, Wang S (2018) Weakly Supervised Facial Action Unit Recognition Through Adversarial Training[C]. IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 2188\u20132196"},{"key":"14527_CR17","first-page":"3703","volume-title":"From Emotions to Action Units with Hidden and Semi-Hidden-Task Learning[C]","author":"A Ruiz","year":"2015","unstructured":"Ruiz A, Van De Weijer J, Binefa X (2015) From Emotions to Action Units with Hidden and Semi-Hidden-Task Learning[C]. IEEE International Conference on Computer Vision (ICCV), pp 3703\u20133711"},{"key":"14527_CR18","volume-title":"Very deep convolutional networks for large-scale image recognition[J]","author":"K Simonyan","year":"2014","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition[J]. Computer Science"},{"key":"14527_CR19","first-page":"1","volume":"2008","author":"Y Tong","year":"2008","unstructured":"Tong Y, Ji Q (2008) Learning bayesian networks with qualitative constraints[C]. IEEE Conference on Computer Vision and Pattern Recognition 2008:1\u20138","journal-title":"IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"14527_CR20","volume-title":"DeepCoder: semi-parametric Variational autoencoders for automatic facial action coding[C]","author":"DL Tran","year":"2017","unstructured":"Tran DL, Walecki R, Rudovic O et al (2017) DeepCoder: semi-parametric Variational autoencoders for automatic facial action coding[C]. IEEE International Conference on Computer Vision"},{"key":"14527_CR21","volume-title":"Deep learning for facial action unit detection under large head poses[M]","author":"Z Ts\u00e9r","year":"2016","unstructured":"Ts\u00e9r Z, Jeni LA, Lrincz A et al (2016) Deep learning for facial action unit detection under large head poses[M]. Springer International Publishing"},{"key":"14527_CR22","unstructured":"Valstar M, Pantic M (2010) Induced disgust, happiness and surprise: an addition to the mmi facial expression database[J]. procinternworkshop on emotion corpora for research on emotion & affect"},{"key":"14527_CR23","doi-asserted-by":"publisher","DOI":"10.1109\/TSMCB.2011.2163710","volume-title":"Fully automatic recognition of the temporal phases of facial actions[J]","author":"MF Valstar","year":"2012","unstructured":"Valstar MF, Pantic M (2012) Fully automatic recognition of the temporal phases of facial actions[J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS"},{"key":"14527_CR24","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.605","volume-title":"Deep Structured Learning for Facial Action Unit Intensity Estimation[C]","author":"R Walecki","year":"2017","unstructured":"Walecki R, Rudovic O, Pavlovic V et al (2017) Deep Structured Learning for Facial Action Unit Intensity Estimation[C]. IEEE Conference on Computer Vision and Pattern Recognition (CVPR)"},{"key":"14527_CR25","first-page":"56","volume":"2019","author":"C Wang","year":"2019","unstructured":"Wang C, Zeng J, Shan S et al (2019) Multi-task learning of emotion recognition and facial action unit detection with adaptively weights sharing network[C]. IEEE International Conference on Image Processing (ICIP) 2019:56\u201360","journal-title":"IEEE International Conference on Image Processing (ICIP)"},{"key":"14527_CR26","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.410","volume-title":"Capturing global semantic relationships for facial action unit recognition[C]","author":"Z Wang","year":"2013","unstructured":"Wang Z, Li Y, Wang S et al (2013) Capturing global semantic relationships for facial action unit recognition[C]. IEEE International Conference on Computer Vision"},{"key":"14527_CR27","first-page":"318","volume-title":"Facial Action Units Recognition using Deep Learning Model with Bottleneck Features[C]","author":"K Wongpatikaseree","year":"2021","unstructured":"Wongpatikaseree K, Hnoohom N, Yuenyong S et al (2021) Facial Action Units Recognition using Deep Learning Model with Bottleneck Features[C]. 25th International Computer Science and Engineering Conference (ICSEC), pp 318\u2013323"},{"key":"14527_CR28","volume-title":"Weakly supervised regional and temporal learning for facial action unit recognition[J]","author":"J Yan","year":"2022","unstructured":"Yan J, Wang J, Li Q, et al. Weakly supervised regional and temporal learning for facial action unit recognition[J], 2022."},{"key":"14527_CR29","doi-asserted-by":"publisher","first-page":"643","DOI":"10.1016\/j.neucom.2017.08.043","volume":"273","author":"N Zeng","year":"2018","unstructured":"Zeng N, Zhang H, Song B et al (2018) Facial expression recognition via learning deep sparse autoencoders[J]. Neurocomputing 273:643\u2013649","journal-title":"Neurocomputing"},{"key":"14527_CR30","unstructured":"Zeng Z, Pantic M, Roisman GI et al (2009) A survey of affect recognition methods[J]. IEEE Trans Pattern Anal Mach Intell"},{"issue":"10","key":"14527_CR31","doi-asserted-by":"publisher","first-page":"1499","DOI":"10.1109\/LSP.2016.2603342","volume":"23","author":"KP Zhang","year":"2016","unstructured":"Zhang KP, Zhang ZP, Li ZF et al (2016) Joint face detection and alignment using multitask cascaded convolutional networks[J]. Ieee Signal Processing Letters 23(10):1499\u20131503","journal-title":"Ieee Signal Processing Letters"},{"key":"14527_CR32","first-page":"1","volume":"2017","author":"S Zhang","year":"2017","unstructured":"Zhang S, Zhu X, Lei Z et al (2017) FaceBoxes: a CPU real-time face detector with high accuracy[C]. IEEE International Joint Conference on Biometrics (IJCB) 2017:1\u20139","journal-title":"IEEE International Joint Conference on Biometrics (IJCB)"},{"key":"14527_CR33","first-page":"2314","volume":"2018","author":"Y Zhang","year":"2018","unstructured":"Zhang Y, Dong W, Hu BG et al (2018) Weakly-supervised deep convolutional neural network learning for facial action unit intensity estimation[C]. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 2018:2314\u20132323","journal-title":"IEEE\/CVF Conference on Computer Vision and Pattern Recognition"},{"key":"14527_CR34","first-page":"794","volume-title":"ML-CNN: A novel deep learning based disease named entity recognition architecture[C]","author":"Z Zhao","year":"2016","unstructured":"Zhao Z, Yang Z, Luo L et al (2016) ML-CNN: A novel deep learning based disease named entity recognition architecture[C]. IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp 794\u2013794"},{"key":"14527_CR35","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2017.112","volume-title":"Pose-independent facial action unit intensity regression based on multi-task deep transfer learning[C]","author":"Y Zhou","year":"2017","unstructured":"Zhou Y, Pi J, Shi BE (2017) Pose-independent facial action unit intensity regression based on multi-task deep transfer learning[C]. IEEE International Conference on Automatic Face & Gesture Recognition"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-14527-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-14527-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-14527-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,6,23]],"date-time":"2023-06-23T20:04:59Z","timestamp":1687550699000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-14527-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,3,2]]},"references-count":35,"journal-issue":{"issue":"17","published-print":{"date-parts":[[2023,7]]}},"alternative-id":["14527"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-14527-6","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"type":"print","value":"1380-7501"},{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2023,3,2]]},"assertion":[{"value":"13 August 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 June 2022","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 January 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 March 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}