{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T20:47:45Z","timestamp":1742935665448,"version":"3.40.3"},"publisher-location":"Cham","reference-count":24,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030057091"},{"type":"electronic","value":"9783030057107"}],"license":[{"start":{"date-parts":[[2018,12,8]],"date-time":"2018-12-08T00:00:00Z","timestamp":1544227200000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-030-05710-7_19","type":"book-chapter","created":{"date-parts":[[2018,12,7]],"date-time":"2018-12-07T12:48:55Z","timestamp":1544186935000},"page":"227-238","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Two-Level Attention with Multi-task Learning for Facial Emotion Estimation"],"prefix":"10.1007","author":[{"given":"Xiaohua","family":"Wang","sequence":"first","affiliation":[]},{"given":"Muzi","family":"Peng","sequence":"additional","affiliation":[]},{"given":"Lijuan","family":"Pan","sequence":"additional","affiliation":[]},{"given":"Min","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Chunhua","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Fuji","family":"Ren","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2018,12,8]]},"reference":[{"key":"19_CR1","doi-asserted-by":"crossref","unstructured":"Jung, H., Lee, S., Yim, J., Park, S., Kim, J.: Joint fine-tuning in deep neural networks for facial expression recognition. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2983\u20132991 (2015)","DOI":"10.1109\/ICCV.2015.341"},{"key":"19_CR2","doi-asserted-by":"crossref","unstructured":"Kim, B.K., Dong, S.Y., Roh, J., Kim, G., Lee, S.Y.: Fusing aligned and non-aligned face information for automatic affect recognition in the wild: a deep learning approach. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 48\u201357 (2016)","DOI":"10.1109\/CVPRW.2016.187"},{"issue":"9","key":"19_CR3","doi-asserted-by":"publisher","first-page":"4193","DOI":"10.1109\/TIP.2017.2689999","volume":"26","author":"K Zhang","year":"2017","unstructured":"Zhang, K., Huang, Y., Du, Y., Wang, L.: Facial expression recognition based on deep evolutional spatial-temporal networks. IEEE Trans. Image Process. 26(9), 4193\u20134203 (2017)","journal-title":"IEEE Trans. Image Process."},{"key":"19_CR4","doi-asserted-by":"crossref","unstructured":"Chen, S., Jin, Q., Zhao, J., Wang, S.: Multimodal multi-task learning for dimensional and continuous emotion recognition. In: Proceedings of the 7th Annual Workshop on Audio\/Visual Emotion Challenge, pp. 19\u201326. ACM (2017)","DOI":"10.1145\/3133944.3133949"},{"key":"19_CR5","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1109\/TAFFC.2015.2512598","volume":"1","author":"R Xia","year":"2017","unstructured":"Xia, R., Liu, Y.: A multi-task learning framework for emotion recognition using 2D continuous space. IEEE Trans. Affect. Comput. 1, 3\u201314 (2017)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"6","key":"19_CR6","doi-asserted-by":"publisher","first-page":"1161","DOI":"10.1037\/h0077714","volume":"39","author":"JA Russell","year":"1980","unstructured":"Russell, J.A.: A circumplex model of affect. J. Pers. Socialpsychol. 39(6), 1161 (1980)","journal-title":"J. Pers. Socialpsychol."},{"key":"19_CR7","unstructured":"Mollahosseini, A., Hasani, B., Mahoor, M.H.: AffectNet: a database for facial expression, valence, and arousal computing in the wild. arXiv preprint arXiv:1708.03985 (2017)"},{"key":"19_CR8","doi-asserted-by":"crossref","unstructured":"Liu, X., Kumar, B.V., You, J., Jia, P.: Adaptive deep metric learning for identity-aware facial expression recognition. In: CVPR Workshops, pp. 522\u2013531 (2017)","DOI":"10.1109\/CVPRW.2017.79"},{"key":"19_CR9","doi-asserted-by":"crossref","unstructured":"Li, S., Deng, W., Du, J.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2584\u20132593. IEEE (2017)","DOI":"10.1109\/CVPR.2017.277"},{"key":"19_CR10","doi-asserted-by":"publisher","first-page":"12","DOI":"10.1016\/j.neucom.2018.03.034","volume":"296","author":"W Sun","year":"2018","unstructured":"Sun, W., Zhao, H., Jin, Z.: A visual attention based ROI detection method for facial expression recognition. Neurocomputing 296, 12\u201322 (2018)","journal-title":"Neurocomputing"},{"key":"19_CR11","unstructured":"Wang, F., et al.: Residual attention network for image classification. arXiv preprint arXiv:1704.06904 (2017)"},{"key":"19_CR12","doi-asserted-by":"crossref","unstructured":"Chang, W.Y., Hsu, S.H., Chien, J.H.: FATAUVA-Net: an integrated deep learning framework for facial attribute recognition, action unit (au) detection, and valence-arousal estimation. In: Proceedings of the IEEE Conference onComputer Vision and Pattern Recognition Workshop (2017)","DOI":"10.1109\/CVPRW.2017.246"},{"key":"19_CR13","unstructured":"Kollias, D., Zafeiriou, S.: A multi-component CNN-RNN approach for dimensional emotion recognition in-the-wild. arXiv preprint arXiv:1805.01452 (2018)"},{"key":"19_CR14","unstructured":"Zhou, F., Kong, S., Fowlkes, C., Chen, T., Lei, B.: Fine-grained facial expression analysis using dimensional emotion model. arXiv preprint arXiv:1805.01024 (2018)"},{"key":"19_CR15","unstructured":"Mnih, V., Heess, N., Graves, A., et al.: Recurrent models of visual attention. In: Advances in Neural Information Processing Systems, pp. 2204\u20132212 (2014)"},{"key":"19_CR16","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. arXiv preprint arXiv:1709.01507 7 (2017)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"19_CR17","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017)"},{"key":"19_CR18","doi-asserted-by":"publisher","first-page":"90","DOI":"10.1016\/j.cviu.2017.10.001","volume":"163","author":"A Das","year":"2017","unstructured":"Das, A., Agrawal, H., Zitnick, L., Parikh, D., Batra, D.: Human attention in visual question answering: do humans and deep networks look at the same regions? Comput. Vis. Image Underst. 163, 90\u2013100 (2017)","journal-title":"Comput. Vis. Image Underst."},{"key":"19_CR19","unstructured":"Xu, K., et al.: Show, attend and tell: neural image caption generation with visual attention. In: International Conference on Machine Learning, pp. 2048\u20132057 (2015)"},{"key":"19_CR20","doi-asserted-by":"crossref","unstructured":"Chang, J., Scherer, S.: Learning representations of emotional speech with deep convolutional generative adversarial networks. In: 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2746\u20132750. IEEE (2017)","DOI":"10.1109\/ICASSP.2017.7952656"},{"key":"19_CR21","unstructured":"Duan, M., Li, K., Tian, Q.: A novel multi-task tensor correlation neural network for facial attribute prediction. arXiv preprint arXiv:1804.02810 (2018)"},{"issue":"1","key":"19_CR22","doi-asserted-by":"publisher","first-page":"57","DOI":"10.1007\/BF00131148","volume":"19","author":"MJ Black","year":"1996","unstructured":"Black, M.J., Rangarajan, A.: On the unification of line processes, outlier rejection, and robust statistics with applications in early vision. Int. J. Comput. Vis. 19(1), 57\u201391 (1996)","journal-title":"Int. J. Comput. Vis."},{"issue":"10","key":"19_CR23","doi-asserted-by":"publisher","first-page":"1499","DOI":"10.1109\/LSP.2016.2603342","volume":"23","author":"K Zhang","year":"2016","unstructured":"Zhang, K., Zhang, Z., Li, Z., Qiao, Y.: Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Sig. Process. Lett. 23(10), 1499\u20131503 (2016)","journal-title":"IEEE Sig. Process. Lett."},{"key":"19_CR24","unstructured":"Mahoor, M.H.: AffectNet. http:\/\/mohammadmahoor.com\/affectnet\/ . Accessed 27 July 2018"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-05710-7_19","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,9,7]],"date-time":"2022-09-07T21:17:26Z","timestamp":1662585446000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-05710-7_19"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,12,8]]},"ISBN":["9783030057091","9783030057107"],"references-count":24,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-05710-7_19","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2018,12,8]]},"assertion":[{"value":"8 December 2018","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Thessaloniki","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Greece","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2019","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 January 2019","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 January 2019","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2019","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/mmm2019.iti.gr\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double blind for full papers and workshop papers, single blind for other paper types","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"204","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"96","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"47% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"2.67","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"6 demonstration papers, 5 industry papers, 6 workshop papers, and 6 Video Browser Showdown papers were also accepted.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}}]}}