{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T14:50:03Z","timestamp":1768920603546,"version":"3.49.0"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030805678","type":"print"},{"value":"9783030805685","type":"electronic"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-80568-5_38","type":"book-chapter","created":{"date-parts":[[2021,6,23]],"date-time":"2021-06-23T17:04:53Z","timestamp":1624467893000},"page":"465-476","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["A Multi-modal Audience Analysis System for Predicting Popularity of Online Videos"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9738-1553","authenticated-orcid":false,"given":"Alexandros","family":"Vrochidis","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6650-7758","authenticated-orcid":false,"given":"Nikolaos","family":"Dimitriou","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9666-7023","authenticated-orcid":false,"given":"Stelios","family":"Krinidis","sequence":"additional","affiliation":[]},{"given":"Savvas","family":"Panagiotidis","sequence":"additional","affiliation":[]},{"given":"Stathis","family":"Parcharidis","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6915-6722","authenticated-orcid":false,"given":"Dimitrios","family":"Tzovaras","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,7,1]]},"reference":[{"key":"38_CR1","doi-asserted-by":"crossref","unstructured":"Symeonidis, P., et al.: Recommending the video to watch next: an offline and online evaluation at YOUTV.de. In: Forteenth ACM Conference on Recommender Systems, pp. 299\u2013308 (2020)","DOI":"10.1145\/3383313.3412257"},{"key":"38_CR2","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"286","DOI":"10.1007\/978-3-030-00066-0_25","volume-title":"Digital Libraries for Open Knowledge","author":"J Medrek","year":"2018","unstructured":"Medrek, J., Otto, C., Ewerth, R.: Recommending scientific videos based on metadata enrichment using linked open data. In: M\u00e9ndez, E., Crestani, F., Ribeiro, C., David, G., Lopes, J.C. (eds.) TPDL 2018. LNCS, vol. 11057, pp. 286\u2013292. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-00066-0_25"},{"key":"38_CR3","doi-asserted-by":"crossref","unstructured":"Zhu, Q., Shyu, M., Wang, H.: VideoTopic: content-based video recommendation using a topic model. In: IEEE International Symposium on Multimedia, pp. 219\u2013222, (2013)","DOI":"10.1109\/ISM.2013.41"},{"key":"38_CR4","doi-asserted-by":"crossref","unstructured":"Yunyu, S., Haisheng, Y., Ming, G., Xiang, L., Yongxiang, X.: A fast and robust key frame extraction method for video copyright protection. J. Electr. Comput. Eng. 2017, 1\u20137 (2017). Article ID 1231794","DOI":"10.1155\/2017\/1231794"},{"issue":"1","key":"38_CR5","doi-asserted-by":"publisher","first-page":"2150002","DOI":"10.1142\/S021812662150002X","volume":"30","author":"B Tang","year":"2021","unstructured":"Tang, B., Chen, W.: A description scheme for video overview based on scene detection and face clustering. J. Circuits Syst. Comput. 30(1), 2150002 (2021)","journal-title":"J. Circuits Syst. Comput."},{"key":"38_CR6","doi-asserted-by":"publisher","first-page":"24549","DOI":"10.1109\/ACCESS.2020.2970063","volume":"8","author":"T Yan","year":"2020","unstructured":"Yan, T., Ra, I., Wen, H., Weng, M., Zhang, Q., Che, Y.: CTU layer rate control algorithm in scene change video for free-viewpoint video. IEEE Access 8, 24549\u201324560 (2020)","journal-title":"IEEE Access"},{"key":"38_CR7","doi-asserted-by":"publisher","first-page":"128","DOI":"10.1016\/j.neucom.2018.02.110","volume":"395","author":"G Guo","year":"2019","unstructured":"Guo, G., Wang, H., Yan, Y., Zheng, J., Li, B.: A fast face detection method via convolutional neural network. Neurocomputing 395, 128\u2013137 (2019)","journal-title":"Neurocomputing"},{"key":"38_CR8","unstructured":"Jimenez, M., Kalogeiton, V., Suarez, P., Zisserman, A.: LAEO-Net: revisiting people looking at each other in videos. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, pp. 3472\u20133480 (2019)"},{"key":"38_CR9","doi-asserted-by":"publisher","unstructured":"Soler, R., Cuevas, F., Salinas, R., Jimenez, J.: RealHePoNet: a robust single-stage ConvNet for head pose estimation in the wild. Neural Comp. and App. (2020). https:\/\/doi.org\/10.1007\/s00521-020-05511-4","DOI":"10.1007\/s00521-020-05511-4"},{"issue":"2","key":"38_CR10","doi-asserted-by":"publisher","first-page":"266","DOI":"10.1109\/83.902291","volume":"10","author":"TF Chan","year":"2001","unstructured":"Chan, T.F., Vese, L.A.: Active contours without edges. IEEE Trans. Image Process. 10(2), 266\u2013277 (2001)","journal-title":"IEEE Trans. Image Process."},{"issue":"15","key":"38_CR11","doi-asserted-by":"publisher","first-page":"11253","DOI":"10.1007\/s00521-019-04564-4","volume":"32","author":"S Jaiswal","year":"2019","unstructured":"Jaiswal, S., Nandi, G.C.: Robust real-time emotion detection system using CNN architecture. Neural Comput. Appl. 32(15), 11253\u201311262 (2019). https:\/\/doi.org\/10.1007\/s00521-019-04564-4","journal-title":"Neural Comput. Appl."},{"key":"38_CR12","doi-asserted-by":"publisher","unstructured":"Goodfellow, J., Erthan, D., Carrier, P., Cournville, et al.: Challenges in representation learning: a report on three machine learning contests. Lecture Notes in Computer Science, vol. 8228 (2013). https:\/\/doi.org\/10.1007\/978-3-642-42051-1_16","DOI":"10.1007\/978-3-642-42051-1_16"},{"key":"38_CR13","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The extended Cohn-Kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition \u2013 Workshops, San Francisco, CA, pp. 94\u2013101 (2010)","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"38_CR14","unstructured":"Dachapally, P.: Facial emotion detection using convolutional neural networks and representational autoencoder units. arXiv, abs\/1706.01509 (2017)"},{"issue":"3","key":"38_CR15","first-page":"152","volume":"17","author":"I Answar","year":"2017","unstructured":"Answar, I., Islam, N.: Learned features are better for ethnicity classification. Cybern. Inf. Technol. 17(3), 152\u2013164 (2017)","journal-title":"Cybern. Inf. Technol."},{"key":"38_CR16","doi-asserted-by":"publisher","first-page":"22","DOI":"10.1016\/j.neunet.2020.06.015","volume":"130","author":"M Deng","year":"2020","unstructured":"Deng, M., Meng, T., Cao, J., Wang, S., Zhang, J., Fan, H.: Heart sound classification based on improved MFCC features and convolutional recurrent neural networks. Neural Netw. 130, 22\u201332 (2020)","journal-title":"Neural Netw."},{"issue":"3","key":"38_CR17","doi-asserted-by":"publisher","first-page":"1694","DOI":"10.1121\/1.4977749","volume":"141","author":"S Ntalampiras","year":"2017","unstructured":"Ntalampiras, S.: A transfer learning framework for predicting the emotional content of generalized sound events. J. Acoust. Soc. Am. 141(3), 1694\u20131701 (2017)","journal-title":"J. Acoust. Soc. Am."},{"key":"38_CR18","doi-asserted-by":"publisher","first-page":"7717","DOI":"10.1109\/ACCESS.2018.2888882","volume":"7","author":"A Khamparia","year":"2019","unstructured":"Khamparia, A., Gupta, D., Nguyen, N., Khanna, A., Pandey, B., Tiwari, P.: Sound classification using convolutional neural network and tensor deep stacking network. IEEE Access 7, 7717\u20137727 (2019)","journal-title":"IEEE Access"},{"key":"38_CR19","doi-asserted-by":"crossref","unstructured":"Piczak, K.: ESC: dataset for environmental sound classification. In: Proceedings of the 23rd ACM International Conference on Multimedia (2015)","DOI":"10.1145\/2733373.2806390"},{"key":"38_CR20","doi-asserted-by":"crossref","unstructured":"Pinto, H., Almeida, J.M., Goncalves, M.A.: Using early view patterns to predict the popularity of Youtube videos. In: Proceedings of the Sixth ACM International Conference on Web Search and Data Mining, pp. 365\u2013374 (2013)","DOI":"10.1145\/2433396.2433443"},{"key":"38_CR21","doi-asserted-by":"crossref","unstructured":"Chen, X., Chen, J., Ma, L., Liu, W., Luo, J., Zhang, T.: Fine-grained video attractiveness prediction using multimodal deep learning on a large real-world dataset. In: Companion Proceedings of the Web Conference, pp. 671\u2013678 (2018)","DOI":"10.1145\/3184558.3186584"},{"key":"38_CR22","unstructured":"Li, Y., Eng, K., Zhang, L.: Youtube Videos Prediction: Will this video be popular? Stanford University (2019)"},{"key":"38_CR23","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Dollar, P.: Focal loss for dense object detection. In: Proceedings of the IEEE, (ICCV), pp. 2980\u20132988 (2017)","DOI":"10.1109\/ICCV.2017.324"},{"key":"38_CR24","doi-asserted-by":"crossref","unstructured":"Ruiz, N., Chong, E., Rehg, J.M.: Fine-grained head pose estimation without keypoints. In: IEEE Computer Vision and Pattern Recognition Workshops (2018)","DOI":"10.1109\/CVPRW.2018.00281"},{"issue":"2","key":"38_CR25","first-page":"151","volume":"4","author":"S Mohammad","year":"2012","unstructured":"Mohammad, S., Mohammad, H., Barlet, K., Trinh, P., Cohn, F.: DISFA: a spontaneous facial action intensity database. IEEE Trans. Affect. Comput. 4(2), 151\u2013160 (2012)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"38_CR26","doi-asserted-by":"publisher","first-page":"321","DOI":"10.1007\/s11263-020-01378-z","volume":"129","author":"Z Shao","year":"2021","unstructured":"Shao, Z., Liu, Z., Cail, J., Ma, L.: JAA-Net: joint facial action unit detection and face alignment via adaptive attention. Int. J. Comput. Vision 129, 321\u2013340 (2021)","journal-title":"Int. J. Comput. Vision"},{"key":"38_CR27","unstructured":"Vafeiadis, A., et al.: Acoustic scene classification: from a hybrid classifier to deep learning. In: DCASE Workshop, Munich, Germany, pp. 123\u2013127 (2017)"},{"key":"38_CR28","doi-asserted-by":"crossref","unstructured":"Koestinger, M., Wohlhart, P., Roth, P., Bischof, H.: Annotated facial landmarks in the wild: a largescale, real-world database for facial landmark localization. In: IEEE International Conference on Computer Vision Workshops, Barcelona, pp. 2144\u20132151 (2011)","DOI":"10.1109\/ICCVW.2011.6130513"},{"key":"38_CR29","unstructured":"Sasaki, Y.: The truth of the F-measure. Teach Tutor Master (2007)"},{"key":"38_CR30","doi-asserted-by":"publisher","first-page":"79","DOI":"10.3354\/cr030079","volume":"30","author":"CJ Willmott","year":"2005","unstructured":"Willmott, C.J., Matsuura, K.: Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance. Climate Res. 30, 79\u201382 (2005)","journal-title":"Climate Res."}],"container-title":["Proceedings of the International Neural Networks Society","Proceedings of the 22nd Engineering Applications of Neural Networks Conference"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-80568-5_38","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,28]],"date-time":"2022-06-28T07:08:48Z","timestamp":1656400128000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-80568-5_38"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030805678","9783030805685"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-80568-5_38","relation":{},"ISSN":["2661-8141","2661-815X"],"issn-type":[{"value":"2661-8141","type":"print"},{"value":"2661-815X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"1 July 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; nor in the decision to publish the results.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of Interest"}},{"value":"EANN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Engineering Applications of Neural Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Crete","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Greece","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25 June 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 June 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eann2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.eann2021.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}