{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T06:40:09Z","timestamp":1750228809646,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":16,"publisher":"ACM","license":[{"start":{"date-parts":[[2015,6,3]],"date-time":"2015-06-03T00:00:00Z","timestamp":1433289600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2015,6,3]]},"DOI":"10.1145\/2745197.2755512","type":"proceedings-article","created":{"date-parts":[[2015,6,2]],"date-time":"2015-06-02T05:35:02Z","timestamp":1433223302000},"page":"125-130","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":6,"title":["SensiTV"],"prefix":"10.1145","author":[{"given":"Diana","family":"Affi","sequence":"first","affiliation":[{"name":"HumanTech Institute, Fribourg, Switzerland"}]},{"given":"Jo\u00ebl","family":"Dumoulin","sequence":"additional","affiliation":[{"name":"HumanTech Institute, Fribourg, Switzerland"}]},{"given":"Marco","family":"Bertini","sequence":"additional","affiliation":[{"name":"MICC, Florence, Italy"}]},{"given":"Elena","family":"Mugellini","sequence":"additional","affiliation":[{"name":"University of Applied Sciences of Western Switzerland - Fribourg, Fribourg, Switzerland"}]},{"given":"Omar","family":"Abou Khaled","sequence":"additional","affiliation":[{"name":"University of Applied Sciences of Western Switzerland, Fribourg, Switzerland"}]},{"given":"Alberto","family":"Del Bimbo","sequence":"additional","affiliation":[{"name":"MICC, Florence, Italy"}]}],"member":"320","published-online":{"date-parts":[[2015,6,3]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"1\n  .  C. Becker S. Kopp and I. Wachsmuth. 2001. Why emotions should be integrated into conversational agents. (2001).  1. C. Becker S. Kopp and I. Wachsmuth. 2001. Why emotions should be integrated into conversational agents. (2001)."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/2502081.2502282"},{"volume-title":"Proc. of IEEE ICIIS. 381--386","author":"De Silva J.","key":"e_1_3_2_1_3_1","unstructured":"3 . J. De Silva and P. S. Haddela . 2013. A term weighting method for identifying emotions from text content . In Proc. of IEEE ICIIS. 381--386 . 3. J. De Silva and P. S. Haddela. 2013. A term weighting method for identifying emotions from text content. In Proc. of IEEE ICIIS. 381--386."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"crossref","unstructured":"4\n  .  A. Dhall R. Goecke J. Joshi and T. Gedeon. 2014. Emotion Recognition In The Wild Challenge 2014 : Baseline Data and Protocol Categories and Subject Descriptors. (2014).  4. A. Dhall R. Goecke J. Joshi and T. Gedeon. 2014. Emotion Recognition In The Wild Challenge 2014 : Baseline Data and Protocol Categories and Subject Descriptors. (2014).","DOI":"10.1145\/2663204.2666275"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/MMUL.2012.26"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.1469-7610.2004.00272.x"},{"key":"e_1_3_2_1_7_1","unstructured":"7\n  .  P. Ekman. 2003. Emotions Revealed.  7. P. Ekman. 2003. Emotions Revealed."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/2502081.2502224"},{"volume-title":"Proc. of IEEE ICCIS. 1068--1072","author":"Firdus S.","key":"e_1_3_2_1_9_1","unstructured":"9 . S. Firdus , W. Fatimah , W. Ahmad , and J. B. Janier . 2012. Development of Audio Video Describer Using Narration to Visualize Movie Film for Blind and Visually Impaired Children . In Proc. of IEEE ICCIS. 1068--1072 . 9. S. Firdus, W. Fatimah, W. Ahmad, and J. B. Janier. 2012. Development of Audio Video Describer Using Narration to Visualize Movie Film for Blind and Visually Impaired Children. In Proc. of IEEE ICCIS. 1068--1072."},{"volume-title":"Proc. of IEEE CBAIVL. 50--57","author":"Hanjalic A.","key":"e_1_3_2_1_10_1","unstructured":"10 . A. Hanjalic , L. Xu , C. D. Delft , A. Park , M. Heath , and I. Ip . 2001. User-oriented Affective Video Content Analysis . In Proc. of IEEE CBAIVL. 50--57 . 10. A. Hanjalic, L. Xu, C. D. Delft, A. Park, M. Heath, and I. Ip. 2001. User-oriented Affective Video Content Analysis. In Proc. of IEEE CBAIVL. 50--57."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2004.840618"},{"key":"e_1_3_2_1_12_1","first-page":"3","article-title":"2006. Touch communicates distinct emotions. Emotion (Washington","volume":"6","author":"Hertenstein M. J.","year":"2006","unstructured":"12 . M. J. Hertenstein , D. Keltner , B. App , B. a. Bulleit , and A. R. Jaskolka . 2006. Touch communicates distinct emotions. Emotion (Washington , D.C.) 6 , 3 ( Aug. 2006 ), 528--33. DOI:http:\/\/dx.doi.org\/10.1037\/1528--3542. 6.3.528 10.1037\/1528--3542 12. M. J. Hertenstein, D. Keltner, B. App, B. a. Bulleit, and A. R. Jaskolka. 2006. Touch communicates distinct emotions. Emotion (Washington, D.C.) 6, 3 (Aug. 2006), 528--33. DOI:http:\/\/dx.doi.org\/10.1037\/1528--3542. 6.3.528","journal-title":"D.C.)"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSMC.2006.384366"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/2663204.2666274"},{"volume-title":"Proc. of ICALT. 2--4.","author":"Shipsey J.","key":"e_1_3_2_1_15_1","unstructured":"15 . J. Ohene-djan and R. Shipsey . 2006. E-Subtitles : Emotional Subtitles as a Technology to assist the Deaf and Hearing-Impaired when Learning from Television and Film . In Proc. of ICALT. 2--4. 15. J. Ohene-djan and R. Shipsey. 2006. E-Subtitles : Emotional Subtitles as a Technology to assist the Deaf and Hearing-Impaired when Learning from Television and Film. In Proc. of ICALT. 2--4."},{"volume-title":"Proc. Interspeech. 2794--2797","author":"Schuller B.","key":"e_1_3_2_1_16_1","unstructured":"16 . B. Schuller , S. Steidl , A. Batliner , F. Burkhardt , L. Devillers , M. Christian , and S. Narayanan . 2010. The INTERSPEECH 2010 Paralinguistic Challenge . In Proc. Interspeech. 2794--2797 . 16. B. Schuller, S. Steidl, A. Batliner, F. Burkhardt, L. Devillers, M. Christian, and S. Narayanan. 2010. The INTERSPEECH 2010 Paralinguistic Challenge. In Proc. Interspeech. 2794--2797."}],"event":{"name":"TVX'15: ACM International Conference on Interactive Experiences for TV and Online Video","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"],"location":"Brussels Belgium","acronym":"TVX'15"},"container-title":["Proceedings of the ACM International Conference on Interactive Experiences for TV and Online Video"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/2745197.2755512","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/2745197.2755512","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T06:12:33Z","timestamp":1750227153000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/2745197.2755512"}},"subtitle":["Smart EmotioNal System for Impaired People's TV"],"short-title":[],"issued":{"date-parts":[[2015,6,3]]},"references-count":16,"alternative-id":["10.1145\/2745197.2755512","10.1145\/2745197"],"URL":"https:\/\/doi.org\/10.1145\/2745197.2755512","relation":{},"subject":[],"published":{"date-parts":[[2015,6,3]]},"assertion":[{"value":"2015-06-03","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}