{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,23]],"date-time":"2025-09-23T12:37:53Z","timestamp":1758631073712,"version":"3.40.5"},"reference-count":71,"publisher":"Cambridge University Press (CUP)","issue":"4","license":[{"start":{"date-parts":[[2019,12,4]],"date-time":"2019-12-04T00:00:00Z","timestamp":1575417600000},"content-version":"unspecified","delay-in-days":33,"URL":"https:\/\/www.cambridge.org\/core\/terms"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["AIEDAM"],"published-print":{"date-parts":[[2019,11]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Universal access on equal terms to audiovisual content is a key point for the full inclusion of people with disabilities in activities of daily life. As a real challenge for the current Information Society, it has been detected but not achieved in an efficient way, due to the fact that current access solutions are mainly based in the traditional television standard and other not automated high-cost solutions. The arrival of new technologies within the hybrid television environment together with the application of different artificial intelligence techniques over the content will assure the deployment of innovative solutions for enhancing the user experience for all. In this paper, a set of different tools for image enhancement based on the combination between deep learning and computer vision algorithms will be presented. These tools will provide automatic descriptive information of the media content based on face detection for magnification and character identification. The fusion of this information will be finally used to provide a customizable description of the visual information with the aim of improving the accessibility level of the content, allowing an efficient and reduced cost solution for all.<\/jats:p>","DOI":"10.1017\/s0890060419000350","type":"journal-article","created":{"date-parts":[[2019,12,4]],"date-time":"2019-12-04T01:48:24Z","timestamp":1575424104000},"page":"399-415","source":"Crossref","is-referenced-by-count":2,"title":["New access services in HbbTV based on a deep learning approach for media content analysis"],"prefix":"10.1017","volume":"33","author":[{"given":"Silvia","family":"Uribe","sequence":"first","affiliation":[]},{"given":"Alberto","family":"Belmonte","sequence":"additional","affiliation":[]},{"given":"Francisco","family":"Moreno","sequence":"additional","affiliation":[]},{"given":"\u00c1lvaro","family":"Llorente","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5400-2747","authenticated-orcid":false,"given":"Juan Pedro","family":"L\u00f3pez","sequence":"additional","affiliation":[]},{"given":"Federico","family":"\u00c1lvarez","sequence":"additional","affiliation":[]}],"member":"56","published-online":{"date-parts":[[2019,12,4]]},"reference":[{"unstructured":"CNMC (2017) Informe sobre el seguimiento de las obligaciones impuestas en materia de accesibilidad correspondiente al a\u00f1o 2016. Available at https:\/\/www.cnmc.es\/sites\/default\/files\/1855187_9.pdf","key":"S0890060419000350_ref10"},{"unstructured":"NIELSEN b (2017) The Nielsen comparable metrics report, Q2-2016. Available at https:\/\/www.nielsen.com\/us\/en\/insights\/reports\/2016\/the-comparable-metrics-report-q2-2016.html","key":"S0890060419000350_ref43"},{"unstructured":"Simonyan K and Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. ArXiv Preprint ArXiv:1409.1556.","key":"S0890060419000350_ref57"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref66","DOI":"10.1109\/ISDA.2011.6121761"},{"unstructured":"Ren S , He K , Girshick RB and Sun J (2015) Faster R-CNN: towards real-time object detection with region proposal networks. CoRR, abs\/1506.0. Available at http:\/\/arxiv.org\/abs\/1506.01497.","key":"S0890060419000350_ref53"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref48","DOI":"10.1109\/ISCAS.2017.8050867"},{"unstructured":"Broad A , Jones M and Lee T-Y (2018) Recurrent multi-frame single shot detector for video object detection. British Machine Video Conference (BMVC), Newcastle, UK.","key":"S0890060419000350_ref5"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref37","DOI":"10.1109\/MCE.2013.2251760"},{"unstructured":"Messer K , Matas J , Kittler J , Luettin J and Maitre G (1999) XM2VTSDB: The extended M2VTS database. Second International Conference on Audio and Video-Based Biometric Person Authentication, Washington, DC, USA, Vol. 964, pp. 965\u2013966.","key":"S0890060419000350_ref40"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref49","DOI":"10.1109\/BMSB.2015.7177252"},{"unstructured":"Lin T-Y , Goyal P , Girshick RB , He K and Doll\u00e1r P (2017) Focal loss for dense object detection. CoRR, abs\/1708.0. Available at http:\/\/arxiv.org\/abs\/1708.02002","key":"S0890060419000350_ref32"},{"unstructured":"S\u00e1ez Trigueros D , Meng L and Hartnett M (2018) Face recognition: from traditional to deep learning methods. CoRR, abs\/1811.00116.","key":"S0890060419000350_ref55"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref71","DOI":"10.1109\/ICCE-Berlin.2013.6697990"},{"unstructured":"Prosperity4All Project (n.d.) Prosperity 4All project website. Available at http:\/\/www.prosperity4all.eu\/","key":"S0890060419000350_ref51"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref54","DOI":"10.1007\/s11263-016-0940-3"},{"unstructured":"NIELSEN c (2017) The Nielsen comparable metrics report, Q3-2016. Available at https:\/\/www.nielsen.com\/us\/en\/insights\/reports\/2017\/the-comparable-metrics-report-q3-2016.html","key":"S0890060419000350_ref44"},{"unstructured":"European Commission (2010) European disability strategy 2010-2020: a renewed commitment to a barrier-free Europe. Available at https:\/\/eur-lex.europa.eu\/LexUriServ\/LexUriServ.do?uri=COM:2010:0636:FIN:en:PDF","key":"S0890060419000350_ref18"},{"unstructured":"Zhu X and Ramanan D (2012) Face detection, pose estimation, and landmark localization in the wild. 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Provicende, Rhode Island, USA, pp. 2879\u20132886.","key":"S0890060419000350_ref70"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref33","DOI":"10.1109\/ICIP.2011.6116689"},{"unstructured":"Redmon J , Divvala SK , Girshick RB and Farhadi A (2015) You only look once: unified, real-time object detection. CoRR, abs\/1506.0. Available at http:\/\/arxiv.org\/abs\/1506.02640","key":"S0890060419000350_ref52"},{"unstructured":"eMarketer (2017) US simultaneous media users: eMarketer's estimates for 2017. Available at https:\/\/www.emarketer.com\/Report\/US-Simultaneous-Media-Users-eMarketers-Estimates-2017\/2002163","key":"S0890060419000350_ref16"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref24","DOI":"10.1109\/CVPR.2016.90"},{"unstructured":"Henriques JF , Caseiro R , Martins P and Batista J (2014) High-speed tracking with kernelized correlation filters. CoRR, abs\/1404.7. Available at http:\/\/arxiv.org\/abs\/1404.7584","key":"S0890060419000350_ref27"},{"key":"S0890060419000350_ref50","first-page":"466","article-title":"Evaluation of Haar cascade classifiers designed for face detection","volume":"6","author":"Padilla","year":"2012","journal-title":"World Academy of Science, Engineering and Technology International Journal of Computer and Information Engineering"},{"unstructured":"Yuheng S and Hao Y (2017) Image segmentation algorithms overview. CoRR, abs\/1707.0. Available at http:\/\/arxiv.org\/abs\/1707.02051","key":"S0890060419000350_ref67"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref2","DOI":"10.1109\/SITIS.2017.62"},{"unstructured":"EasyTV Project (n.d.) EasyTV project website. Available at https:\/\/easytvproject.eu\/","key":"S0890060419000350_ref15"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref23","DOI":"10.1007\/978-3-319-28854-3_2"},{"unstructured":"Matamala A , Orero P , Rovira-Esteva S , Casas Tost H , Morales Morante F , Soler Vilageliu O and Tor-Carroggio I (2018) User-centric approaches in access services evaluation: profiling the end user. Proceedings of the Eleventh International Conference on Language Resources Evaluation (LREC 2018), Miyazaki, Japan, pp. 1\u20137.","key":"S0890060419000350_ref38"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref14","DOI":"10.1109\/TBC.2017.2755403"},{"unstructured":"NIELSEN e (2018) The Nielsen comparable metrics report, Q1-2017. Available at https:\/\/www.nielsen.com\/us\/en\/insights\/reports\/2017\/the-nielsen-comparable-metrics-report-q1-2017.html","key":"S0890060419000350_ref46"},{"doi-asserted-by":"crossref","unstructured":"Agustsson E , Timofte R , Escalera S , Baro X , Guyon I and Rothe R (2017) Apparent and real age estimation in still images with deep residual regressors on APPA-REAL database. 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), Washington, DC, USA, pp. 87\u201394.","key":"S0890060419000350_ref1","DOI":"10.1109\/FG.2017.20"},{"unstructured":"Immersive Accessibility Project (n.d.) Immersive accessibility project website. Available at http:\/\/www.imac-project.eu\/","key":"S0890060419000350_ref29"},{"unstructured":"Held D , Thrun S and Savarese S (2016) Learning to track at 100 {FPS} with deep regression networks. CoRR, abs\/1604.0. Available at http:\/\/arxiv.org\/abs\/1604.01802","key":"S0890060419000350_ref26"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref3","DOI":"10.1109\/TPAMI.2013.23"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref4","DOI":"10.1007\/978-3-319-48881-3_56"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref8","DOI":"10.5121\/ijit.2014.3202"},{"unstructured":"Cao Z , Simon T , Wei S-E and Sheikh Y (2016) Realtime multi-person 2D pose estimation using part affinity fields. CoRR, abs\/1611.0. Available at http:\/\/arxiv.org\/abs\/1611.08050","key":"S0890060419000350_ref6"},{"key":"S0890060419000350_ref9","doi-asserted-by":"crossref","first-page":"52","DOI":"10.1109\/MSPEC.2012.6361764","article-title":"The broadcast empire strikes back","volume":"49","author":"Claudy","year":"2012","journal-title":"IEEE Spectrum"},{"unstructured":"CSA (2017) L'accessibilit\u00e9 des programmes de t\u00e9l\u00e9vision aux personnes handicap\u00e9es et la repr\u00e9sentation du h\u00e1ndicap \u00e0 l'antenne. Conseil Sup\u00e9rieur de L'audiovisuel. Rapport annuel 2016.","key":"S0890060419000350_ref11"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref12","DOI":"10.1109\/ICEMI.2017.8265863"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref13","DOI":"10.5244\/C.28.65"},{"unstructured":"ETSI (2016) Hybrid broadcast broadband TV ETSI standard TS 102 796 2016. Available at https:\/\/www.etsi.org\/deliver\/etsi_ts\/102700_102799\/102796\/01.04.01_60\/ts_102796v010401p.pdf","key":"S0890060419000350_ref17"},{"unstructured":"Feichtenhofer C , Pinz A and Zisserman A (2017) Detect to track and track to detect. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, Hawai, USA, pp. 3038\u20133046.","key":"S0890060419000350_ref19"},{"unstructured":"Castellano B (2018) Pyscenedetect. Available at https:\/\/pyscenedetect.readthedocs.io","key":"S0890060419000350_ref7"},{"unstructured":"Fiaz M , Mahmood A and Jung SK (2018) Tracking noisy targets: a review of recent object tracking approaches. ArXiv Preprint ArXiv:1802.03098.","key":"S0890060419000350_ref20"},{"unstructured":"Gordon D , Farhadi A and Fox D (2017) Re3: real-time recurrent regression networks for object tracking. CoRR, abs\/1705.0. Available at http:\/\/arxiv.org\/abs\/1705.06368","key":"S0890060419000350_ref21"},{"unstructured":"G\u00fcler RA , Neverova N and Kokkinos I (2018) DensePose: dense human pose estimation in the wild. CoRR, abs\/1802.0. Available at http:\/\/arxiv.org\/abs\/1802.00434","key":"S0890060419000350_ref22"},{"doi-asserted-by":"crossref","unstructured":"He K , Gkioxari G , Doll\u00e1r P and Girshick RB (2017) Mask {R-CNN}. CoRR, abs\/1703.0. Available at http:\/\/arxiv.org\/abs\/1703.06870","key":"S0890060419000350_ref25","DOI":"10.1109\/ICCV.2017.322"},{"unstructured":"Howard AG , Zhu M , Chen B , Kalenichenko D , Wang W , Weyand T , Andreetto M , Adam H (2017) Mobilenets: efficient convolutional neural networks for mobile vision applications. ArXiv Preprint ArXiv:1704.04861.","key":"S0890060419000350_ref28"},{"unstructured":"Jain V and Learned-Miller E (2010) FDDB: a benchmark for face detection in unconstrained settings.","key":"S0890060419000350_ref30"},{"unstructured":"Liu W , Anguelov D , Erhan D , Szegedy C , Reed SE , Fu C-Y and Berg AC (2015) SSD: Single Shot MultiBox Detector. CoRR, abs\/1512.0. Available at http:\/\/arxiv.org\/abs\/1512.02325.","key":"S0890060419000350_ref34"},{"unstructured":"Lukezic A , Vojir T , Cehovin L , Matas J and Kristan M (2016) Discriminative correlation filter with channel and spatial reliability. CoRR, abs\/1611.0. Available at http:\/\/arxiv.org\/abs\/1611.08461","key":"S0890060419000350_ref35"},{"unstructured":"Luo W , Xing J , Milan A , Zhang X , Liu W , Zhao X and Kim T-K (2014) Multiple object tracking: a literature review. ArXiv Preprint ArXiv:1409.7618.","key":"S0890060419000350_ref36"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref39","DOI":"10.1145\/3077548.3077555"},{"unstructured":"NIELSEN a (2017) The Nielsen comparable metrics report, Q1-2016. Available at https:\/\/www.nielsen.com\/us\/en\/insights\/reports\/2016\/the-comparable-metrics-report-q1-2016.html","key":"S0890060419000350_ref42"},{"unstructured":"NIELSEN d (2017) The Nielsen comparable metrics report, Q4-2016. Available at https:\/\/www.nielsen.com\/us\/en\/insights\/reports\/2017\/the-comparable-metrics-report-q4-2016.html","key":"S0890060419000350_ref45"},{"unstructured":"NIELSEN f (2018) The Nielsen comparable metrics report, Q2-2017. Available at https:\/\/www.nielsen.com\/us\/en\/insights\/reports\/2017\/the-nielsen-comparable-metrics-report-q2-2017.html","key":"S0890060419000350_ref47"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref41","DOI":"10.1109\/ICPR.2010.1145"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref56","DOI":"10.1016\/j.imavis.2016.01.002"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref58","DOI":"10.1109\/MMUL.2011.71"},{"unstructured":"Statista (2017) Smart TV shipments worldwide. Available at https:\/\/www.statista.com\/statistics\/461561\/smart-tv-shipments-worldwide-by-region\/","key":"S0890060419000350_ref59"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref60","DOI":"10.1145\/2212776.2212832"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref61","DOI":"10.1155\/2018\/7068349"},{"unstructured":"Wang M and Deng W (2018) Deep face recognition: a survey. ArXiv Preprint ArXiv:1804.06655.","key":"S0890060419000350_ref62"},{"unstructured":"Wolf L , Hassner T and Maoz I (2011) Face recognition in unconstrained videos with matched background similarity. 2011 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, pp. 529\u2013534.","key":"S0890060419000350_ref63"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref64","DOI":"10.1111\/j.1475-1313.2011.00833.x"},{"unstructured":"World Health Organization and others (2013) Universal eye health: a global action plan 2014-2019.","key":"S0890060419000350_ref65"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref68","DOI":"10.5244\/C.30.87"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref69","DOI":"10.1109\/CVPR.2017.463"},{"doi-asserted-by":"publisher","key":"S0890060419000350_ref31","DOI":"10.1007\/978-3-642-33712-3_49"}],"container-title":["Artificial Intelligence for Engineering Design, Analysis and Manufacturing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.cambridge.org\/core\/services\/aop-cambridge-core\/content\/view\/S0890060419000350","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2019,12,31]],"date-time":"2019-12-31T03:31:30Z","timestamp":1577763090000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.cambridge.org\/core\/product\/identifier\/S0890060419000350\/type\/journal_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,11]]},"references-count":71,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2019,11]]}},"alternative-id":["S0890060419000350"],"URL":"https:\/\/doi.org\/10.1017\/s0890060419000350","relation":{},"ISSN":["0890-0604","1469-1760"],"issn-type":[{"type":"print","value":"0890-0604"},{"type":"electronic","value":"1469-1760"}],"subject":[],"published":{"date-parts":[[2019,11]]}}}