{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T21:24:27Z","timestamp":1776115467830,"version":"3.50.1"},"reference-count":59,"publisher":"Springer Science and Business Media LLC","issue":"11","license":[{"start":{"date-parts":[[2024,1,19]],"date-time":"2024-01-19T00:00:00Z","timestamp":1705622400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,19]],"date-time":"2024-01-19T00:00:00Z","timestamp":1705622400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100012165","name":"Key Technologies Research and Development Program","doi-asserted-by":"publisher","award":["2022YFB3303301"],"award-info":[{"award-number":["2022YFB3303301"]}],"id":[{"id":"10.13039\/501100012165","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No. 62107035"],"award-info":[{"award-number":["No. 62107035"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No. 62006208"],"award-info":[{"award-number":["No. 62006208"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,11]]},"DOI":"10.1007\/s00371-023-03239-5","type":"journal-article","created":{"date-parts":[[2024,1,19]],"date-time":"2024-01-19T16:02:17Z","timestamp":1705680137000},"page":"8345-8361","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Hearing with the eyes: modulating lyrics typography for music visualization"],"prefix":"10.1007","volume":"40","author":[{"given":"Kaixin","family":"Han","sequence":"first","affiliation":[]},{"given":"Weitao","family":"You","sequence":"additional","affiliation":[]},{"given":"Shuhui","family":"Shi","sequence":"additional","affiliation":[]},{"given":"Lingyun","family":"Sun","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,19]]},"reference":[{"key":"3239_CR1","doi-asserted-by":"crossref","unstructured":"Fadillah, C., Rahayu, R. R. A.: Sound Visualization Using Typography Composition Based GIF. In:\u00a02019 International Conference on Sustainable Engineering and Creative Computing (ICSECC)\u00a0pp. 309\u2013314. IEEE. (2019)","DOI":"10.1109\/ICSECC.2019.8907207"},{"issue":"7","key":"3239_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3461835","volume":"54","author":"HB Lima","year":"2021","unstructured":"Lima, H.B., Santos, C.G.D., Meiguins, B.S.: A survey of music visualization techniques. ACM Comput Surv (CSUR) 54(7), 1\u201329 (2021)","journal-title":"ACM Comput Surv (CSUR)"},{"key":"3239_CR3","doi-asserted-by":"crossref","unstructured":"Castro, J. C. E., Martins, P., Boavida, A., and Machado, P.: \u00abM\u00e1quina de Ouver\u00bb-from sound to type: finding the visual representation of speech by mapping sound features to typographic variables. In:\u00a0Proceedings of the 9th International Conference on Digital and Interactive Arts\u00a0pp. 1\u20138. (2019)","DOI":"10.1145\/3359852.3359892"},{"issue":"1","key":"3239_CR4","doi-asserted-by":"publisher","first-page":"6","DOI":"10.1109\/TAFFC.2022.3174721","volume":"14","author":"C de Lacerda Pataca","year":"2022","unstructured":"de Lacerda Pataca, C., Costa, P.D.P.: Hidden bawls, whispers, and yelps: can text convey the sound of speech, beyond words? IEEE Trans. Affect. Comput. 14(1), 6\u201316 (2022)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"3239_CR5","doi-asserted-by":"crossref","unstructured":"de Lacerda Pataca, C., Costa, P. D. P.: Speech modulated typography: towards an affective representation model. In:\u00a0Proceedings of the 25th International Conference on Intelligent User Interfaces\u00a0pp. 139\u2013143. (2020)","DOI":"10.1145\/3377325.3377526"},{"key":"3239_CR6","doi-asserted-by":"crossref","unstructured":"Parente, J., Martins, T., Bicker, J., Machado, P.: Speechtyper: From speech to typographic composition. In\u00a0International Conference on Computational Intelligence in Music, Sound, Art and Design (Part of EvoStar)\u00a0pp. 212\u2013227. Cham: Springer International Publishing (2022)","DOI":"10.1007\/978-3-031-03789-4_14"},{"key":"3239_CR7","doi-asserted-by":"crossref","unstructured":"Matthias, W., Tim, S., and Angelo, S.: Voice driven type design. In: International Conference on Speech Technology and Human-Computer Dialogue, SpeD 2015, Bucharest, Romania, October 14\u201317, IEEE, pp. 1\u20139 (2015)","DOI":"10.1109\/SPED.2015.7343095"},{"key":"3239_CR8","unstructured":"de Lacerda Pataca, C., Watkins, M., Peiris, R., Lee, S., Huenerfauth, M.: Visualization of Speech Prosody and Emotion in Captions: Accessibility for Deaf and Hard-of-Hearing Users [n.d.]"},{"issue":"2022","key":"3239_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.jretconser.2021.102724","volume":"64","author":"Y Jun","year":"2022","unstructured":"Jun, Y., Lee, H.: A sound brand identity design: The interplay between sound symbolism and typography on brand attitude and memory. J. Retail. Consum. Serv. 64(2022), 102724 (2022)","journal-title":"J. Retail. Consum. Serv."},{"key":"3239_CR10","doi-asserted-by":"crossref","unstructured":"Schlippe, T., Alessai, S., El-Taweel, G., W\u00f6lfel, M., Zaghouani, W.: Visualizing voice characteristics with type design in closed captions for arabic. In:\u00a02020 International Conference on Cyberworlds (CW)\u00a0pp. 196\u2013203. IEEE. (2020)","DOI":"10.1109\/CW49994.2020.00039"},{"key":"3239_CR11","doi-asserted-by":"crossref","unstructured":"Kato, J., Nakano, T., Goto, M.: TextAlive: Integrated design environment for kinetic typography. In:\u00a0Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems\u00a0pp. 3403-3412. (2015)","DOI":"10.1145\/2702123.2702140"},{"key":"3239_CR12","doi-asserted-by":"crossref","unstructured":"Miniukovich, A., Scaltritti, M., Sulpizio, S., & De Angeli, A.: Guideline-based evaluation of web readability. In:\u00a0Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems\u00a0pp. 1\u201312. (2019)","DOI":"10.1145\/3290605.3300738"},{"key":"3239_CR13","doi-asserted-by":"crossref","unstructured":"Zhiquan, Y., Scott, E.H.: KTE2: an engine for kinetic typography. In: Proceedings of the 27th International Conference on Human Factors in Computing Systems, CHI 2009, Extended Abstracts Volume, Boston, MA, USA, April 4\u20139, 2009, Dan R. Olsen Jr., Richard B. Arthur, Ken Hinckley, Meredith Ringel Morris, Scott E. Hudson, and Saul Greenberg (Eds.). ACM, pp. 3413\u20133418 (2009)","DOI":"10.1145\/1520340.1520495"},{"key":"3239_CR14","doi-asserted-by":"crossref","unstructured":"Lee, J., Hayashi, M., Nosaka, M.: The effect of a visual element on musical sensitivity. In:\u00a0HCI International 2011\u2013Posters\u2019 Extended Abstracts: International Conference, HCI International 2011, Orlando, FL, USA, July 9-14, 2011, Proceedings, Part I 14\u00a0(pp 303\u2013307). Springer Berlin Heidelberg. (2011)","DOI":"10.1007\/978-3-642-22098-2_61"},{"issue":"6","key":"3239_CR15","doi-asserted-by":"publisher","first-page":"922","DOI":"10.1162\/jocn_a_01395","volume":"31","author":"C Fassnidge","year":"2019","unstructured":"Fassnidge, C., Ball, D., Kazaz, Z., et al.: Hearing through your eyes: neural basis of audiovisual cross-activation, revealed by transcranial alternating current stimulation[J]. J. Cogn. Neurosci. 31(6), 922\u2013935 (2019)","journal-title":"J. Cogn. Neurosci."},{"key":"3239_CR16","first-page":"61","volume":"8","author":"AZM Raden","year":"2019","unstructured":"Raden, A.Z.M., Qeis, M.I.: Song and typography: expressing the lyrics visually through lyrical typography[J]. Int. J. Sci. Technol. Res. 8, 61\u201364 (2019)","journal-title":"Int. J. Sci. Technol. Res."},{"key":"3239_CR17","unstructured":"Lee, S. W., Essl, G.: Web-based temporal typography for musical expression and performance. In:\u00a0NIME\u00a0pp. 65\u201369 (2015)"},{"issue":"1\u20132","key":"3239_CR18","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1080\/00293652.2021.1951830","volume":"54","author":"E Mol","year":"2021","unstructured":"Mol, E.: \u2018Trying to hear with the eyes\u2019: slow looking and ontological difference in archaeological object analysis[J]. Nor. Archaeol. Rev. 54(1\u20132), 80\u201399 (2021)","journal-title":"Nor. Archaeol. Rev."},{"issue":"3","key":"3239_CR19","doi-asserted-by":"publisher","first-page":"51","DOI":"10.18848\/1835-2014\/CGP\/v12i03\/51-64","volume":"12","author":"M Eikelenboom","year":"2019","unstructured":"Eikelenboom, M., Wattel, R., De Vet, M.: Listening with your eyes: an accessible museum for deaf and hard-of-hearing visitors[J]. Int J Inclus Museum 12(3), 51 (2019)","journal-title":"Int J Inclus Museum"},{"issue":"3","key":"3239_CR20","doi-asserted-by":"publisher","first-page":"228","DOI":"10.1111\/j.0956-7976.2005.00808.x","volume":"16","author":"SE Guttman","year":"2005","unstructured":"Guttman, S.E., Gilroy, L.A., Blake, R.: Hearing what the eyes see: Auditory encoding of visual temporal sequences[J]. Psychol. Sci. 16(3), 228\u2013235 (2005)","journal-title":"Psychol. Sci."},{"key":"3239_CR21","doi-asserted-by":"crossref","unstructured":"Pentimalli, B., Gobo, G.: \u201cHearing with the eyes\u201d visual hearing in (a trio) music rehearsals[J]. Sens. Soc., 1\u201319 (2023)","DOI":"10.1080\/17458927.2023.2232621"},{"issue":"4","key":"3239_CR22","doi-asserted-by":"publisher","first-page":"511","DOI":"10.1080\/03007760802184065","volume":"31","author":"M Katz","year":"2008","unstructured":"Katz, M.: Hearing through our eyes: musical archives and authentic performance[J]. Pop. Music Soc. 31(4), 511\u2013527 (2008)","journal-title":"Pop. Music Soc."},{"key":"3239_CR23","doi-asserted-by":"crossref","unstructured":"Rosenkvist, A., Eriksen, D.S., Koehlert, J., et al.: Hearing with eyes in virtual reality[C]. In: 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR). IEEE, pp. 1349\u20131350 (2019)","DOI":"10.1109\/VR.2019.8797903"},{"issue":"10","key":"3239_CR24","doi-asserted-by":"publisher","first-page":"1155","DOI":"10.1068\/p3020","volume":"29","author":"J MacDonald","year":"2000","unstructured":"MacDonald, J., Andersen, S., Bachmann, T.: Hearing by eye: how much spatial degradation can be tolerated?[J]. Perception 29(10), 1155\u20131168 (2000)","journal-title":"Perception"},{"key":"3239_CR25","unstructured":"Ludden, D.: Hearing with our eyes, seeing with our ears[J]. Psychology Today, 19 (2015)"},{"key":"3239_CR26","doi-asserted-by":"crossref","unstructured":"Agrawal, C., Peiris, R.L.: I see what you\u2019re saying: A literature review of eye tracking research in communication of deaf or hard of hearing users[C] In: Proceedings of the 23rd International ACM SIGACCESS Conference on Computers and Accessibility. pp. 1\u201313 (2021)","DOI":"10.1145\/3441852.3471209"},{"key":"3239_CR27","unstructured":"Toledo, J., Torres, J., Alonso, S., et al.: SLOAS: hearing with the eyes[C] In: Proceedings of the 4th WSEAS international conference on electronic, signal processing and control. 27 (2005)"},{"key":"3239_CR28","first-page":"1","volume":"2020","author":"JA Deja","year":"2020","unstructured":"Deja, J.A., Dela Torre, A., Lee, H.J., et al.: Vitune: a visualizer tool to allow the deaf and hard of hearing to see music with their eyes[C]\/\/extended abstracts of the. CHI Conf Human Factor Comput Syst. 2020, 1\u20138 (2020)","journal-title":"CHI Conf Human Factor Comput Syst."},{"issue":"1","key":"3239_CR29","doi-asserted-by":"publisher","first-page":"161","DOI":"10.1109\/TVCG.2009.63","volume":"16","author":"WY Chan","year":"2009","unstructured":"Chan, W.Y., Qu, H., Mak, W.H.: Visualizing the semantic structure in classical music works. IEEE Trans Visualizat Comput Graph 16(1), 161\u2013173 (2009)","journal-title":"IEEE Trans Visualizat Comput Graph"},{"issue":"4","key":"3239_CR30","first-page":"1","volume":"3","author":"E Chew","year":"2005","unstructured":"Chew, E., Fran\u00e7ois, A.R.: Interactive multi-scale visualizations of tonal evolution in MuSA. RT Opus 2. Comput Entertain (CIE) 3(4), 1\u201316 (2005)","journal-title":"Comput Entertain (CIE)"},{"key":"3239_CR31","doi-asserted-by":"crossref","unstructured":"Lima, H., Santos, C., Meiguins, B.: Visualizing the semantics of music. In\u00a02019 23rd International Conference Information Visualisation (IV)\u00a0pp. 352\u2013357. IEEE. (2019)","DOI":"10.1109\/IV.2019.00066"},{"key":"3239_CR32","doi-asserted-by":"crossref","unstructured":"Ciuha, P., Klemenc, B., & Solina, F.: Visualization of concurrent tones in music with colours. In:\u00a0Proceedings of the 18th ACM international conference on Multimedia\u00a0pp. 1677\u20131680. (2010)","DOI":"10.1145\/1873951.1874320"},{"key":"3239_CR33","doi-asserted-by":"crossref","unstructured":"Hiraga, R., Mizaki, R., Fujishiro, I.: Performance visualization: a new challenge to music through visualization. In:\u00a0Proceedings of the tenth ACM international conference on Multimedia\u00a0pp. 239\u2013242. (2002)","DOI":"10.1145\/641007.641054"},{"key":"3239_CR34","doi-asserted-by":"crossref","unstructured":"Malandrino, D., Pirozzi, D., Zaccagnino, G., Zaccagnino, R.: A color-based visualization approach to understand harmonic structures of musical compositions. In:\u00a02015 19th International Conference on Information Visualisation\u00a0pp. 56\u201361. IEEE. (2015)","DOI":"10.1109\/iV.2015.21"},{"issue":"2","key":"3239_CR35","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1177\/1473871616655468","volume":"16","author":"R De Prisco","year":"2017","unstructured":"De Prisco, R., Malandrino, D., Pirozzi, D., Zaccagnino, G., Zaccagnino, R.: Understanding the structure of musical compositions: Is visualization an effective approach? Inf Visualizat 16(2), 139\u2013152 (2017)","journal-title":"Inf Visualizat"},{"key":"3239_CR36","doi-asserted-by":"crossref","unstructured":"Kim, J., Ananthanarayan, S., Yeh, T.: Seen music: ambient music data visualization for children with hearing impairments. In:\u00a0Proceedings of the 14th International Conference on Interaction Design and Children\u00a0pp. 426\u2013429. (2015)","DOI":"10.1145\/2771839.2771870"},{"key":"3239_CR37","doi-asserted-by":"crossref","unstructured":"Lehtiniemi, A., Holm, J.: Using animated mood pictures in music recommendation. In\u00a02012 16th International Conference on Information Visualisation\u00a0pp. 143\u2013150. IEEE. (2012)","DOI":"10.1109\/IV.2012.34"},{"issue":"6","key":"3239_CR38","doi-asserted-by":"publisher","first-page":"472","DOI":"10.1016\/j.jvlc.2013.10.002","volume":"24","author":"JH Fonteles","year":"2013","unstructured":"Fonteles, J.H., Rodrigues, M.A.F., Basso, V.E.D.: Creating and evaluating a particle system for music visualization. J Visual Languages Comput 24(6), 472\u2013482 (2013)","journal-title":"J Visual Languages Comput"},{"issue":"11","key":"3239_CR39","doi-asserted-by":"publisher","first-page":"2238","DOI":"10.1109\/TMM.2016.2614226","volume":"18","author":"GD Cantareira","year":"2016","unstructured":"Cantareira, G.D., Nonato, L.G., Paulovich, F.V.: Moshviz: A detail+ overview approach to visualize music elements. IEEE Trans Multimed 18(11), 2238\u20132246 (2016)","journal-title":"IEEE Trans Multimed"},{"issue":"2","key":"3239_CR40","doi-asserted-by":"publisher","first-page":"218","DOI":"10.1145\/965103.807447","volume":"13","author":"JB Mitroo","year":"1979","unstructured":"Mitroo, J.B., Herman, N., Badler, N.I.: Movies from music: visualizing musical compositions. ACM SIGGRAPH Comput Graph 13(2), 218\u2013225 (1979)","journal-title":"ACM SIGGRAPH Comput Graph"},{"key":"3239_CR41","doi-asserted-by":"crossref","unstructured":"Bergstrom, T., Karahalios, K., & Hart, J. C.: Isochords: visualizing structure in music. In:\u00a0Proceedings of Graphics Interface 2007\u00a0pp. 297-304 (2007)","DOI":"10.1145\/1268517.1268565"},{"key":"3239_CR42","doi-asserted-by":"crossref","unstructured":"Fujishiro, I., Haga, N., Nakayama, M.: SeeGroove: Supporting groove learning through visualization. In:\u00a02015 International Conference on Cyberworlds (CW)\u00a0pp. 189\u2013192. IEEE. (2015)","DOI":"10.1109\/CW.2015.65"},{"key":"3239_CR43","doi-asserted-by":"crossref","unstructured":"Machida, W., Itoh, T.: Lyricon: A visual music selection interface featuring multiple icons. In:\u00a02011 15th International Conference on Information Visualisation\u00a0pp. 145\u2013150 IEEE. (2011)","DOI":"10.1109\/IV.2011.62"},{"key":"3239_CR44","doi-asserted-by":"crossref","unstructured":"Gumulia, A., Puzon, B., Kosugi, N.: Music visualization: predicting the perceived speed of a composition--misual project--. In:\u00a0Proceedings of the 19th ACM international conference on Multimedia\u00a0pp. 949\u2013952 (2011)","DOI":"10.1145\/2072298.2071910"},{"key":"3239_CR45","doi-asserted-by":"crossref","unstructured":"Snydal, J., Hearst, M.: ImproViz: visual explorations of jazz improvisations. In:\u00a0CHI'05 extended abstracts on Human factors in computing systems\u00a0pp. 1805\u20131808 (2005)","DOI":"10.1145\/1056808.1057027"},{"key":"3239_CR46","doi-asserted-by":"crossref","unstructured":"Hiraga, R., Watanabe, F., Fujishiro, I.: Music learning through visualization. In:\u00a0Second International Conference on Web Delivering of Music, 2002. WEDELMUSIC 2002. Proceedings.\u00a0pp. 101\u2013108 IEEE (2002)","DOI":"10.1109\/WDM.2002.1176199"},{"key":"3239_CR47","unstructured":"Isaacson, E. J.: What You See Is What You Get: on Visualizing Music. In:\u00a0ISMIR\u00a0(pp. 389-395). (2005)"},{"key":"3239_CR48","doi-asserted-by":"crossref","unstructured":"Aoki, T., Chujo, R., Matsui, K., Choi, S., and Hautasaari, A.: Emoballoon-conveying emotional arousal in text chats with speech balloons. In:\u00a0Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems\u00a0pp. 1-16 (2022)","DOI":"10.1145\/3491102.3501920"},{"key":"3239_CR49","unstructured":"Mardirossian, A., Chew, E.: Visualizing Music: Tonal Progressions and Distributions. In:\u00a0ISMIR\u00a0pp. 189\u2013194 (2007)"},{"key":"3239_CR50","doi-asserted-by":"publisher","first-page":"503","DOI":"10.1007\/978-3-642-55038-6_79","volume-title":"Future Information Technology, James J","author":"S Lim","year":"2014","unstructured":"Lim, S., Kim, S.: Implementation of Kinetic Typography by Motion Recognition Sensor. In: Jong Hyuk) Park, Yi Pan, Cheon-Shik Kim, and Yun Yang, (ed.) Future Information Technology, James J, pp. 503\u2013507. Springer, Berlin Heidelberg, Berlin, Heidelberg (2014)"},{"key":"3239_CR51","doi-asserted-by":"publisher","first-page":"583","DOI":"10.1016\/j.sbspro.2012.08.209","volume":"51","author":"OP Turgut","year":"2012","unstructured":"Turgut, O.P.: Kinetic typography in movie title sequences. Procedia-Soc Behav Sci 51, 583\u2013588 (2012)","journal-title":"Procedia-Soc Behav Sci"},{"key":"3239_CR52","doi-asserted-by":"crossref","unstructured":"\u00d6zden, P.T.: Calligraphic forms in contemporary typographic design. Procedia\u2014Social and Behavioral Sciences 122, 40\u201345. 2nd World Conference on Design, Arts and Education (DAE-2013). (2014)","DOI":"10.1016\/j.sbspro.2014.01.1300"},{"key":"3239_CR53","unstructured":"Fellows, K. S.: Typecast: the voice of typography\u00a0(Doctoral dissertation, University of Iowa). (2009)"},{"key":"3239_CR54","doi-asserted-by":"crossref","unstructured":"Kadner, F., Keller, Y., Rothkopf, C.: Adaptifont: Increasing individuals\u2019 reading speed with a generative font model and Bayesian optimization. In:\u00a0Proceedings of the 2021 chi conference on human factors in computing systems\u00a0pp. 1\u201311 (2021)","DOI":"10.1145\/3411764.3445140"},{"key":"3239_CR55","unstructured":"Ekstrom, M.P.: Digital image processing techniques[M]. Academic Press, (2012)"},{"issue":"4","key":"3239_CR56","doi-asserted-by":"publisher","first-page":"755","DOI":"10.1080\/10447318.2022.2049081","volume":"39","author":"C So","year":"2023","unstructured":"So, C.: Measuring aesthetic preferences of neural style transfer: more precision with the two-alternative-forced-choice task[J]. Int J Human-Comput Interact 39(4), 755\u2013775 (2023)","journal-title":"Int J Human-Comput Interact"},{"key":"3239_CR57","unstructured":"Koch, B. E.: Human emotion response to typographic design\u00a0(Doctoral dissertation, University of Minnesota). (2011)"},{"issue":"1","key":"3239_CR58","first-page":"122","volume":"10","author":"S Lim","year":"2022","unstructured":"Lim, S.: A study on the interactive expression of human emotions in typography. Int J Adv Cult Technol 10(1), 122\u2013130 (2022)","journal-title":"Int J Adv Cult Technol"},{"key":"3239_CR59","doi-asserted-by":"crossref","unstructured":"Ma\u00e7\u00e3s, C., Palma, D., Rebelo, A.: Typem: A generative typeface that represents the emotion of the text. In:\u00a0Proceedings of the 9th international conference on digital and interactive arts\u00a0pp. 1\u201310 (2019)","DOI":"10.1145\/3359852.3359874"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03239-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-03239-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03239-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,6]],"date-time":"2024-11-06T16:14:22Z","timestamp":1730909662000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-03239-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,19]]},"references-count":59,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2024,11]]}},"alternative-id":["3239"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-03239-5","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,1,19]]},"assertion":[{"value":"14 December 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 January 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}