{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,3]],"date-time":"2025-08-03T01:11:52Z","timestamp":1754183512542,"version":"3.41.2"},"reference-count":26,"publisher":"Institute of Electronics, Information and Communications Engineers (IEICE)","issue":"8","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEICE Trans. Inf. &amp; Syst."],"published-print":{"date-parts":[[2025,8,1]]},"DOI":"10.1587\/transinf.2024edp7262","type":"journal-article","created":{"date-parts":[[2025,2,17]],"date-time":"2025-02-17T17:12:47Z","timestamp":1739812367000},"page":"958-966","source":"Crossref","is-referenced-by-count":0,"title":["Exploring Sentence-Level Text-Font Retrieval via Contrastive Learning"],"prefix":"10.1587","volume":"E108.D","author":[{"given":"Qinghua","family":"SUN","sequence":"first","affiliation":[{"name":"Shanghai Jiao Tong University"}]},{"given":"Jia","family":"CUI","sequence":"additional","affiliation":[{"name":"South China University of Technology"}]},{"given":"Zhenyu","family":"GU","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University"}]}],"member":"532","reference":[{"key":"1","doi-asserted-by":"publisher","unstructured":"[1] P. O\u2019Donovan, J. L\u0131\u0304beks, A. Agarwala, and A. Hertzmann, \u201cExploratory font selection using crowdsourced attributes,\u201d ACM Trans. Graph., vol.33, no.4, pp.92:1-92:9, July 2014. 10.1145\/2601097.2601110","DOI":"10.1145\/2601097.2601110"},{"key":"2","doi-asserted-by":"publisher","unstructured":"[2] S. Choi and K. Aizawa, \u201cEmotype: Expressing emotions by changing typeface in mobile messenger texting,\u201d Multimedia Tools and Applications, vol.78, no.11, pp.14155-14172, June 2019. 10.1007\/s11042-018-6753-3","DOI":"10.1007\/s11042-018-6753-3"},{"key":"3","unstructured":"[3] T. Kulahcioglu and G. de Melo, \u201cFontLex: A typographical lexicon based on affective associations,\u201d LREC 2018 - 11th International Conference on Language Resources and Evaluation, pp.62-69, European Language Resources Association (ELRA), 2019."},{"key":"4","doi-asserted-by":"crossref","unstructured":"[4] S. Choi, S. Matsumura, and K. Aizawa, \u201cAssist users\u2019 interactions in font search with unexpected but useful concepts generated by multimodal learning,\u201d Proc. 2019 on International Conference on Multimedia Retrieval, ICMR\u2006\u201919, New York, NY, USA, pp.235-243, Association for Computing Machinery, June 2019. 10.1145\/3323873.3325037","DOI":"10.1145\/3323873.3325037"},{"key":"5","doi-asserted-by":"crossref","unstructured":"[5] A. Shirani, F. Dernoncourt, J. Echevarria, P. Asente, N. Lipka, and T. Solorio, \u201cLet me choose: From verbal context to font selection,\u201d Proc. 58th Annual Meeting of the Association for Computational Linguistics, Online, pp.8607-8613, Association for Computational Linguistics, 2020. 10.18653\/v1\/2020.acl-main.762","DOI":"10.18653\/v1\/2020.acl-main.762"},{"key":"6","unstructured":"[6] Y. Zhang, H. Jiang, Y. Miura, C.D. Manning, and C.P. Langlotz, \u201cContrastive learning of medical visual representations from paired images and text,\u201d Proc. 7th Machine Learning for Healthcare Conference, pp.2-25, PMLR, Dec. 2022."},{"key":"7","unstructured":"[7] A. Radford, J.W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark, G. Krueger, and I. Sutskever, \u201cLearning transferable visual models from natural language supervision,\u201d Proc. 38th International Conference on Machine Learning, pp.8748-8763, PMLR, July 2021."},{"key":"8","unstructured":"[8] S. Shen, L.H. Li, H. Tan, M. Bansal, A. Rohrbach, K.-W. Chang, Z. Yao, and K. Keutzer, \u201cHow much can CLIP benefit vision-and-language tasks?,\u201d arXiv:2107.06383, July 2021. 10.48550\/arXiv.2107.06383"},{"key":"9","doi-asserted-by":"publisher","unstructured":"[9] K. Wang and J.V. Nickerson, \u201cA literature review on individual creativity support systems,\u201d Computers in Human Behavior, vol.74, pp.139-151, Sept. 2017. 10.1016\/j.chb.2017.04.035","DOI":"10.1016\/j.chb.2017.04.035"},{"key":"10","doi-asserted-by":"crossref","unstructured":"[10] Y. Shinahara, T. Karamatsu, D. Harada, K. Yamaguchi, and S. Uchida, \u201cSerif or sans: Visual font analytics on book covers and online advertisements,\u201d 2019 International Conference on Document Analysis and Recognition (ICDAR), pp.1041-1046, Sept. 2019. 10.1109\/icdar.2019.00170","DOI":"10.1109\/ICDAR.2019.00170"},{"key":"11","doi-asserted-by":"crossref","unstructured":"[11] N. Yasukochi, H. Hayashi, D. Haraguchi, and S. Uchida, \u201cAnalyzing font style usage and contextual factors in real images,\u201d arXiv:2306.12050, June 2023. 10.48550\/arXiv.2306.12050","DOI":"10.1007\/978-3-031-41682-8_21"},{"key":"12","doi-asserted-by":"crossref","unstructured":"[12] T.-R. Chou and J. Lin, \u201cChinese font recommendation based on emotional adjectives,\u201d Proc. 2020 the 4th International Conference on Innovation in Artificial Intelligence, ICIAI 2020, New York, NY, USA, pp.155-158, Association for Computing Machinery, May 2020. 10.1145\/3390557.3394131","DOI":"10.1145\/3390557.3394131"},{"key":"13","doi-asserted-by":"crossref","unstructured":"[13] T. Kulahcioglu and G. de Melo, \u201cFonts like this but happier: A new way to discover fonts,\u201d Proc. 28th ACM International Conference on Multimedia, New York, NY, USA, pp.2973-2981, Association for Computing Machinery, Oct. 2020. 10.1145\/3394171.3413534","DOI":"10.1145\/3394171.3413534"},{"key":"14","doi-asserted-by":"crossref","unstructured":"[14] J. Kang, D. Haraguchi, S. Matsuda, A. Kimura, and S. Uchida, \u201cShared latent space of font shapes and their noisy impressions,\u201d MultiMedia Modeling, ed. B. P\u00f3r J\u00f3nsson, C. Gurrin, M.-T. Tran, D.-T. Dang-Nguyen, A.M.-C. Hu, B. Huynh Thi Thanh, and B. Huet, Lecture Notes in Computer Science, vol.13142, pp.146-157, Springer International Publishing, Cham, 2022. 10.1007\/978-3-030-98355-0_13","DOI":"10.1007\/978-3-030-98355-0_13"},{"key":"15","doi-asserted-by":"publisher","unstructured":"[15] J. Wang, K.C.K. Chan, and C.C. Loy, \u201cExploring CLIP for assessing the look and feel of images,\u201d Proc. AAAI Conference on Artificial Intelligence, vol.37, no.2, pp.2555-2563, June 2023. 10.1609\/aaai.v37i2.25353","DOI":"10.1609\/aaai.v37i2.25353"},{"key":"16","doi-asserted-by":"crossref","unstructured":"[16] C. Grover, I.D. Mastan, and D. Gupta, \u201cContextCLIP: Contextual alignment of image-text pairs on CLIP visual representations,\u201d Proc. Thirteenth Indian Conference on Computer Vision, Graphics and Image Processing, ICVGIP\u2006\u201922, New York, NY, USA, Article No. 51, pp.1-10, Association for Computing Machinery, May 2023. 10.1145\/3571600.3571653","DOI":"10.1145\/3571600.3571653"},{"key":"17","doi-asserted-by":"crossref","unstructured":"[17] M. Hendriksen, M. Bleeker, S. Vakulenko, N. van Noord, E. Kuiper, and M. de Rijke, \u201cExtending CLIP for category-to-image retrieval in e-commerce,\u201d Advances in Information Retrieval, ed. M. Hagen, S. Verberne, C. Macdonald, C. Seifert, K. Balog, K. N\u00f8rv\u00e5g, and V. Setty, Lecture Notes in Computer Science, vol.13185, pp.289-303, Springer International Publishing, Cham, 2022. 10.1007\/978-3-030-99736-6_20","DOI":"10.1007\/978-3-030-99736-6_20"},{"key":"18","doi-asserted-by":"publisher","unstructured":"[18] I.-C. Shen, F.-Y. Cherng, T. Igarashi, W.-C. Lin, and B.-Y. Chen, \u201cEvIcon: Designing high-usability icon with human-in-the-loop exploration and IconCLIP,\u201d Computer Graphics Forum, vol.42, no.6, e14924, Sept. 2023. 10.1111\/cgf.14924","DOI":"10.1111\/cgf.14924"},{"key":"19","doi-asserted-by":"publisher","unstructured":"[19] N. Zhao, Y. Cao, and R.W.H. Lau, \u201cModeling fonts in context: Font prediction on web designs,\u201d Computer Graphics Forum, vol.37, no.7, pp.385-395, 2018. 10.1111\/cgf.13576","DOI":"10.1111\/cgf.13576"},{"key":"20","doi-asserted-by":"crossref","unstructured":"[20] T. Kulahcioglu and G. de Melo, \u201cPredicting semantic signatures of fonts,\u201d 2018 IEEE 12th International Conference on Semantic Computing (ICSC), pp.115-122, Jan. 2018. 10.1109\/icsc.2018.00025","DOI":"10.1109\/ICSC.2018.00025"},{"key":"21","doi-asserted-by":"crossref","unstructured":"[21] J. Camacho-collados, K. Rezaee, T. Riahi, A. Ushio, D. Loureiro, D. Antypas, J. Boisson, L. Espinosa Anke, F. Liu, and E. Mart\u00ednez C\u00e1mara, \u201cTweetNLP: Cutting-edge natural language processing for social media,\u201d Proc. 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Abu Dhabi, UAE, pp.38-49, Association for Computational Linguistics, Dec. 2022. 10.18653\/v1\/2022.emnlp-demos.5","DOI":"10.18653\/v1\/2022.emnlp-demos.5"},{"key":"22","unstructured":"[22] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby, \u201cAn image is worth 16x16 words: Transformers for image recognition at scale,\u201d arXiv:2010.11929, Oct. 2020. 10.48550\/arXiv.2010.11929"},{"key":"23","unstructured":"[23] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, \u201cBERT: Pre-training of deep bidirectional transformers for language understanding,\u201d arXiv:1810.04805, May 2019. 10.48550\/arXiv.1810.04805"},{"key":"24","doi-asserted-by":"crossref","unstructured":"[24] L. McInnes, J. Healy, and J. Melville, \u201cUMAP: Uniform manifold approximation and projection for dimension reduction,\u201d arXiv:1802.03426, Feb. 2018. 10.48550\/arXiv.1802.03426","DOI":"10.21105\/joss.00861"},{"key":"25","doi-asserted-by":"crossref","unstructured":"[25] S. Choi, K. Aizawa, and N. Sebe, \u201cFontMatcher: Font image paring for harmonious digital graphic design,\u201d 23rd International Conference on Intelligent User Interfaces, IUI\u2006\u201918, New York, NY, USA, pp.37-41, Association for Computing Machinery, March 2018. 10.1145\/3172944.3173001","DOI":"10.1145\/3172944.3173001"},{"key":"26","unstructured":"[26] M. D\u017ebor, Design Problems, Frames and Innovative Solutions, IOS Press, 2009. 10.3233\/978-1-60750-067-4-i"}],"container-title":["IEICE Transactions on Information and Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transinf\/E108.D\/8\/E108.D_2024EDP7262\/_pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,2]],"date-time":"2025-08-02T03:29:07Z","timestamp":1754105347000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transinf\/E108.D\/8\/E108.D_2024EDP7262\/_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,1]]},"references-count":26,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2025]]}},"URL":"https:\/\/doi.org\/10.1587\/transinf.2024edp7262","relation":{},"ISSN":["0916-8532","1745-1361"],"issn-type":[{"type":"print","value":"0916-8532"},{"type":"electronic","value":"1745-1361"}],"subject":[],"published":{"date-parts":[[2025,8,1]]},"article-number":"2024EDP7262"}}