{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T05:05:51Z","timestamp":1750309551378,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":40,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,4,25]],"date-time":"2025-04-25T00:00:00Z","timestamp":1745539200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,4,26]]},"DOI":"10.1145\/3706599.3720044","type":"proceedings-article","created":{"date-parts":[[2025,4,23]],"date-time":"2025-04-23T20:15:12Z","timestamp":1745439312000},"page":"1-8","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Foundation Model ColorMap: A Framework for Extracting and Visualizing the Foundation Models\u2019 Color Knowledge"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2327-5464","authenticated-orcid":false,"given":"Eric","family":"Languenou","sequence":"first","affiliation":[{"name":"Computer Dpt, Nantes Universit\u00e9, Nantes, France and LS2N, Laboratoire des Sciences du Num\u00e9rique de Nantes, Nantes, France"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4318-6717","authenticated-orcid":false,"given":"Christophe","family":"Hurter","sequence":"additional","affiliation":[{"name":"Singapore Lab, Ecole Nationale de l'Aviation Civile, Toulouse, France"}]}],"member":"320","published-online":{"date-parts":[[2025,4,25]]},"reference":[{"key":"e_1_3_3_3_2_2","doi-asserted-by":"crossref","unstructured":"Mostafa Abdou Artur Kulmizev Daniel Hershcovich Stella Frank Ellie Pavlick and Anders S\u00f8gaard. 2021. Can Language Models Encode Perceptual Structure Without Grounding? A Case Study in Color. arxiv:https:\/\/arXiv.org\/abs\/2109.06129\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2109.06129","DOI":"10.18653\/v1\/2021.conll-1.9"},{"key":"e_1_3_3_3_3_2","first-page":"1858","volume-title":"Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"Bexte Marie","year":"2024","unstructured":"Marie Bexte, Andrea Horbach, and Torsten Zesch. 2024. Rainbow - A Benchmark for Systematic Testing of How Sensitive Visio-Linguistic Models are to Color Naming. In Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), Yvette Graham and Matthew Purver (Eds.). Association for Computational Linguistics, St. Julian\u2019s, Malta, 1858\u20131875. https:\/\/aclanthology.org\/2024.eacl-long.112\/"},{"key":"e_1_3_3_3_4_2","doi-asserted-by":"crossref","unstructured":"Zeyd Boukhers Timo Hartmann and Jan J\u00fcrjens. 2022. COIN: Counterfactual Image Generation for Visual Question Answering Interpretation. Sensors 22 6 (2022) 2245. http:\/\/dblp.uni-trier.de\/db\/journals\/sensors\/sensors22.html#BoukhersHJ22","DOI":"10.3390\/s22062245"},{"key":"e_1_3_3_3_5_2","unstructured":"Han-Cheol Cho Won\u00a0Young Jhoo Wooyoung Kang and Byungseok Roh. 2023. Open-Vocabulary Object Detection using Pseudo Caption Labels. arxiv:https:\/\/arXiv.org\/abs\/2303.13040\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2303.13040"},{"key":"e_1_3_3_3_6_2","doi-asserted-by":"publisher","unstructured":"Jason Chuang Maureen Stone and Pat Hanrahan. 2008. A Probabilistic Model of the Categorical Association Between Colors. Final Program and Proceedings - IS and T\/SID Color Imaging Conference 16 (01 2008) 6\u201311. 10.2352\/CIC.2008.16.1.art00002","DOI":"10.2352\/CIC.2008.16.1.art00002"},{"key":"e_1_3_3_3_7_2","doi-asserted-by":"publisher","DOI":"10.1109\/VisGuides57787.2022.00008"},{"key":"e_1_3_3_3_8_2","unstructured":"Bommasani et al. 2021. On the Opportunities and Risks of Foundation Models. arxiv:https:\/\/arXiv.org\/abs\/2108.07258\u00a0[cs.LG] https:\/\/arxiv.org\/abs\/2108.07258"},{"key":"e_1_3_3_3_9_2","unstructured":"Percy et al. 2023. Holistic Evaluation of Language Models. arxiv:https:\/\/arXiv.org\/abs\/2211.09110\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2211.09110"},{"key":"e_1_3_3_3_10_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.393"},{"key":"e_1_3_3_3_11_2","unstructured":"Chaoyou Fu Peixian Chen Yunhang Shen Yulei Qin Mengdan Zhang Xu Lin Jinrui Yang Xiawu Zheng Ke Li Xing Sun Yunsheng Wu and Rongrong Ji. 2024. MME: A Comprehensive Evaluation Benchmark for Multimodal Large Language Models. arxiv:https:\/\/arXiv.org\/abs\/2306.13394\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2306.13394"},{"key":"e_1_3_3_3_12_2","unstructured":"Gemini Team. 2024. Gemini: A Family of Highly Capable Multimodal Models. arxiv:https:\/\/arXiv.org\/abs\/2312.11805\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2312.11805"},{"key":"e_1_3_3_3_13_2","doi-asserted-by":"crossref","unstructured":"Demis Hassabis Dharshan Kumaran Christopher Summerfield and Matthew Botvinick. 2017. Neuroscience-inspired artificial intelligence. Neuron 95 2 (2017) 245\u2013258.","DOI":"10.1016\/j.neuron.2017.06.011"},{"key":"e_1_3_3_3_14_2","doi-asserted-by":"publisher","DOI":"10.1145\/2207676.2208547"},{"key":"e_1_3_3_3_15_2","unstructured":"Danqing Huang Jiaqi Guo Shizhao Sun Hanling Tian Jieru Lin Zheng Hu Chin-Yew Lin Jian-Guang Lou and Dongmei Zhang. 2023. A Survey for Graphic Design Intelligence. arxiv:https:\/\/arXiv.org\/abs\/2309.01371\u00a0[cs.HC] https:\/\/arxiv.org\/abs\/2309.01371"},{"key":"e_1_3_3_3_16_2","unstructured":"Jieru Lin Danqing Huang Tiejun Zhao Dechen Zhan and Chin-Yew Lin. 2024. DesignProbe: A Graphic Design Benchmark for Multimodal Large Language Models. arxiv:https:\/\/arXiv.org\/abs\/2404.14801\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2404.14801"},{"key":"e_1_3_3_3_17_2","doi-asserted-by":"crossref","unstructured":"Delwin\u00a0T Lindsey and Angela\u00a0M Brown. 2006. Universality of color names. Proceedings of the National Academy of Sciences 103 44 (2006) 16608\u201316613.","DOI":"10.1073\/pnas.0607708103"},{"key":"e_1_3_3_3_18_2","doi-asserted-by":"crossref","unstructured":"Zachary\u00a0C Lipton. 2018. The mythos of model interpretability: In machine learning the concept of interpretability is both important and slippery. Queue 16 3 (2018) 31\u201357.","DOI":"10.1145\/3236386.3241340"},{"key":"e_1_3_3_3_19_2","unstructured":"Llama team. 2024. The Llama 3 Herd of Models. arxiv:https:\/\/arXiv.org\/abs\/2407.21783\u00a0[cs.AI] https:\/\/arxiv.org\/abs\/2407.21783"},{"key":"e_1_3_3_3_20_2","first-page":"13","volume-title":"ViLBERT: pretraining task-agnostic visiolinguistic representations for vision-and-language tasks","author":"Lu Jiasen","year":"2019","unstructured":"Jiasen Lu, Dhruv Batra, Devi Parikh, and Stefan Lee. 2019. ViLBERT: pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. Curran Associates Inc., 57 Morehouse Lane, Red Hook, NY, United States, 13\u201323."},{"key":"e_1_3_3_3_21_2","doi-asserted-by":"crossref","unstructured":"Adrian Mayorga and Michael Gleicher. 2013. Splatterplots: Overcoming Overdraw in Scatter Plots. IEEE transactions on visualization and computer graphics 19 9 (2013) 1526\u201338. https:\/\/api.semanticscholar.org\/CorpusID:268110470","DOI":"10.1109\/TVCG.2013.65"},{"key":"e_1_3_3_3_22_2","doi-asserted-by":"publisher","unstructured":"W.\u00a0R. Merrifield. 1971. Brent Berlin and Paul Kay Basic color terms: their universality and evolution. Journal of Linguistics 7 2 (1971) 259\u2013268. 10.1017\/S0022226700002966","DOI":"10.1017\/S0022226700002966"},{"key":"e_1_3_3_3_23_2","unstructured":"Randall Munroe. 2010. Color Survey Results. https:\/\/blog.xkcd.com\/2010\/05\/03\/color-survey-results\/"},{"key":"e_1_3_3_3_24_2","unstructured":"Humza Naveed Asad\u00a0Ullah Khan Shi Qiu Muhammad Saqib Saeed Anwar Muhammad Usman Naveed Akhtar Nick Barnes and Ajmal Mian. 2023. A comprehensive overview of large language models."},{"key":"e_1_3_3_3_25_2","doi-asserted-by":"crossref","unstructured":"Chris Olah Alexander Mordvintsev and Ludwig Schubert. 2017. Feature visualization. Distill 2 11 (2017) e7.","DOI":"10.23915\/distill.00007"},{"key":"e_1_3_3_3_26_2","unstructured":"OpenAI. 2024. GPT-4 Technical Report. arxiv:https:\/\/arXiv.org\/abs\/2303.08774\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2303.08774"},{"key":"e_1_3_3_3_27_2","doi-asserted-by":"publisher","DOI":"10.3115\/1073083.1073135"},{"key":"e_1_3_3_3_28_2","series-title":"Proceedings of Machine Learning Research","first-page":"8748","volume-title":"Proceedings of the 38th International Conference on Machine Learning","volume":"139","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong\u00a0Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever. 2021. Learning Transferable Visual Models From Natural Language Supervision. In Proceedings of the 38th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol.\u00a0139), Marina Meila and Tong Zhang (Eds.). PMLR, Cambridge, Massachusetts, USA, 8748\u20138763. https:\/\/proceedings.mlr.press\/v139\/radford21a.html"},{"key":"e_1_3_3_3_29_2","unstructured":"Colin Raffel Noam Shazeer Adam Roberts Katherine Lee Sharan Narang Michael Matena Yanqi Zhou Wei Li and Peter\u00a0J Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of machine learning research 21 140 (2020) 1\u201367."},{"key":"e_1_3_3_3_30_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D16-1264"},{"key":"e_1_3_3_3_31_2","doi-asserted-by":"publisher","unstructured":"Terry Regier Paul Kay and Richard\u00a0S. Cook. 2005. Focal colors are universal after all. Proceedings of the National Academy of Sciences 102 23 (2005) 8386\u20138391. 10.1073\/pnas.0503281102","DOI":"10.1073\/pnas.0503281102"},{"key":"e_1_3_3_3_32_2","unstructured":"Joshua Robinson Christopher\u00a0Michael Rytting and David Wingate. 2023. Leveraging Large Language Models for Multiple Choice Question Answering. arxiv:https:\/\/arXiv.org\/abs\/2210.12353\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2210.12353"},{"key":"e_1_3_3_3_33_2","doi-asserted-by":"publisher","unstructured":"Emmanuelle Salin Badreddine Farah St\u00e9phane Ayache and Benoit Favre. 2022. Are Vision-Language Transformers Learning Multimodal Representations? A Probing Perspective. Proceedings of the AAAI Conference on Artificial Intelligence 36 10 (Jun. 2022) 11248\u201311257. 10.1609\/aaai.v36i10.21375","DOI":"10.1609\/aaai.v36i10.21375"},{"key":"e_1_3_3_3_34_2","doi-asserted-by":"crossref","unstructured":"Sakib Shahriar Brady Lund Nishith\u00a0Reddy Mannuru Muhammad\u00a0Arbab Arshad Kadhim Hayawi Ravi Varma\u00a0Kumar Bevara Aashrith Mannuru and Laiba Batool. 2024. Putting GPT-4o to the Sword: A Comprehensive Evaluation of Language Vision Speech and Multimodal Proficiency. arxiv:https:\/\/arXiv.org\/abs\/2407.09519\u00a0[cs.AI] https:\/\/arxiv.org\/abs\/2407.09519","DOI":"10.20944\/preprints202406.1635.v1"},{"key":"e_1_3_3_3_35_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4899-3324-9"},{"key":"e_1_3_3_3_36_2","unstructured":"Hugo Touvron. 2023. Llama 2: Open Foundation and Fine-Tuned Chat Models. arxiv:https:\/\/arXiv.org\/abs\/2307.09288\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2307.09288"},{"key":"e_1_3_3_3_37_2","unstructured":"Hugo Touvron Thibaut Lavril Gautier Izacard Xavier Martinet Marie-Anne Lachaux Timoth\u00e9e Lacroix Baptiste Rozi\u00e8re Naman Goyal Eric Hambro Faisal Azhar Aurelien Rodriguez Armand Joulin Edouard Grave and Guillaume Lample. 2023. LLaMA: Open and Efficient Foundation Language Models. ArXiv abs\/2302.13971 (Feb. 2023). https:\/\/api.semanticscholar.org\/CorpusID:257219404"},{"key":"e_1_3_3_3_38_2","unstructured":"A Vaswani. 2017. Attention is all you need."},{"key":"e_1_3_3_3_39_2","volume-title":"Information Visualization: Perception for Design","author":"Ware C.","year":"2013","unstructured":"C. Ware. 2013. Information Visualization: Perception for Design. Elsevier Science, Customer Service Department, P.O. Box 211, Amsterdam, Netherlands. https:\/\/books.google.fr\/books?id=qFmS95vf6H8C"},{"key":"e_1_3_3_3_40_2","unstructured":"Zicheng Zhang Haoning Wu Erli Zhang Guangtao Zhai and Weisi Lin. 2024. Q-Bench+: A Benchmark for Multi-modal Foundation Models on Low-level Vision from Single Images to Pairs. arxiv:https:\/\/arXiv.org\/abs\/2402.07116\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2402.07116"},{"key":"e_1_3_3_3_41_2","doi-asserted-by":"publisher","DOI":"10.1109\/ITCA52113.2020.00069"}],"event":{"name":"CHI EA '25: Extended Abstracts of the CHI Conference on Human Factors in Computing Systems","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"],"location":"Yokohama Japan","acronym":"CHI EA '25"},"container-title":["Proceedings of the Extended Abstracts of the CHI Conference on Human Factors in Computing Systems"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3706599.3720044","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3706599.3720044","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:47Z","timestamp":1750295927000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3706599.3720044"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,25]]},"references-count":40,"alternative-id":["10.1145\/3706599.3720044","10.1145\/3706599"],"URL":"https:\/\/doi.org\/10.1145\/3706599.3720044","relation":{},"subject":[],"published":{"date-parts":[[2025,4,25]]},"assertion":[{"value":"2025-04-25","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}