{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,23]],"date-time":"2025-08-23T20:40:10Z","timestamp":1755981610241,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":26,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,5,6]],"date-time":"2025-05-06T00:00:00Z","timestamp":1746489600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-sa\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,5,6]]},"DOI":"10.1145\/3722573.3727826","type":"proceedings-article","created":{"date-parts":[[2025,5,3]],"date-time":"2025-05-03T01:04:09Z","timestamp":1746234249000},"page":"1-9","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Simulation vs. Hallucination: Assessing Vision-Language Model Question Answering Capabilities in Engineering Simulations"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-4003-1522","authenticated-orcid":false,"given":"Jessica","family":"Ezemba","sequence":"first","affiliation":[{"name":"Carnegie Mellon University, Pittsburgh, Pennsylvania, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5024-7701","authenticated-orcid":false,"given":"Christopher","family":"McComb","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, Pittsburgh, Pennsylvania, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5365-0240","authenticated-orcid":false,"given":"Conrad","family":"Tucker","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, Pittsburgh, Pennsylvania, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,5,6]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Nguyen Bach, Amit Bahree, Arash Bakhtiari, Jianmin Bao, Harkirat Behl, et al.","author":"Abdin Marah","year":"2024","unstructured":"Marah Abdin, Jyoti Aneja, Hany Awadalla, Ahmed Awadallah, Ammar Ahmad Awan, Nguyen Bach, Amit Bahree, Arash Bakhtiari, Jianmin Bao, Harkirat Behl, et al. 2024. Phi-3 technical report: A highly capable language model locally on your phone. arXiv preprint arXiv:2404.14219 (2024)."},{"key":"e_1_3_2_1_2_1","volume-title":"Datasets in Engineering Design. Journal of Mechanical Design","author":"Ahmed Faez","year":"2025","unstructured":"Faez Ahmed, Cyril Picard, Wei Chen, Christopher Mccomb, Pingfeng Wang, Ikjin Lee, Tino Stankovic, Douglas Allaire, and Stefan Menzel. 2025. Datasets in Engineering Design. Journal of Mechanical Design (2025), 1--8."},{"key":"e_1_3_2_1_3_1","volume-title":"Hisham Cholakkal, Mubarak Shah, Ming-Hsuan Yang, and Fahad Shahbaz Khan.","author":"Awais Muhammad","year":"2025","unstructured":"Muhammad Awais, Muzammal Naseer, Salman Khan, Rao Muhammad Anwer, Hisham Cholakkal, Mubarak Shah, Ming-Hsuan Yang, and Fahad Shahbaz Khan. 2025. Foundation Models Defining a New Era in Vision: a Survey and Outlook. IEEE Transactions on Pattern Analysis and Machine Intelligence (2025)."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11831-024-10117-3"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1115\/1.4063894"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1115\/1.4067333"},{"key":"e_1_3_2_1_7_1","first-page":"499","article-title":"Drivaernet+ +: A large-scale multimodal car dataset with computational fluid dynamics simulations and deep learning benchmarks","volume":"37","author":"Elrefaie Mohamed","year":"2025","unstructured":"Mohamed Elrefaie, Florin Morar, Angela Dai, and Faez Ahmed. 2025. Drivaernet+ +: A large-scale multimodal car dataset with computational fluid dynamics simulations and deep learning benchmarks. Advances in Neural Information Processing Systems 37 (2025), 499--536.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00370"},{"key":"e_1_3_2_1_9_1","volume-title":"Vision-Language Models for Design Concept Generation: An Actor-Critic Framework. Journal of Mechanical Design","author":"Ghasemi Parisa","year":"2025","unstructured":"Parisa Ghasemi and Mohsen Moghaddam. 2025. Vision-Language Models for Design Concept Generation: An Actor-Critic Framework. Journal of Mechanical Design (2025), 1--28."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1080\/00207543.2023.2276811"},{"key":"e_1_3_2_1_11_1","volume-title":"A review of digital twin applications in various sectors. Transforming industry using digital twin technology","author":"Kanaga Priya P","year":"2024","unstructured":"P Kanaga Priya and A Reethika. 2024. A review of digital twin applications in various sectors. Transforming industry using digital twin technology (2024), 239--258."},{"volume-title":"Structural analysis","author":"Kassimali Aslam","key":"e_1_3_2_1_12_1","unstructured":"Aslam Kassimali, GV Ramana, and Germ\u00e1n Rojas Orozco. 2015. Structural analysis. Cengage Learning Stamford, CT."},{"key":"e_1_3_2_1_13_1","article-title":"Deep learning in computational design synthesis: a comprehensive review","volume":"24","author":"Singh Shubhendu Kumar","year":"2024","unstructured":"Shubhendu Kumar Singh, Rahul Rai, Raj Pradip Khawale, Darshil Patel, Dustin Bielecki, Ryan Nguyen, Jun Wang, and Zhibo Zhang. 2024. Deep learning in computational design synthesis: a comprehensive review. Journal of Computing and Information Science in Engineering 24, 4 (2024).","journal-title":"Journal of Computing and Information Science in Engineering"},{"key":"e_1_3_2_1_14_1","volume-title":"VLMaterial: Procedural Material Generation with Large Vision-Language Models. arXiv preprint arXiv:2501.18623","author":"Li Beichen","year":"2025","unstructured":"Beichen Li, Rundi Wu, Armando Solar-Lezama, Changxi Zheng, Liang Shi, Bernd Bickel, and Wojciech Matusik. 2025. VLMaterial: Procedural Material Generation with Large Vision-Language Models. arXiv preprint arXiv:2501.18623 (2025)."},{"key":"e_1_3_2_1_15_1","unstructured":"Bo Li Yuanhan Zhang Dong Guo Renrui Zhang Feng Li Hao Zhang Kaichen Zhang Yanwei Li Ziwei Liu and Chunyuan Li. 2024. LLaVA-OneVision: Easy Visual Task Transfer. arXiv:2408.03326 [cs.CV] https:\/\/arxiv.org\/abs\/2408.03326"},{"key":"e_1_3_2_1_16_1","volume-title":"International conference on machine learning. PMLR","author":"Li Junnan","year":"2023","unstructured":"Junnan Li, Dongxu Li, Silvio Savarese, and Steven Hoi. 2023. Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. In International conference on machine learning. PMLR, 19730--19742."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3664647.3681529"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3672758.3672824"},{"key":"e_1_3_2_1_19_1","volume-title":"Revolutionizing edge AI and vision with open, customizable models. Meta AI Blog. Retrieved December 20","author":"Meta AI","year":"2024","unstructured":"AI Meta. 2024. Llama 3.2: Revolutionizing edge AI and vision with open, customizable models. Meta AI Blog. Retrieved December 20 (2024), 2024."},{"key":"e_1_3_2_1_20_1","unstructured":"OpenAI. 2024. Hello GPT-4o. https:\/\/openai.com\/index\/hello-gpt-4o\/ Accessed: 03\/06."},{"key":"e_1_3_2_1_21_1","volume-title":"Kosmos-2: Grounding Multimodal Large Language Models to the World. ArXiv abs\/2306","author":"Peng Zhiliang","year":"2023","unstructured":"Zhiliang Peng, Wenhui Wang, Li Dong, Yaru Hao, Shaohan Huang, Shuming Ma, and Furu Wei. 2023. Kosmos-2: Grounding Multimodal Large Language Models to the World. ArXiv abs\/2306 (2023)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00799-022-00329-y"},{"key":"e_1_3_2_1_23_1","unstructured":"Qwen Team. 2025. Qwen2.5-VL. https:\/\/qwenlm.github.io\/blog\/qwen2.5-vl\/"},{"key":"e_1_3_2_1_24_1","volume-title":"Chaoqun Liu, and Yao Tao.","author":"Tu Jiyuan","year":"2023","unstructured":"Jiyuan Tu, Guan Heng Yeoh, Chaoqun Liu, and Yao Tao. 2023. Computational fluid dynamics: a practical approach. Elsevier."},{"key":"e_1_3_2_1_25_1","volume-title":"Mm-vet v2: A challenging benchmark to evaluate large multimodal models for integrated capabilities. arXiv preprint arXiv:2408.00765","author":"Yu Weihao","year":"2024","unstructured":"Weihao Yu, Zhengyuan Yang, Lingfeng Ren, Linjie Li, Jianfeng Wang, Kevin Lin, Chung-Ching Lin, Zicheng Liu, Lijuan Wang, and Xinchao Wang. 2024. Mm-vet v2: A challenging benchmark to evaluate large multimodal models for integrated capabilities. arXiv preprint arXiv:2408.00765 (2024)."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00637"}],"event":{"name":"SenSys '25: The 23rd ACM Conference on Embedded Networked Sensor Systems","sponsor":["SIGMETRICS ACM Special Interest Group on Measurement and Evaluation","SIGMOBILE ACM Special Interest Group on Mobility of Systems, Users, Data and Computing","SIGOPS ACM Special Interest Group on Operating Systems","SIGBED ACM Special Interest Group on Embedded Systems","SIGARCH ACM Special Interest Group on Computer Architecture"],"location":"Irvine CA USA","acronym":"SenSys '25"},"container-title":["Proceedings of the 7th Workshop on Design Automation for CPS and IoT"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3722573.3727826","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3722573.3727826","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,23]],"date-time":"2025-08-23T20:17:16Z","timestamp":1755980236000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3722573.3727826"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,6]]},"references-count":26,"alternative-id":["10.1145\/3722573.3727826","10.1145\/3722573"],"URL":"https:\/\/doi.org\/10.1145\/3722573.3727826","relation":{},"subject":[],"published":{"date-parts":[[2025,5,6]]},"assertion":[{"value":"2025-05-06","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}