{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,7]],"date-time":"2025-10-07T00:30:22Z","timestamp":1759797022885,"version":"build-2065373602"},"publisher-location":"New York, NY, USA","reference-count":15,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,9,28]]},"DOI":"10.1145\/3746058.3758449","type":"proceedings-article","created":{"date-parts":[[2025,9,27]],"date-time":"2025-09-27T14:33:09Z","timestamp":1758983589000},"page":"1-3","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["MermaidLLM: Dataflow Diagrams for Explainable Skill Formalization and Real-time Support with Multimodal LLMs"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-4371-4584","authenticated-orcid":false,"given":"Sotaro","family":"Yokoi","sequence":"first","affiliation":[{"name":"Sony CSL Kyoto, Kyoto, Japan and The University of Tokyo, Tokyo, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3629-2514","authenticated-orcid":false,"given":"Jun","family":"Rekimoto","sequence":"additional","affiliation":[{"name":"Sony CSL Kyoto, Kyoto, Japan and The University of Tokyo, Tokyo, Japan"}]}],"member":"320","published-online":{"date-parts":[[2025,9,27]]},"reference":[{"key":"e_1_3_3_1_2_2","unstructured":"[n. d.]. Mermaid: Diagramming and charting tool. https:\/\/mermaid.js.org\/"},{"key":"e_1_3_3_1_3_2","doi-asserted-by":"publisher","DOI":"10.1145\/3654777.3676350"},{"key":"e_1_3_3_1_4_2","unstructured":"Aaron Bangor Philip Kortum and James Miller. 2009. Determining what individual SUS scores mean: adding an adjective rating scale. J. Usability Studies 4 3 (May 2009) 114\u2013123."},{"key":"e_1_3_3_1_5_2","unstructured":"John Brooke et\u00a0al. 1996. SUS-A quick and dirty usability scale. Usability evaluation in industry 189 194 (1996) 4\u20137."},{"key":"e_1_3_3_1_6_2","doi-asserted-by":"publisher","unstructured":"Haodong Chen Niloofar Zendehdel Ming\u00a0C. Leu and Zhaozheng Yin. 2025. A gaze-driven manufacturing assembly assistant system with integrated step recognition repetition analysis and real-time feedback. Engineering Applications of Artificial Intelligence 144 (2025) 110076. 10.1016\/j.engappai.2025.110076","DOI":"10.1016\/j.engappai.2025.110076"},{"key":"e_1_3_3_1_7_2","unstructured":"Jakob Engel et\u00a0al. 2023. Project Aria: A New Tool for Egocentric Multi-Modal AI Research. arXiv:https:\/\/arXiv.org\/abs\/2308.13561\u00a0[cs.HC] https:\/\/arxiv.org\/abs\/2308.13561"},{"key":"e_1_3_3_1_8_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2011.5995444"},{"key":"e_1_3_3_1_9_2","unstructured":"Kristen Grauman et\u00a0al. 2024. Ego-Exo4D: Understanding Skilled Human Activity from First- and Third-Person Perspectives. arXiv:https:\/\/arXiv.org\/abs\/2311.18259\u00a0[cs.CV] https:\/\/arxiv.org\/abs\/2311.18259"},{"key":"e_1_3_3_1_10_2","volume-title":"CEUR","author":"Herwanto Guntur\u00a0Budi","year":"2024","unstructured":"Guntur\u00a0Budi Herwanto. 2024. Automating Data Flow Diagram Generation from User Stories Using Large Language Models. In CEUR. Winterthur, Switzerland. https:\/\/hal.science\/hal-04525925"},{"key":"e_1_3_3_1_11_2","unstructured":"Binbin Li Tianxin Meng Xiaoming Shi Jie Zhai and Tong Ruan. 2023. MedDM: LLM-executable clinical guidance tree for clinical decision-making. ArXiv abs\/2312.02441 (2023). https:\/\/api.semanticscholar.org\/CorpusID:265658947"},{"key":"e_1_3_3_1_12_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA55743.2025.11127365"},{"key":"e_1_3_3_1_13_2","unstructured":"pupil labs. [n. d.]. NEON: Eye tracking for research and beyond. https:\/\/pupil-labs.com\/products\/neon"},{"key":"e_1_3_3_1_14_2","series-title":"(AHs \u201925)","volume-title":"Proceedings of Augmented Humans 2025","author":"Rekimoto Jun","year":"2025","unstructured":"Jun Rekimoto. 2025. GazeLLM: Multimodal LLMs Incorporating Human Visual Attention. In Proceedings of Augmented Humans 2025 (Abu Dhavi, UAE) (AHs \u201925)."},{"key":"e_1_3_3_1_15_2","unstructured":"Ryunosuke Takebayashi Vitor\u00a0Hideyo Isume Takuya Kiyokawa Weiwei Wan and Kensuke Harada. 2025. Cooking Task Planning using LLM and Verified by Graph Network. arXiv:https:\/\/arXiv.org\/abs\/2503.21564\u00a0[cs.RO] https:\/\/arxiv.org\/abs\/2503.21564"},{"key":"e_1_3_3_1_16_2","doi-asserted-by":"publisher","unstructured":"Yuheng Zhao Junjie Wang Linbin Xiang Xiaowen Zhang Zifei Guo Cagatay Turkay Yu Zhang and Siming Chen. 2024. LightVA: Lightweight Visual Analytics with LLM Agent-Based Task Planning and Execution. IEEE Transactions on Visualization and Computer Graphics (2024) 1\u201313. 10.1109\/TVCG.2024.3496112","DOI":"10.1109\/TVCG.2024.3496112"}],"event":{"name":"UIST '25: The 38th Annual ACM Symposium on User Interface Software and Technology","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction","SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"],"location":"Busan Republic of Korea","acronym":"UIST Adjunct '25"},"container-title":["Adjunct Proceedings of the 38th Annual ACM Symposium on User Interface Software and Technology"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3746058.3758449","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,6]],"date-time":"2025-10-06T10:07:09Z","timestamp":1759745229000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3746058.3758449"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,27]]},"references-count":15,"alternative-id":["10.1145\/3746058.3758449","10.1145\/3746058"],"URL":"https:\/\/doi.org\/10.1145\/3746058.3758449","relation":{},"subject":[],"published":{"date-parts":[[2025,9,27]]},"assertion":[{"value":"2025-09-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}