{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,23]],"date-time":"2026-02-23T00:42:24Z","timestamp":1771807344535,"version":"3.50.1"},"reference-count":18,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"5","license":[{"start":{"date-parts":[[2024,5,1]],"date-time":"2024-05-01T00:00:00Z","timestamp":1714521600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,5,1]],"date-time":"2024-05-01T00:00:00Z","timestamp":1714521600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,5,1]],"date-time":"2024-05-01T00:00:00Z","timestamp":1714521600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"JST ACT-X","award":["JPMJAX20A5"],"award-info":[{"award-number":["JPMJAX20A5"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Robot. Autom. Lett."],"published-print":{"date-parts":[[2024,5]]},"DOI":"10.1109\/lra.2024.3375257","type":"journal-article","created":{"date-parts":[[2024,3,11]],"date-time":"2024-03-11T18:21:57Z","timestamp":1710181317000},"page":"4059-4066","source":"Crossref","is-referenced-by-count":13,"title":["Continuous Object State Recognition for Cooking Robots Using Pre-Trained Vision-Language Models and Black-Box Optimization"],"prefix":"10.1109","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7464-7187","authenticated-orcid":false,"given":"Kento","family":"Kawaharazuka","sequence":"first","affiliation":[{"name":"Department of Mechano-Informatics, Graduate School of Information Science and Technology, The University of Tokyo, Bunkyo-ku, Tokyo, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-3527-3087","authenticated-orcid":false,"given":"Naoaki","family":"Kanazawa","sequence":"additional","affiliation":[{"name":"Department of Mechano-Informatics, Graduate School of Information Science and Technology, The University of Tokyo, Bunkyo-ku, Tokyo, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1429-4401","authenticated-orcid":false,"given":"Yoshiki","family":"Obinata","sequence":"additional","affiliation":[{"name":"Department of Mechano-Informatics, Graduate School of Information Science and Technology, The University of Tokyo, Bunkyo-ku, Tokyo, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6606-6692","authenticated-orcid":false,"given":"Kei","family":"Okada","sequence":"additional","affiliation":[{"name":"Department of Mechano-Informatics, Graduate School of Information Science and Technology, The University of Tokyo, Bunkyo-ku, Tokyo, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1273-1567","authenticated-orcid":false,"given":"Masayuki","family":"Inaba","sequence":"additional","affiliation":[{"name":"Department of Mechano-Informatics, Graduate School of Information Science and Technology, The University of Tokyo, Bunkyo-ku, Tokyo, Japan"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1145\/6462.6464"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1016\/j.autcon.2017.10.016"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48891.2023.10160390"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/Humanoids.2011.6100855"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2020.2965418"},{"key":"ref6","article-title":"Classifying cooking objects state using a tuned VGG convolutional neural network","author":"Paul","year":"2018"},{"key":"ref7","article-title":"Identifying object states in cooking-related images","author":"Jelodar","year":"2018"},{"key":"ref8","article-title":"Cooking objects state identification without using pretrained model","author":"Sakib","year":"2021"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/IROS47612.2022.9981280"},{"key":"ref10","article-title":"Vision-language intelligence: Tasks, representation learning, and large models","author":"Li","year":"2022"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/humanoids57100.2023.10375211"},{"key":"ref12","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. 38th Int. Conf. Mach. Learn.","author":"Radford","year":"2021"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01457"},{"key":"ref14","first-page":"23318","article-title":"OFA: Unifying architectures, tasks, and modalities through a simple sequence-to-sequence learning framework","volume-title":"Proc. 39th Int. Conf. Mach. Learn.","author":"Wang","year":"2022"},{"key":"ref15","first-page":"1","article-title":"Recognition of heat-induced food state changes by time-series use of vision-language model for cooking robot","volume-title":"Proc. 18th Int. Conf. Intell. Auton. Syst.","author":"Kanazawa","year":"2023"},{"key":"ref16","first-page":"2171","article-title":"DEAP: Evolutionary algorithms made easy","volume":"13","author":"Fortin","year":"2012","journal-title":"J. Mach. Learn. Res."},{"key":"ref17","first-page":"1877","article-title":"Language models are few-shot learners","volume-title":"Adv. Neural Inf. Process. Syst.","author":"Brown","year":"2020"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TSMC.2023.3282950"}],"container-title":["IEEE Robotics and Automation Letters"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7083369\/10474437\/10465607.pdf?arnumber=10465607","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,23]],"date-time":"2024-12-23T19:28:09Z","timestamp":1734982089000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10465607\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5]]},"references-count":18,"journal-issue":{"issue":"5"},"URL":"https:\/\/doi.org\/10.1109\/lra.2024.3375257","relation":{},"ISSN":["2377-3766","2377-3774"],"issn-type":[{"value":"2377-3766","type":"electronic"},{"value":"2377-3774","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,5]]}}}