{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T18:09:50Z","timestamp":1776881390637,"version":"3.51.2"},"reference-count":27,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,11,26]],"date-time":"2024-11-26T00:00:00Z","timestamp":1732579200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,11,26]],"date-time":"2024-11-26T00:00:00Z","timestamp":1732579200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,11,26]]},"DOI":"10.1109\/fllm63129.2024.10852491","type":"proceedings-article","created":{"date-parts":[[2025,1,28]],"date-time":"2025-01-28T18:35:23Z","timestamp":1738089323000},"page":"392-397","source":"Crossref","is-referenced-by-count":17,"title":["LLM-BRAIn: AI-driven Fast Generation of Robot Behaviour Tree based on Large Language Model"],"prefix":"10.1109","author":[{"given":"Artem","family":"Lykov","sequence":"first","affiliation":[{"name":"Skolkovo Institute of Science and Technology,Intelligent Space Robotics Laboratory,Moscow,Russia"}]},{"given":"Dzmitry","family":"Tsetserukou","sequence":"additional","affiliation":[{"name":"Skolkovo Institute of Science and Technology,Intelligent Space Robotics Laboratory,Moscow,Russia"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Attention Is All You Need","author":"Vaswani","year":"2017"},{"key":"ref2","article-title":"Language Models are Unsupervised Multitask Learners","author":"Radford","year":"2018"},{"key":"ref3","article-title":"Language Models are Few-Shot Learners","author":"Brown","year":"2020"},{"key":"ref4","article-title":"Introducing ChatGPT","year":"2022"},{"key":"ref5","article-title":"Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer","author":"Raffel","year":"2020"},{"key":"ref6","article-title":"LLaMA: Open and Efficient Foundation Language Models","author":"Touvron","year":"2023"},{"key":"ref7","article-title":"The Falcon Series of Open Language Models","author":"Almazrouei","year":"2023"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-55560-2_5"},{"key":"ref9","article-title":"Alpaca: A Strong, Replicable Instruction-Following Model","author":"Taori","year":"2023"},{"key":"ref10","article-title":"Alpaca-LoRA","author":"Wang","year":"2023"},{"key":"ref11","article-title":"PaLM-E: An Embodied Multimodal Language Model","author":"Driess","year":"2023"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1055\/a-1858-6495"},{"key":"ref13","article-title":"Rt-2: Vision-language-action models transfer web knowledge to robotic control","author":"Brohan","year":"2023"},{"key":"ref14","article-title":"ChatGPT injected into Boston Dynamics\u2019 Spot","volume-title":"Cybernews","author":"Petkauskas","year":"2023"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/3610978.3641080"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/SMC54092.2024.10831380"},{"key":"ref17","article-title":"Robot Behavior-Tree-Based Task Generation with Large Language Models","author":"Cao","year":"2023"},{"key":"ref18","article-title":"Learning to summarize from human feedback","author":"Stiennon","year":"2022"},{"key":"ref19","article-title":"Self-Instruct: Aligning Language Model with Self Generated Instructions","author":"Wang","year":"2022"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1126\/scirobotics.abm6074"},{"key":"ref21","article-title":"BehaviorTree.CPP","author":"Faconti","year":"2019"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/iros.2014.6942752"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1201\/9780429489105"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/IROS45743.2020.9341562"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.14311\/app.2016.6.0011"},{"key":"ref26","article-title":"Empirical Analysis of the Strengths and Weaknesses of PEFT Techniques for LLMs","author":"Pu","year":"2023"},{"key":"ref27","article-title":"LoRA: Low-Rank Adaptation of Large Language Models","author":"Hu","year":"2021"}],"event":{"name":"2024 2nd International Conference on Foundation and Large Language Models (FLLM)","location":"Dubai, United Arab Emirates","start":{"date-parts":[[2024,11,26]]},"end":{"date-parts":[[2024,11,29]]}},"container-title":["2024 2nd International Conference on Foundation and Large Language Models (FLLM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10852419\/10852420\/10852491.pdf?arnumber=10852491","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,29]],"date-time":"2025-01-29T06:47:13Z","timestamp":1738133233000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10852491\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,26]]},"references-count":27,"URL":"https:\/\/doi.org\/10.1109\/fllm63129.2024.10852491","relation":{},"subject":[],"published":{"date-parts":[[2024,11,26]]}}}