{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T16:25:58Z","timestamp":1775665558360,"version":"3.50.1"},"reference-count":55,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,10,30]],"date-time":"2023-10-30T00:00:00Z","timestamp":1698624000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,10,30]],"date-time":"2023-10-30T00:00:00Z","timestamp":1698624000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100006751","name":"U.S. Army","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100006751","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000185","name":"Defense Advanced Research Projects Agency","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100000185","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,10,30]]},"DOI":"10.1109\/milcom58377.2023.10356344","type":"proceedings-article","created":{"date-parts":[[2023,12,25]],"date-time":"2023-12-25T19:37:01Z","timestamp":1703533021000},"page":"156-161","source":"Crossref","is-referenced-by-count":4,"title":["Challenges and Opportunities in Neuro-Symbolic Composition of Foundation Models"],"prefix":"10.1109","author":[{"given":"Susmit","family":"Jha","sequence":"first","affiliation":[{"name":"SRI International,Computer Science Laboratory,Menlo Park,CA,USA"}]},{"given":"Anirban","family":"Roy","sequence":"additional","affiliation":[{"name":"SRI International,Computer Science Laboratory,Menlo Park,CA,USA"}]},{"given":"Adam","family":"Cobb","sequence":"additional","affiliation":[{"name":"SRI International,Computer Science Laboratory,Menlo Park,CA,USA"}]},{"given":"Alexander","family":"Berenbeim","sequence":"additional","affiliation":[{"name":"United States Military Academy,Army Cyber Institute,West Point,NY,USA"}]},{"given":"Nathaniel D.","family":"Bastian","sequence":"additional","affiliation":[{"name":"United States Military Academy,Army Cyber Institute,West Point,NY,USA"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"crossref","DOI":"10.24963\/ijcai.2022\/89","article-title":"Detecting out-of-context objects using graph context reasoning network","volume-title":"IJCAI","author":"Acharya"},{"key":"ref2","first-page":"23716","article-title":"Flamingo: a visual language model for few-shot learning","volume":"35","author":"Alayrac","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1126\/science.177.4047.393"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00032"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.2139\/ssrn.4314839"},{"key":"ref6","article-title":"Scaling transformer to 1m tokens and beyond with rmt","author":"Bulatov","year":"2023"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01750"},{"key":"ref8","article-title":"Evaluating large language models trained on code","author":"Chen","year":"2021"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58577-8_7"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1016\/j.patrec.2011.12.004"},{"key":"ref11","article-title":"Scaling instruction-finetuned language models","author":"Chung","year":"2022"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.4324\/9780203849408"},{"key":"ref13","article-title":"Instructblip: Towards general-purpose vision-language models with instruction tuning","author":"Dai","year":"2023"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00269"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533229"},{"key":"ref16","article-title":"From images to textual prompts: Zero-shot vqa with frozen large language models","author":"Guo","year":"2022"},{"key":"ref17","article-title":"Training compute-optimal large language models","author":"Hoffmann","year":"2022"},{"key":"ref18","first-page":"9118","article-title":"Language models as zero-shot planners: Extracting actionable knowledge for embodied agents","volume-title":"International Conference on Machine Learning","author":"Huang"},{"key":"ref19","article-title":"Inner monologue: Embodied reasoning through planning with language models","author":"Huang","year":"2022"},{"key":"ref20","article-title":"Responsible reasoning with large language models and the impact of proper nouns","volume-title":"Workshop on Trustworthy and Socially Responsible ML, NeurIPS","author":"Jha"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1145\/1806799.1806833"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICAA58325.2023.00029"},{"key":"ref23","first-page":"4904","article-title":"Scaling up visual and vision-language representation learning with noisy text supervision","volume-title":"International Conference on Machine Learning","author":"Jia"},{"key":"ref24","article-title":"Scaling laws for neural language models","author":"Kaplan","year":"2020"},{"key":"ref25","article-title":"Large language models are zero-shot reasoners","author":"Kojima","year":"2022"},{"key":"ref26","article-title":"Mixout: Effective regularization to finetune large-scale pretrained language models","author":"Lee","year":"2019"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.488"},{"key":"ref28","article-title":"Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models","author":"Li","year":"2023"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58577-8_8"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE48619.2023.00207"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.eacl-main.185"},{"key":"ref32","article-title":"The next decade in ai: four steps towards robust artificial intelligence","author":"Marcus","year":"2020"},{"key":"ref33","article-title":"Experiments testing gpt-3\u2019s ability at commonsense reasoning: results","author":"Marcus","year":"2020"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.naacl-main.113"},{"key":"ref35","article-title":"Codegen: An open large language model for code with multi-turn program synthesis","author":"Nijkamp","year":"2022"},{"key":"ref36","year":"2023","journal-title":"Gpt-4 system card"},{"key":"ref37","first-page":"27730","article-title":"Training language models to follow instructions with human feedback","volume":"35","author":"Ouyang","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref38","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"International conference on machine learning","author":"Radford"},{"key":"ref39","article-title":"Improving language understanding by generative pre-training","author":"Radford","year":"2018"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1097\/01.back.0000532708.46105.ed"},{"key":"ref42","article-title":"Large language models are strong zero-shot retriever","author":"Shen","year":"2023"},{"key":"ref43","article-title":"Llm-planner: Few-shot grounded planning for embodied agents with large language models","author":"Song","year":"2022"},{"key":"ref44","article-title":"Beyond the imitation game: Quantifying and extrapolating the capabilities of language models","author":"Srivastava","year":"2022"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.findings-emnlp.67"},{"key":"ref46","article-title":"Llama: Open and efficient foundation language models","author":"Touvron","year":"2023"},{"key":"ref47","article-title":"Large language models still can\u2019t plan (a benchmark for llms on planning and reasoning about change)","author":"Valmeekam","year":"2022"},{"key":"ref48","article-title":"On the planning abilities of large language models (a critical investigation with a proposed benchmark)","author":"Valmeekam","year":"2023"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2017.2754499"},{"key":"ref50","article-title":"Preserving in-context learning ability in large language model fine-tuning","author":"Wang","year":"2022"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.685"},{"key":"ref52","article-title":"Emergent abilities of large language models","author":"Wei","year":"2022"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.346"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr46437.2021.00553"},{"key":"ref55","article-title":"Opt: Open pre-trained transformer language models","author":"Zhang","year":"2022"}],"event":{"name":"MILCOM 2023 - 2023 IEEE Military Communications Conference (MILCOM)","location":"Boston, MA, USA","start":{"date-parts":[[2023,10,30]]},"end":{"date-parts":[[2023,11,3]]}},"container-title":["MILCOM 2023 - 2023 IEEE Military Communications Conference (MILCOM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10356123\/10356124\/10356344.pdf?arnumber=10356344","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,12]],"date-time":"2024-01-12T20:20:04Z","timestamp":1705090804000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10356344\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,30]]},"references-count":55,"URL":"https:\/\/doi.org\/10.1109\/milcom58377.2023.10356344","relation":{},"subject":[],"published":{"date-parts":[[2023,10,30]]}}}