{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T20:07:15Z","timestamp":1769544435667,"version":"3.49.0"},"reference-count":37,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"am","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["2322056"],"award-info":[{"award-number":["2322056"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["2423068"],"award-info":[{"award-number":["2423068"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["2520136"],"award-info":[{"award-number":["2520136"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000199","name":"U.S. Department of Agriculture","doi-asserted-by":"publisher","award":["2023-67021-39072"],"award-info":[{"award-number":["2023-67021-39072"]}],"id":[{"id":"10.13039\/100000199","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000199","name":"U.S. Department of Agriculture","doi-asserted-by":"publisher","award":["2024-67021-42878"],"award-info":[{"award-number":["2024-67021-42878"]}],"id":[{"id":"10.13039\/100000199","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Robot."],"published-print":{"date-parts":[[2026]]},"DOI":"10.1109\/tro.2025.3636302","type":"journal-article","created":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T19:03:15Z","timestamp":1764010995000},"page":"204-223","source":"Crossref","is-referenced-by-count":0,"title":["DiffOG: Differentiable Policy Trajectory Optimization With Generalizability"],"prefix":"10.1109","volume":"42","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2789-1910","authenticated-orcid":false,"given":"Zhengtong","family":"Xu","sequence":"first","affiliation":[{"name":"School of Industrial Engineering, Purdue University, West Lafayette, IN, USA"}]},{"given":"Zichen","family":"Miao","sequence":"additional","affiliation":[{"name":"School of Industrial Engineering, Purdue University, West Lafayette, IN, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2610-3502","authenticated-orcid":false,"given":"Qiang","family":"Qiu","sequence":"additional","affiliation":[{"name":"School of Industrial Engineering, Purdue University, West Lafayette, IN, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-0204-0571","authenticated-orcid":false,"given":"Zhe","family":"Zhang","sequence":"additional","affiliation":[{"name":"School of Industrial Engineering, Purdue University, West Lafayette, IN, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5914-3573","authenticated-orcid":false,"given":"Yu","family":"She","sequence":"additional","affiliation":[{"name":"School of Industrial Engineering, Purdue University, West Lafayette, IN, USA"}]}],"member":"263","reference":[{"key":"ref1","first-page":"136","article-title":"OptNet: Differentiable optimization as a layer in neural networks","volume-title":"Proc. 34th Int. Conf. Mach. Learn.","author":"Amos","year":"2017"},{"key":"ref2","first-page":"8299","article-title":"Differentiable MPC for end-to-end planning and control","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Amos","year":"2018"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.15607\/rss.2025.xxi.010"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-44051-0_26"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2023.XIX.026"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2024.XX.045"},{"key":"ref7","article-title":"Implicit behavioral cloning","volume-title":"Proc. Conf. Robot Learn.","author":"Florence","year":"2022"},{"key":"ref8","first-page":"4066","article-title":"Mobile ALOHA: Learning bimanual mobile manipulation with low-cost whole-body teleoperation","volume-title":"Proc. Conf. Robot Learn.","author":"Fu","year":"2024"},{"key":"ref9","first-page":"6840","article-title":"Denoising diffusion probabilistic models","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Ho","year":"2020"},{"key":"ref10","first-page":"26991","article-title":"Behavior generation with latent actions","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Lee","year":"2024"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.15607\/rss.2021.xvii.092"},{"key":"ref12","article-title":"RDT-1B: A diffusion foundation model for bimanual manipulation","volume-title":"Proc. 13th Int. Conf. Learn. Representations","author":"Liu","year":"2024"},{"key":"ref13","first-page":"1678","article-title":"What matters in learning from offline human demonstrations for robot manipulation","volume-title":"Proc. Conf. Robot Learn.","author":"Mandlekar","year":"2022"},{"key":"ref14","article-title":"Octo: An open-source generalist robot policy","volume-title":"Proc. Robot.: Sci. Syst. Conf.","year":"2024"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA57147.2024.10611477"},{"key":"ref16","first-page":"305","article-title":"ALVINN: An autonomous land vehicle in a neural network","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Pomerleau","year":"1988"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/IROS51168.2021.9636097"},{"key":"ref18","article-title":"Riemannian motion policies","author":"Ratliff","year":"2018"},{"key":"ref19","article-title":"Koopman constrained policy optimization: A Koopman operator theoretic method for differentiable optimal control in robotics","volume-title":"Proc. Int. Conf. Mach. Learn. workshop","author":"Retchin"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1177\/0278364914528132"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2025.3535182"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.2307\/j.ctt1ffjjxf.10"},{"key":"ref23","article-title":"Denoising diffusion implicit models","volume-title":"Proc. Int. Conf. Learn. Represent.","author":"Song"},{"key":"ref24","article-title":"DiffTORI: Differentiable trajectory optimization for deep reinforcement and imitation learning","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Wan"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2024.XX.043"},{"key":"ref26","article-title":"Equivariant diffusion policy","volume-title":"Proc. Conf. Robot Learn.","author":"Wang","year":"2024"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2024.XX.127"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2024.XX.092"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/IROS58592.2024.10801581"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/tro.2023.3249564"},{"key":"ref31","article-title":"Flow as the cross-domain manipulation interface","volume-title":"Proc. Conf. Robot Learn.","author":"Xu","year":"2024"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/tro.2024.3463470"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/tase.2024.3486542"},{"key":"ref34","article-title":"Meta-world: A benchmark and evaluation for multi-task and meta reinforcement learning","volume-title":"Proc. Conf. Robot Learn.","author":"Yu","year":"2020"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2024.XX.067"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.15607\/rss.2023.xix.016"},{"key":"ref37","article-title":"VIOLA: Imitation learning for vision-based manipulation with object proposal priors","volume-title":"Proc. Conf. Robot Learn.","author":"Zhu","year":"2023"}],"container-title":["IEEE Transactions on Robotics"],"original-title":[],"link":[{"URL":"https:\/\/ieeexplore.ieee.org\/ielam\/8860\/11297026\/11267071-aam.pdf","content-type":"application\/pdf","content-version":"am","intended-application":"syndication"},{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/8860\/11297026\/11267071.pdf?arnumber=11267071","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T06:09:01Z","timestamp":1769494141000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11267071\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"references-count":37,"URL":"https:\/\/doi.org\/10.1109\/tro.2025.3636302","relation":{},"ISSN":["1552-3098","1941-0468"],"issn-type":[{"value":"1552-3098","type":"print"},{"value":"1941-0468","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]}}}