{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T20:22:36Z","timestamp":1740169356679,"version":"3.37.3"},"reference-count":35,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"name":"Shaanxi University of Technology","award":["SLGRC202416"],"award-info":[{"award-number":["SLGRC202416"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2024]]},"DOI":"10.1109\/access.2024.3404463","type":"journal-article","created":{"date-parts":[[2024,5,23]],"date-time":"2024-05-23T18:14:11Z","timestamp":1716488051000},"page":"73218-73227","source":"Crossref","is-referenced-by-count":0,"title":["ASDeM: Augmenting SAM With Decoupled Memory for Video Object Segmentation"],"prefix":"10.1109","volume":"12","author":[{"given":"Xiaohu","family":"Liu","sequence":"first","affiliation":[{"name":"Trine Engineering Institute, Shaanxi University of Technology, Hanzhong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2064-8037","authenticated-orcid":false,"given":"Yichuang","family":"Luo","sequence":"additional","affiliation":[{"name":"Department of Intelligent Science and Engineering, Xi&#x2019;an Peihua University, Xi&#x2019;an, China"}]},{"given":"Wei","family":"Sun","sequence":"additional","affiliation":[{"name":"School of Mechanical Engineering, Shaanxi University of Technology, Hanzhong, China"}]}],"member":"263","reference":[{"key":"ref1","first-page":"11781","article-title":"Rethinking space-time networks with improved memory coverage for efficient video object segmentation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Cheng"},{"key":"ref2","first-page":"2491","article-title":"Associating objects with transformers for video object segmentation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Yang"},{"key":"ref3","first-page":"36324","article-title":"Decoupling features in hierarchical propagation for video object segmentation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Yang"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19815-1_37"},{"key":"ref5","article-title":"The 2017 Davis challenge on video object segmentation","author":"Pont-Tuset","year":"2017","journal-title":"arXiv:1704.00675"},{"key":"ref6","article-title":"YouTube-VOS: A large-scale video object segmentation benchmark","author":"Xu","year":"2018","journal-title":"arXiv:1809.03327"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"ref8","article-title":"Track anything: Segment anything meets videos","author":"Yang","year":"2023","journal-title":"arXiv:2304.11968"},{"key":"ref9","article-title":"Segment and track anything","author":"Cheng","year":"2023","journal-title":"arXiv:2305.06558"},{"key":"ref10","article-title":"Tracking anything in high quality","author":"Zhu","year":"2023","journal-title":"arXiv:2307.13974"},{"key":"ref11","article-title":"Segment anything in high quality","author":"Ke","year":"2023","journal-title":"arXiv:2306.01567"},{"key":"ref12","article-title":"Segment everything everywhere all at once","author":"Zou","year":"2023","journal-title":"arXiv:2304.06718"},{"key":"ref13","article-title":"An image is worth 16\u00d716 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020","journal-title":"arXiv:2010.11929"},{"key":"ref14","article-title":"Segment anything meets point tracking","author":"Raji\u010d","year":"2023","journal-title":"arXiv:2307.01197"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-20047-2_4"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00932"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58542-6_38"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00770"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58558-7_20"},{"key":"ref20","article-title":"Adam: A method for stochastic optimization","author":"Kingma","year":"2014","journal-title":"arXiv:1412.6980"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.85"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00551"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2021\/662"},{"key":"ref24","first-page":"31265","article-title":"MinVIS: A minimal video instance segmentation framework without video-based training","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Huang"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00135"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00399"},{"key":"ref27","first-page":"13352","article-title":"Video instance segmentation using inter-frame communication transformers","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Hwang"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00863"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00976"},{"key":"ref30","article-title":"Mask2Former for video instance segmentation","author":"Cheng","year":"2021","journal-title":"arXiv:2112.10764"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00617"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01334"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00127"},{"key":"ref34","first-page":"1192","article-title":"Prototypical cross-attention networks for multiple object tracking and segmentation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Ke"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3081597"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/10380310\/10537204.pdf?arnumber=10537204","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,31]],"date-time":"2024-05-31T04:44:31Z","timestamp":1717130671000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10537204\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"references-count":35,"URL":"https:\/\/doi.org\/10.1109\/access.2024.3404463","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2024]]}}}