{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,7]],"date-time":"2024-09-07T12:22:22Z","timestamp":1725711742522},"reference-count":28,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,10,1]],"date-time":"2023-10-01T00:00:00Z","timestamp":1696118400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,10,1]],"date-time":"2023-10-01T00:00:00Z","timestamp":1696118400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100007129","name":"Natural Science Foundation of Shandong Province","doi-asserted-by":"publisher","award":["ZR2020MF041,ZR2022MF237"],"award-info":[{"award-number":["ZR2020MF041,ZR2022MF237"]}],"id":[{"id":"10.13039\/501100007129","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["11901325"],"award-info":[{"award-number":["11901325"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,10,1]]},"DOI":"10.1109\/smc53992.2023.10394607","type":"proceedings-article","created":{"date-parts":[[2024,1,29]],"date-time":"2024-01-29T18:32:04Z","timestamp":1706553124000},"page":"666-671","source":"Crossref","is-referenced-by-count":0,"title":["Exemplar-Free Continual Learning in Vision Transformers via Feature Attention Distillation"],"prefix":"10.1109","volume":"abs\/2005.14165","author":[{"given":"Xiaoyu","family":"Dai","sequence":"first","affiliation":[{"name":"Qilu University of Technology (Shandong Academy of Sciences),Faculty of Computer Science and Technology,Jinan,China"}]},{"given":"Jinyong","family":"Cheng","sequence":"additional","affiliation":[{"name":"Qilu University of Technology (Shandong Academy of Sciences),Faculty of Computer Science and Technology,Jinan,China"}]},{"given":"Zhonghe","family":"Wei","sequence":"additional","affiliation":[{"name":"Qilu University of Technology (Shandong Academy of Sciences),Faculty of Computer Science and Technology,Jinan,China"}]},{"given":"Baoyu","family":"Du","sequence":"additional","affiliation":[{"name":"Qilu University of Technology (Shandong Academy of Sciences),Faculty of Computer Science and Technology,Jinan,China"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Language models are fewshot learners","volume":"abs\/2005.14165","author":"Brown","year":"2020","journal-title":"ArXiv"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00353"},{"key":"ref3","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","volume":"abs\/2010.11929","author":"Dosovitskiy","year":"2020","journal-title":"ArXiv"},{"key":"ref4","article-title":"Efficientnet: Rethinking model scaling for convolutional neural networks","volume":"abs\/1905.11946","author":"Tan","year":"2019","journal-title":"ArXiv"},{"key":"ref5","article-title":"Training data-efficient image transformers & distillation through attention","volume":"abs\/2012.12877","author":"Touvron","year":"2020","journal-title":"ArXiv"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/IGARSS46834.2022.9883983"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00863"},{"key":"ref8","article-title":"Online continual learning through mutual information maximization","volume-title":"International Conference on Machine Learning","author":"Guo","year":"2022"},{"key":"ref9","article-title":"Gradient episodic memory for continual learning","author":"Lopez-Paz","year":"2017","journal-title":"NIPS"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3057446"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1038\/s41583-021-00548-3"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1611835114"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01252-6_33"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00228"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2017.2773081"},{"key":"ref16","first-page":"3987","article-title":"Continual learning through synaptic intelligence","volume-title":"Proceedings of machine learning research","volume":"70","author":"Zenke","year":"2017"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP46576.2022.9897554"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW56347.2022.00427"},{"key":"ref19","article-title":"Early convolutions help transformers see better","author":"Xiao","year":"2021","journal-title":"Neural Information Processing Systems"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref21","article-title":"Attention is all you need","volume":"abs\/1706.03762","author":"Vaswani","year":"2017","journal-title":"ArXiv"},{"key":"ref22","article-title":"Self-paced contrastive learning with hybrid memory for domain adaptive object re-id","volume":"abs\/2006.02713","author":"Ge","year":"2020","journal-title":"ArXiv"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00381"},{"key":"ref24","article-title":"Attention distillation: self-supervised vision transformer students need more guidance","volume-title":"British Machine Vision Conference","author":"Wang","year":"2022"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1186\/s40537-019-0197-0"},{"key":"ref26","article-title":"Continual normalization: Rethinking batch normalization for online continual learning","volume":"abs\/2203.16102","author":"Pham","year":"2022","journal-title":"ArXiv"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.587"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58565-5_6"}],"event":{"name":"2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC)","start":{"date-parts":[[2023,10,1]]},"location":"Honolulu, Oahu, HI, USA","end":{"date-parts":[[2023,10,4]]}},"container-title":["2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10391856\/10393862\/10394607.pdf?arnumber=10394607","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,1]],"date-time":"2024-02-01T13:18:01Z","timestamp":1706793481000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10394607\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,1]]},"references-count":28,"URL":"https:\/\/doi.org\/10.1109\/smc53992.2023.10394607","relation":{},"subject":[],"published":{"date-parts":[[2023,10,1]]}}}