{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,8,9]],"date-time":"2024-08-09T07:50:41Z","timestamp":1723189841699},"reference-count":20,"publisher":"Institute of Electronics, Information and Communications Engineers (IEICE)","issue":"12","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEICE Trans. Fundamentals"],"published-print":{"date-parts":[[2022,12,1]]},"DOI":"10.1587\/transfun.2022eal2008","type":"journal-article","created":{"date-parts":[[2022,5,23]],"date-time":"2022-05-23T22:10:22Z","timestamp":1653343822000},"page":"1621-1625","source":"Crossref","is-referenced-by-count":1,"title":["Vehicle Re-Identification Based on Quadratic Split Architecture and Auxiliary Information Embedding"],"prefix":"10.1587","volume":"E105.A","author":[{"given":"Tongwei","family":"LU","sequence":"first","affiliation":[{"name":"Hubei Key Laboratory of Intelligent Robot (Wuhan Institute of Technology)"},{"name":"School of Computer Science &amp; Engineering, Wuhan Institute of Technology"}]},{"given":"Hao","family":"ZHANG","sequence":"additional","affiliation":[{"name":"Hubei Key Laboratory of Intelligent Robot (Wuhan Institute of Technology)"},{"name":"School of Computer Science &amp; Engineering, Wuhan Institute of Technology"}]},{"given":"Feng","family":"MIN","sequence":"additional","affiliation":[{"name":"Hubei Key Laboratory of Intelligent Robot (Wuhan Institute of Technology)"},{"name":"School of Computer Science &amp; Engineering, Wuhan Institute of Technology"}]},{"given":"Shihai","family":"JIA","sequence":"additional","affiliation":[{"name":"Hubei Key Laboratory of Intelligent Robot (Wuhan Institute of Technology)"},{"name":"School of Computer Science &amp; Engineering, Wuhan Institute of Technology"}]}],"member":"532","reference":[{"key":"1","doi-asserted-by":"crossref","unstructured":"[1] X. Liu, S. Zhang, Q. Huang, and W. Gao, \u201cRam: A region-aware deep model for vehicle re-identification,\u201d 2018 IEEE International Conference on Multimedia and Expo (ICME), pp.1-6, IEEE, 2018. 10.1109\/icme.2018.8486589","DOI":"10.1109\/ICME.2018.8486589"},{"key":"2","doi-asserted-by":"publisher","unstructured":"[2] J. Zhu, H. Zeng, J. Huang, S. Liao, Z. Lei, C. Cai, and L. Zheng, \u201cVehicle re-identification using quadruple directional deep learning features,\u201d IEEE Trans. Intell. Transp. Syst., vol.21, no.1, pp.410-420, 2019. 10.1109\/tits.2019.2901312","DOI":"10.1109\/TITS.2019.2901312"},{"key":"3","doi-asserted-by":"crossref","unstructured":"[3] B. He, J. Li, Y. Zhao, and Y. Tian, \u201cPart-regularized near-duplicate vehicle re-identification,\u201d Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.3997-4005, 2019. 10.1109\/cvpr.2019.00412","DOI":"10.1109\/CVPR.2019.00412"},{"key":"4","doi-asserted-by":"crossref","unstructured":"[4] I.O. de Oliveira, K.V. Fonseca, and R. Minetto, \u201cA two-stream siamese neural network for vehicle re-identification by using non-overlapping cameras,\u201d 2019 IEEE International Conference on Image Processing (ICIP), pp.669-673, IEEE, 2019. 10.1109\/icip.2019.8803810","DOI":"10.1109\/ICIP.2019.8803810"},{"key":"5","doi-asserted-by":"crossref","unstructured":"[5] D. Meng, L. Li, X. Liu, Y. Li, S. Yang, Z.J. Zha, X. Gao, S. Wang, and Q. Huang, \u201cParsing-based view-aware embedding network for vehicle re-identification,\u201d Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.7103-7112, 2020. 10.1109\/cvpr42600.2020.00713","DOI":"10.1109\/CVPR42600.2020.00713"},{"key":"6","doi-asserted-by":"crossref","unstructured":"[6] H. Chen, Y. Wang, T. Guo, C. Xu, Y. Deng, Z. Liu, S. Ma, C. Xu, C. Xu, and W. Gao, \u201cPre-trained image processing transformer,\u201d Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.12299-12310, 2021. 10.1109\/cvpr46437.2021.01212","DOI":"10.1109\/CVPR46437.2021.01212"},{"key":"7","unstructured":"[7] X. Zhu, W. Su, L. Lu, B. Li, X. Wang, and J. Dai, \u201cDeformable detr: Deformable transformers for end-to-end object detection,\u201d arXiv preprint arXiv:2010.04159, 2020. 10.48550\/arXiv.2010.04159"},{"key":"8","unstructured":"[8] H. Touvron, M. Cord, M. Douze, F. Massa, A. Sablayrolles, and H. J\u00e9gou, \u201cTraining data-efficient image transformers &amp; distillation through attention,\u201d International Conference on Machine Learning, pp.10347-10357, PMLR, 2021."},{"key":"9","unstructured":"[9] B. Wu, C. Xu, X. Dai, A. Wan, P. Zhang, Z. Yan, M. Tomizuka, J. Gonzalez, K. Keutzer, and P. Vajda, \u201cVisual transformers: Token-based image representation and processing for computer vision,\u201d arXiv preprint arXiv:2006.03677, 2020. 10.48550\/arXiv.2006.03677"},{"key":"10","unstructured":"[10] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby, \u201cAn image is worth 16x16 words: Transformers for image recognition at scale,\u201d International Conference on Learning Representations, 2021."},{"key":"11","doi-asserted-by":"crossref","unstructured":"[11] S. He, H. Luo, P. Wang, F. Wang, H. Li, and W. Jiang, \u201cTransReID: Transformer-based object re-identification,\u201d Proc. IEEE\/CVF International Conference on Computer Vision, pp.15013-15022, 2021. 10.1109\/iccv48922.2021.01474","DOI":"10.1109\/ICCV48922.2021.01474"},{"key":"12","doi-asserted-by":"crossref","unstructured":"[12] X. Liu, W. Liu, H. Ma, and H. Fu, \u201cLarge-scale vehicle re-identification in urban surveillance videos,\u201d 2016 IEEE international conference on multimedia and expo (ICME), pp.1-6, IEEE, 2016. 10.1109\/icme.2016.7553002","DOI":"10.1109\/ICME.2016.7553002"},{"key":"13","doi-asserted-by":"crossref","unstructured":"[13] H. Liu, Y. Tian, Y. Yang, L. Pang, and T. Huang, \u201cDeep relative distance learning: Tell the difference between similar vehicles,\u201d Proc. IEEE Conference on Computer Vision and Pattern Recognition, pp.2167-2175, 2016. 10.1109\/cvpr.2016.238","DOI":"10.1109\/CVPR.2016.238"},{"key":"14","doi-asserted-by":"crossref","unstructured":"[14] P. Khorramshahi, A. Kumar, N. Peri, S.S. Rambhatla, J.C. Chen, and R. Chellappa, \u201cA dual-path model with adaptive attention for vehicle re-identification,\u201d Proc. IEEE\/CVF International Conference on Computer Vision, pp.6132-6141, 2019. 10.1109\/iccv.2019.00623","DOI":"10.1109\/ICCV.2019.00623"},{"key":"15","doi-asserted-by":"crossref","unstructured":"[15] R. Chu, Y. Sun, Y. Li, Z. Liu, C. Zhang, and Y. Wei, \u201cVehicle re-identification with viewpoint-aware metric learning,\u201d Proc. IEEE\/CVF International Conference on Computer Vision, pp.8282-8291, 2019. 10.1109\/iccv.2019.00837","DOI":"10.1109\/ICCV.2019.00837"},{"key":"16","doi-asserted-by":"publisher","unstructured":"[16] H. Guo, C. Zhao, Z. Liu, J. Wang, and H. Lu, \u201cLearning coarse-to-fine structured feature embedding for vehicle re-identification,\u201d Proc. Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence, vol.32, no.1, pp.6853-6860, 2018. 10.1609\/aaai.v32i1.12237","DOI":"10.1609\/aaai.v32i1.12237"},{"key":"17","doi-asserted-by":"crossref","unstructured":"[17] W. Lin, Y. Li, X. Yang, P. Peng, and J. Xing, \u201cMulti-view learning for vehicle re-identification,\u201d 2019 IEEE international conference on multimedia and expo (ICME), pp.832-837, IEEE, 2019. 10.1109\/icme.2019.00148","DOI":"10.1109\/ICME.2019.00148"},{"key":"18","doi-asserted-by":"crossref","unstructured":"[18] X. Zhang, R. Zhang, J. Cao, D. Gong, M. You, and C. Shen, \u201cPart-guided attention learning for vehicle re-identification,\u201d arXiv preprint arXiv:1909.06023, vol.2, no.8, 2019. 10.48550\/arXiv.1909.06023","DOI":"10.5260\/chara.21.2.8"},{"key":"19","doi-asserted-by":"crossref","unstructured":"[19] P. Khorramshahi, N. Peri, J.C. Chen, and R. Chellappa, \u201cThe devil is in the details: Self-supervised attention for vehicle re-identification,\u201d European Conference on Computer Vision, pp.369-386, Springer, 2020. 10.1007\/978-3-030-58568-6_22","DOI":"10.1007\/978-3-030-58568-6_22"},{"key":"20","unstructured":"[20] A. Suprem and C. Pu, \u201cLooking glamorous: Vehicle re-id in heterogeneous cameras networks with global and local attention,\u201d arXiv preprint arXiv:2002.02256, 2020. 10.48550\/arXiv.2002.02256"}],"container-title":["IEICE Transactions on Fundamentals of Electronics, Communications and Computer Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transfun\/E105.A\/12\/E105.A_2022EAL2008\/_pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,10]],"date-time":"2024-05-10T05:07:28Z","timestamp":1715317648000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transfun\/E105.A\/12\/E105.A_2022EAL2008\/_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,12,1]]},"references-count":20,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2022]]}},"URL":"https:\/\/doi.org\/10.1587\/transfun.2022eal2008","relation":{},"ISSN":["0916-8508","1745-1337"],"issn-type":[{"value":"0916-8508","type":"print"},{"value":"1745-1337","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,12,1]]},"article-number":"2022EAL2008"}}