{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,9]],"date-time":"2025-05-09T14:48:47Z","timestamp":1746802127767,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":33,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819985548"},{"type":"electronic","value":"9789819985555"}],"license":[{"start":{"date-parts":[[2023,12,28]],"date-time":"2023-12-28T00:00:00Z","timestamp":1703721600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,28]],"date-time":"2023-12-28T00:00:00Z","timestamp":1703721600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8555-5_39","type":"book-chapter","created":{"date-parts":[[2023,12,27]],"date-time":"2023-12-27T07:02:36Z","timestamp":1703660556000},"page":"495-507","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Stereo3DMOT: Stereo Vision Based 3D Multi-object Tracking with\u00a0Multimodal ReID"],"prefix":"10.1007","author":[{"given":"Chen","family":"Mao","sequence":"first","affiliation":[]},{"given":"Chong","family":"Tan","sequence":"additional","affiliation":[]},{"given":"Hong","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jingqi","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Min","family":"Zheng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,28]]},"reference":[{"key":"39_CR1","doi-asserted-by":"crossref","unstructured":"Shi, S., Wang, X., Li, H.: Pointrcnn: 3d object proposal generation and detection from point cloud. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 770\u2013779 (2019)","DOI":"10.1109\/CVPR.2019.00086"},{"key":"39_CR2","doi-asserted-by":"crossref","unstructured":"Sun, J., et al.: DISP R-CNN: stereo 3d object detection via shape prior guided instance disparity estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10548\u201310557 (2020)","DOI":"10.1109\/CVPR42600.2020.01056"},{"key":"39_CR3","doi-asserted-by":"crossref","unstructured":"Li, P., Chen, X., Shen, S.: Stereo R-CNN based 3d object detection for autonomous driving. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7644\u20137652 (2019)","DOI":"10.1109\/CVPR.2019.00783"},{"key":"39_CR4","doi-asserted-by":"crossref","unstructured":"Weng, X., Wang, J., Held, D., Kitani, K.: Ab3dmot: a baseline for 3d multi-object tracking and new evaluation metrics. arXiv preprint arXiv:2008.08063 (2020)","DOI":"10.1109\/IROS45743.2020.9341164"},{"key":"39_CR5","doi-asserted-by":"publisher","unstructured":"Pang, Z., Li, Z., Wang, N.: SimpleTrack: understanding and rethinking 3D multi-object tracking. In: Karlinsky, L., Michaeli, T., Nishino, K. (eds.) Computer - ECCV 2022 Workshops. ECCV 2022, Part I, LNCS, vol. 13801, pp. 680\u2013696. Springer, Cham (2023). https:\/\/doi.org\/10.1007\/978-3-031-25056-9_43","DOI":"10.1007\/978-3-031-25056-9_43"},{"key":"39_CR6","doi-asserted-by":"crossref","unstructured":"Benbarka, N., Schr\u00f6der, J., Zell, A.: Score refinement for confidence-based 3d multi-object tracking. In: 2021 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 8083\u20138090. IEEE (2021)","DOI":"10.1109\/IROS51168.2021.9636032"},{"issue":"6","key":"39_CR7","doi-asserted-by":"publisher","first-page":"5668","DOI":"10.1109\/TITS.2021.3055616","volume":"23","author":"H Wu","year":"2021","unstructured":"Wu, H., Han, W., Wen, C., Li, X., Wang, C.: 3d multi-object tracking in point clouds based on prediction confidence-guided data association. IEEE Trans. Intell. Transp. Syst. 23(6), 5668\u20135677 (2021)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"39_CR8","unstructured":"Wang, X., He, J., Fu, C., Meng, T., Huang, M.: You only need two detectors to achieve multi-modal 3d multi-object tracking. arXiv preprint arXiv:2304.08709 (2023)"},{"key":"39_CR9","doi-asserted-by":"crossref","unstructured":"Huang, K., Hao, Q.: Joint multi-object detection and tracking with camera-lidar fusion for autonomous driving. In: 2021 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 6983\u20136989. IEEE (2021)","DOI":"10.1109\/IROS51168.2021.9636311"},{"key":"39_CR10","doi-asserted-by":"publisher","first-page":"11241","DOI":"10.1109\/JSEN.2022.3226490","volume":"23","author":"X Wang","year":"2022","unstructured":"Wang, X., Fu, C., He, J., Wang, S., Wang, J.: Strongfusionmot: a multi-object tracking method based on lidar-camera fusion. IEEE Sens. J. 23, 11241\u201311252 (2022)","journal-title":"IEEE Sens. J."},{"issue":"4","key":"39_CR11","doi-asserted-by":"publisher","first-page":"874","DOI":"10.3390\/rs15040874","volume":"15","author":"K Zhang","year":"2023","unstructured":"Zhang, K., Liu, Y., Mei, F., Jin, J., Wang, Y.: Boost correlation features with 3D-MiIoU-based camera-LiDAR fusion for MODT in autonomous driving. Remote Sens. 15(4), 874 (2023)","journal-title":"Remote Sens."},{"key":"39_CR12","doi-asserted-by":"crossref","unstructured":"Baser, E., Balasubramanian, V., Bhattacharyya, P., Czarnecki, K.: Fantrack: 3d multi-object tracking with feature association network. In: 2019 IEEE Intelligent Vehicles Symposium (IV), pp. 1426\u20131433. IEEE (2019)","DOI":"10.1109\/IVS.2019.8813779"},{"key":"39_CR13","doi-asserted-by":"crossref","unstructured":"Marinello, N., Proesmans, M., Van Gool, L.: Triplettrack: 3d object tracking using triplet embeddings and LSTM. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4500\u20134510 (2022)","DOI":"10.1109\/CVPRW56347.2022.00496"},{"key":"39_CR14","doi-asserted-by":"crossref","unstructured":"Weng, X., Wang, Y., Man, Y., Kitani, K.M.: Gnn3dmot: Graph neural network for 3d multi-object tracking with 2d\u20133d multi-feature learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6499\u20136508 (2020)","DOI":"10.1109\/CVPR42600.2020.00653"},{"key":"39_CR15","doi-asserted-by":"crossref","unstructured":"Kim, A., O\u0161ep, A., Leal-Taix\u00e9, L.: Eagermot: 3d multi-object tracking via sensor fusion. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 11315\u201311321. IEEE (2021)","DOI":"10.1109\/ICRA48506.2021.9562072"},{"issue":"3","key":"39_CR16","doi-asserted-by":"publisher","first-page":"8260","DOI":"10.1109\/LRA.2022.3187264","volume":"7","author":"X Wang","year":"2022","unstructured":"Wang, X., Fu, C., Li, Z., Lai, Y., He, J.: DeepFusionMOT: a 3d multi-object tracking framework based on camera-lidar fusion with deep association. IEEE Robot. Autom. Lett. 7(3), 8260\u20138267 (2022)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"39_CR17","doi-asserted-by":"crossref","unstructured":"Kuma, R., Weill, E., Aghdasi, F., Sriram, P.: Vehicle re-identification: an efficient baseline using triplet embedding. In: 2019 International Joint Conference on Neural Networks (IJCNN), pp. 1\u20139. IEEE (2019)","DOI":"10.1109\/IJCNN.2019.8852059"},{"key":"39_CR18","doi-asserted-by":"crossref","unstructured":"Hao, Y., Wang, N., Li, J., Gao, X.: HSmE: hypersphere manifold embedding for visible thermal person re-identification. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 8385\u20138392 (2019)","DOI":"10.1609\/aaai.v33i01.33018385"},{"key":"39_CR19","doi-asserted-by":"crossref","unstructured":"Li, Y.J., Chen, Y.C., Lin, Y.Y., Du, X., Wang, Y.C.F.: Recover and identify: a generative dual model for cross-resolution person re-identification. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8090\u20138099 (2019)","DOI":"10.1109\/ICCV.2019.00818"},{"issue":"7","key":"39_CR20","doi-asserted-by":"publisher","first-page":"1770","DOI":"10.1109\/TPAMI.2019.2903058","volume":"42","author":"M Li","year":"2019","unstructured":"Li, M., Zhu, X., Gong, S.: Unsupervised tracklet person re-identification. IEEE Trans. Pattern Anal. Mach. Intell. 42(7), 1770\u20131782 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"39_CR21","doi-asserted-by":"crossref","unstructured":"Luo, H., Gu, Y., Liao, X., Lai, S., Jiang, W.: Bag of tricks and a strong baseline for deep person re-identification. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2019)","DOI":"10.1109\/CVPRW.2019.00190"},{"issue":"6","key":"39_CR22","doi-asserted-by":"publisher","first-page":"2872","DOI":"10.1109\/TPAMI.2021.3054775","volume":"44","author":"M Ye","year":"2021","unstructured":"Ye, M., Shen, J., Lin, G., Xiang, T., Shao, L., Hoi, S.C.: Deep learning for person re-identification: a survey and outlook. IEEE Trans. Pattern Anal. Mach. Intell. 44(6), 2872\u20132893 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"39_CR23","unstructured":"He, L., Liao, X., Liu, W., Liu, X., Cheng, P., Mei, T.: Fastreid: a pytorch toolbox for general instance re-identification. arXiv preprint arXiv:2006.02631 (2020)"},{"issue":"6","key":"39_CR24","doi-asserted-by":"publisher","first-page":"2588","DOI":"10.1109\/TIP.2017.2675201","volume":"26","author":"A Wu","year":"2017","unstructured":"Wu, A., Zheng, W.S., Lai, J.H.: Robust depth-based person re-identification. IEEE Trans. Image Process. 26(6), 2588\u20132603 (2017)","journal-title":"IEEE Trans. Image Process."},{"key":"39_CR25","unstructured":"Karianakis, N., Liu, Z., Chen, Y., Soatto, S.: Person depth REID: robust person re-identification with commodity depth sensors. arXiv preprint arXiv:1705.09882 (2017)"},{"key":"39_CR26","doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? the kitti vision benchmark suite. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 3354\u20133361. IEEE (2012)","DOI":"10.1109\/CVPR.2012.6248074"},{"key":"39_CR27","doi-asserted-by":"crossref","unstructured":"Howard, A., et al.: Searching for mobilenetv3. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1314\u20131324 (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"key":"39_CR28","doi-asserted-by":"crossref","unstructured":"Xu, G., Wang, Y., Cheng, J., Tang, J., Yang, X.: Accurate and efficient stereo matching via attention concatenation volume. arXiv preprint arXiv:2209.12699 (2022)","DOI":"10.1109\/TPAMI.2023.3335480"},{"key":"39_CR29","unstructured":"Qi, C.R., Su, H., Mo, K., Guibas, L.J.: Pointnet: deep learning on point sets for 3d classification and segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 652\u2013660 (2017)"},{"key":"39_CR30","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818\u20132826 (2016)","DOI":"10.1109\/CVPR.2016.308"},{"key":"39_CR31","doi-asserted-by":"crossref","unstructured":"Shenoi, A., et al.: JRMOT: a real-time 3d multi-object tracker and a new large-scale dataset. In: 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 10335\u201310342. IEEE (2020)","DOI":"10.1109\/IROS45743.2020.9341635"},{"issue":"2","key":"39_CR32","doi-asserted-by":"publisher","first-page":"1803","DOI":"10.1109\/LRA.2020.2969183","volume":"5","author":"J Luiten","year":"2020","unstructured":"Luiten, J., Fischer, T., Leibe, B.: Track to reconstruct and reconstruct to track. IEEE Robot. Autom. Lett. 5(2), 1803\u20131810 (2020)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"39_CR33","doi-asserted-by":"publisher","unstructured":"Kim, A., Bras\u00f3, G., O\u0161ep, A., Leal-Taix\u00e9, L. (2022). PolarMOT: how far can geometric relations take us in 3D multi-object tracking?. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision - ECCV 2022. ECCV 2022. LNCS, Part XXII, vol. 13682, pp. 41\u201358. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20047-2_3","DOI":"10.1007\/978-3-031-20047-2_3"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8555-5_39","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,27]],"date-time":"2023-12-27T07:10:52Z","timestamp":1703661052000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8555-5_39"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,28]]},"ISBN":["9789819985548","9789819985555"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8555-5_39","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023,12,28]]},"assertion":[{"value":"28 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Xiamen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/prcv2023.xmu.edu.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1420","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"532","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"37% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,78","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,69","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}