{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,30]],"date-time":"2026-03-30T10:17:41Z","timestamp":1774865861230,"version":"3.50.1"},"reference-count":33,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2023,7,9]],"date-time":"2023-07-09T00:00:00Z","timestamp":1688860800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,7,9]],"date-time":"2023-07-09T00:00:00Z","timestamp":1688860800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"National Program on Key Basic Research Project","award":["2014CB744903"],"award-info":[{"award-number":["2014CB744903"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61673270"],"award-info":[{"award-number":["61673270"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Artificial Intelligence Key Laboratory of Sichuan Province","award":["2022RZY02"],"award-info":[{"award-number":["2022RZY02"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,4]]},"DOI":"10.1007\/s00371-023-02983-y","type":"journal-article","created":{"date-parts":[[2023,7,9]],"date-time":"2023-07-09T14:01:12Z","timestamp":1688911272000},"page":"2761-2773","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":14,"title":["MotionTrack: rethinking the motion cue for multiple object tracking in USV videos"],"prefix":"10.1007","volume":"40","author":[{"given":"Zhenqi","family":"Liang","sequence":"first","affiliation":[]},{"given":"Gang","family":"Xiao","sequence":"additional","affiliation":[]},{"given":"Jianqiu","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Jingshi","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Chunshan","family":"Ding","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,7,9]]},"reference":[{"key":"2983_CR1","doi-asserted-by":"publisher","unstructured":"Ramanan, D., Forsyth, D.A.: Finding and tracking people from the bottom up. In: 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings., Vol.\u00a02, IEEE (2003). https:\/\/doi.org\/10.1007\/10.1109\/CVPR.2003.1211504","DOI":"10.1007\/10.1109\/CVPR.2003.1211504"},{"key":"2983_CR2","doi-asserted-by":"publisher","unstructured":"Bewley, A., Ge, Z., Ott, L., Ramos, F., Upcroft, B.: Simple online and realtime tracking. In: IEEE International Conference on Image Processing (ICIP). IEEE 2016, 3464\u20133468 (2016). https:\/\/doi.org\/10.1109\/ICIP.2016.7533003","DOI":"10.1109\/ICIP.2016.7533003"},{"key":"2983_CR3","doi-asserted-by":"publisher","unstructured":"Zhang, Y., Sun, P., Jiang, Y., Yu, D., Weng, F., Yuan, Z., Luo, P., Liu, W., Wang, X.: Bytetrack: multi-object tracking by associating every detection box. In: Computer Vision\u2013ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23\u201327, 2022, Proceedings, Part XXII, Springer, (2022), pp. 1\u201321. https:\/\/doi.org\/10.1007\/978-3-031-20047-2_1","DOI":"10.1007\/978-3-031-20047-2_1"},{"key":"2983_CR4","doi-asserted-by":"publisher","unstructured":"Cao, J., Weng, X., Khirodkar, R., Pang, J., Kitani, K.: Observation-centric sort: rethinking sort for robust multi-object tracking (2022). https:\/\/doi.org\/10.48550\/arXiv.2203.14360. arXiv preprint arXiv:2203.14360","DOI":"10.48550\/arXiv.2203.14360"},{"key":"2983_CR5","doi-asserted-by":"publisher","unstructured":"Kalman, R.E.: A new approach to linear filtering and prediction problems (1960). https:\/\/doi.org\/10.1115\/1.3662552","DOI":"10.1115\/1.3662552"},{"key":"2983_CR6","doi-asserted-by":"publisher","unstructured":"Wojke, N., Bewley, A., Paulus, D.: Simple online and realtime tracking with a deep association metric. In: IEEE International Conference on Image Processing (ICIP). IEEE 2017, 3645\u20133649 (2017). https:\/\/doi.org\/10.1109\/ICIP.2017.8296962","DOI":"10.1109\/ICIP.2017.8296962"},{"key":"2983_CR7","doi-asserted-by":"publisher","unstructured":"Chen, L., Ai, H., Zhuang, Z., Shang, C.: Real-time multiple people tracking with deeply learned candidate selection and person re-identification. In: IEEE International Conference on Multimedia and Expo (ICME). IEEE 2018, 1\u20136 (2018). https:\/\/doi.org\/10.1109\/ICME.2018.8486597","DOI":"10.1109\/ICME.2018.8486597"},{"key":"2983_CR8","doi-asserted-by":"publisher","unstructured":"Voigtlaender, P., Krause, M., Osep, A., Luiten, J., Sekar, B.B.G., Geiger, A., Leibe, B.: Mots: multi-object tracking and segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7942\u20137951 (2019). https:\/\/doi.org\/10.1109\/CVPR.2019.00813","DOI":"10.1109\/CVPR.2019.00813"},{"key":"2983_CR9","doi-asserted-by":"publisher","unstructured":"Wang, Z., Zheng, L., Liu, Y., Li, Y., Wang, S.: Towards real-time multi-object tracking. In: European Conference on Computer Vision. Springer, pp. 107\u2013122 (2020). https:\/\/doi.org\/10.1007\/978-3-030-58621-8_7","DOI":"10.1007\/978-3-030-58621-8_7"},{"issue":"11","key":"2983_CR10","doi-asserted-by":"publisher","first-page":"3069","DOI":"10.1007\/s11263-021-01513-4","volume":"129","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Wang, C., Wang, X., Zeng, W., Liu, W.: Fairmot: on the fairness of detection and re-identification in multiple object tracking. Int. J. Comput. Vis. 129(11), 3069\u20133087 (2021). https:\/\/doi.org\/10.1007\/s11263-021-01513-4","journal-title":"Int. J. Comput. Vis."},{"key":"2983_CR11","doi-asserted-by":"publisher","first-page":"1089","DOI":"10.1007\/s00371-020-01854-0","volume":"37","author":"X Zhang","year":"2021","unstructured":"Zhang, X., Wang, X., Gu, C.: Online multi-object tracking with pedestrian re-identification and occlusion processing. Vis. Comput. 37, 1089\u20131099 (2021). https:\/\/doi.org\/10.1007\/s00371-020-01854-0","journal-title":"Vis. Comput."},{"key":"2983_CR12","doi-asserted-by":"publisher","unstructured":"Patel, A.S., Vyas, R., Vyas, O., Ojha, M., Tiwari, V.: Motion-compensated online object tracking for activity detection and crowd behavior analysis. Vis. Comput. 1\u201321 (2022).https:\/\/doi.org\/10.1007\/s00371-022-02469-3","DOI":"10.1007\/s00371-022-02469-3"},{"key":"2983_CR13","doi-asserted-by":"publisher","unstructured":"Pang, J., Qiu, L., Li, X., Chen, H., Li, Q., Darrell, T., Yu, F.: Quasi-dense similarity learning for multiple object tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 164\u2013173 (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.00023","DOI":"10.1109\/CVPR46437.2021.00023"},{"issue":"3","key":"2983_CR14","doi-asserted-by":"publisher","first-page":"1051","DOI":"10.1007\/s00371-021-02067-9","volume":"38","author":"G Chen","year":"2022","unstructured":"Chen, G., Qin, H.: Class-discriminative focal loss for extreme imbalanced multiclass object detection towards autonomous driving. Vis. Comput. 38(3), 1051\u20131063 (2022). https:\/\/doi.org\/10.1007\/s00371-021-02067-9","journal-title":"Vis. Comput."},{"key":"2983_CR15","unstructured":"https:\/\/github.com\/hjq3659\/Jari-Maritime-Tracking-Dataset.git"},{"issue":"1\u20132","key":"2983_CR16","doi-asserted-by":"publisher","first-page":"83","DOI":"10.1002\/nav.3800020109","volume":"2","author":"HW Kuhn","year":"1955","unstructured":"Kuhn, H.W.: The Hungarian method for the assignment problem. Nav. Res. Logist. Q. 2(1\u20132), 83\u201397 (1955). https:\/\/doi.org\/10.1002\/nav.3800020109","journal-title":"Nav. Res. Logist. Q."},{"key":"2983_CR17","doi-asserted-by":"publisher","unstructured":"Redmon, J., Farhadi, A.: Yolov3: an incremental improvement (2018). https:\/\/doi.org\/10.48550\/arXiv.1804.02767. arXiv preprint arXiv:1804.02767","DOI":"10.48550\/arXiv.1804.02767"},{"key":"2983_CR18","doi-asserted-by":"publisher","unstructured":"Bochkovskiy, A., Wang, C.-Y., Liao, H.-Y.M.: Yolov4: optimal speed and accuracy of object detection (2020). https:\/\/doi.org\/10.48550\/arXiv.2004.10934. arXiv preprint arXiv:2004.10934","DOI":"10.48550\/arXiv.2004.10934"},{"key":"2983_CR19","doi-asserted-by":"publisher","unstructured":"Ge, Z., Liu, S., Wang, F., Li, Z., Sun, J.: Yolox: exceeding yolo series in 2021 (2021). https:\/\/doi.org\/10.48550\/arXiv.2107.08430. arXiv preprint arXiv:2107.08430","DOI":"10.48550\/arXiv.2107.08430"},{"key":"2983_CR20","doi-asserted-by":"publisher","unstructured":"Wang, C.-Y., Bochkovskiy, A., Liao, H.-Y.M.: Yolov7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors (2022). https:\/\/doi.org\/10.48550\/arXiv.2207.02696. arXiv preprint arXiv:2207.02696","DOI":"10.48550\/arXiv.2207.02696"},{"issue":"6","key":"2983_CR21","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2016","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137\u20131149 (2016). https:\/\/doi.org\/10.1109\/TPAMI.2016.2577031","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2983_CR22","doi-asserted-by":"publisher","unstructured":"Guo, Z., Shuai, H., Liu, G., Zhu, Y., Wang, W.: Multi-level feature fusion pyramid network for object detection. Vis. Comput. 1\u201311 (2022).https:\/\/doi.org\/10.1007\/s00371-022-02589-w","DOI":"10.1007\/s00371-022-02589-w"},{"issue":"6","key":"2983_CR23","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1109\/MCS.2009.934469","volume":"29","author":"Y Bar-Shalom","year":"2009","unstructured":"Bar-Shalom, Y., Daum, F., Huang, J.: The probabilistic data association filter. IEEE Control Syst. Mag. 29(6), 82\u2013100 (2009). https:\/\/doi.org\/10.1109\/MCS.2009.934469","journal-title":"IEEE Control Syst. Mag."},{"key":"2983_CR24","unstructured":"Lehmann, E.L., Casella, G.: Theory of Point Estimation. Springer, Berlin (2006)"},{"key":"2983_CR25","doi-asserted-by":"publisher","unstructured":"Rabiner, L., Juang, B.: An introduction to hidden Markov models. IEEE ASSP Mag. 3(1) 4\u201316 (1986). https:\/\/doi.org\/10.1109\/MASSP.1986.1165342","DOI":"10.1109\/MASSP.1986.1165342"},{"key":"2983_CR26","doi-asserted-by":"publisher","first-page":"3182","DOI":"10.1109\/TIP.2022.3165376","volume":"31","author":"C Liang","year":"2022","unstructured":"Liang, C., Zhang, Z., Zhou, X., Li, B., Zhu, S., Hu, W.: Rethinking the competition between detection and reid in multiobject tracking. IEEE Trans. Image Process. 31, 3182\u20133196 (2022). https:\/\/doi.org\/10.1109\/TIP.2022.3165376","journal-title":"IEEE Trans. Image Process."},{"key":"2983_CR27","doi-asserted-by":"publisher","unstructured":"Liang, C., Zhang, Z., Zhou, X., Li, B., Hu, W.: One more check: making \u201cfake background\u201d be tracked again. In: Proceedings of the AAAI Conference on Artificial Intelligence, Vol.\u00a036, pp. 1546\u20131554 (2022). https:\/\/doi.org\/10.1609\/aaai.v36i2.20045","DOI":"10.1609\/aaai.v36i2.20045"},{"issue":"8","key":"2983_CR28","doi-asserted-by":"publisher","first-page":"081101","DOI":"10.3788\/COL202220.081101","volume":"20","author":"Z Liang","year":"2022","unstructured":"Liang, Z., Wang, J., Xiao, G., Zeng, L.: Faanet: feature-aligned attention network for real-time multiple object tracking in UAV videos. Chin. Opt. Lett. 20(8), 081101 (2022). https:\/\/doi.org\/10.3788\/COL202220.081101","journal-title":"Chin. Opt. Lett."},{"key":"2983_CR29","unstructured":"Anonymous, Designing network design strategies (2022)"},{"key":"2983_CR30","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2008\/246309","volume":"2008","author":"K Bernardin","year":"2008","unstructured":"Bernardin, K., Stiefelhagen, R.: Evaluating multiple object tracking performance: the clear mot metrics. EURASIP J. Image Video Process. 2008, 1\u201310 (2008). https:\/\/doi.org\/10.1155\/2008\/246309","journal-title":"EURASIP J. Image Video Process."},{"key":"2983_CR31","doi-asserted-by":"publisher","unstructured":"Ristani, E., Solera, F., Zou, R., Cucchiara, R., Tomasi, C.: Performance measures and a data set for multi-target, multi-camera tracking. In: European Conference on Computer Vision. Springer, pp. 17\u201335 (2016). https:\/\/doi.org\/10.1007\/978-3-319-48881-3_2","DOI":"10.1007\/978-3-319-48881-3_2"},{"key":"2983_CR32","doi-asserted-by":"publisher","unstructured":"Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft coco: common objects in context. In: European Conference on Computer Vision. Springer, pp. 740\u2013755 (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"2983_CR33","doi-asserted-by":"publisher","unstructured":"Micikevicius, P., Narang, S., Alben, J., Diamos, G., Elsen, E., Garcia, D., Ginsburg, B., Houston, M., Kuchaiev, O., Venkatesh, G., et\u00a0al.: Mixed precision training, (2017). https:\/\/doi.org\/10.48550\/arXiv.1710.03740. arXiv preprint arXiv:1710.03740","DOI":"10.48550\/arXiv.1710.03740"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02983-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-02983-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02983-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,4,5]],"date-time":"2024-04-05T17:08:23Z","timestamp":1712336903000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-02983-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,9]]},"references-count":33,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["2983"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-02983-y","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,7,9]]},"assertion":[{"value":"31 March 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 July 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no potential conflict of interests. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}