{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T11:40:47Z","timestamp":1742989247273,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":27,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819985548"},{"type":"electronic","value":"9789819985555"}],"license":[{"start":{"date-parts":[[2023,12,28]],"date-time":"2023-12-28T00:00:00Z","timestamp":1703721600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,28]],"date-time":"2023-12-28T00:00:00Z","timestamp":1703721600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8555-5_20","type":"book-chapter","created":{"date-parts":[[2023,12,27]],"date-time":"2023-12-27T07:02:36Z","timestamp":1703660556000},"page":"251-264","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["FAFVTC: A Real-Time Network for\u00a0Vehicle Tracking and\u00a0Counting"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7701-8909","authenticated-orcid":false,"given":"Zhiwen","family":"Wang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7682-8478","authenticated-orcid":false,"given":"Kai","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4678-1936","authenticated-orcid":false,"given":"Fei","family":"Gao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,28]]},"reference":[{"key":"20_CR1","doi-asserted-by":"crossref","unstructured":"Henriques, J.F., et al.: High-speed tracking with kernelized correlation filters. TPAMI 37(3), 583\u2013596 (2014)","DOI":"10.1109\/TPAMI.2014.2345390"},{"key":"20_CR2","doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: Fairmot: on the fairness of detection and re-identification in multiple object tracking. IJCV 129(11), 3069\u20133087 (2021)","DOI":"10.1007\/s11263-021-01513-4"},{"key":"20_CR3","doi-asserted-by":"crossref","unstructured":"Li, S., et al.: FAFMOTS: a fast and anchor free method for online joint multi-object tracking and segmentation. In: ISMARW, pp. 465\u2013470 (2022)","DOI":"10.1109\/ISMAR-Adjunct57072.2022.00098"},{"key":"20_CR4","doi-asserted-by":"crossref","unstructured":"Bewley, A., et al.: Simple online and realtime tracking. In: ICIP, pp. 3464\u20133468 (2016)","DOI":"10.1109\/ICIP.2016.7533003"},{"key":"20_CR5","doi-asserted-by":"crossref","unstructured":"Wojke, N., et al.: Simple online and realtime tracking with a deep association metric. In: ICIP, pp. 3645\u20133649 (2017)","DOI":"10.1109\/ICIP.2017.8296962"},{"key":"20_CR6","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/978-3-031-20047-2_1","volume-title":"Computer Vision - ECCV 2022","author":"Y Zhang","year":"2022","unstructured":"Zhang, Y., et al.: Bytetrack: multi-object tracking by associating every detection box. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13682, pp. 1\u201321. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20047-2_1"},{"key":"20_CR7","doi-asserted-by":"crossref","unstructured":"Guo, S., et al.: Online multiple object tracking with cross-task synergy. In: CVPR, pp. 8136\u20138145 (2021)","DOI":"10.1109\/CVPR46437.2021.00804"},{"key":"20_CR8","doi-asserted-by":"crossref","unstructured":"Li, W., et al.: Simultaneous multi-person tracking and activity recognition based on cohesive cluster search. CVIU 214, 103301, 1\u201313 (2022)","DOI":"10.1016\/j.cviu.2021.103301"},{"key":"20_CR9","doi-asserted-by":"crossref","unstructured":"Wang, H., et al.: STURE: spatial-temporal mutual representation learning for robust data association in online multi-object tracking. CVIU 220, 1\u201310 (2022)","DOI":"10.1016\/j.cviu.2022.103433"},{"key":"20_CR10","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"474","DOI":"10.1007\/978-3-030-58548-8_28","volume-title":"Computer Vision \u2013 ECCV 2020","author":"X Zhou","year":"2020","unstructured":"Zhou, X., Koltun, V., Kr\u00e4henb\u00fchl, P.: Tracking objects as points. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12349, pp. 474\u2013490. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58548-8_28"},{"key":"20_CR11","doi-asserted-by":"crossref","unstructured":"Wu, J., et al.: Track to detect and segment: an online multi-object tracker. In: CVPR, pp. 12352\u201312361 (2021)","DOI":"10.1109\/CVPR46437.2021.01217"},{"key":"20_CR12","doi-asserted-by":"crossref","unstructured":"Chu, P., et al.: Famnet: joint learning of feature, affinity and multi-dimensional assignment for online multiple object tracking. In: ICCV, pp. 6172\u20136181 (2019)","DOI":"10.1109\/ICCV.2019.00627"},{"key":"20_CR13","doi-asserted-by":"crossref","unstructured":"Wang, G., et al.: Track without appearance: Learn box and tracklet embedding with local and global motion patterns for vehicle tracking. In: ICCV, pp. 9876\u20139886 (2021)","DOI":"10.1109\/ICCV48922.2021.00973"},{"key":"20_CR14","doi-asserted-by":"crossref","unstructured":"Hu, J., et al.: Squeeze-and-excitation networks. In: CVPR, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"20_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-01234-2_1","volume-title":"Computer Vision \u2013 ECCV 2018","author":"S Woo","year":"2018","unstructured":"Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3\u201319. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01234-2_1"},{"key":"20_CR16","doi-asserted-by":"crossref","unstructured":"Ruan, D., et al.: Linear context transform block. In: AAAI, vol. 34, no. 4, pp. 5553\u20135560 (2020)","DOI":"10.1609\/aaai.v34i04.6007"},{"key":"20_CR17","doi-asserted-by":"crossref","unstructured":"Ruan, D., et al.: Gaussian context transformer. In: CVPR, pp. 15129\u201315138 (2021)","DOI":"10.1109\/CVPR46437.2021.01488"},{"key":"20_CR18","doi-asserted-by":"crossref","unstructured":"Qin, Z., et al.: Fcanet: frequency channel attention networks. In: ICCV, pp. 783\u2013792 (2021)","DOI":"10.1109\/ICCV48922.2021.00082"},{"key":"20_CR19","doi-asserted-by":"crossref","unstructured":"Su, B., et al.: CFCAnet: a complete frequency channel attention network for SAR image scene classification. In: IEEE J-STARS, vol. 14, pp. 11750\u201311763 (2021)","DOI":"10.1109\/JSTARS.2021.3125107"},{"key":"20_CR20","doi-asserted-by":"crossref","unstructured":"Guo, M.H., et al.: Beyond self-attention: external attention using two linear layers for visual tasks. TPAMI 45(5), 5436\u20135447 (2022)","DOI":"10.1109\/TPAMI.2022.3211006"},{"key":"20_CR21","doi-asserted-by":"crossref","unstructured":"Amato, G., et al.: Counting vehicles with deep learning in onboard UAV imagery. In: ISCC, pp. 1\u20136 (2019)","DOI":"10.1109\/ISCC47284.2019.8969620"},{"key":"20_CR22","doi-asserted-by":"crossref","unstructured":"Zhang, B., et al.: A traffic surveillance system for obtaining comprehensive information of the passing vehicles based on instance segmentation. TITS 22(11), 7040\u20137055 (2021)","DOI":"10.1109\/TITS.2020.3001154"},{"key":"20_CR23","doi-asserted-by":"crossref","unstructured":"Gomaa, A., et al.: Faster CNN-based vehicle detection and counting strategy for fixed camera scenes. MTA 81(18), 25443\u201325471 (2022)","DOI":"10.1007\/s11042-022-12370-9"},{"key":"20_CR24","doi-asserted-by":"crossref","unstructured":"Ciampi, L., et al.: Multi-camera vehicle counting using edge-AI. ESWA 207, 117929, 1\u20139 (2022)","DOI":"10.1016\/j.eswa.2022.117929"},{"key":"20_CR25","doi-asserted-by":"crossref","unstructured":"Xu, H., et al.: Efficient CityCam-to-edge cooperative learning for vehicle counting in ITS. TITS 23(9), 16600\u201316611 (2022)","DOI":"10.1109\/TITS.2022.3149657"},{"key":"20_CR26","doi-asserted-by":"crossref","unstructured":"Yu, F., et al.: Deep layer aggregation. In: CVPR, pp. 2403\u20132412 (2018)","DOI":"10.1109\/CVPR.2018.00255"},{"key":"20_CR27","doi-asserted-by":"crossref","unstructured":"Wen, L., et al.: UA-DETRAC: a new benchmark and protocol for multi-object detection and tracking. CVIU 193, 102907, 1\u20139 (2020)","DOI":"10.1016\/j.cviu.2020.102907"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8555-5_20","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,27]],"date-time":"2023-12-27T07:06:55Z","timestamp":1703660815000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8555-5_20"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,28]]},"ISBN":["9789819985548","9789819985555"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8555-5_20","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023,12,28]]},"assertion":[{"value":"28 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Xiamen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/prcv2023.xmu.edu.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1420","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"532","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"37% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,78","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,69","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}