{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,5]],"date-time":"2025-06-05T04:54:04Z","timestamp":1749099244920,"version":"3.40.3"},"publisher-location":"Cham","reference-count":21,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031594618"},{"type":"electronic","value":"9783031594625"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-59462-5_3","type":"book-chapter","created":{"date-parts":[[2024,5,2]],"date-time":"2024-05-02T07:02:12Z","timestamp":1714633332000},"page":"36-51","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["An Approach for\u00a0Object Recognition in\u00a0Videos for\u00a0Vocabulary Extraction"],"prefix":"10.1007","author":[{"given":"Anh Bao Nguyen","family":"Le","sequence":"first","affiliation":[]},{"given":"Chi Bao","family":"Nguyen","sequence":"additional","affiliation":[]},{"given":"Quoc Cuong","family":"Dang","sequence":"additional","affiliation":[]},{"given":"Be Hai","family":"Danh","sequence":"additional","affiliation":[]},{"given":"Huynh Nhu","family":"Le","sequence":"additional","affiliation":[]},{"given":"Huong Hoang","family":"Luong","sequence":"additional","affiliation":[]},{"given":"Hai Thanh","family":"Nguyen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,3]]},"reference":[{"key":"3_CR1","unstructured":"Liu, H., Aderon, C., Wagon, N., Liu, H., MacCall, S., Gan, Y.: Deep learning-based automatic player identification and logging in American football videos. arXiv preprint arXiv:2204.13809 (2022)"},{"key":"3_CR2","doi-asserted-by":"crossref","unstructured":"Zou, S., et al.: TOD-CNN: an effective convolutional neural network for tiny object detection in sperm videos. arXiv preprint arXiv:2204.08166 (2022)","DOI":"10.1016\/j.compbiomed.2022.105543"},{"key":"3_CR3","unstructured":"Zhao, W., et al.: A survey of semen quality evaluation in microscopic videos using computer assisted sperm analysis. arXiv preprint arXiv:2202.07820 (2022)"},{"key":"3_CR4","unstructured":"Gu, Y., Liao, X., Qin, X.: YouTube-GDD: a challenging gun detection dataset with rich contextual information. arXiv preprint arXiv:2203.04129 (2022)"},{"key":"3_CR5","doi-asserted-by":"publisher","unstructured":"Yin, Q., et al.: Detecting and tracking small and dense moving objects in satellite videos: a benchmark. IEEE Trans. Geosci. Remote Sens. 60, 1\u201318 (2022). https:\/\/doi.org\/10.1109\/TGRS.2021.3130436","DOI":"10.1109\/TGRS.2021.3130436"},{"key":"3_CR6","doi-asserted-by":"crossref","unstructured":"Zhu, X., Dai, J., Yuan, L., Wei, Y.: Towards high performance video object detection. arXiv preprint arXiv:1711.11577 (2017)","DOI":"10.1109\/CVPR.2018.00753"},{"key":"3_CR7","unstructured":"Tang, P., Wang, C., Wang, X., Liu, W., Zeng, W., Wang, J.: Object detection in videos by high quality object linking. arXiv preprint arXiv:1801.09823 (2018)"},{"key":"3_CR8","doi-asserted-by":"publisher","unstructured":"He, F., Gao, N., Jia, J., Zhao, X., Huang, K.: QueryProp: object query propagation for high-performance video object detection. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, no. 1, pp. 834\u2013842 (2022). https:\/\/doi.org\/10.1609\/aaai.v36i1.19965","DOI":"10.1609\/aaai.v36i1.19965"},{"key":"3_CR9","doi-asserted-by":"crossref","unstructured":"Han, M., Wang, Y., Chang, X., Qiao, Y.: Mining inter-video proposal relations for video object detection (2020). https:\/\/www.ecva.net\/papers\/eccv_2020\/papers_ECCV\/papers\/123660426.pdf","DOI":"10.1007\/978-3-030-58589-1_26"},{"key":"3_CR10","unstructured":"Kolarova, S.T.V., et al.: Autonomous driving (2016). https:\/\/www.ifmo.de\/files\/publications_content\/2016\/ifmo_2016_Autonomous_Driving_2035_en.pdf"},{"key":"3_CR11","unstructured":"Advantech\u00a0Co., Ltd.: The future of intelligent surveillance (2012). https:\/\/advcloudfiles.advantech.com\/ecatalog\/MyAdvantech\/MyAdvantech_No_11_eng.pdf"},{"key":"3_CR12","unstructured":"Han, H., et al.: Real-time robust video object detection system against physical-world adversarial attacks. arXiv preprint arXiv:2208.09195 (2022)"},{"key":"3_CR13","doi-asserted-by":"crossref","unstructured":"Schofield, D., et al.: Chimpanzee face recognition from videos in the wild using deep learning. Sci. Adv. 5(9), eaaw0736 (2019). https:\/\/www.science.org\/doi\/abs\/10.1126\/sciadv.aaw0736","DOI":"10.1126\/sciadv.aaw0736"},{"key":"3_CR14","doi-asserted-by":"crossref","unstructured":"Ardianto, S., Hang, H.M., Cheng, W.H.: Fast vehicle detection and tracking on fisheye traffic monitoring video using CNN and bounding box propagation. arXiv preprint arXiv:2207.01183 (2022), to be published in International Conference on Image Processing (ICIP) 2022, Bordeaux, France","DOI":"10.1109\/ICIP46576.2022.9897160"},{"key":"3_CR15","doi-asserted-by":"publisher","unstructured":"Raskar, P.S., Shah, S.K.: Real time object-based video forgery detection using YOLO (V2) (2021). https:\/\/doi.org\/10.1016\/j.forsciint.2021.110979","DOI":"10.1016\/j.forsciint.2021.110979"},{"key":"3_CR16","doi-asserted-by":"publisher","unstructured":"Jiang, C., et al.: Object detection from UAV thermal infrared images and videos using YOLO models (2022). https:\/\/doi.org\/10.1016\/j.jag.2022.102912","DOI":"10.1016\/j.jag.2022.102912"},{"key":"3_CR17","unstructured":"Torresani, G.B.L., Shi, J.: Object detection in video with spatiotemporal sampling networks (2018). https:\/\/openaccess.thecvf.com\/content_ECCV_2018\/papers\/Gedas_Bertasius_Object_Detection_in_ECCV_2018_paper.pdf"},{"key":"3_CR18","doi-asserted-by":"crossref","unstructured":"Deng, H., et al.: Object guided external memory network for video object detection (2019). https:\/\/ieeexplore.ieee.org\/document\/9011008","DOI":"10.1109\/ICCV.2019.00678"},{"key":"3_CR19","doi-asserted-by":"crossref","unstructured":"Oh, S.W., University, Y., Lee, J.Y., Research, A., Xu, N., Research, A., Kim, S.J., University, Y.: Video object segmentation using space-time memory networks (2019). https:\/\/openaccess.thecvf.com\/content_ICCV_2019\/papers\/Oh_Video_Object_Segmentation_Using_Space-Time_Memory_Networks_ICCV_2019_paper.pdf","DOI":"10.1109\/ICCV.2019.00932"},{"key":"3_CR20","doi-asserted-by":"crossref","unstructured":"Fan, Q., Tang, C.K., Tai, Y.W.: Few-shot video object detection (2021). https:\/\/www.researchgate.net\/publication\/351278547_Few-Shot_Video_Object_Detection#pf9","DOI":"10.1007\/978-3-031-20044-1_5"},{"key":"3_CR21","unstructured":"Ultralytics: Ultralytics yolov5. https:\/\/github.com\/ultralytics\/yolov5. Accessed 27 Sep 2023"}],"container-title":["Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering","Nature of Computation and Communication"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-59462-5_3","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,2]],"date-time":"2024-05-02T07:04:08Z","timestamp":1714633448000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-59462-5_3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031594618","9783031594625"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-59462-5_3","relation":{},"ISSN":["1867-8211","1867-822X"],"issn-type":[{"type":"print","value":"1867-8211"},{"type":"electronic","value":"1867-822X"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"3 May 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICTCC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Nature of Computation and Communication","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Ho Chi Minh City","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vietnam","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ictcc2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Confy +","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"30","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"12","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"40% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}