{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T13:47:14Z","timestamp":1765547234006,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":21,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819703753"},{"type":"electronic","value":"9789819703760"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-0376-0_6","type":"book-chapter","created":{"date-parts":[[2024,2,11]],"date-time":"2024-02-11T09:02:09Z","timestamp":1707642129000},"page":"69-81","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Computational Analysis of Table Tennis Matches from Real-Time Videos Using Deep Learning"],"prefix":"10.1007","author":[{"given":"Hong","family":"Zhou","sequence":"first","affiliation":[]},{"given":"Minh","family":"Nguyen","sequence":"additional","affiliation":[]},{"given":"Wei Qi","family":"Yan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,12]]},"reference":[{"key":"6_CR1","doi-asserted-by":"crossref","unstructured":"Xiao, B., Nguyen, M., Yan, W.Q.: Fruit ripeness identification using YOLOv8 model. Multimedia Tools Appl. 1\u201318 (2023)","DOI":"10.1007\/s11042-023-16570-9"},{"key":"6_CR2","doi-asserted-by":"publisher","unstructured":"Yan, W.: Computational methods for Deep Learning: Theory, Algorithms, and Implementations. Springer, Singapore (2023). https:\/\/doi.org\/10.1007\/978-981-99-4823-9","DOI":"10.1007\/978-981-99-4823-9"},{"key":"6_CR3","doi-asserted-by":"crossref","unstructured":"Tian, B., Zhang, D., Zhang, C.: High-speed tiny tennis ball detection based on deep convolutional neural networks. In International Conference on Anti-counterfeiting, Security, and Identification (ASID), pp. 30\u201333 (2020)","DOI":"10.1109\/ASID50160.2020.9271695"},{"key":"6_CR4","doi-asserted-by":"crossref","unstructured":"Cai, G.L.: A method for prediction the trajectory of table tennis in multirotation state based on binocular vision. Comput. Intell. Neurosci. (2022)","DOI":"10.1155\/2022\/8274202"},{"key":"6_CR5","doi-asserted-by":"crossref","unstructured":"Huang, Y., Liao, I., Chen, C., \u0130k, T., Peng, W.: TrackNet: a deep learning network for tracking high-speed and tiny objects in sports applications. In: IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pp. 1\u20138 (2019)","DOI":"10.1109\/AVSS.2019.8909871"},{"key":"6_CR6","doi-asserted-by":"crossref","unstructured":"Moshayedi, A.J., Chen, Z., Liao, L., Li, S.: Kinect based virtual referee for table tennis game: TTV (Table Tennis Var System). In: International Conference on Information Science and Control Engineering (ICISCE), pp. 354\u2013359 (2019)","DOI":"10.1109\/ICISCE48695.2019.00077"},{"key":"6_CR7","doi-asserted-by":"crossref","unstructured":"Shi, J., Xu, L., Jia, J.: Discriminative blur detection features. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2965\u20132972 (2014)","DOI":"10.1109\/CVPR.2014.379"},{"key":"6_CR8","doi-asserted-by":"publisher","first-page":"208554","DOI":"10.1109\/ACCESS.2020.3038913","volume":"8","author":"Y Wu","year":"2020","unstructured":"Wu, Y., Zhang, H., Li, Y., Yang, Y., Yuan, D.: Video object detection guided by object blur evaluation. IEEE Access 8, 208554\u2013208565 (2020)","journal-title":"IEEE Access"},{"key":"6_CR9","doi-asserted-by":"crossref","unstructured":"Zheng, W., Liu, X., Yin, L.: Research on image classification method based on improved multi-scale relational network. J. Comput. Sci. 7, 613 (2012)","DOI":"10.7717\/peerj-cs.613"},{"key":"6_CR10","doi-asserted-by":"publisher","unstructured":"Yan, W.: Introduction to Intelligent Surveillance: Surveillance Data Capture, Transmission, and Analytics. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-10713-0","DOI":"10.1007\/978-3-030-10713-0"},{"key":"6_CR11","doi-asserted-by":"crossref","unstructured":"Xiao, B., Nguyen, M., Yan, W.Q.: Apple ripeness identification from digital images using transformers. Multimedia Tools Appl. 1\u201315 (2023)","DOI":"10.1007\/s11042-023-15938-1"},{"key":"6_CR12","doi-asserted-by":"publisher","first-page":"34","DOI":"10.1016\/j.robot.2017.12.004","volume":"101","author":"Y Liu","year":"2018","unstructured":"Liu, Y., Liu, L.: Accurate real-time ball trajectory estimation with onboard stereo camera system for humanoid ping-pong robot. Robot. Auton. Syst. 101, 34\u201344 (2018)","journal-title":"Robot. Auton. Syst."},{"key":"6_CR13","doi-asserted-by":"crossref","unstructured":"Mehtab, S., Yan, W.: Flexible neural network for fast and accurate road scene perception. Multimedia Tools Appl. (2022)","DOI":"10.1007\/s11042-022-11933-0"},{"key":"6_CR14","doi-asserted-by":"crossref","unstructured":"Mehtab, S., Yan, W.: FlexiNet: fast and accurate vehicle detection for autonomous vehicles-2D vehicle detection using deep neural network. In: International Conference on Control and Computer Vision (2021)","DOI":"10.1145\/3484274.3484282"},{"key":"6_CR15","doi-asserted-by":"crossref","unstructured":"Xiang, Y., Yan, W.: Fast-moving coin recognition using deep learning. Multimedia Tools Appl. (2021)","DOI":"10.1007\/s11042-021-10857-5"},{"key":"6_CR16","doi-asserted-by":"crossref","unstructured":"Qi, J., Nguyen, M., Yan, W.: Small visual object detection in smart waste classification using transformers with deep learning. In: IVCNZ (2022)","DOI":"10.1007\/978-3-031-25825-1_22"},{"key":"6_CR17","doi-asserted-by":"publisher","unstructured":"Liu, J., Pan, C., Yan, W.: Litter Detection From Digital Images Using Deep Learning. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/s42979-022-01568-1","DOI":"10.1007\/s42979-022-01568-1"},{"issue":"27\u201328","key":"6_CR18","doi-asserted-by":"publisher","first-page":"19925","DOI":"10.1007\/s11042-020-08866-x","volume":"79","author":"C Pan","year":"2020","unstructured":"Pan, C., Yan, W.: Object detection based on saturation of visual perception. Multimedia Tools Appl. 79(27\u201328), 19925\u201319944 (2020)","journal-title":"Multimedia Tools Appl."},{"key":"6_CR19","doi-asserted-by":"crossref","unstructured":"Pan, C., Yan, W.: A learning-based positive feedback in salient object detection. In: IVCNZ (2018)","DOI":"10.1109\/IVCNZ.2018.8634717"},{"key":"6_CR20","doi-asserted-by":"crossref","unstructured":"Pan, C., Liu, J., Yan, W., Zhou, Y.: Salient object detection based on visual perceptual saturation and two-stream hybrid networks. IEEE Trans. Image Process. (2021)","DOI":"10.1109\/TIP.2021.3074796"},{"key":"6_CR21","doi-asserted-by":"crossref","unstructured":"Qi, J., Nguyen, M., Yan, W.: CISO: co-iteration semi-supervised learning for visual object detection. Multimedia Tools Appl. (2023)","DOI":"10.1007\/s11042-023-16915-4"}],"container-title":["Lecture Notes in Computer Science","Image and Video Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-0376-0_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,11]],"date-time":"2024-02-11T09:03:29Z","timestamp":1707642209000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-0376-0_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819703753","9789819703760"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-0376-0_6","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"12 February 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PSIVT","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Pacific-Rim Symposium on Image and Video Technology","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Auckland","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Zealand","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 November 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 November 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"psivt2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/psivt2023.aut.ac.nz\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easy Chair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"75","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"34","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}