{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,9]],"date-time":"2025-09-09T22:20:36Z","timestamp":1757456436985,"version":"3.40.5"},"reference-count":24,"publisher":"IGI Global","issue":"2","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,7]]},"abstract":"<jats:p>Due to the increase in video content being generated from surveillance cameras and filming, videos analysis becomes imperative. Sometimes it becomes tedious to watch a video captured by a surveillance camera for hours, just to find out the desired footage. Current state of-the-art video analysis methods do not address the problem of searching and localizing a particular object in a video using the name of the object as a query and to return only a segment of the video clip showing the instances of that object. In this research the authors make use of combined implementations from existing work and also applied the dropping frames algorithm to produce a shorter, trimmed video clip showing the target object specified by the search tag. The resulting video is short and specific to the object of interest.<\/jats:p>","DOI":"10.4018\/ijictrame.2019070102","type":"journal-article","created":{"date-parts":[[2019,6,11]],"date-time":"2019-06-11T15:46:51Z","timestamp":1560268011000},"page":"18-31","source":"Crossref","is-referenced-by-count":5,"title":["Searching Objects in a Video Footage"],"prefix":"10.4018","volume":"8","author":[{"given":"Tapiwanashe Miranda","family":"Sanyanga","sequence":"first","affiliation":[{"name":"Chinhoyi University of Technology, Chinhoyi, Zimbabwe"}]},{"given":"Munyaradzi Sydney","family":"Chinzvende","sequence":"additional","affiliation":[{"name":"Chinhoyi University of Technology, Chinhoyi, Zimbabwe"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9479-1143","authenticated-orcid":true,"given":"Tatenda Duncan","family":"Kavu","sequence":"additional","affiliation":[{"name":"Chinhoyi University of Technology, Chinhoyi, Zimbabwe"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2374-6337","authenticated-orcid":true,"given":"John","family":"Batani","sequence":"additional","affiliation":[{"name":"Chinhoyi University of Technology, Chinhoyi, Zimbabwe"}]}],"member":"2432","reference":[{"key":"IJICTRAME.2019070102-0","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2015.2473295"},{"key":"IJICTRAME.2019070102-1","unstructured":"Elkan, C. (2014). Maximum Likelihood, Logistic Regression, and Stochastic Gradient Training. Retrieved from https:\/\/pdfs.semanticscholar.org\/a42a\/76c854c8e3bd72447c1d3fb0b12d391ea342.pdf"},{"key":"IJICTRAME.2019070102-2","doi-asserted-by":"crossref","unstructured":"Erhan, D., Szegedy, C., Toshev, A., & Anguelov, D. (2014). Scalable Object Detection using Deep Neural Networks. Retrieved from https:\/\/www.cv-foundation.org\/openaccess\/content_cvpr_2014\/html\/Erhan_Scalable_Object_Detection_2014_CVPR_paper.html","DOI":"10.1109\/CVPR.2014.276"},{"key":"IJICTRAME.2019070102-3","unstructured":"Fathi, A., Korattikara, A., Sun, C., Fischer, I., Huang, J., & Murphy, K. \u2026 Wojna, Z. (2017). Speed and accuracy trade-offs for modern convolutional object detectors. Retrieved from https:\/\/research.google.com\/pubs\/pub46259.html"},{"key":"IJICTRAME.2019070102-4","unstructured":"Fernando, C., Banarse, D., Blundell, C., Zwols, Y., Ha, D., & Rusu, A. A. \u2026 Wierstra, D. (2017). PathNet: Evolution Channels Gradient Descent in Super Neural Networks. Retrieved from http:\/\/arxiv.org\/abs\/1701.08734"},{"key":"IJICTRAME.2019070102-5","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2017.7952132"},{"key":"IJICTRAME.2019070102-6","unstructured":"Howard, A. G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T. \u2026 Adam, H. (2017). MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. Retrieved from http:\/\/arxiv.org\/abs\/1704.04861"},{"key":"IJICTRAME.2019070102-7","doi-asserted-by":"publisher","DOI":"10.1016\/j.nima.2015.05.028"},{"key":"IJICTRAME.2019070102-8","unstructured":"Ioffe, S., & Szegedy, C. (2015, June 1). Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. Retrieved from http:\/\/proceedings.mlr.press\/v37\/ioffe15.html"},{"key":"IJICTRAME.2019070102-9","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2017.2736553"},{"key":"IJICTRAME.2019070102-10","first-page":"18","article-title":"An Effective Object Detection Video Surveillance and Alert System.","author":"M. V.Khadse","year":"2016","journal-title":"National Conference on Advances in Computing, Communication and Networking (ACCNet \u2013 2016)"},{"key":"IJICTRAME.2019070102-11","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"IJICTRAME.2019070102-12","doi-asserted-by":"crossref","unstructured":"Mansencal, B., Benois-pineau, J., Vieux, R., Domenger, J., Mansencal, B., Benois-pineau, J., \u2026 Benois-Pineau, J. (2012). Search of objects of interest in videos To cite this version\u202f: HAL Id\u202f: hal-00718360 Search of objects of interest in videos.","DOI":"10.1109\/CBMI.2012.6269809"},{"key":"IJICTRAME.2019070102-13","doi-asserted-by":"crossref","unstructured":"Mao, J., Huang, J., Toshev, A., Camburu, O., Yuille, A. L., & Murphy, K. (2016). Generation and Comprehension of Unambiguous Object Descriptions. Retrieved from https:\/\/www.cv-foundation.org\/openaccess\/content_cvpr_2016\/html\/Mao_Generation_and_Comprehension_CVPR_2016_paper.html","DOI":"10.1109\/CVPR.2016.9"},{"key":"IJICTRAME.2019070102-14","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2015.2500734"},{"key":"IJICTRAME.2019070102-15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2011.5995586"},{"key":"IJICTRAME.2019070102-16","unstructured":"Ott, J., Lin, Z., Zhang, Y., Liu, S.-C., & Bengio, Y. (2016). Recurrent Neural Networks With Limited Numerical Precision. Retrieved from http:\/\/arxiv.org\/abs\/1608.06902"},{"key":"IJICTRAME.2019070102-17","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2537320"},{"key":"IJICTRAME.2019070102-18","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2577031"},{"key":"IJICTRAME.2019070102-19","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2017.2705222"},{"key":"IJICTRAME.2019070102-20","unstructured":"Simonyan, K., & Zisserman, A. (2014). Very Deep Convolutional Networks for Large-Scale Image Recognition. Retrieved from http:\/\/arxiv.org\/abs\/1409.1556"},{"key":"IJICTRAME.2019070102-21","unstructured":"Szegedy, C., Ioffe, S., Vanhoucke, V., & Alemi, A. (2016). Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. Retrieved from http:\/\/arxiv.org\/abs\/1602.07261"},{"key":"IJICTRAME.2019070102-22","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-017-4440-4"},{"key":"IJICTRAME.2019070102-23","unstructured":"Yu, H., & Siskind, J. M. (2015). Sentence Directed Video Object Codetection. Retrieved from http:\/\/arxiv.org\/abs\/1506.02059"}],"container-title":["International Journal of ICT Research in Africa and the Middle East"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.igi-global.com\/viewtitle.aspx?TitleId=231626","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,6]],"date-time":"2022-05-06T00:23:27Z","timestamp":1651796607000},"score":1,"resource":{"primary":{"URL":"http:\/\/services.igi-global.com\/resolvedoi\/resolve.aspx?doi=10.4018\/IJICTRAME.2019070102"}},"subtitle":["Dropping Frames and Object Detection Approach"],"short-title":[],"issued":{"date-parts":[[2019,7]]},"references-count":24,"journal-issue":{"issue":"2"},"URL":"https:\/\/doi.org\/10.4018\/ijictrame.2019070102","relation":{},"ISSN":["2472-3126","2472-3134"],"issn-type":[{"type":"print","value":"2472-3126"},{"type":"electronic","value":"2472-3134"}],"subject":[],"published":{"date-parts":[[2019,7]]}}}