{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,3]],"date-time":"2026-03-03T16:28:32Z","timestamp":1772555312900,"version":"3.50.1"},"reference-count":55,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"DOI":"10.13039\/501100004761","name":"Natural Science Foundation of Hainan Province","doi-asserted-by":"publisher","award":["2019CXTD400"],"award-info":[{"award-number":["2019CXTD400"]}],"id":[{"id":"10.13039\/501100004761","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004761","name":"Natural Science Foundation of Hainan Province","doi-asserted-by":"publisher","award":["617079"],"award-info":[{"award-number":["617079"]}],"id":[{"id":"10.13039\/501100004761","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["SQ2020YFF0423852"],"award-info":[{"award-number":["SQ2020YFF0423852"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Hainan Key Research and Development Program","award":["620RC554"],"award-info":[{"award-number":["620RC554"]}]},{"name":"Hainan Key Research and Development Program","award":["ZDYF2019115"],"award-info":[{"award-number":["ZDYF2019115"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62175054"],"award-info":[{"award-number":["62175054"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61865005"],"award-info":[{"award-number":["61865005"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61762033"],"award-info":[{"award-number":["61762033"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["2020WNLOKF001"],"award-info":[{"award-number":["2020WNLOKF001"]}]},{"name":"National Key Technology Support Program","award":["2015BAH55F04"],"award-info":[{"award-number":["2015BAH55F04"]}]},{"name":"National Key Technology Support Program","award":["2015BAH55F01"],"award-info":[{"award-number":["2015BAH55F01"]}]},{"DOI":"10.13039\/501100013072","name":"Major Science and Technology Project of Hainan Province","doi-asserted-by":"publisher","award":["ZDKJ2016015"],"award-info":[{"award-number":["ZDKJ2016015"]}],"id":[{"id":"10.13039\/501100013072","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005693","name":"Scientific Research Staring Foundation of Hainan University","doi-asserted-by":"publisher","award":["KYQD(ZR)1882"],"award-info":[{"award-number":["KYQD(ZR)1882"]}],"id":[{"id":"10.13039\/501100005693","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2021]]},"DOI":"10.1109\/access.2021.3110798","type":"journal-article","created":{"date-parts":[[2021,9,7]],"date-time":"2021-09-07T19:31:01Z","timestamp":1631043061000},"page":"124847-124860","source":"Crossref","is-referenced-by-count":25,"title":["Generative Adversarial Networks for Abnormal Event Detection in Videos Based on Self-Attention Mechanism"],"prefix":"10.1109","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3232-7801","authenticated-orcid":false,"given":"Weichao","family":"Zhang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5458-9509","authenticated-orcid":false,"given":"Guanjun","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5709-703X","authenticated-orcid":false,"given":"Mengxing","family":"Huang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0224-3156","authenticated-orcid":false,"given":"Hongyu","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Shaoping","family":"Wen","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","first-page":"694","article-title":"Perceptual losses for real-time style transfer and super-resolution","author":"johnson","year":"2016","journal-title":"Proc ECCV"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00931"},{"key":"ref33","article-title":"The relativistic discriminator: A key element missing from standard GAN","author":"jolicoeur-martineau","year":"2018","journal-title":"arXiv 1807 00734"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2017.2654543"},{"key":"ref30","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref37","first-page":"169","article-title":"LiteFlowNet3: Resolving correspondence ambiguity for more accurate optical flow estimation","author":"hui","year":"2020","journal-title":"Proc ECCV"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00936"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.316"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00813"},{"key":"ref28","article-title":"Attention is all you need","author":"vaswani","year":"2017","journal-title":"arXiv 1706 03762"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D16-1053"},{"key":"ref29","first-page":"7354","article-title":"Self-attention generative adversarial networks","author":"zhang","year":"2019","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00678"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00684"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2017.8296547"},{"key":"ref22","first-page":"1","article-title":"ESRGAN: Enhanced super-resolution generative adversarial networks","author":"wang","year":"2018","journal-title":"Proc ECCV"},{"key":"ref21","first-page":"234","article-title":"U-Net: Convolutional networks for biomedical image segmentation","author":"ronneberger","year":"2015","journal-title":"Medical Image Computing and Computer-Assisted Intervention"},{"key":"ref24","first-page":"1","article-title":"Image super-resolution as sparse representation of raw image","author":"yang","year":"2008","journal-title":"Proc CVPR"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1175\/1520-0450(1979)018<1016:LFIOAT>2.0.CO;2"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.19"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2015.2439281"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00179"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2979869"},{"key":"ref55","doi-asserted-by":"crossref","first-page":"938","DOI":"10.1609\/aaai.v35i2.16177","article-title":"Appearance-motion memory consistency network for video anomaly detection","author":"cai","year":"2021","journal-title":"Proc AAAI"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2993373"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/ICMEW46912.2020.9105987"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3077577"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.510"},{"key":"ref11","first-page":"568","article-title":"Two-stream convolutional networks for action recognition in videos","author":"simonyan","year":"2014","journal-title":"Proc NIPS"},{"key":"ref40","article-title":"Very deep convolutional networks for large-scale image recognition","author":"simonyan","year":"2015","journal-title":"arXiv 1409 1556"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.620"},{"key":"ref13","first-page":"1","article-title":"Motion-aware feature for improved video anomaly detection","author":"zhu","year":"2019","journal-title":"Proc BMVC"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00133"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.338"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2011.5995524"},{"key":"ref17","first-page":"189","article-title":"Abnormal event detection in videos using spatiotemporal autoencoder","author":"chong","year":"2017","journal-title":"Proc ISNN"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.86"},{"key":"ref19","article-title":"Anomaly detection in video using predictive convolutional long short-term memory networks","author":"medel","year":"2016","journal-title":"arXiv 1612 00390"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/6979.880968"},{"key":"ref3","first-page":"3","article-title":"Angry crowds: Detecting violent events in videos","author":"mohammadi","year":"2016","journal-title":"Proc ECCV"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2005.177"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2010.88"},{"key":"ref8","first-page":"428","article-title":"Human detection using oriented histograms of flow and appearance","author":"dalal","year":"2006","journal-title":"Proc ECCV"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.5244\/C.22.99"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01438"},{"key":"ref9","first-page":"51","article-title":"A survey of video abnormal behavior detection based on deep learning","volume":"58","author":"peng","year":"2021","journal-title":"Laser Optoelectron Prog"},{"key":"ref46","first-page":"779","article-title":"Deep appearance features for abnormal behavior detection in video","author":"smeureanu","year":"2017","journal-title":"Proc ICIAP"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.315"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00803"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.45"},{"key":"ref42","first-page":"334","article-title":"A discriminative framework for anomaly detection in large videos","author":"giorno","year":"2016","journal-title":"Proc ECCV"},{"key":"ref41","article-title":"Deep multi-scale video prediction beyond mean square error","author":"mathieu","year":"2015","journal-title":"arXiv 1511 05440"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206569"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2017.8019325"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/9312710\/09530576.pdf?arnumber=9530576","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,8]],"date-time":"2023-01-08T20:09:48Z","timestamp":1673208588000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9530576\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"references-count":55,"URL":"https:\/\/doi.org\/10.1109\/access.2021.3110798","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021]]}}}