{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T00:11:54Z","timestamp":1730247114150,"version":"3.28.0"},"reference-count":30,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,10,1]],"date-time":"2020-10-01T00:00:00Z","timestamp":1601510400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,10,1]],"date-time":"2020-10-01T00:00:00Z","timestamp":1601510400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,10,1]],"date-time":"2020-10-01T00:00:00Z","timestamp":1601510400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,10]]},"DOI":"10.1109\/icip40778.2020.9191186","type":"proceedings-article","created":{"date-parts":[[2020,9,30]],"date-time":"2020-09-30T20:45:18Z","timestamp":1601498718000},"page":"2915-2919","source":"Crossref","is-referenced-by-count":5,"title":["Implicit Saliency In Deep Neural Networks"],"prefix":"10.1109","author":[{"given":"Yutong","family":"Sun","sequence":"first","affiliation":[]},{"given":"Mohit","family":"Prabhushankar","sequence":"additional","affiliation":[]},{"given":"Ghassan","family":"AlRegib","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref30","first-page":"740","article-title":"Microsoft coco: Common objects in context","author":"lin","year":"2014","journal-title":"European Conference on Computer Vision"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.3758\/s13414-016-1102-y"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1111\/1467-9280.00488"},{"key":"ref12","article-title":"Semantic and contrast-aware saliency","author":"sun","year":"2018","journal-title":"arXiv preprint arXiv 1811 03736"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.tics.2007.06.010"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1038\/381520a0"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1152\/jn.00777.2002"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref17","article-title":"Very deep convolutional networks for large-scale image recognition","author":"simonyan","year":"2015","journal-title":"International Conference on Learning Representations"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref19","article-title":"Faster R-CNN: Towards real-time object detection with region proposal networks","author":"ren","year":"2015","journal-title":"Advances in Neural Information Processing Systems (NIPS)"},{"key":"ref28","article-title":"Striving for simplicity: The all convolutional net","author":"springenberg","year":"2014","journal-title":"arXiv preprint arXiv 1412 6806"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2018.00250"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.74"},{"key":"ref3","article-title":"Deepgaze II: reading fixations from deep features trained on object recognition","volume":"abs 1610 1563","author":"k\u00fcmmerer","year":"2016","journal-title":"CoRR"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2018.03.005"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298710"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2016.7900174"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.tics.2009.06.003"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/j.actpsy.2010.12.002"},{"key":"ref2","article-title":"Uncertainty Estimation of Visual Attention Models Using Spatiotemporal Analysis","author":"alshawi","year":"2018","journal-title":"Thesis Georgia Institute of Technology"},{"key":"ref9","article-title":"Stimulus conflict and stimulus novelty trigger saliency signals in locus coeruleus and anterior cingulate cortex","volume":"114","author":"krebs","year":"2012","journal-title":"Front Hum Neurosci Conference Abstract Belgian Brain Council"},{"journal-title":"Salience network of the human brain","year":"2016","author":"uddin","key":"ref1"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-009-0275-4"},{"key":"ref22","article-title":"Deepgaze ii: Reading fixations from deep features trained on object recognition","author":"k\u00fcmmerer","year":"2016","journal-title":"arXiv preprint arXiv 1610 01292"},{"key":"ref21","article-title":"Salgan: Visual saliency prediction with generative adversarial networks","author":"pan","year":"2017","journal-title":"arXiv preprint arXiv 1701 01887"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2019.8803228"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.71"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2018.2815601"},{"journal-title":"MIT Saliency Benchmark","year":"0","author":"bylinskii","key":"ref25"}],"event":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","start":{"date-parts":[[2020,10,25]]},"location":"Abu Dhabi, United Arab Emirates","end":{"date-parts":[[2020,10,28]]}},"container-title":["2020 IEEE International Conference on Image Processing (ICIP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9184803\/9190635\/09191186.pdf?arnumber=9191186","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,28]],"date-time":"2022-06-28T00:09:51Z","timestamp":1656374991000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9191186\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,10]]},"references-count":30,"URL":"https:\/\/doi.org\/10.1109\/icip40778.2020.9191186","relation":{},"subject":[],"published":{"date-parts":[[2020,10]]}}}