{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,24]],"date-time":"2026-02-24T17:10:35Z","timestamp":1771953035778,"version":"3.50.1"},"publisher-location":"Cham","reference-count":49,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030695439","type":"print"},{"value":"9783030695446","type":"electronic"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-69544-6_6","type":"book-chapter","created":{"date-parts":[[2021,2,25]],"date-time":"2021-02-25T11:05:00Z","timestamp":1614251100000},"page":"83-100","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["DoFNet: Depth of Field Difference Learning for Detecting Image Forgery"],"prefix":"10.1007","author":[{"given":"Yonghyun","family":"Jeong","sequence":"first","affiliation":[]},{"given":"Jongwon","family":"Choi","sequence":"additional","affiliation":[]},{"given":"Doyeon","family":"Kim","sequence":"additional","affiliation":[]},{"given":"Sehyeon","family":"Park","sequence":"additional","affiliation":[]},{"given":"Minki","family":"Hong","sequence":"additional","affiliation":[]},{"given":"Changhyun","family":"Park","sequence":"additional","affiliation":[]},{"given":"Seungjai","family":"Min","sequence":"additional","affiliation":[]},{"given":"Youngjune","family":"Gwon","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,2,26]]},"reference":[{"key":"6_CR1","unstructured":"Foran, P.: This rental listing scam is on the rise and catching people off guard (2020). https:\/\/toronto.ctvnews.ca\/this-rental-listing-scam-is-on-the-rise-and-catching-people-off-guard-1.4995168. Accessed 22 June 2020"},{"key":"6_CR2","unstructured":"Marcellin, F.: Tackling rail fraud in the UK (2020). https:\/\/www.railway-technology.com\/features\/rail-fraud-in-the-uk\/. Accessed 28 Jan 2020"},{"key":"6_CR3","unstructured":"Nguyen, T.T., Nguyen, C.M., Nguyen, D.T., Nguyen, D.T., Nahavandi, S.: Deep learning for deepfakes creation and detection. arXiv preprint arXiv:1909.11573 (2019)"},{"key":"6_CR4","unstructured":"Cozzolino, D., Thies, J., R\u00f6ssler, A., Riess, C., Nie\u00dfner, M., Verdoliva, L.: ForensicTransfer: weakly-supervised domain adaptation for forgery detection. arXiv (2018)"},{"key":"6_CR5","doi-asserted-by":"crossref","unstructured":"Zhang, X., Karaman, S., Chang, S.: Detecting and simulating artifacts in GAN fake images. In: IEEE International Workshop on Information Forensics and Security, pp. 1\u20136 (2019)","DOI":"10.1109\/WIFS47025.2019.9035107"},{"key":"6_CR6","doi-asserted-by":"crossref","unstructured":"Durall, R., Keuper, M., Keuper, J.: Watch your up-convolution: CNN based generative deep neural networks are failing to reproduce spectral distributions. In: IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, United States (2020)","DOI":"10.1109\/CVPR42600.2020.00791"},{"key":"6_CR7","unstructured":"Frank, J., Eisenhofer, T., Sch\u00f6nherr, L., Fischer, A., Kolossa, D., Holz, T.: Leveraging frequency analysis for deep fake image recognition. arXiv preprint arXiv:2003.08685 (2020)"},{"key":"6_CR8","doi-asserted-by":"crossref","unstructured":"Wang, S.Y., Wang, O., Zhang, R., Owens, A., Efros, A.A.: CNN-generated images are surprisingly easy to spot...for now. In: IEEE Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.00872"},{"key":"6_CR9","unstructured":"Company, S.F.M.A.I.: State farm \u00ae mobile app (2020). https:\/\/www.statefarm.com\/customer-care\/download-mobile-apps\/state-farm-mobile-app. Accessed 7 July 2020"},{"key":"6_CR10","unstructured":"Metz, J.: How to file a car insurance claim from your couch (2020). https:\/\/www.forbes.com\/advisor\/car-insurance\/virtual-claims\/. Accessed 8 May 2020"},{"key":"6_CR11","unstructured":"Smith, R.: Allstate to move away from physical inspections (2017). https:\/\/www.insurancebusinessmag.com\/us\/news\/breaking-news\/allstate-to-move-away-from-physical-inspections-66880.aspx\/. Accesesed 5 May 2017"},{"key":"6_CR12","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of GANs for improved quality, stability, and variation. In: International Conference on Learning Representations (2018)"},{"key":"6_CR13","unstructured":"Brock, A., Donahue, J., Simonyan, K.: Large scale GAN training for high fidelity natural image synthesis. In: International Conference on Learning Representations (2019)"},{"key":"6_CR14","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: IEEE International Conference on Computer Vision (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"6_CR15","doi-asserted-by":"crossref","unstructured":"Choi, Y., Choi, M., Kim, M., Ha, J.W., Kim, S., Choo, J.: StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: IEEE Conference on Computer Vision and Pattern Recognition (2018)","DOI":"10.1109\/CVPR.2018.00916"},{"key":"6_CR16","doi-asserted-by":"crossref","unstructured":"Choi, Y., Uh, Y., Yoo, J., Ha, J.W.: StarGAN v2: diverse image synthesis for multiple domains. In: IEEE Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"6_CR17","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 4401\u20134410 (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"6_CR18","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of StyleGAN. CoRR abs\/1912.04958 (2019)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"6_CR19","doi-asserted-by":"crossref","unstructured":"Yang, X., Li, Y., Lyu, S.: Exposing deep fakes using inconsistent head poses. In: IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 8261\u20138265 (2019)","DOI":"10.1109\/ICASSP.2019.8683164"},{"key":"6_CR20","doi-asserted-by":"crossref","unstructured":"Li, Y., Chang, M., Lyu, S.: In Ictu Oculi: exposing AI created fake videos by detecting eye blinking. In: 2018 IEEE International Workshop on Information Forensics and Security (WIFS), pp. 1\u20137 (2018)","DOI":"10.1109\/WIFS.2018.8630787"},{"key":"6_CR21","doi-asserted-by":"crossref","unstructured":"Ciftci, U.A., Demir, I.: FakeCatcher: detection of synthetic portrait videos using biological signals. arXiv preprint arXiv:1901.02212 (2019)","DOI":"10.1109\/TPAMI.2020.3009287"},{"key":"6_CR22","doi-asserted-by":"crossref","unstructured":"Kee, E., Farid, H.: Exposing digital forgeries from 3-D lighting environments. In: IEEE International Workshop on Information Forensics and Security, pp. 1\u20136. IEEE (2010)","DOI":"10.1109\/WIFS.2010.5711437"},{"key":"6_CR23","doi-asserted-by":"crossref","unstructured":"Carvalho, T., Farid, H., Kee, E.R.: Exposing photo manipulation from user-guided 3D lighting analysis. In: Media Watermarking, Security, and Forensics 2015, vol. 9409, p. 940902. International Society for Optics and Photonics (2015)","DOI":"10.1117\/12.2075544"},{"key":"6_CR24","unstructured":"Peng, B., Wang, W., Dong, J., Tan, T.: Improved 3D lighting environment estimation for image forgery detection. In: IEEE International Workshop on Information Forensics and Security (WIFS), 1\u20136. IEEE (2015)"},{"key":"6_CR25","doi-asserted-by":"publisher","first-page":"479","DOI":"10.1109\/TIFS.2016.2623589","volume":"12","author":"B Peng","year":"2016","unstructured":"Peng, B., Wang, W., Dong, J., Tan, T.: Optimized 3D lighting environment estimation for image forgery detection. IEEE Trans. Inf. Forensics Secur. 12, 479\u2013494 (2016)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"6_CR26","doi-asserted-by":"crossref","unstructured":"Ye, S., Sun, Q., Chang, E.C.: Detecting digital image forgeries by measuring inconsistencies of blocking artifact. In: IEEE International Conference on Multimedia and Expo, pp. 12\u201315. IEEE (2007)","DOI":"10.1109\/ICME.2007.4284574"},{"key":"6_CR27","unstructured":"Tralic, D., Petrovic, J., Grgic, S.: JPEG image tampering detection using blocking artifacts. In: International Conference on Systems, Signals and Image Processing, pp. 5\u20138. IEEE (2012)"},{"key":"6_CR28","unstructured":"Agarwal, S., Farid, H., Gu, Y., He, M., Nagano, K., Li, H.: Protecting world leaders against deep fakes. In: IEEE Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, p. 8. IEEE (2019)"},{"key":"6_CR29","doi-asserted-by":"crossref","unstructured":"Matern, F., Riess, C., Stamminger, M.: Exploiting visual artifacts to expose Deepfakes and face manipulations. In: IEEE Winter Applications of Computer Vision Workshops, pp. 83\u201392. IEEE (2019)","DOI":"10.1109\/WACVW.2019.00020"},{"key":"6_CR30","unstructured":"Li, Y., Lyu, S.: Exposing DeepFake videos by detecting face warping artifacts. In: IEEE Conference on Computer Vision and Pattern Recognition Workshops (2019)"},{"key":"6_CR31","doi-asserted-by":"crossref","unstructured":"Montserrat, D.M., et al.: Deepfakes detection with automatic face weighting. arXiv preprint arXiv:2004.12027 (2020)","DOI":"10.1109\/CVPRW50498.2020.00342"},{"key":"6_CR32","doi-asserted-by":"crossref","unstructured":"Bayar, B., Stamm, M.C.: A deep learning approach to universal image manipulation detection using a new convolutional layer. In: ACM Workshop on Information Hiding and Multimedia Security, pp. 5\u201310 (2016)","DOI":"10.1145\/2909827.2930786"},{"key":"6_CR33","doi-asserted-by":"crossref","unstructured":"Kirchner, M.: Fast and reliable resampling detection by spectral analysis of fixed linear predictor residue. In: ACM Workshop on Multimedia and Security, pp. 11\u201320 (2008)","DOI":"10.1145\/1411328.1411333"},{"key":"6_CR34","doi-asserted-by":"publisher","first-page":"1509","DOI":"10.1007\/s11042-015-3152-x","volume":"76","author":"DY Huang","year":"2017","unstructured":"Huang, D.Y., Huang, C.N., Hu, W.C., Chou, C.H.: Robustness of copy-move forgery detection under high jpeg compression artifacts. Multimed. Tools Appl. 76, 1509\u20131530 (2017)","journal-title":"Multimed. Tools Appl."},{"key":"6_CR35","doi-asserted-by":"crossref","unstructured":"Marra, F., Gragnaniello, D., Verdoliva, L., Poggi, G.: Do GANs leave artificial fingerprints? In: IEEE Conference on Multimedia Information Processing and Retrieval, pp. 506\u2013511. IEEE (2019)","DOI":"10.1109\/MIPR.2019.00103"},{"key":"6_CR36","doi-asserted-by":"publisher","first-page":"3286","DOI":"10.1109\/TIP.2019.2895466","volume":"28","author":"JH Bappy","year":"2019","unstructured":"Bappy, J.H., Simons, C., Nataraj, L., Manjunath, B., Roy-Chowdhury, A.K.: Hybrid LSTM and encoder-decoder architecture for detection of image forgeries. IEEE Trans. Image Process. 28, 3286\u20133300 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"6_CR37","unstructured":"Durall, R., Keuper, M., Pfreundt, F.J., Keuper, J.: Unmasking DeepFakes with simple features. arXiv preprint arXiv:1911.00686 (2019)"},{"key":"6_CR38","first-page":"U390","volume":"1","author":"J Demers","year":"2004","unstructured":"Demers, J.: Depth of field: a survey of techniques. GPU Gems 1, U390 (2004)","journal-title":"GPU Gems"},{"key":"6_CR39","doi-asserted-by":"publisher","first-page":"704","DOI":"10.1364\/OPTICA.5.000704","volume":"5","author":"Y Wu","year":"2018","unstructured":"Wu, Y., et al.: Extended depth-of-field in holographic imaging using deep-learning-based autofocusing and phase recovery. Optica 5, 704\u2013710 (2018)","journal-title":"Optica"},{"key":"6_CR40","unstructured":"Mirza, M., Osindero, S.: Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784 (2014)"},{"key":"6_CR41","doi-asserted-by":"publisher","first-page":"1982","DOI":"10.1109\/TMM.2019.2895292","volume":"21","author":"X Guo","year":"2019","unstructured":"Guo, X., Nie, R., Cao, J., Zhou, D., Mei, L., He, K.: FuseGAN: learning to fuse multi-focus image via conditional generative adversarial network. IEEE Trans. Multimedia 21, 1982\u20131996 (2019)","journal-title":"IEEE Trans. Multimedia"},{"key":"6_CR42","doi-asserted-by":"crossref","unstructured":"Zhang, J., Liao, Q., Liu, S., Ma, H., Yang, W., Xue, J.h.: Real-MFF dataset: a large realistic multi-focus image dataset with ground truth. arXiv preprint arXiv:2003.12779 (2020)","DOI":"10.1016\/j.patrec.2020.08.002"},{"key":"6_CR43","doi-asserted-by":"crossref","unstructured":"Cheng, Z., Bai, F., Xu, Y., Zheng, G., Pu, S., Zhou, S.: Focusing attention: towards accurate text recognition in natural images. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 5076\u20135084 (2017)","DOI":"10.1109\/ICCV.2017.543"},{"key":"6_CR44","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2014.10.004","volume":"25","author":"M Nejati","year":"2015","unstructured":"Nejati, M., Samavi, S., Shirani, S.: Multi-focus image fusion using dictionary-based sparse representation. Inf. Fusion 25, 72\u201384 (2015)","journal-title":"Inf. Fusion"},{"key":"6_CR45","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu, Y., Chen, X., Peng, H., Wang, Z.: Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 36, 191\u2013207 (2017)","journal-title":"Inf. Fusion"},{"key":"6_CR46","doi-asserted-by":"crossref","unstructured":"Ma, H., Zhang, J., Liu, S., Liao, Q.: Boundary aware multi-focus image fusion using deep neural network. In: 2019 IEEE International Conference on Multimedia and Expo (ICME), pp. 1150\u20131155. IEEE (2019)","DOI":"10.1109\/ICME.2019.00201"},{"key":"6_CR47","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"6_CR48","doi-asserted-by":"publisher","first-page":"533","DOI":"10.1038\/323533a0","volume":"323","author":"DE Rumelhart","year":"1986","unstructured":"Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning representations by back-propagating errors. Nature 323, 533\u2013536 (1986)","journal-title":"Nature"},{"key":"6_CR49","unstructured":"Kingma, D., Ba, J.: Adam: a method for stochastic optimization. In: International Conference on Learning Representations (2014)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ACCV 2020"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-69544-6_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,12,18]],"date-time":"2022-12-18T23:46:02Z","timestamp":1671407162000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-69544-6_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030695439","9783030695446"],"references-count":49,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-69544-6_6","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"26 February 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ACCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Asian Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kyoto","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 November 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 December 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"accv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/accv2020.kyoto\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"768","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"254","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"33% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}