{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T05:12:47Z","timestamp":1755839567705,"version":"3.40.3"},"publisher-location":"Cham","reference-count":38,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031365737"},{"type":"electronic","value":"9783031365744"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-36574-4_1","type":"book-chapter","created":{"date-parts":[[2023,7,15]],"date-time":"2023-07-15T13:01:58Z","timestamp":1689426118000},"page":"3-18","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Image-to-Image Translation Generative Adversarial Networks for Video Source Camera Falsification"],"prefix":"10.1007","author":[{"given":"Maryna","family":"Veksler","sequence":"first","affiliation":[]},{"given":"Clara","family":"Caspard","sequence":"additional","affiliation":[]},{"given":"Kemal","family":"Akkaya","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,7,16]]},"reference":[{"key":"1_CR1","doi-asserted-by":"publisher","unstructured":"Auto-painter: cartoon image generation from sketch by using conditional Wasserstein generative adversarial networks. Neurocomputing 311, 78\u201387 (2018). https:\/\/doi.org\/10.1016\/j.neucom.2018.05.045","DOI":"10.1016\/j.neucom.2018.05.045"},{"key":"1_CR2","doi-asserted-by":"publisher","unstructured":"GANs for medical image analysis. Artif. Intell. Med. 109, 101938 (2020). https:\/\/doi.org\/10.1016\/j.artmed.2020.101938","DOI":"10.1016\/j.artmed.2020.101938"},{"key":"1_CR3","doi-asserted-by":"publisher","unstructured":"Super-resolution using GANs for medical imaging. Proc. Comput. Sci. 173, 28\u201335 (2020). https:\/\/doi.org\/10.1016\/j.procs.2020.06.005. International Conference on Smart Sustainable Intelligent Computing and Applications Under ICITETM 2020","DOI":"10.1016\/j.procs.2020.06.005"},{"issue":"2","key":"1_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3487891","volume":"55","author":"N Aldausari","year":"2022","unstructured":"Aldausari, N., Sowmya, A., Marcus, N., Mohammadi, G.: Video generative adversarial networks: a review. ACM Comput. Surv. 55(2), 1\u201325 (2022)","journal-title":"ACM Comput. Surv."},{"key":"1_CR5","doi-asserted-by":"publisher","unstructured":"Barni, M., Chen, Z., Tondi, B.: Adversary-aware, data-driven detection of double JPEG compression: how to make counter-forensics harder. In: 2016 IEEE International Workshop on Information Forensics and Security (WIFS), pp. 1\u20136 (2016). https:\/\/doi.org\/10.1109\/WIFS.2016.7823902","DOI":"10.1109\/WIFS.2016.7823902"},{"key":"1_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11042-020-09011-4","volume":"80","author":"C Chen","year":"2021","unstructured":"Chen, C., Stamm, M.: Robust camera model identification using demosaicing residual features. Multimed. Tools Appl. 80, 1\u201329 (2021). https:\/\/doi.org\/10.1007\/s11042-020-09011-4","journal-title":"Multimed. Tools Appl."},{"key":"1_CR7","doi-asserted-by":"publisher","unstructured":"Chen, C., Zhao, X., Stamm, M.C.: MISLGAN: an anti-forensic camera model falsification framework using a generative adversarial network. In: 2018 25th IEEE International Conference on Image Processing (ICIP), pp. 535\u2013539 (2018). https:\/\/doi.org\/10.1109\/ICIP.2018.8451503","DOI":"10.1109\/ICIP.2018.8451503"},{"key":"1_CR8","doi-asserted-by":"publisher","unstructured":"Chen, C., Zhao, X., Stamm, M.C.: Generative adversarial attacks against deep-learning-based camera model identification. IEEE Trans. Inf. Forensics Secur. PP, 1 (2019). https:\/\/doi.org\/10.1109\/TIFS.2019.2945198","DOI":"10.1109\/TIFS.2019.2945198"},{"key":"1_CR9","unstructured":"Cozzolino, D., Thies, J., R\u00f6ssler, A., Nie\u00dfner, M., Verdoliva, L.: SpoC: spoofing camera fingerprints (2019)"},{"key":"1_CR10","doi-asserted-by":"publisher","unstructured":"Cozzolino, D., Verdoliva, L.: Multimedia forensics before the deep learning era. In: Rathgeb, C., Tolosana, R., Vera-Rodriguez, R., Busch, C. (eds.) Handbook of Digital Face Manipulation and Detection. ACVPR, pp. 45\u201367. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-030-87664-7_3","DOI":"10.1007\/978-3-030-87664-7_3"},{"issue":"8","key":"1_CR11","doi-asserted-by":"publisher","first-page":"135","DOI":"10.3390\/jimaging7080135","volume":"7","author":"D Dal Cortivo","year":"2021","unstructured":"Dal Cortivo, D., Mandelli, S., Bestagini, P., Tubaro, S.: CNN-based multi-modal camera model identification on video sequences. J. Imag. 7(8), 135 (2021)","journal-title":"J. Imag."},{"key":"1_CR12","unstructured":"Damiani, J.: A voice deepfake was used to scam a CEO out of \\$243,000 (2019). https:\/\/www.forbes.com\/sites\/jessedamiani\/2019\/09\/03\/a-voice-deepfake-was-used-to-scam-a-ceo-out-of-243000\/?sh=34e8298a2241"},{"issue":"24","key":"1_CR13","doi-asserted-by":"publisher","first-page":"31835","DOI":"10.1007\/s11042-018-6170-7","volume":"77","author":"TK Das","year":"2018","unstructured":"Das, T.K.: Anti-forensics of JPEG compression detection schemes using approximation of DCT coefficients. Multimed. Tools Appl. 77(24), 31835\u201331854 (2018)","journal-title":"Multimed. Tools Appl."},{"key":"1_CR14","unstructured":"Duan, B., Wang, W., Tang, H., Latapie, H., Yan, Y.: Cascade attention guided residue learning GAN for cross-modal translation (2019)"},{"key":"1_CR15","doi-asserted-by":"publisher","unstructured":"Flor, E., Aygun, R., Mercan, S., Akkaya, K.: PRNU-based source camera identification for multimedia forensics. In: 2021 IEEE 22nd International Conference on Information Reuse and Integration for Data Science (IRI), pp. 168\u2013175 (2021). https:\/\/doi.org\/10.1109\/IRI51335.2021.00029","DOI":"10.1109\/IRI51335.2021.00029"},{"key":"1_CR16","unstructured":"Gauthier, J.: Conditional generative adversarial nets for convolutional face generation (2015)"},{"key":"1_CR17","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. Advances in Neural Information Processing Systems, vol. 27 (2014)"},{"key":"1_CR18","doi-asserted-by":"publisher","unstructured":"Hosler, B., et al.: A video camera model identification system using deep learning and fusion. In: ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 8271\u20138275 (2019). https:\/\/doi.org\/10.1109\/ICASSP.2019.8682608","DOI":"10.1109\/ICASSP.2019.8682608"},{"key":"1_CR19","doi-asserted-by":"crossref","unstructured":"Jeong, S., Lee, J., Sohn, K.: Multi-domain unsupervised image-to-image translation with appearance adaptive convolution (2022)","DOI":"10.1109\/ICASSP43922.2022.9746500"},{"key":"1_CR20","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of GANs for improved quality, stability, and variation (2017)"},{"issue":"4","key":"1_CR21","doi-asserted-by":"publisher","first-page":"582","DOI":"10.1109\/TIFS.2008.2008214","volume":"3","author":"M Kirchner","year":"2008","unstructured":"Kirchner, M., Bohme, R.: Hiding traces of resampling in digital images. IEEE Trans. Inf. Forensics Secur. 3(4), 582\u2013592 (2008)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"1_CR22","doi-asserted-by":"crossref","unstructured":"Korshunova, I., Shi, W., Dambre, J., Theis, L.: Fast face-swap using convolutional neural networks (2016)","DOI":"10.1109\/ICCV.2017.397"},{"key":"1_CR23","doi-asserted-by":"crossref","unstructured":"Li, Y., Min, M.R., Shen, D., Carlson, D., Carin, L.: Video generation from text. In: Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence, AAAI 2018\/IAAI 2018\/EAAI 2018. AAAI Press (2018)","DOI":"10.1609\/aaai.v32i1.12233"},{"key":"1_CR24","doi-asserted-by":"crossref","unstructured":"Mayer, O., Stamm, M.C.: Countering anti-forensics of lateral chromatic aberration. Association for Computing Machinery, New York, NY, USA (2017)","DOI":"10.1145\/3082031.3083242"},{"key":"1_CR25","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TCSVT.2022.3177238","volume":"32","author":"F Peng","year":"2022","unstructured":"Peng, F., Yin, L., Long, M.: BDC-GAN: bidirectional conversion between computer-generated and natural facial images for anti-forensics. IEEE Trans. Circ. Syst. Video Technol. 32, 1 (2022). https:\/\/doi.org\/10.1109\/TCSVT.2022.3177238","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"issue":"1","key":"1_CR26","first-page":"23","volume":"7","author":"D Rong","year":"2021","unstructured":"Rong, D., Wang, Y., Sun, Q.: Video source forensics for IoT devices based on convolutional neural networks. Open J. Internet Things (OJIOT) 7(1), 23\u201331 (2021)","journal-title":"Open J. Internet Things (OJIOT)"},{"key":"1_CR27","doi-asserted-by":"crossref","unstructured":"Sharma, S., Ravi, H., Subramanyam, A., Emmanuel, S.: Anti-forensics of median filtering and contrast enhancement. J. Vis. Commun. Image Represent. 66(C), 102682 (2020)","DOI":"10.1016\/j.jvcir.2019.102682"},{"key":"1_CR28","doi-asserted-by":"publisher","unstructured":"Shullani, D., Fontani, M., Iuliani, M., Alshaya, O., Piva, A.: Vision: a video and image dataset for source identification. EURASIP J. Inf. Secur. 2017, 15 (2017). https:\/\/doi.org\/10.1186\/s13635-017-0067-2","DOI":"10.1186\/s13635-017-0067-2"},{"issue":"4","key":"1_CR29","doi-asserted-by":"publisher","first-page":"1315","DOI":"10.1109\/TIFS.2012.2205568","volume":"7","author":"MC Stamm","year":"2012","unstructured":"Stamm, M.C., Lin, W.S., Liu, K.J.R.: Temporal forensics and anti-forensics for motion compensated video. IEEE Trans. Inf. Forensics Secur. 7(4), 1315\u20131329 (2012). https:\/\/doi.org\/10.1109\/TIFS.2012.2205568","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"1_CR30","doi-asserted-by":"crossref","unstructured":"Thies, J., Zollh\u00f6fer, M., Stamminger, M., Theobalt, C., Nie\u00dfner, M.: Face2Face: real-time face capture and reenactment of RGB videos, vol. 62, no. 1 (2018)","DOI":"10.1145\/3292039"},{"key":"1_CR31","doi-asserted-by":"crossref","unstructured":"Tulyakov, S., Liu, M.Y., Yang, X., Kautz, J.: MoCoGAN: decomposing motion and content for video generation (2017)","DOI":"10.1109\/CVPR.2018.00165"},{"key":"1_CR32","doi-asserted-by":"crossref","unstructured":"Veksler, M., Aygun, R., Akkaya, K., Iyengar, S.: Video origin camera identification using ensemble CNNs of positional patches. In: 2022 IEEE 5th International Conference on Multimedia Information Processing and Retrieval (IEEE MIPR) (2022). (in Press)","DOI":"10.1109\/MIPR54900.2022.00015"},{"key":"1_CR33","doi-asserted-by":"crossref","unstructured":"Venkatesh, S., Zhang, H., Ramachandra, R., Raja, K., Damer, N., Busch, C.: Can GAN generated morphs threaten face recognition systems equally as landmark based morphs? - vulnerability and detection (2020)","DOI":"10.1109\/IWBF49977.2020.9107970"},{"key":"1_CR34","unstructured":"Villegas, R., Yang, J., Hong, S., Lin, X., Lee, H.: Decomposing motion and content for natural video sequence prediction. ArXiv abs\/1706.08033 (2017)"},{"key":"1_CR35","unstructured":"Vondrick, C., Pirsiavash, H., Torralba, A.: Generating videos with scene dynamics. In: NIPS 2016, pp. 613\u2013621. Curran Associates Inc., Red Hook, NY, USA (2016)"},{"key":"1_CR36","doi-asserted-by":"crossref","unstructured":"Yu, J., Xue, H., Liu, B., Wang, Y., Zhu, S., Ding, M.: GAN-based differential private image privacy protection framework for the internet of multimedia things. Sensors 21(1), 58 (2021)","DOI":"10.3390\/s21010058"},{"key":"1_CR37","doi-asserted-by":"publisher","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2242\u20132251 (2017). https:\/\/doi.org\/10.1109\/ICCV.2017.244","DOI":"10.1109\/ICCV.2017.244"},{"key":"1_CR38","doi-asserted-by":"crossref","unstructured":"Zou, H., Yang, P., Ni, R., Zhao, Y., Zhou, N.: Anti-forensics of image contrast enhancement based on generative adversarial network (2021)","DOI":"10.1155\/2021\/6663486"}],"container-title":["Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering","Digital Forensics and Cyber Crime"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-36574-4_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,15]],"date-time":"2023-07-15T13:02:10Z","timestamp":1689426130000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-36574-4_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031365737","9783031365744"],"references-count":38,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-36574-4_1","relation":{},"ISSN":["1867-8211","1867-822X"],"issn-type":[{"type":"print","value":"1867-8211"},{"type":"electronic","value":"1867-822X"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"16 July 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICDF2C","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Digital Forensics and Cyber Crime","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Boston, MA","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"USA","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 November 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 November 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icdf2c2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Confy plus","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"80","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"35% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}