{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T19:11:22Z","timestamp":1772910682932,"version":"3.50.1"},"publisher-location":"Cham","reference-count":43,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197895","type":"print"},{"value":"9783031197901","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19790-1_23","type":"book-chapter","created":{"date-parts":[[2022,10,23]],"date-time":"2022-10-23T11:02:44Z","timestamp":1666522964000},"page":"373-389","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":21,"title":["Learning Object Placement via\u00a0Dual-Path Graph Completion"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6337-1130","authenticated-orcid":false,"given":"Siyuan","family":"Zhou","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8035-6610","authenticated-orcid":false,"given":"Liu","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1970-8634","authenticated-orcid":false,"given":"Li","family":"Niu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7597-8503","authenticated-orcid":false,"given":"Liqing","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,24]]},"reference":[{"key":"23_CR1","doi-asserted-by":"publisher","first-page":"2570","DOI":"10.1007\/s11263-020-01336-9","volume":"128","author":"S Azadi","year":"2020","unstructured":"Azadi, S., Pathak, D., Ebrahimi, S., Darrell, T.: Compositional GAN: learning image-conditional binary composition. Int. J. Comput. Vis. 128, 2570\u20132585 (2020)","journal-title":"Int. J. Comput. Vis."},{"key":"23_CR2","doi-asserted-by":"crossref","unstructured":"Chen, B.C., Kae, A.: Toward realistic image compositing with adversarial learning. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00861"},{"key":"23_CR3","first-page":"1","volume":"28","author":"T Chen","year":"2009","unstructured":"Chen, T., Cheng, M.M., Tan, P., Shamir, A., Hu, S.M.: Sketch2Photo: Internet image montage. ACM Trans. Graph. (TOG) 28, 1\u201310 (2009)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"23_CR4","doi-asserted-by":"crossref","unstructured":"Cong, W., Niu, L., Zhang, J., Liang, J., Zhang, L.: BargainNet: background-guided domain translation for image harmonization. In: ICME (2021)","DOI":"10.1109\/ICME51207.2021.9428394"},{"key":"23_CR5","doi-asserted-by":"crossref","unstructured":"Cong, W., et al.: High-resolution image harmonization via collaborative dual transformations. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01792"},{"key":"23_CR6","doi-asserted-by":"crossref","unstructured":"Cong, W., et al.: DoveNet: deep image harmonization via domain verification. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00842"},{"key":"23_CR7","doi-asserted-by":"crossref","unstructured":"Dai, Z., Yang, Z., Yang, Y., Carbonell, J., Le, Q.V., Salakhutdinov, R.: Transformer-XL: attentive language models beyond a fixed-length context (2019)","DOI":"10.18653\/v1\/P19-1285"},{"key":"23_CR8","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: CVPR (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"23_CR9","unstructured":"Gehring, J., Auli, M., Grangier, D., Yarats, D., Dauphin, Y.N.: Convolutional sequence to sequence learning. In: ICML (2017)"},{"key":"23_CR10","doi-asserted-by":"crossref","unstructured":"Georgakis, G., Mousavian, A., Berg, A.C., Kosecka, J.: Synthesizing training data for object detection in indoor scenes (2017)","DOI":"10.15607\/RSS.2017.XIII.043"},{"key":"23_CR11","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. NIPS (2014)"},{"key":"23_CR12","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local Nash equilibrium. In: NeurIPS (2017)"},{"key":"23_CR13","doi-asserted-by":"crossref","unstructured":"Hong, Y., Niu, L., Zhang, J.: Shadow generation for composite image in real-world scenes. In: AAAI (2022)","DOI":"10.1609\/aaai.v36i1.19974"},{"key":"23_CR14","doi-asserted-by":"crossref","unstructured":"Johnson, J., Gupta, A., Fei-Fei, L.: Image generation from scene graphs. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00133"},{"key":"23_CR15","unstructured":"Kingma, D.P., Welling, M.: Auto-encoding variational bayes (2014)"},{"key":"23_CR16","doi-asserted-by":"crossref","unstructured":"Lalonde, J.F., Efros, A.A.: Using color compatibility for assessing image realism. In: ICCV (2007)","DOI":"10.1109\/ICCV.2007.4409107"},{"key":"23_CR17","doi-asserted-by":"crossref","unstructured":"Lalonde, J.F., Hoiem, D., Efros, A.A., Rother, C., Winn, J., Criminisi, A.: Photo clip art. ACM Trans. Graph. (TOG) 26, 3-es (2007)","DOI":"10.1145\/1239451.1239454"},{"key":"23_CR18","unstructured":"Lee, D., Liu, S., Gu, J., Liu, M.Y., Yang, M.H., Kautz, J.: Context-aware synthesis and placement of object instances (2018)"},{"key":"23_CR19","doi-asserted-by":"crossref","unstructured":"Li, X., Liu, S., Kim, K., Wang, X., Yang, M.H., Kautz, J.: Putting humans in a scene: learning affordance in 3D indoor environments. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01265"},{"key":"23_CR20","doi-asserted-by":"crossref","unstructured":"Lin, C.H., Yumer, E., Wang, O., Shechtman, E., Lucey, S.: ST-GAN: spatial transformer generative adversarial networks for image compositing. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00985"},{"key":"23_CR21","doi-asserted-by":"crossref","unstructured":"Liu, D., Long, C., Zhang, H., Yu, H., Dong, X., Xiao, C.: ARShadowGAN: shadow generative adversarial network for augmented reality in single light scenes. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00816"},{"key":"23_CR22","unstructured":"Liu, L., Zhang, B., Li, J., Niu, L., Liu, Q., Zhang, L.: OPA: object placement assessment dataset. arXiv preprint arXiv:2107.01889 (2021)"},{"key":"23_CR23","unstructured":"Liu, X., Yu, H.F., Dhillon, I., Hsieh, C.J.: Learning to encode position for transformer with continuous dynamical model. In: ICML (2020)"},{"key":"23_CR24","unstructured":"Niu, L., et al.: Making images real again: a comprehensive survey on deep image composition. arXiv preprint arXiv:2106.14490 (2021)"},{"key":"23_CR25","unstructured":"Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer (2020)"},{"key":"23_CR26","doi-asserted-by":"crossref","unstructured":"Schuster, M.J., Okerman, J., Nguyen, H., Rehg, J.M., Kemp, C.C.: Perceiving clutter and surfaces for object placement in indoor environments. In: ICHR (2010)","DOI":"10.1109\/ICHR.2010.5686328"},{"key":"23_CR27","doi-asserted-by":"crossref","unstructured":"Shaw, P., Uszkoreit, J., Vaswani, A.: Self-attention with relative position representations (2018)","DOI":"10.18653\/v1\/N18-2074"},{"key":"23_CR28","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition (2015)"},{"key":"23_CR29","doi-asserted-by":"crossref","unstructured":"Smith, A.R., Blinn, J.F.: Blue screen matting. In: SIGGRAPH (1996)","DOI":"10.1145\/237170.237263"},{"key":"23_CR30","doi-asserted-by":"crossref","unstructured":"Tan, F., Bernier, C., Cohen, B., Ordonez, V., Barnes, C.: Where and who? Automatic semantic-aware person composition. In: WACV (2018)","DOI":"10.1109\/WACV.2018.00170"},{"key":"23_CR31","doi-asserted-by":"crossref","unstructured":"Tripathi, S., Chandra, S., Agrawal, A., Tyagi, A., Rehg, J.M., Chari, V.: Learning to generate synthetic data via compositing. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00055"},{"key":"23_CR32","doi-asserted-by":"crossref","unstructured":"Tsai, Y.H., Shen, X., Lin, Z., Sunkavalli, K., Lu, X., Yang, M.H.: Deep image harmonization. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.299"},{"key":"23_CR33","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NeurIPS (2017)"},{"key":"23_CR34","doi-asserted-by":"crossref","unstructured":"Wang, T.C., Liu, M.Y., Zhu, J.Y., Tao, A., Kautz, J., Catanzaro, B.: High-resolution image synthesis and semantic manipulation with conditional GANs. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00917"},{"key":"23_CR35","doi-asserted-by":"crossref","unstructured":"Weng, S., Li, W., Li, D., Jin, H., Shi, B.: MISC: multi-condition injection and spatially-adaptive compositing for conditional person image synthesis. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00776"},{"key":"23_CR36","doi-asserted-by":"crossref","unstructured":"Wu, H., Zheng, S., Zhang, J., Huang, K.: GP-GAN: towards realistic high-resolution image blending. In: ACM Multimedia (2019)","DOI":"10.1145\/3343031.3350944"},{"key":"23_CR37","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2185520.2185580","volume":"31","author":"S Xue","year":"2012","unstructured":"Xue, S., Agarwala, A., Dorsey, J., Rushmeier, H.: Understanding and improving the realism of image composites. ACM Trans. Graph. (TOG) 31, 1\u201310 (2012)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"23_CR38","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"566","DOI":"10.1007\/978-3-030-58601-0_34","volume-title":"Computer Vision \u2013 ECCV 2020","author":"L Zhang","year":"2020","unstructured":"Zhang, L., Wen, T., Min, J., Wang, J., Han, D., Shi, J.: Learning object placement by inpainting for compositional data augmentation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12358, pp. 566\u2013581. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58601-0_34"},{"key":"23_CR39","doi-asserted-by":"crossref","unstructured":"Zhang, L., Wen, T., Shi, J.: Deep image blending. In: WACV (2020)","DOI":"10.1109\/WACV45572.2020.9093632"},{"key":"23_CR40","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"issue":"1","key":"23_CR41","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1007\/s41095-020-0158-8","volume":"6","author":"S-H Zhang","year":"2020","unstructured":"Zhang, S.-H., Zhou, Z.-P., Liu, B., Dong, X., Hall, P.: What and where: a context-based recommendation system for object insertion. Comput. Vis. Media 6(1), 79\u201393 (2020). https:\/\/doi.org\/10.1007\/s41095-020-0158-8","journal-title":"Comput. Vis. Media"},{"key":"23_CR42","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Krahenbuhl, P., Shechtman, E., Efros, A.A.: Learning a discriminative model for the perception of realism in composite images. In: ICCV, pp. 3943\u20133951 (2015)","DOI":"10.1109\/ICCV.2015.449"},{"key":"23_CR43","unstructured":"Zhu, J.Y., et al.: Multimodal image-to-image translation by enforcing bi-cycle consistency. In: NeurIPS (2017)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19790-1_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,12]],"date-time":"2024-03-12T15:08:48Z","timestamp":1710256128000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19790-1_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197895","9783031197901"],"references-count":43,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19790-1_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"24 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}