{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,21]],"date-time":"2025-12-21T21:17:50Z","timestamp":1766351870773,"version":"3.40.3"},"publisher-location":"Cham","reference-count":51,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031200496"},{"type":"electronic","value":"9783031200502"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-20050-2_28","type":"book-chapter","created":{"date-parts":[[2022,10,27]],"date-time":"2022-10-27T22:09:58Z","timestamp":1666908598000},"page":"477-493","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":16,"title":["Neural-Sim: Learning to\u00a0Generate Training Data with\u00a0NeRF"],"prefix":"10.1007","author":[{"given":"Yunhao","family":"Ge","sequence":"first","affiliation":[]},{"given":"Harkirat","family":"Behl","sequence":"additional","affiliation":[]},{"given":"Jiashu","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Suriya","family":"Gunasekar","sequence":"additional","affiliation":[]},{"given":"Neel","family":"Joshi","sequence":"additional","affiliation":[]},{"given":"Yale","family":"Song","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Laurent","family":"Itti","sequence":"additional","affiliation":[]},{"given":"Vibhav","family":"Vineet","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,28]]},"reference":[{"key":"28_CR1","unstructured":"Jahanian, A., Lucy Chai, P.I.: On the \"steerability\" of generative adversarial networks. CoRR (2019)"},{"key":"28_CR2","unstructured":"Barbu, A., et al.: Objectnet: A large-scale bias-controlled dataset for pushing the limits of object recognition models. In: Wallach, H., Larochelle, H., Beygelzimer, A., d\u2019Alch\u00e9-Buc, F., Fox, E., Garnett, R. (eds.) Advances in Neural Information Processing Systems. vol. 32. Curran Associates, Inc. (2019), https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/97af07a14cacba681feacf3012730892-Paper.pdf"},{"key":"28_CR3","unstructured":"Barbu, A., et al.: Objectnet: A large-scale bias-controlled dataset for pushing the limits of object recognition models. In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"28_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"255","DOI":"10.1007\/978-3-030-58542-6_16","volume-title":"Computer Vision \u2013 ECCV 2020","author":"HS Behl","year":"2020","unstructured":"Behl, H.S., Baydin, A.G., Gal, R., Torr, P.H.S., Vineet, V.: Autosimulate: (quickly) learning synthetic data generation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12367, pp. 255\u2013271. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58542-6_16"},{"key":"28_CR5","unstructured":"Bi, S., et al.: Neural reflectance fields for appearance acquisition. arXiv preprint arXiv:2008.03824 (2020)"},{"key":"28_CR6","unstructured":"Brock, A., Donahue, J., Simonyan, K.: Large scale GAN training for high fidelity natural image synthesis. In: International Conference on Learning Representations (2019). https:\/\/openreview.net\/forum?id=B1xsqj09Fm"},{"key":"28_CR7","doi-asserted-by":"crossref","unstructured":"Calli, B., Walsman, A., Singh, A., Srinivasa, S., Abbeel, P., Dollar, A.M.: Benchmarking in manipulation research: The ycb object and model set and benchmarking protocols. arXiv preprint arXiv:1502.03143 (2015)","DOI":"10.1109\/MRA.2015.2448951"},{"issue":"1","key":"28_CR8","doi-asserted-by":"publisher","first-page":"235","DOI":"10.1007\/s10479-007-0176-2","volume":"153","author":"B Colson","year":"2007","unstructured":"Colson, B., Marcotte, P., Savard, G.: An overview of bilevel optimization. Ann. Oper. Res. 153(1), 235\u2013256 (2007)","journal-title":"Ann. Oper. Res."},{"key":"28_CR9","unstructured":"Danilo Jimenez Rezende, S.M.: Variational inference with normalizing flows. In: ICML (2015)"},{"key":"28_CR10","unstructured":"Denninger, M., et al.: Blenderproc. arXiv preprint arXiv:1911.01911 (2019)"},{"key":"28_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"715","DOI":"10.1007\/978-3-030-58520-4_42","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Devaranjan","year":"2020","unstructured":"Devaranjan, J., Kar, A., Fidler, S.: Meta-Sim2: Unsupervised learning of scene structure for synthetic data generation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12362, pp. 715\u2013733. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58520-4_42"},{"key":"28_CR12","unstructured":"Diederik Kingma, M.W.: Autoencoding variational bayes. In: ICLR (2014)"},{"key":"28_CR13","unstructured":"Doersch, C., Zisserman, A.: Sim2real transfer learning for 3d human pose estimation: motion to the rescue. In: NeurIPS (2019)"},{"key":"28_CR14","doi-asserted-by":"crossref","unstructured":"Dwibedi, D., Misra, I., Hebert, M.: Cut, paste and learn: Surprisingly easy synthesis for instance detection. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.146"},{"key":"28_CR15","unstructured":"Franceschi, L., Frasconi, P., Salzo, S., Grazzi, R., Pontil, M.: Bilevel programming for hyperparameter optimization and meta-learning. In: International Conference on Machine Learning, pp. 1568\u20131577. PMLR (2018)"},{"key":"28_CR16","doi-asserted-by":"crossref","unstructured":"Gafni, G., Thies, J., Zollhofer, M., Nie\u00dfner, M.: Dynamic neural radiance fields for monocular 4d facial avatar reconstruction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8649\u20138658 (2021)","DOI":"10.1109\/CVPR46437.2021.00854"},{"key":"28_CR17","unstructured":"Ganin, Y., Kulkarni, T., Babuschkin, I., Eslami, S.M.A., Vinyals, O.: Synthesizing programs for images using reinforced adversarial learning. In: ICML (2018)"},{"key":"28_CR18","unstructured":"Ge, Y., Abu-El-Haija, S., Xin, G., Itti, L.: Zero-shot synthesis with group-supervised learning. arXiv preprint arXiv:2009.06586 (2020)"},{"key":"28_CR19","unstructured":"Ge, Y., Xu, J., Zhao, B.N., Itti, L., Vineet, V.: Dall-e for detection: Language-driven context image synthesis for object detection. arXiv preprint arXiv:2206.09592 (2022)"},{"key":"28_CR20","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"138","DOI":"10.1007\/978-3-030-58604-1_9","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Ge","year":"2020","unstructured":"Ge, Y., Zhao, J., Itti, L.: Pose augmentation: Class-agnostic object pose transformation for object recognition. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12373, pp. 138\u2013155. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58604-1_9"},{"key":"28_CR21","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3182160","volume":"37","author":"I Georgiev","year":"2018","unstructured":"Georgiev, I., et al.: Arnold: A brute-force production path tracer. TOG 37, 1\u201312 (2018)","journal-title":"TOG"},{"key":"28_CR22","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: Advances in neural information processing systems, pp. 2672\u20132680 (2014)"},{"key":"28_CR23","doi-asserted-by":"crossref","unstructured":"Handa, A., Patraucean, V., Badrinarayanan, V., Stent, S., Cipolla, R.: Understanding real world indoor scenes with synthetic data. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.442"},{"key":"28_CR24","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask r-cnn. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.322"},{"key":"28_CR25","unstructured":"Higgins, I., et al.: beta-vae: Learning basic visual concepts with a constrained variational framework. In: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France (2017)"},{"key":"28_CR26","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1007\/978-3-030-01249-6_2","volume-title":"Computer Vision \u2013 ECCV 2018","author":"T Hoda\u0148","year":"2018","unstructured":"Hoda\u0148, T., et al.: BOP: Benchmark for 6d object pose estimation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11214, pp. 19\u201335. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01249-6_2"},{"key":"28_CR27","doi-asserted-by":"crossref","unstructured":"Hoda\u0148, T., et al.: Photorealistic image synthesis for object instance detection. In: ICIP (2019)","DOI":"10.1109\/ICIP.2019.8803821"},{"key":"28_CR28","doi-asserted-by":"publisher","unstructured":"Ilg, E., Mayer, N., Saikia, T., Keuper, M., Dosovitskiy, A., Brox, T.: Flownet 2.0: Evolution of optical flow estimation with deep networks. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21\u201326 July 2017, pp. 1647\u20131655 (2017). https:\/\/doi.org\/10.1109\/CVPR.2017.179","DOI":"10.1109\/CVPR.2017.179"},{"key":"28_CR29","doi-asserted-by":"crossref","unstructured":"Jang, W., Agapito, L.: Codenerf: Disentangled neural radiance fields for object categories. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 12949\u201312958 (2021)","DOI":"10.1109\/ICCV48922.2021.01271"},{"key":"28_CR30","doi-asserted-by":"crossref","unstructured":"Kar, A., et al.: Meta-sim: Learning to generate synthetic datasets. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00465"},{"key":"28_CR31","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Doll\u00e1r, P.: Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2980\u20132988 (2017)","DOI":"10.1109\/ICCV.2017.324"},{"key":"28_CR32","unstructured":"Louppe, G., Cranmer, K.: Adversarial variational optimization of non-differentiable simulators. In: AISTATS (2019)"},{"key":"28_CR33","doi-asserted-by":"crossref","unstructured":"Martin-Brualla, R., Radwan, N., Sajjadi, M.S., Barron, J.T., Dosovitskiy, A., Duckworth, D.: Nerf in the wild: Neural radiance fields for unconstrained photo collections. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7210\u20137219 (2021)","DOI":"10.1109\/CVPR46437.2021.00713"},{"key":"28_CR34","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: Representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"28_CR35","unstructured":"Ng, A.: Mlops: From model-centric to data-centric ai. https:\/\/www.deeplearning.ai\/wp-content\/uploads\/2021\/06\/MLOps-From-Model-centric-to-Data-centric-AI.pdf"},{"key":"28_CR36","doi-asserted-by":"crossref","unstructured":"Park, K., et al.: Nerfies: Deformable neural radiance fields. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00581"},{"key":"28_CR37","doi-asserted-by":"crossref","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: towards real-time object detection with region proposal networks. In: PAMI (2017)","DOI":"10.1109\/TPAMI.2016.2577031"},{"key":"28_CR38","doi-asserted-by":"crossref","unstructured":"Richter, S.R., Hayder, Z., Koltun, V.: Playing for benchmarks. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.243"},{"key":"28_CR39","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1007\/978-3-319-46475-6_7","volume-title":"Computer Vision \u2013 ECCV 2016","author":"SR Richter","year":"2016","unstructured":"Richter, S.R., Vineet, V., Roth, S., Koltun, V.: Playing for data: Ground truth from computer games. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 102\u2013118. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_7"},{"key":"28_CR40","doi-asserted-by":"crossref","unstructured":"Ros, G., Sellart, L., Materzynska, J., V\u00e1zquez, D., L\u00f3pez, A.M.: The SYNTHIA dataset: A large collection of synthetic images for semantic segmentation of urban scenes. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.352"},{"key":"28_CR41","unstructured":"Ruiz, N., Schulter, S., Chandraker, M.: Learning to simulate. In: ICLR (2019)"},{"key":"28_CR42","doi-asserted-by":"crossref","unstructured":"Shelhamer, E., Long, J., Darrell, T.: Fully convolutional networks for semantic segmentation. In: PAMI (2017)","DOI":"10.1109\/TPAMI.2016.2572683"},{"key":"28_CR43","doi-asserted-by":"crossref","unstructured":"Srinivasan, P.P., Deng, B., Zhang, X., Tancik, M., Mildenhall, B., Barron, J.T.: Nerv: Neural reflectance and visibility fields for relighting and view synthesis. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00741"},{"key":"28_CR44","doi-asserted-by":"crossref","unstructured":"Tremblay, J., To, T., Birchfield, S.: Falling things: A synthetic dataset for 3d object detection and pose estimation. In: CVPR (2018)","DOI":"10.1109\/CVPRW.2018.00275"},{"key":"28_CR45","doi-asserted-by":"publisher","first-page":"229","DOI":"10.1007\/BF00992696","volume":"8","author":"RJ Williams","year":"1992","unstructured":"Williams, R.J.: Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach. Learn. 8, 229\u2013256 (1992)","journal-title":"Mach. Learn."},{"key":"28_CR46","doi-asserted-by":"crossref","unstructured":"Xiang, Y., Schmidt, T., Narayanan, V., Fox, D.: Posecnn: A convolutional neural network for 6d object pose estimation in cluttered scenes. arXiv preprint arXiv:1711.00199 (2017)","DOI":"10.15607\/RSS.2018.XIV.019"},{"key":"28_CR47","unstructured":"Xiaogang, X.u., Ying-Cong Chen, J.J.: View independent generative adversarial network for novel view synthesis. In: ICCV (2019)"},{"key":"28_CR48","doi-asserted-by":"crossref","unstructured":"Yang, D., Deng, J.: Learning to generate synthetic 3d training data through hybrid gradient. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00086"},{"key":"28_CR49","unstructured":"Yen-Chen, L.: Nerf-pytorch. https:\/\/github.com\/yenchenlin\/nerf-pytorch\/ (2020)"},{"issue":"6","key":"28_CR50","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3478513.3480500","volume":"40","author":"X Zhang","year":"2021","unstructured":"Zhang, X., Srinivasan, P.P., Deng, B., Debevec, P., Freeman, W.T., Barron, J.T.: Nerfactor: Neural factorization of shape and reflectance under an unknown illumination. ACM Trans. Graph. (TOG) 40(6), 1\u201318 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"28_CR51","doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: Physically-based rendering for indoor scene understanding using convolutional neural networks. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.537"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-20050-2_28","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,27]],"date-time":"2022-10-27T22:27:53Z","timestamp":1666909673000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-20050-2_28"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031200496","9783031200502"],"references-count":51,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-20050-2_28","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"28 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}