{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,20]],"date-time":"2026-02-20T18:44:32Z","timestamp":1771613072207,"version":"3.50.1"},"reference-count":36,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2019,8,13]],"date-time":"2019-08-13T00:00:00Z","timestamp":1565654400000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2019,8,13]],"date-time":"2019-08-13T00:00:00Z","timestamp":1565654400000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100003176","name":"Ministerio de Educaci\u00f3n, Cultura y Deporte","doi-asserted-by":"publisher","award":["FPU15\/04516"],"award-info":[{"award-number":["FPU15\/04516"]}],"id":[{"id":"10.13039\/501100003176","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003176","name":"Ministerio de Educaci\u00f3n, Cultura y Deporte","doi-asserted-by":"publisher","award":["FPU17\/00166"],"award-info":[{"award-number":["FPU17\/00166"]}],"id":[{"id":"10.13039\/501100003176","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003359","name":"Generalitat Valenciana","doi-asserted-by":"publisher","award":["ACIF\/2018\/197"],"award-info":[{"award-number":["ACIF\/2018\/197"]}],"id":[{"id":"10.13039\/501100003359","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100010198","name":"Ministerio de Econom\u00eda, Industria y Competitividad, Gobierno de Espa\u00f1a","doi-asserted-by":"crossref","award":["TIN2016- 76515-R"],"award-info":[{"award-number":["TIN2016- 76515-R"]}],"id":[{"id":"10.13039\/501100010198","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100003359","name":"Generalitat Valenciana","doi-asserted-by":"publisher","award":["GV\/2018\/022"],"award-info":[{"award-number":["GV\/2018\/022"]}],"id":[{"id":"10.13039\/501100003359","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Virtual Reality"],"published-print":{"date-parts":[[2020,6]]},"DOI":"10.1007\/s10055-019-00399-5","type":"journal-article","created":{"date-parts":[[2019,8,13]],"date-time":"2019-08-13T09:02:42Z","timestamp":1565686962000},"page":"271-288","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":69,"title":["UnrealROX: an extremely photorealistic virtual reality environment for robotics simulations and synthetic data generation"],"prefix":"10.1007","volume":"24","author":[{"given":"Pablo","family":"Martinez-Gonzalez","sequence":"first","affiliation":[]},{"given":"Sergiu","family":"Oprea","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9575-6403","authenticated-orcid":false,"given":"Alberto","family":"Garcia-Garcia","sequence":"additional","affiliation":[]},{"given":"Alvaro","family":"Jover-Alvarez","sequence":"additional","affiliation":[]},{"given":"Sergio","family":"Orts-Escolano","sequence":"additional","affiliation":[]},{"given":"Jose","family":"Garcia-Rodriguez","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,8,13]]},"reference":[{"key":"399_CR1","unstructured":"Bhoi A (2019) Monocular depth estimation: a survey. arXiv preprint \narXiv:1901-09402"},{"key":"399_CR2","unstructured":"Bousmalis K, Irpan A, Wohlhart P, Bai Y, Kelcey M, Kalakrishnan M, Downs L, Ibarz J, Pastor P, Konolige K et\u00a0al (2017) Using simulation and domain adaptation to improve efficiency of deep robotic grasping. arXiv preprint \narXiv:1709.07857"},{"key":"399_CR3","unstructured":"Brodeur S, Perez E, Anand A, Golemo F, Celotti L, Strub F, Rouat J, Larochelle H, Courville A (2017) Home: a household multimodal environment. arXiv preprint \narXiv:1711.11017"},{"key":"399_CR4","doi-asserted-by":"crossref","unstructured":"Butler DJ, Wulff J, Stanley GB, Black MJ (2012) A naturalistic open source movie for optical flow evaluation. In: Proceedings of the European conference on computer vision (ECCV), pp 611\u2013625","DOI":"10.1007\/978-3-642-33783-3_44"},{"key":"399_CR5","doi-asserted-by":"crossref","unstructured":"Eigen D, Fergus R (2015) Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture. In: Proceedings of the IEEE international conference on computer vision (ICCV), pp 2650\u20132658","DOI":"10.1109\/ICCV.2015.304"},{"key":"399_CR6","unstructured":"Eigen D, Puhrsch C, Fergus R (2014) Depth map prediction from a single image using a multi-scale deep network. In: Advances in neural information processing systems (NIPS), pp 2366\u20132374"},{"key":"399_CR7","unstructured":"Gaidon A, Wang Q, Cabon Y, Vig E (2016) Virtual worlds as proxy for multi-object tracking analysis. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 4340\u20134349"},{"key":"399_CR8","doi-asserted-by":"crossref","unstructured":"He K, Gkioxari G, Doll\u00e1r P, Girshick R (2017) Mask r-cnn. In: Proceedings of the IEEE international conference on computer vision (CVPR), pp 2961\u20132969","DOI":"10.1109\/ICCV.2017.322"},{"key":"399_CR9","unstructured":"Kolve E, Mottaghi R, Gordon D, Zhu Y, Gupta A, Farhadi A (2017) Ai2-thor: an interactive 3d environment for visual ai. arXiv preprint \narXiv:1712.05474"},{"key":"399_CR10","doi-asserted-by":"crossref","unstructured":"Laina I, Rupprecht C, Belagiannis V, Tombari F, Navab N (2016) Deeper depth prediction with fully convolutional residual networks. In: Proceedings of the IEEE conference on 3D vision (3DV), pp 239\u2013248","DOI":"10.1109\/3DV.2016.32"},{"issue":"7553","key":"399_CR11","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436","journal-title":"Nature"},{"issue":"4\u20135","key":"399_CR12","doi-asserted-by":"publisher","first-page":"705","DOI":"10.1177\/0278364914549607","volume":"34","author":"I Lenz","year":"2015","unstructured":"Lenz I, Lee H, Saxena A (2015) Deep learning for detecting robotic grasps. Int J Robot Res 34(4\u20135):705\u2013724","journal-title":"Int J Robot Res"},{"issue":"4\u20135","key":"399_CR13","doi-asserted-by":"publisher","first-page":"421","DOI":"10.1177\/0278364917710318","volume":"37","author":"S Levine","year":"2018","unstructured":"Levine S, Pastor P, Krizhevsky A, Ibarz J, Quillen D (2018) Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection. Int J Robot Res 37(4\u20135):421\u2013436","journal-title":"Int J Robot Res"},{"key":"399_CR14","doi-asserted-by":"crossref","unstructured":"Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 3431\u20133440","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"399_CR15","unstructured":"Looman T (2017) Vr template. \nhttps:\/\/wiki.unrealengine.com\/VR_Template\n\n. Accessed 1 Sept 2018"},{"key":"399_CR16","unstructured":"Mahler J, Liang J, Niyaz S, Laskey M, Doan R, Liu X, Ojea JA, Goldberg K (2017) Dex-net 2.0: deep learning to plan robust grasps with synthetic point clouds and analytic grasp metrics. arXiv preprint \narXiv:1703.09312"},{"key":"399_CR17","unstructured":"McCormac J, Handa A, Leutenegger S, Davison AJ (2016) Scenenet rgb-d: 5m photorealistic images of synthetic indoor trajectories with ground truth. arXiv preprint \narXiv:1612.05079"},{"key":"399_CR18","unstructured":"Oculus (2017a) Distance grab sample now available in oculus unity sample framework. \nhttps:\/\/developer.oculus.com\/blog\/distance-grab-sample-now-available-in-oculus-unity-sample-framework\/\n\n. Accessed 1 Sept 2018"},{"key":"399_CR19","unstructured":"Oculus (2017b) Oculus first contact. \nhttps:\/\/www.oculus.com\/experiences\/rift\/1217155751659625\/\n\n. Accessed 1 Sept 2018"},{"key":"399_CR20","doi-asserted-by":"crossref","unstructured":"Pashevich A, Strudel R, Kalevatykh I, Laptev I, Schmid C (2019) Learning to augment synthetic images for sim2real policy transfer. arXiv preprint \narXiv:1903.07740","DOI":"10.1109\/IROS40897.2019.8967622"},{"key":"399_CR21","doi-asserted-by":"crossref","unstructured":"Qiu W, Yuille A (2016) Unrealcv: connecting computer vision to unreal engine. In: Proceedings of the European conference on computer vision (ECCV), pp 909\u2013916","DOI":"10.1007\/978-3-319-49409-8_75"},{"key":"399_CR22","doi-asserted-by":"crossref","unstructured":"Qiu W, Zhong F, Zhang Y, Qiao S, Xiao Z, Kim TS, Wang Y (2017) Unrealcv: virtual worlds for computer vision. In: Proceedings of the 2017 ACM on multimedia conference (ACMMM), pp 1221\u20131224","DOI":"10.1145\/3123266.3129396"},{"key":"399_CR23","doi-asserted-by":"publisher","unstructured":"Redmon J, Farhadi A (2017) YOLO9000: better, faster, stronger. In: Proceedings\u201430th IEEE conference on computer vision and pattern recognition, CVPR 2017 2017-Janua, pp 6517\u20136525. \nhttps:\/\/doi.org\/10.1109\/CVPR.2017.690","DOI":"10.1109\/CVPR.2017.690"},{"key":"399_CR24","doi-asserted-by":"publisher","unstructured":"Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: unified, real-time object detection. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition. \nhttps:\/\/doi.org\/10.1109\/CVPR.2016.91","DOI":"10.1109\/CVPR.2016.91"},{"key":"399_CR25","doi-asserted-by":"crossref","unstructured":"Ros G, Sellart L, Materzynska J, Vazquez D, Lopez AM (2016) The synthia dataset: a large collection of synthetic images for semantic segmentation of urban scenes. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 3234\u20133243","DOI":"10.1109\/CVPR.2016.352"},{"key":"399_CR26","unstructured":"Savva M, Chang AX, Dosovitskiy A, Funkhouser T, Koltun V (2017) Minos: multimodal indoor simulator for navigation in complex environments. arXiv preprint \narXiv:1712.03931"},{"key":"399_CR27","doi-asserted-by":"crossref","unstructured":"Silberman N, Hoiem D, Kohli P, Fergus R (2012) Indoor segmentation and support inference from rgbd images. In: Proceedings of the European conference on computer vision (ECCV), pp 746\u2013760","DOI":"10.1007\/978-3-642-33715-4_54"},{"key":"399_CR28","doi-asserted-by":"publisher","unstructured":"Tekin B, Sinha SN, Fua P (2018) Real-time seamless single shot 6d object pose prediction. In: 2018 IEEE\/CVF conference on computer vision and pattern recognition, pp 292\u2013301. \nhttps:\/\/doi.org\/10.1109\/CVPR.2018.00038","DOI":"10.1109\/CVPR.2018.00038"},{"key":"399_CR29","unstructured":"To T, Tremblay J, McKay D, Yamaguchi Y, Leung K, Balanon A, Cheng J, Birchfield S (2018) NDDS: NVIDIA deep learning dataset synthesizer. \nhttps:\/\/github.com\/NVIDIA\/Dataset_Synthesizer"},{"key":"399_CR30","doi-asserted-by":"crossref","unstructured":"Tobin J, Fong R, Ray A, Schneider J, Zaremba W, Abbeel P (2017a) Domain randomization for transferring deep neural networks from simulation to the real world. In: Proceedings of the IEEE international conference on intelligent robots and systems (IROS), pp 23\u201330","DOI":"10.1109\/IROS.2017.8202133"},{"key":"399_CR31","doi-asserted-by":"crossref","unstructured":"Tobin J, Zaremba W, Abbeel P (2017b) Domain randomization and generative models for robotic grasping. arXiv preprint \narXiv:1710.06425","DOI":"10.1109\/IROS.2018.8593933"},{"key":"399_CR32","doi-asserted-by":"crossref","unstructured":"Tremblay J, Prakash A, Acuna D, Brophy M, Jampani V, Anil C, To T, Cameracci E, Boochoon S, Birchfield S (2018) Training deep networks with synthetic data: bridging the reality gap by domain randomization. arXiv preprint \narXiv:1804.06516","DOI":"10.1109\/CVPRW.2018.00143"},{"key":"399_CR33","doi-asserted-by":"crossref","unstructured":"Ummenhofer B, Zhou H, Uhrig J, Mayer N, Ilg E, Dosovitskiy A, Brox T (2017) Demon: depth and motion network for learning monocular stereo. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 5038\u20135047","DOI":"10.1109\/CVPR.2017.596"},{"key":"399_CR34","doi-asserted-by":"crossref","unstructured":"Xia F, Zamir RA, He ZY, Sax A, Malik J, Savarese S (2018) Gibson env: real-world perception for embodied agents. In: Proceedings of the IEEE computer vision and pattern recognition (CVPR)","DOI":"10.1109\/CVPR.2018.00945"},{"key":"399_CR35","doi-asserted-by":"crossref","unstructured":"Xu D, Wang W, Tang H, Liu H, Sebe N, Ricci E (2018) Structured attention guided convolutional neural fields for monocular depth estimation. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 3917\u20133925","DOI":"10.1109\/CVPR.2018.00412"},{"key":"399_CR36","unstructured":"Yan C, Misra D, Bennnett A, Walsman A, Bisk Y, Artzi Y (2018) Chalet: cornell house agent learning environment. arXiv preprint \narXiv:1801.07357"}],"container-title":["Virtual Reality"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s10055-019-00399-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s10055-019-00399-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s10055-019-00399-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2020,8,11]],"date-time":"2020-08-11T23:30:46Z","timestamp":1597188646000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s10055-019-00399-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,8,13]]},"references-count":36,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2020,6]]}},"alternative-id":["399"],"URL":"https:\/\/doi.org\/10.1007\/s10055-019-00399-5","relation":{},"ISSN":["1359-4338","1434-9957"],"issn-type":[{"value":"1359-4338","type":"print"},{"value":"1434-9957","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019,8,13]]},"assertion":[{"value":"19 September 2018","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 August 2019","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 August 2019","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}