{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,29]],"date-time":"2025-11-29T08:01:23Z","timestamp":1764403283886,"version":"3.37.3"},"reference-count":31,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2023,8,2]],"date-time":"2023-08-02T00:00:00Z","timestamp":1690934400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,8,2]],"date-time":"2023-08-02T00:00:00Z","timestamp":1690934400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100002347","name":"Bundesministerium f\u00fcr Bildung und Forschung","doi-asserted-by":"publisher","award":["13GW0389C"],"award-info":[{"award-number":["13GW0389C"]}],"id":[{"id":"10.13039\/501100002347","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002347","name":"Bundesministerium f\u00fcr Bildung und Forschung","doi-asserted-by":"publisher","award":["13GW0389C"],"award-info":[{"award-number":["13GW0389C"]}],"id":[{"id":"10.13039\/501100002347","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100011699","name":"Siemens Healthineers","doi-asserted-by":"publisher","award":["13GW0389B"],"award-info":[{"award-number":["13GW0389B"]}],"id":[{"id":"10.13039\/501100011699","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Image Video Proc."],"abstract":"<jats:title>Abstract<\/jats:title><jats:sec>\n                <jats:title>Purpose<\/jats:title>\n                <jats:p>The availability of real data from areas with high privacy requirements, such as the medical intervention space is low and the acquisition complex in terms of data protection. To enable research for assistance systems in the medical intervention room, new methods for data generation for these areas must be researched. Therefore, this work presents a way to create a synthetic dataset for the medical context, using medical clothing object detection as an example. The goal is to close the reality gap between the synthetic and real data.<\/jats:p>\n              <\/jats:sec><jats:sec>\n                <jats:title>Methods<\/jats:title>\n                <jats:p>Methods of 3D-scanned clothing and designed clothing are compared in a Domain-Randomization and Structured-Domain-Randomization scenario using two different rendering engines. Additionally, a Mixed-Reality dataset in front of a greenscreen and a target domain dataset were used while the latter is used to evaluate the different datasets. The experiments conducted are to show whether scanned clothing or designed clothing produce better results in Domain Randomization and Structured Domain Randomization. Likewise, a baseline will be generated using the mixed reality data. In a further experiment it is investigated whether the combination of real, synthetic and mixed reality image data improves the accuracy compared to real data only.<\/jats:p>\n              <\/jats:sec><jats:sec>\n                <jats:title>Results<\/jats:title>\n                <jats:p>Our experiments show, that Structured-Domain-Randomization of designed clothing together with Mixed-Reality data provide a baseline achieving 72.0% mAP on the test dataset of the clinical target domain. When additionally using 15% (99 images) of available target domain train data, the gap towards 100% (660 images) target domain train data could be nearly closed 80.05% mAP (81.95% mAP). Finally, we show that when additionally using 100% target domain train data the accuracy could be increased to 83.35% mAP.<\/jats:p>\n              <\/jats:sec><jats:sec>\n                <jats:title>Conclusion<\/jats:title>\n                <jats:p>In conclusion, it can be stated that the presented modeling of health professionals is a promising methodology to address the challenge of missing datasets from medical intervention rooms. We will further investigate it on various tasks, like assistance systems, in the medical domain.<\/jats:p>\n              <\/jats:sec>","DOI":"10.1186\/s13640-023-00612-1","type":"journal-article","created":{"date-parts":[[2023,8,2]],"date-time":"2023-08-02T16:02:19Z","timestamp":1690992139000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Comparison of synthetic dataset generation methods for medical intervention rooms using medical clothing detection as an example"],"prefix":"10.1186","volume":"2023","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9792-0038","authenticated-orcid":false,"given":"Patrick","family":"Sch\u00fclein","sequence":"first","affiliation":[]},{"given":"Hannah","family":"Teufel","sequence":"additional","affiliation":[]},{"given":"Ronja","family":"Vorpahl","sequence":"additional","affiliation":[]},{"given":"Indira","family":"Emter","sequence":"additional","affiliation":[]},{"given":"Yannick","family":"Bukschat","sequence":"additional","affiliation":[]},{"given":"Marcus","family":"Pfister","sequence":"additional","affiliation":[]},{"given":"Nils","family":"Rathmann","sequence":"additional","affiliation":[]},{"given":"Steffen","family":"Diehl","sequence":"additional","affiliation":[]},{"given":"Marcus","family":"Vetter","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,8,2]]},"reference":[{"key":"612_CR1","doi-asserted-by":"publisher","unstructured":"V. Belagiannis, X. Wang, H. Beny Ben\u00a0Shitrit, K. Hashimoto, R. Stauder, Y. Aoki, M. Kranzfelder, A. Schneider, P. Fua, S. Ilic, H. Feussner, N. Navab, Parsing human skeletons in an operating room. Mach. Vis. Appl. (2016). https:\/\/doi.org\/10.1007\/s00138-016-0792-4","DOI":"10.1007\/s00138-016-0792-4"},{"key":"612_CR2","unstructured":"A. Bochkovskiy, C.-Y. Wang, H.-Y.M. Liao, Yolov4: Optimal speed and accuracy of object detection. https:\/\/arxiv.org\/pdf\/2004.10934.pdf. Accessed 24 Nov 2022"},{"key":"612_CR3","unstructured":"S. Borkman, A. Crespi, S. Dhakad, S. Ganguly, J. Hogins, Y.-C. Jhang, M. Kamalzadeh, B. Li, S. Leal, P. Parisi, C. Romero, W. Smith, A. Thaman, S. Warren, N. Yadav, Unity perception: generate synthetic data for computer vision. http:\/\/arxiv.org\/pdf\/2107.04259v2.pdf. Accessed 24 Nov 2022"},{"key":"612_CR4","doi-asserted-by":"publisher","unstructured":"K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016), IEEE. https:\/\/doi.org\/10.1109\/cvpr.2016.90","DOI":"10.1109\/cvpr.2016.90"},{"issue":"7","key":"612_CR5","doi-asserted-by":"publisher","first-page":"1325","DOI":"10.1109\/TPAMI.2013.248","volume":"36","author":"C Ionescu","year":"2014","unstructured":"C. Ionescu, D. Papava, V. Olaru, C. Sminchisescu, Human3.6m: large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE Trans. Pattern Anal. Mach. Intell. 36(7), 1325\u20131339 (2014). https:\/\/doi.org\/10.1109\/TPAMI.2013.248","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"612_CR6","unstructured":"S. James, A.J. Davison, E. Johns, Transferring end-to-end visuomotor control from simulation to real world for a multi-stage task. CoRR abs\/1707.02267 (2017). http:\/\/arxiv.org\/pdf\/1707.02267.pdf. Accessed 24 Nov 2022"},{"key":"612_CR7","doi-asserted-by":"publisher","unstructured":"Y. LeCun, 1.1 deep learning hardware: past, present, and future. In IEEE International Solid- State Circuits Conference\u2014(ISSCC) (2019). IEEE (2019). https:\/\/doi.org\/10.1109\/isscc.2019.8662396","DOI":"10.1109\/isscc.2019.8662396"},{"key":"612_CR8","unstructured":"T. Lin, M. Maire, S.J. Belongie, L.D. Bourdev, R.B. Girshick, J. Hays, P. Perona, D. Ramanan, P. Doll\u00e1r, C.L. Zitnick, Microsoft COCO: common objects in context. CoRR abs\/1405.0312 (2014). http:\/\/arxiv.org\/abs\/1405.0312.\u00a0Accessed 24 Nov 2022"},{"issue":"2","key":"612_CR9","doi-asserted-by":"publisher","first-page":"318","DOI":"10.1109\/tpami.2018.2858826","volume":"42","author":"T-Y Lin","year":"2020","unstructured":"T.-Y. Lin, P. Goyal, R. Girshick, K. He, P. Dollar, Focal loss for dense object detection. IEEE Trans. Pattern Anal. Mach. Intell. 42(2), 318\u2013327 (2020). https:\/\/doi.org\/10.1109\/tpami.2018.2858826","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"612_CR10","doi-asserted-by":"publisher","unstructured":"M. Loper, N. Mahmood, J. Romero, G. Pons-Moll, M.J. Black, SMPL: A skinned multi-person linear model. ACM Trans. Graphics (Proc. SIGGRAPH Asia) 34(6), 24:81\u201324:816 (2015). https:\/\/doi.org\/10.1145\/2816795.2818013","DOI":"10.1145\/2816795.2818013"},{"key":"612_CR11","doi-asserted-by":"publisher","unstructured":"A.A.A. Osman, T. Bolkart, M.J. Black, STAR: A sparse trained articulated human body regressor. In European Conference on Computer Vision (ECCV), pp.\u00a0598\u2013613 (2020). https:\/\/doi.org\/10.1007\/978-3-030-58539-6_36","DOI":"10.1007\/978-3-030-58539-6_36"},{"key":"612_CR12","doi-asserted-by":"publisher","unstructured":"A. Prakash, S. Boochoon, M. Brophy, D. Acuna, E. Cameracci, G. State, O. Shapira, S. Birchfield, Structured domain randomization: Bridging the reality gap by context-aware synthetic data. In 2019 International Conference on Robotics and Automation (ICRA) (2019), IEEE. https:\/\/doi.org\/10.1109\/icra.2019.8794443","DOI":"10.1109\/icra.2019.8794443"},{"key":"612_CR13","unstructured":"K. Robinette, S. Blackwell, H. Daanen, M. Boehmer, S. Fleming, Civilian American and European surface anthropometry resource (caesar), final report. volume 1. summary. 74. https:\/\/www.humanics-es.com\/CAESARvol1.pdf. Accessed 24 Nov 2022"},{"key":"612_CR14","doi-asserted-by":"publisher","unstructured":"V.F. Rodrigues, R.S. Antunes, L.A. Seewald, R. Bazo, E.S. dos Reis, U.J. dos Santos, R. da\u00a0R.\u00a0Righi, L.G. da S., C.A. da Costa, F.L. Bertollo, A. Maier, B. Eskofier, T. Horz, M. Pfister, R. Fahrig, A multi-sensor architecture combining human pose estimation and real-time location systems for workflow monitoring on hybrid operating suites. Future Gener. Comput. Syst. 135, 283\u2013298 (2022). https:\/\/doi.org\/10.1016\/j.future.2022.05.006","DOI":"10.1016\/j.future.2022.05.006"},{"key":"612_CR15","doi-asserted-by":"publisher","unstructured":"J. Romero, D. Tzionas, M.J. Black, Embodied hands: Modeling and capturing hands and bodies together. ACM Trans. Graph. (Proc. SIGGRAPH Asia) 36, 6 (2017). https:\/\/doi.org\/10.1145\/3130800.3130883","DOI":"10.1145\/3130800.3130883"},{"issue":"3","key":"612_CR16","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A.C. Berg, L. Fei-Fei, ImageNet large scale visual recognition challenge. Int. J. Comput. Vis. (IJCV) 115(3), 211\u2013252 (2015). https:\/\/doi.org\/10.1007\/s11263-015-0816-y","journal-title":"Int. J. Comput. Vis. (IJCV)"},{"key":"612_CR17","doi-asserted-by":"publisher","unstructured":"F. Sadeghi, S. Levine, CAD2RL: Real single-image flight without a single real image. In Robotics: Science and Systems XIII (2017), Robotics: Science and Systems Foundation. https:\/\/doi.org\/10.15607\/rss.2017.xiii.034","DOI":"10.15607\/rss.2017.xiii.034"},{"key":"612_CR18","doi-asserted-by":"publisher","unstructured":"A. Sharghi, H. Haugerud, D. Oh, O. Mohareri, Automatic operating room surgical activity recognition for robot-assisted surgery. CoRR abs\/2006.16166 (2020). https:\/\/doi.org\/10.1007\/978-3-030-59716-0_37","DOI":"10.1007\/978-3-030-59716-0_37"},{"key":"612_CR19","unstructured":"V. Srivastav, T. Issenhuth, K. Abdolrahim, M. de\u00a0Mathelin, A. Gangi, N. Padoy, Mvor: A multi-view rgb-d operating room dataset for 2d and 3d human pose estimation"},{"key":"612_CR20","doi-asserted-by":"publisher","unstructured":"V. Sze, Y.-H. Chen, J. Emer, A. Suleiman, Z. Hardware. Zhang, for machine learning: Challenges and opportunities. In, IEEE Custom Integrated Circuits Conference (CICC) (apr 2018). IEEE (2018). https:\/\/doi.org\/10.1109\/cicc.2018.8357072","DOI":"10.1109\/cicc.2018.8357072"},{"key":"612_CR21","doi-asserted-by":"publisher","unstructured":"M. Tan, R. Pang, Q.V. Le, EfficientDet: Scalable and efficient object detection. In 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020), IEEE. https:\/\/doi.org\/10.1109\/cvpr42600.2020.01079","DOI":"10.1109\/cvpr42600.2020.01079"},{"key":"612_CR22","unstructured":"U. Technologies, ML-ImageSynthesis, 2017. https:\/\/bitbucket.org\/Unity-Technologies\/ml-imagesynthesis\/src\/master\/. Accessed 24 Nov 2022"},{"key":"612_CR23","unstructured":"T. To, J. Tremblay, D. McKay, Y. Yamaguchi, K. Leung, A. Balanon, J. Cheng, W. Hodge, S. Birchfield, NDDS: NVIDIA Deep Learning Dataset Synthesizer (2018). https:\/\/github.com\/NVIDIA\/Dataset_Synthesizer. Accessed 24 Nov 2022"},{"key":"612_CR24","doi-asserted-by":"publisher","unstructured":"J. Tobin, R. Fong, A. Ray, J. Schneider, W. Zaremba, P. Abbeel, Domain randomization for transferring deep neural networks from simulation to the real world. In 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS) (2017), IEEE. https:\/\/doi.org\/10.1109\/iros.2017.8202133","DOI":"10.1109\/iros.2017.8202133"},{"key":"612_CR25","doi-asserted-by":"crossref","unstructured":"J. Tremblay, A. Prakash, D. Acuna, M. Brophy, V. Jampani, C. Anil, T. To, E. Cameracci, S. Boochoon, S. Birchfield, Training deep networks with synthetic data: Bridging the reality gap by domain randomization. 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 1082\u201310828 (2018). https:\/\/openaccess.thecvf.com\/content_cvpr_2018_workshops\/papers\/w14\/Tremblay_Training_Deep_Networks_CVPR_2018_paper.pdf, Accessed: 24.11.2022","DOI":"10.1109\/CVPRW.2018.00143"},{"key":"612_CR26","doi-asserted-by":"publisher","unstructured":"J. Tremblay, T. To, S. Birchfield, Falling things: A synthetic dataset for 3d object detection and pose estimation. In 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (2018), IEEE. https:\/\/doi.org\/10.1109\/cvprw.2018.00275","DOI":"10.1109\/cvprw.2018.00275"},{"key":"612_CR27","unstructured":"J. Tremblay, T. To, B. Sundaralingam, Y. Xiang, D. Fox, S. Birchfield, Deep object pose estimation for semantic robotic grasping of household objects. arXiv preprint http:\/\/arxiv.org\/abs\/1809.10790 (2018). Accessed 24 Nov 2022"},{"issue":"6","key":"612_CR28","doi-asserted-by":"publisher","first-page":"737","DOI":"10.1007\/s11548-015-1186-1","volume":"10","author":"AP Twinanda","year":"2015","unstructured":"A.P. Twinanda, E.O. Alkan, A. Gangi, M. de Mathelin, N. Padoy, Data-driven spatio-temporal RGBD feature encoding for action recognition in operating rooms. Int. J. Comput. Assist. Radiol. Surg. 10(6), 737\u2013747 (2015). https:\/\/doi.org\/10.1007\/s11548-015-1186-1","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"612_CR29","doi-asserted-by":"crossref","unstructured":"G. Varol, J. Romero, X. Martin, N. Mahmood, M.J. Black, I. Laptev, C. Schmid, Learning from synthetic humans. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 4627\u20134635 (2017). https:\/\/openaccess.thecvf.com\/content_cvpr_2017\/papers\/Varol_Learning_From_Synthetic_CVPR_2017_paper.pdf. Accessed 24 Nov 2022","DOI":"10.1109\/CVPR.2017.492"},{"key":"612_CR30","doi-asserted-by":"crossref","unstructured":"C.-Y. Wang, A. Bochkovskiy, H.-Y.M. Liao, Scaled-YOLOv4: Scaling cross stage partial network. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a013029\u201313038 (2021). https:\/\/openaccess.thecvf.com\/content\/CVPR2021\/papers\/Wang_Scaled-YOLOv4_Scaling_Cross_Stage_Partial_Network_CVPR_2021_paper.pdf. Accessed 24 Nov 2022","DOI":"10.1109\/CVPR46437.2021.01283"},{"key":"612_CR31","unstructured":"F. Zhang, J. Leitner, M. Milford, P. Corke, Sim-to-real transfer of visuo-motor policies for reaching in clutter: Domain randomization and adaptation with modular networks. CoRR abs\/1709.05746v1 (2017). https:\/\/arxiv.org\/pdf\/1709.05746v1.pdf. Accessed 24 Nov 2022"}],"container-title":["EURASIP Journal on Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s13640-023-00612-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1186\/s13640-023-00612-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s13640-023-00612-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,2]],"date-time":"2023-11-02T13:03:28Z","timestamp":1698930208000},"score":1,"resource":{"primary":{"URL":"https:\/\/jivp-eurasipjournals.springeropen.com\/articles\/10.1186\/s13640-023-00612-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,2]]},"references-count":31,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2023,12]]}},"alternative-id":["612"],"URL":"https:\/\/doi.org\/10.1186\/s13640-023-00612-1","relation":{},"ISSN":["1687-5281"],"issn-type":[{"type":"electronic","value":"1687-5281"}],"subject":[],"published":{"date-parts":[[2023,8,2]]},"assertion":[{"value":"14 February 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 July 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 August 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"This article does not contain patient data. Informed consent was obtained from all individual participants included in the study.The methods and information presented in this work are based on research and are not commercially available.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}},{"value":"The authors declare that they have no conflict of interest.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"12"}}