{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T19:17:35Z","timestamp":1742930255398,"version":"3.40.3"},"publisher-location":"Cham","reference-count":37,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031479687"},{"type":"electronic","value":"9783031479694"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-47969-4_25","type":"book-chapter","created":{"date-parts":[[2023,11,30]],"date-time":"2023-11-30T20:02:06Z","timestamp":1701374526000},"page":"315-328","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["UniTorch - Integrating Neural Rendering into\u00a0Unity"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-8950-1790","authenticated-orcid":false,"given":"Laura","family":"Fink","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0000-7643-2094","authenticated-orcid":false,"given":"Daniel","family":"Keitel","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8699-3442","authenticated-orcid":false,"given":"Marc","family":"Stamminger","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1857-3862","authenticated-orcid":false,"given":"Joachim","family":"Keinert","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,1]]},"reference":[{"key":"25_CR1","unstructured":"Game On: Game-Engine Technology Expands Filmmaking Horizons - The American Society of Cinematographers (en-US). https:\/\/tinyurl.com\/2txhdvac"},{"key":"25_CR2","unstructured":"NeuralNetworkInference: Unreal Engine Doc. https:\/\/tinyurl.com\/aa2ryfej"},{"key":"25_CR3","unstructured":"Official - Introducing Unity Muse and Unity Sentis, AI-Powered Creativity. https:\/\/tinyurl.com\/35c8eev8"},{"key":"25_CR4","unstructured":"ONNX. https:\/\/onnx.ai\/"},{"key":"25_CR5","unstructured":"Torch\u2014PyTorch 2.0 documentation. https:\/\/pytorch.org\/docs\/stable\/torch.html"},{"key":"25_CR6","unstructured":"Unreal Engine 5.2 Release Notes. https:\/\/tinyurl.com\/yx5e6ba3"},{"key":"25_CR7","unstructured":"Unity Barracuda. Unity Technologies (2021)"},{"key":"25_CR8","unstructured":"PyTorch Realtime Style Transfer Model in Unreal Engine 5 with ONNX Runtime (2022)"},{"key":"25_CR9","unstructured":"Neural Network Engine (NNE). Course (2023). https:\/\/tinyurl.com\/fejs7pem"},{"key":"25_CR10","unstructured":"Unreal Engine. Wikipedia (2023)"},{"key":"25_CR11","unstructured":"Bo\u017ei\u010d, A., Gladkov, D., Doukakis, L., Lassner, C.: Neural assets: volumetric object capture and rendering for interactive environments (2022)"},{"key":"25_CR12","doi-asserted-by":"crossref","unstructured":"Chen, Z., Funkhouser, T., Hedman, P., Tagliasacchi, A.: MobileNeRF: exploiting the polygon rasterization pipeline for efficient neural field rendering on mobile architectures. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16569\u201316578 (2023)","DOI":"10.1109\/CVPR52729.2023.01590"},{"key":"25_CR13","unstructured":"Frames, W.: Bringing Deep Learning to Unreal Engine 5\u2014Pt. 2 (2022)"},{"key":"25_CR14","doi-asserted-by":"crossref","unstructured":"Fridovich-Keil, S., Yu, A., Tancik, M., Chen, Q., Recht, B., Kanazawa, A.: Plenoxels: radiance fields without neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5501\u20135510 (2022)","DOI":"10.1109\/CVPR52688.2022.00542"},{"key":"25_CR15","unstructured":"Gao, K., Gao, Y., He, H., Lu, D., Xu, L., Li, J.: NeRF: neural radiance field in 3D vision, a comprehensive review (2022)"},{"key":"25_CR16","unstructured":"Guttenberg, N.: Neural networks in unity using native libraries. https:\/\/www.goodai.com\/neural-networks-in-unity-using-native-libraries\/"},{"key":"25_CR17","doi-asserted-by":"crossref","unstructured":"Hedman, P., Srinivasan, P.P., Mildenhall, B., Barron, J.T., Debevec, P.: Baking neural radiance fields for real-time view synthesis. In: IEEE\/CVF International Conference on Computer Vision (ICCV), Virtual, pp. 5875\u20135884 (2021)","DOI":"10.1109\/ICCV48922.2021.00582"},{"key":"25_CR18","doi-asserted-by":"crossref","unstructured":"Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.167"},{"key":"25_CR19","unstructured":"Kipp, J.: MobileNeRF in Unity (2022). https:\/\/t.co\/SslORxUbFJ"},{"key":"25_CR20","unstructured":"Kipp, J.: SNeRG Unity Viewer (2023). https:\/\/github.com\/julienkay\/SNeRG-Unity-Viewer"},{"key":"25_CR21","unstructured":"Lavik, M.: UnityVolumeRendering (2023). https:\/\/tinyurl.com\/bdf8vhp3"},{"key":"25_CR22","doi-asserted-by":"crossref","unstructured":"Li, K., et al.: Bringing instant neural graphics primitives to immersive virtual reality. In: 2023 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW), Shanghai, China, pp. 739\u2013740. IEEE (2023)","DOI":"10.1109\/VRW58643.2023.00212"},{"key":"25_CR23","doi-asserted-by":"crossref","unstructured":"Marschner, S., Shirley, P.: Fundamentals of Computer Graphics. CRC Press, Taylor & Francis Group (2022)","DOI":"10.1201\/9781003050339"},{"key":"25_CR24","doi-asserted-by":"crossref","unstructured":"Marshall, C.S.: Practical machine learning for rendering: from research to deployment. In: ACM SIGGRAPH 2021 Courses, Virtual Event, USA, pp. 1\u2013239. ACM (2021)","DOI":"10.1145\/3450508.3464564"},{"key":"25_CR25","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"25_CR26","unstructured":"Mills, C.: Create a libtorch plugin for unity. https:\/\/christianjmills.com\/posts\/fastai-libtorch-unity-tutorial\/part-1\/"},{"issue":"4","key":"25_CR27","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530127","volume":"41","author":"T M\u00fcller","year":"2022","unstructured":"M\u00fcller, T., Evans, A., Schied, C., Keller, A.: Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. 41(4), 1\u201315 (2022)","journal-title":"ACM Trans. Graph."},{"key":"25_CR28","doi-asserted-by":"crossref","unstructured":"Pokhrel, C., Khatiwada, A.: Deep Q-learning for intelligent non-playable characters in combat games (2023)","DOI":"10.2139\/ssrn.4358026"},{"key":"25_CR29","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"909","DOI":"10.1007\/978-3-319-49409-8_75","volume-title":"Computer Vision \u2013 ECCV 2016 Workshops","author":"W Qiu","year":"2016","unstructured":"Qiu, W., Yuille, A.: UnrealCV: connecting computer vision to unreal engine. In: Hua, G., J\u00e9gou, H. (eds.) ECCV 2016. LNCS, vol. 9915, pp. 909\u2013916. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-49409-8_75"},{"key":"25_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1007\/978-3-319-46475-6_7","volume-title":"Computer Vision \u2013 ECCV 2016","author":"SR Richter","year":"2016","unstructured":"Richter, S.R., Vineet, V., Roth, S., Koltun, V.: Playing for data: ground truth from computer games. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 102\u2013118. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_7"},{"key":"25_CR31","unstructured":"Saint-Denis, A., Vanhoey, K., Deliot, T.: Multi-stylization of video-games in real-time guided by G-buffer information. In: High Performance Graphics 2019, Strasbourg, France (2019)"},{"key":"25_CR32","doi-asserted-by":"publisher","first-page":"844","DOI":"10.1007\/978-3-031-22216-0_56","volume-title":"Intelligent Autonomous Systems 17","author":"P Szl\u0119g","year":"2023","unstructured":"Szl\u0119g, P., Barczyk, P., Maruszczak, B., Zieli\u00f1ski, S., Szyma\u00f1ska, E.: Simulation environment for underwater vehicles testing and training in Unity3D. In: Petrovic, I., Menegatti, E., Markovi\u0107, I. (eds.) IAS 2022. LNNS, vol. 577, pp. 844\u2013853. Springer, Cham (2023). https:\/\/doi.org\/10.1007\/978-3-031-22216-0_56"},{"key":"25_CR33","doi-asserted-by":"crossref","unstructured":"Tewari, A., et al.: State of the art on neural rendering. In: Computer Graphics Forum, vol. 39, no. 2, pp. 701\u2013727 (2020)","DOI":"10.1111\/cgf.14022"},{"issue":"4","key":"25_CR34","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3306346.3323035","volume":"38","author":"J Thies","year":"2019","unstructured":"Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M.: Deferred neural rendering: image synthesis using neural textures. ACM Trans. Graph. 38(4), 1\u201312 (2019)","journal-title":"ACM Trans. Graph."},{"key":"25_CR35","unstructured":"Unity Technologies: Manual: Render pipelines. https:\/\/docs.unity3d.com\/Manual\/render-pipelines.html"},{"key":"25_CR36","unstructured":"NeuralVFX: basic libtorch dll. https:\/\/github.com\/NeuralVFX\/basic-libtorch-dll"},{"key":"25_CR37","unstructured":"Yuan, L.: A Brief History of Deep Learning Frameworks (2021). https:\/\/tinyurl.com\/46zb9yfm"}],"container-title":["Lecture Notes in Computer Science","Advances in Visual Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-47969-4_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,27]],"date-time":"2023-12-27T02:57:19Z","timestamp":1703645839000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-47969-4_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031479687","9783031479694"],"references-count":37,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-47969-4_25","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"1 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ISVC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Symposium on Visual Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Lake Tahoe, NV","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"USA","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"isvc2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.isvc.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"25","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"58","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"232% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"43 (oral), 15 (poster), 25 (special tracks) out of 34 submissions","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}