{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T07:32:00Z","timestamp":1776238320863,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":51,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,9,1]],"date-time":"2023-09-01T00:00:00Z","timestamp":1693526400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-sa\/4.0\/"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CNS 1956018"],"award-info":[{"award-number":["CNS 1956018"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CSN 1956190"],"award-info":[{"award-number":["CSN 1956190"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"name":"KDDI Research, INC"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,9,10]]},"DOI":"10.1145\/3603269.3604876","type":"proceedings-article","created":{"date-parts":[[2023,9,1]],"date-time":"2023-09-01T16:16:29Z","timestamp":1693584989000},"page":"516-532","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":26,"title":["Dragonfly: Higher Perceptual Quality For Continuous 360\u00b0 Video Playback"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2254-9101","authenticated-orcid":false,"given":"Ehab","family":"Ghabashneh","sequence":"first","affiliation":[{"name":"Purdue University, West Lafayette, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-9674-4646","authenticated-orcid":false,"given":"Chandan","family":"Bothra","sequence":"additional","affiliation":[{"name":"Purdue University, West Lafayette, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8311-8853","authenticated-orcid":false,"given":"Ramesh","family":"Govindan","sequence":"additional","affiliation":[{"name":"University of Southern California, Los Angeles, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5403-0940","authenticated-orcid":false,"given":"Antonio","family":"Ortega","sequence":"additional","affiliation":[{"name":"University of Southern California, Los Angeles, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4825-4352","authenticated-orcid":false,"given":"Sanjay","family":"Rao","sequence":"additional","affiliation":[{"name":"Purdue University, West Lafayette, United States of America"}]}],"member":"320","published-online":{"date-parts":[[2023,9]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"360 video on Vimeo. https:\/\/vimeo.com\/channels\/360vr."},{"key":"e_1_3_2_1_2_1","unstructured":"The 360\u00b0 Effect | Understanding Immersive Video Research. https:\/\/magnaglobal.com\/wp-content\/uploads\/2021\/04\/Magna.IPG-Lab-YuMe-The-360-Effect.pdf."},{"key":"e_1_3_2_1_3_1","unstructured":"End-to-end optimizations for dynamic streaming. https:\/\/engineering.fb.com\/2017\/02\/22\/virtual-reality\/end-to-end-optimizations-for-dynamic-streaming\/."},{"key":"e_1_3_2_1_4_1","unstructured":"How to watch Netflix VR? https:\/\/help.netflix.com\/en\/node\/110502."},{"key":"e_1_3_2_1_5_1","unstructured":"Libavcodec Documentation. https:\/\/ffmpeg.org\/libavcodec.html\/."},{"key":"e_1_3_2_1_6_1","unstructured":"Libavformat Documentation. https:\/\/libav.org\/documentation\/doxygen\/master\/avio_8h.html\/."},{"key":"e_1_3_2_1_7_1","unstructured":"Netflix recommended bandwidth. https:\/\/help.netflix.com\/en\/node\/306."},{"key":"e_1_3_2_1_8_1","unstructured":"Oculus Airlink. https:\/\/www.oculus.com\/blog\/introducing-oculus-air-link-a-wireless-way-to-play-pc-vr-games-on-oculus-quest-2-plus-infinite-office-updates-support-for-120-hz-on-quest-2-and-more\/."},{"key":"e_1_3_2_1_9_1","unstructured":"Oculus Quest 2. https:\/\/www.oculus.com\/quest-2\/."},{"key":"e_1_3_2_1_10_1","unstructured":"Oculus Virtual Rality. https:\/\/www.oculus.com\/blog\/what-is-virtual-reality-all-about\/."},{"key":"e_1_3_2_1_11_1","unstructured":"Pano github. https:\/\/github.com\/pkuguanyu\/PanoProject."},{"key":"e_1_3_2_1_12_1","unstructured":"Unity. https:\/\/unity.com\/."},{"key":"e_1_3_2_1_13_1","unstructured":"Vocabulary for performance and quality of service. https:\/\/www.itu.int\/rec\/T-REC-P.10."},{"key":"e_1_3_2_1_14_1","unstructured":"VQMT. https:\/\/github.com\/rolinh\/VQMT."},{"key":"e_1_3_2_1_15_1","volume-title":"Survey: The State of Video Marketing","year":"2018","unstructured":"Wyzowl. Survey: The State of Video Marketing 2018. https:\/\/www.wyzowl.com\/video-marketing-statistics-2018\/."},{"key":"e_1_3_2_1_16_1","unstructured":"YouTube recommended bitrates. https:\/\/support.google.com\/youtube\/answer\/1722171?hl=en#zippy=%2Cvideo-resolution-and-aspect-ratio%2Cbitrate."},{"key":"e_1_3_2_1_17_1","unstructured":"YouTube VR. https:\/\/vr.youtube.com\/."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3230543.3230558"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/BigData.2016.7840720"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM48880.2022.9796904"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICC.2017.7996611"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM41043.2020.9155477"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3458306.3461000"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3341302.3342063"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3210240.3210323"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3523230.3523233"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/2619239.2626296"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/2413176.2413189"},{"key":"e_1_3_2_1_29_1","first-page":"3","article-title":"Next-generation video encoding techniques for 360 video and VR","volume":"21","author":"Kuzyakov Evgeny","year":"2016","unstructured":"Evgeny Kuzyakov and David Pio. Next-generation video encoding techniques for 360 video and VR. Jan, 21:3, 2016.","journal-title":"Jan"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMC.2019.2913364"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3152434.3152443"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3097895.3097901"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3098822.3098843"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3304109.3325812"},{"key":"e_1_3_2_1_35_1","first-page":"417","volume-title":"Proceedings of the USENIX Conference on Usenix Annual Technical Conference, USENIX ATC '15","author":"Netravali Ravi","year":"2015","unstructured":"Ravi Netravali, Anirudh Sivaraman, Somak Das, Ameesh Goyal, Keith Winstein, James Mickens, and Hari Balakrishnan. Mahimahi: Accurate record-and-replay for http. In Proceedings of the USENIX Conference on Usenix Annual Technical Conference, USENIX ATC '15, page 417--429, Santa Clara, CA, 2015."},{"key":"e_1_3_2_1_36_1","first-page":"1","volume-title":"Proceedings of the Conference of the IEEE Computer and Communications Societies, INFOCOM '21","author":"Palash Mijanur","year":"2021","unstructured":"Mijanur Palash, Voicu Popescu, Amit Sheoran, and Sonia Fahmy. Robust 360\u00b0 video streaming via non-linear sampling. In Proceedings of the Conference of the IEEE Computer and Communications Societies, INFOCOM '21, pages 1--10, Vancouver, BC, Canada, 2021."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3123266.3123453"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3241539.3241565"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/2980055.2980056"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/3339825.3394938"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2016.7524428"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISM.2016.0126"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/JETCAS.2019.2898877"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/2413176.2413190"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/LCOMM.2016.2601087"},{"key":"e_1_3_2_1_46_1","first-page":"34","volume-title":"Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video, NOSSDAV '21","author":"Wu Chenglei","year":"2021","unstructured":"Chenglei Wu, Zhi Wang, and Lifeng Sun. Paas: A preference-aware deep reinforcement learning approach for 360\u00b0 video streaming. In Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video, NOSSDAV '21, page 34--41, Istanbul, Turkey, 2021."},{"key":"e_1_3_2_1_47_1","first-page":"495","volume-title":"Proceedings of the 17th USENIX Symposium on Networked Systems Design and Implementation, NSDI '20","author":"Yan Francis Y.","year":"2020","unstructured":"Francis Y. Yan, Hudson Ayers, Chenzhi Zhu, Sadjad Fouladi, James Hong, Keyi Zhang, Philip Levis, and Keith Winstein. Learning in situ: a randomized experiment in video streaming. In Proceedings of the 17th USENIX Symposium on Networked Systems Design and Implementation, NSDI '20, pages 495--511, Santa Clara, CA, 2020."},{"key":"e_1_3_2_1_48_1","first-page":"731","volume-title":"Proceedings of the USENIX Conference on Usenix Annual Technical Conference, USENIX ATC '18","author":"Yan Francis Y.","year":"2018","unstructured":"Francis Y. Yan, Jestin Ma, Greg D. Hill, Deepti Raghavan, Riad S. Wahby, Philip Levis, and Keith Winstein. Pantheon: the training ground for internet congestion-control research. In Proceedings of the USENIX Conference on Usenix Annual Technical Conference, USENIX ATC '18, pages 731--743, Boston, MA, 2018."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/2785956.2787486"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475590"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2018.8486282"}],"event":{"name":"ACM SIGCOMM '23: ACM SIGCOMM 2023 Conference","location":"New York NY USA","acronym":"ACM SIGCOMM '23","sponsor":["SIGCOMM ACM Special Interest Group on Data Communication"]},"container-title":["Proceedings of the ACM SIGCOMM 2023 Conference"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3603269.3604876","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3603269.3604876","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3603269.3604876","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:46:43Z","timestamp":1750178803000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3603269.3604876"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9]]},"references-count":51,"alternative-id":["10.1145\/3603269.3604876","10.1145\/3603269"],"URL":"https:\/\/doi.org\/10.1145\/3603269.3604876","relation":{},"subject":[],"published":{"date-parts":[[2023,9]]},"assertion":[{"value":"2023-09-01","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}