{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,2]],"date-time":"2026-05-02T15:04:12Z","timestamp":1777734252498,"version":"3.51.4"},"reference-count":48,"publisher":"Association for Computing Machinery (ACM)","issue":"5","license":[{"start":{"date-parts":[[2020,5,31]],"date-time":"2020-05-31T00:00:00Z","timestamp":1590883200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Institute of Information 8 Communications Technology Planning 8 Evaluation"},{"name":"Korean government","award":["2017-0-01481"],"award-info":[{"award-number":["2017-0-01481"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Graph."],"published-print":{"date-parts":[[2020,10,31]]},"abstract":"<jats:p>We present a new interactive playback method to enhance 360\u00b0\u00a0viewing experiences. Our method automatically rotates the virtual camera of a 360\u00b0\u00a0 panoramic video (360\u00b0\u00a0video) player during interactive viewing to guide the viewer through the most important regions of the video. With this method, the viewer can watch a 360\u00b0\u00a0video with minimum efforts to find important events in a scene both in interactive (e.g., HMD) and less-interactive (e.g., PC and TV) viewing environments. To estimate the importance of each viewing direction, we combine spatial and temporal saliency with cluster-based weighting. A maximum backward cumulative importance volume (MBCIV) is then constructed by accumulating this importance in the video space. During playback, which uses a forward tracing scheme through the MBCIV, the initial optimal path is found based on the viewer\u2019s viewing direction. A smooth path is then derived using penalized curve fitting. Finally, the virtual camera is rotated to follow the path. The experiments and user studies demonstrate that our method allows the viewer to effectively enjoy 360\u00b0\u00a0videos with minimum interaction efforts, or even through a non-interactive display.<\/jats:p>","DOI":"10.1145\/3183794","type":"journal-article","created":{"date-parts":[[2020,6,1]],"date-time":"2020-06-01T04:31:06Z","timestamp":1590985866000},"page":"1-15","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":16,"title":["Enhanced Interactive 360\u00b0\u00a0Viewing via Automatic Guidance"],"prefix":"10.1145","volume":"39","author":[{"given":"Seunghoon","family":"Cha","sequence":"first","affiliation":[{"name":"Visual Media Lab, KAIST and KAI Inc., Daejeon, Republic of Korea"}]},{"given":"Jungjin","family":"Lee","sequence":"additional","affiliation":[{"name":"KAI Inc., Daejeon, Republic of Korea"}]},{"given":"Seunghwa","family":"Jeong","sequence":"additional","affiliation":[{"name":"Visual Media Lab, KAIST and KAI Inc., Daejeon, Republic of Korea"}]},{"given":"Younghui","family":"Kim","sequence":"additional","affiliation":[{"name":"KAI Inc., Daejeon, Republic of Korea"}]},{"given":"Junyong","family":"Noh","sequence":"additional","affiliation":[{"name":"Visual Media Lab, KAIST, Daejeon, Republic of Korea"}]}],"member":"320","published-online":{"date-parts":[[2020,5,31]]},"reference":[{"key":"e_1_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/2980179.2980257"},{"key":"e_1_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2012.2185937"},{"key":"e_1_2_2_3_1","volume-title":"O\u2019Connor","author":"Assens Marc","year":"2017","unstructured":"Marc Assens , Kevin McGuinness , Xavier Gir\u00f3 , and Noel E . O\u2019Connor . 2017 . SaltiNet: Scan-path prediction on 360 degree images using saliency volumes. CoRR abs\/1707.03123 (2017). Marc Assens, Kevin McGuinness, Xavier Gir\u00f3, and Noel E. O\u2019Connor. 2017. SaltiNet: Scan-path prediction on 360 degree images using saliency volumes. CoRR abs\/1707.03123 (2017)."},{"key":"e_1_2_2_4_1","first-page":"4","article-title":"SUS-A quick and dirty usability scale","volume":"189","author":"\u00a0al John Brooke","year":"1996","unstructured":"John Brooke et \u00a0al . 1996 . SUS-A quick and dirty usability scale . Usab. Eval. Indust. 189 , 194 (1996), 4 -- 7 . John Brooke et\u00a0al. 1996. SUS-A quick and dirty usability scale. Usab. Eval. Indust. 189, 194 (1996), 4--7.","journal-title":"Usab. Eval. Indust."},{"key":"e_1_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/1531326.1531349"},{"key":"e_1_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00154"},{"key":"e_1_2_2_7_1","volume-title":"Proceedings of the AAAI Conference on Artificial Intelligence.","author":"Chou Shih-Han","year":"2018","unstructured":"Shih-Han Chou , Yi-Chun Chen , Kuo-Hao Zeng , Hou-Ning Hu , Jianlong Fu , and Min Sun . 2018 . Self-view grounding given a narrated 360\u00b0 video . In Proceedings of the AAAI Conference on Artificial Intelligence. Shih-Han Chou, Yi-Chun Chen, Kuo-Hao Zeng, Hou-Ning Hu, Jianlong Fu, and Min Sun. 2018. Self-view grounding given a narrated 360\u00b0 video. In Proceedings of the AAAI Conference on Artificial Intelligence."},{"key":"e_1_2_2_8_1","volume-title":"Viewport-adaptive navigable 360-degree video delivery. arXiv preprint arXiv:1609.08042","author":"Corbillon Xavier","year":"2016","unstructured":"Xavier Corbillon , Alisa Devlic , Gwendal Simon , and Jacob Chakareski . 2016. Viewport-adaptive navigable 360-degree video delivery. arXiv preprint arXiv:1609.08042 ( 2016 ). Xavier Corbillon, Alisa Devlic, Gwendal Simon, and Jacob Chakareski. 2016. Viewport-adaptive navigable 360-degree video delivery. arXiv preprint arXiv:1609.08042 (2016)."},{"key":"e_1_2_2_9_1","volume-title":"A deep multi-level network for saliency prediction. CoRR abs\/1609.01064","author":"Cornia Marcella","year":"2016","unstructured":"Marcella Cornia , Lorenzo Baraldi , Giuseppe Serra , and Rita Cucchiara . 2016. A deep multi-level network for saliency prediction. CoRR abs\/1609.01064 ( 2016 ). Marcella Cornia, Lorenzo Baraldi, Giuseppe Serra, and Rita Cucchiara. 2016. A deep multi-level network for saliency prediction. CoRR abs\/1609.01064 (2016)."},{"key":"e_1_2_2_10_1","volume-title":"Proceedings of the 20th Annual Conference on Computer Graphics and Interactive Techniques. ACM, 135--142","author":"Cruz-Neira Carolina","unstructured":"Carolina Cruz-Neira , Daniel J. Sandin , and Thomas A . DeFanti. 1993. Surround-screen projection-based virtual reality: The design and implementation of the CAVE . In Proceedings of the 20th Annual Conference on Computer Graphics and Interactive Techniques. ACM, 135--142 . Carolina Cruz-Neira, Daniel J. Sandin, and Thomas A. DeFanti. 1993. Surround-screen projection-based virtual reality: The design and implementation of the CAVE. In Proceedings of the 20th Annual Conference on Computer Graphics and Interactive Techniques. ACM, 135--142."},{"key":"e_1_2_2_11_1","volume-title":"Proceedings of the 9th International Conference on Quality of Multimedia Experience (QoMEX\u201917)","author":"Abreu Ana De","year":"2017","unstructured":"Ana De Abreu , Cagri Ozcinar , and Aljosa Smolic . 2017 . Look around you: Saliency maps for omnidirectional images in VR applications . In Proceedings of the 9th International Conference on Quality of Multimedia Experience (QoMEX\u201917) . IEEE, 1--6. Ana De Abreu, Cagri Ozcinar, and Aljosa Smolic. 2017. Look around you: Saliency maps for omnidirectional images in VR applications. In Proceedings of the 9th International Conference on Quality of Multimedia Experience (QoMEX\u201917). IEEE, 1--6."},{"key":"e_1_2_2_12_1","volume-title":"Proceedings of the Eurographics Conference. Eurographics Association. DOI:https:\/\/doi.org\/10","author":"Duchowski Andrew","year":"2002","unstructured":"Andrew Duchowski and Gerd Marmitt . 2002 . Modeling visual attention in VR: Measuring the accuracy of predicted scanpaths . In Proceedings of the Eurographics Conference. Eurographics Association. DOI:https:\/\/doi.org\/10 .2312\/egs.20021022 10.2312\/egs.20021022 Andrew Duchowski and Gerd Marmitt. 2002. Modeling visual attention in VR: Measuring the accuracy of predicted scanpaths. In Proceedings of the Eurographics Conference. Eurographics Association. DOI:https:\/\/doi.org\/10.2312\/egs.20021022"},{"key":"e_1_2_2_13_1","volume-title":"Streaming virtual reality content. arXiv preprint arXiv:1612.08350","author":"El-Ganainy Tarek","year":"2016","unstructured":"Tarek El-Ganainy and Mohamed Hefeeda . 2016. Streaming virtual reality content. arXiv preprint arXiv:1612.08350 ( 2016 ). Tarek El-Ganainy and Mohamed Hefeeda. 2016. Streaming virtual reality content. arXiv preprint arXiv:1612.08350 (2016)."},{"key":"e_1_2_2_14_1","unstructured":"Facebook Inc. 2016. Facebook for Media New Publisher Tools for 360 Video. Retrieved from: https:\/\/www.facebook.com\/facebookmedia\/blog\/new-publisher-tools-for-360-video.  Facebook Inc. 2016. Facebook for Media New Publisher Tools for 360 Video. Retrieved from: https:\/\/www.facebook.com\/facebookmedia\/blog\/new-publisher-tools-for-360-video."},{"key":"e_1_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/2822013.2822025"},{"key":"e_1_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/2522628.2522899"},{"key":"e_1_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/2668904.2668936"},{"key":"e_1_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2007.383267"},{"key":"e_1_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.153"},{"key":"e_1_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46487-9_10"},{"key":"e_1_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/2699644"},{"key":"e_1_2_2_22_1","volume-title":"So Kweon, and Katsushi Ikeuchi. 2011. Stabilizing omnidirectional videos using 3D structure and spherical image warping. In Proceedings of the IAPR Conference on Machine Vision Applications. 177--180","author":"Kamali Mostafa","unstructured":"Mostafa Kamali , Atsuhiko Banno , Jean-Charles Bazin , In So Kweon, and Katsushi Ikeuchi. 2011. Stabilizing omnidirectional videos using 3D structure and spherical image warping. In Proceedings of the IAPR Conference on Machine Vision Applications. 177--180 . Mostafa Kamali, Atsuhiko Banno, Jean-Charles Bazin, In So Kweon, and Katsushi Ikeuchi. 2011. Stabilizing omnidirectional videos using 3D structure and spherical image warping. In Proceedings of the IAPR Conference on Machine Vision Applications. 177--180."},{"key":"e_1_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/2980179.2982405"},{"key":"e_1_2_2_24_1","volume-title":"Semantic-driven generation of hyperlapse from 360\u00b0 video. arXiv preprint arXiv:1703.10798","author":"Lai Wei-Sheng","year":"2017","unstructured":"Wei-Sheng Lai , Yujia Huang , Neel Joshi , Chris Buehler , Ming-Hsuan Yang , and Sing Bing Kang . 2017. Semantic-driven generation of hyperlapse from 360\u00b0 video. arXiv preprint arXiv:1703.10798 ( 2017 ). Wei-Sheng Lai, Yujia Huang, Neel Joshi, Chris Buehler, Ming-Hsuan Yang, and Sing Bing Kang. 2017. Semantic-driven generation of hyperlapse from 360\u00b0 video. arXiv preprint arXiv:1703.10798 (2017)."},{"key":"e_1_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/1778765.1778812"},{"key":"e_1_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/2897824.2925983"},{"key":"e_1_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2016.2532327"},{"key":"e_1_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00153"},{"key":"e_1_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/1180639.1180702"},{"key":"e_1_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/1531326.1531350"},{"key":"e_1_2_2_31_1","volume-title":"Comput. Graph. For.","author":"Liu Ligang","unstructured":"Ligang Liu , Renjie Chen , Lior Wolf , and Daniel Cohen-Or . 2010. Optimizing photo composition . In Comput. Graph. For. Vol. 29 . Wiley Online Library , 469--478. Ligang Liu, Renjie Chen, Lior Wolf, and Daniel Cohen-Or. 2010. Optimizing photo composition. In Comput. Graph. For. Vol. 29. Wiley Online Library, 469--478."},{"key":"e_1_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISCCSP.2012.6217836"},{"key":"e_1_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/2502081.2502135"},{"key":"e_1_2_2_34_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2016.02.001"},{"key":"e_1_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2009.06.006"},{"key":"e_1_2_2_36_1","volume-title":"Wright","author":"Nocedal Jorge","year":"2006","unstructured":"Jorge Nocedal and Stephen J . Wright . 2006 . Numerical Optimization (2nd ed.). Springer . Jorge Nocedal and Stephen J. Wright. 2006. Numerical Optimization (2nd ed.). Springer."},{"key":"e_1_2_2_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3126594.3126636"},{"key":"e_1_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/1360612.1360615"},{"key":"e_1_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3072959.3073668"},{"key":"e_1_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2018.2793599"},{"key":"e_1_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.150"},{"key":"e_1_2_2_42_1","volume-title":"Pano2Vid: Automatic cinematography for watching 360 videos. arXiv preprint arXiv:1612.02335","author":"Su Yu-Chuan","year":"2016","unstructured":"Yu-Chuan Su , Dinesh Jayaraman , and Kristen Grauman . 2016. Pano2Vid: Automatic cinematography for watching 360 videos. arXiv preprint arXiv:1612.02335 ( 2016 ). Yu-Chuan Su, Dinesh Jayaraman, and Kristen Grauman. 2016. Pano2Vid: Automatic cinematography for watching 360 videos. arXiv preprint arXiv:1612.02335 (2016)."},{"key":"e_1_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2005.854388"},{"key":"e_1_2_2_44_1","volume-title":"Applied Computer Science","author":"Torbert Shane","unstructured":"Shane Torbert . 2016. Applied Computer Science . Springer . Shane Torbert. 2016. Applied Computer Science. Springer."},{"key":"e_1_2_2_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2001.990517"},{"key":"e_1_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/1778765.1778827"},{"key":"e_1_2_2_47_1","volume-title":"Proceedings of the AAAI Conference on Artificial Intelligence.","author":"Yu Youngjae","year":"2018","unstructured":"Youngjae Yu , Sangho Lee , Joonil Na , Jaeyun Kang , and Gunhee Kim . 2018 . A deep ranking model for spatio-temporal highlight detection from a 360\u00b0video . In Proceedings of the AAAI Conference on Artificial Intelligence. Youngjae Yu, Sangho Lee, Joonil Na, Jaeyun Kang, and Gunhee Kim. 2018. A deep ranking model for spatio-temporal highlight detection from a 360\u00b0video. In Proceedings of the AAAI Conference on Artificial Intelligence."},{"key":"e_1_2_2_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/2964284.2967292"}],"container-title":["ACM Transactions on Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3183794","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3183794","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T01:08:29Z","timestamp":1750208909000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3183794"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,5,31]]},"references-count":48,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2020,10,31]]}},"alternative-id":["10.1145\/3183794"],"URL":"https:\/\/doi.org\/10.1145\/3183794","relation":{},"ISSN":["0730-0301","1557-7368"],"issn-type":[{"value":"0730-0301","type":"print"},{"value":"1557-7368","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,5,31]]},"assertion":[{"value":"2017-11-01","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2020-03-01","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2020-05-31","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}