{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T00:23:07Z","timestamp":1769041387735,"version":"3.49.0"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031533013","type":"print"},{"value":"9783031533020","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-53302-0_15","type":"book-chapter","created":{"date-parts":[[2024,1,28]],"date-time":"2024-01-28T09:02:09Z","timestamp":1706432529000},"page":"202-215","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["An Integrated System for\u00a0Spatio-temporal Summarization of\u00a0360-Degrees Videos"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-5311-8167","authenticated-orcid":false,"given":"Ioannis","family":"Kontostathis","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5376-7158","authenticated-orcid":false,"given":"Evlampios","family":"Apostolidis","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0121-4364","authenticated-orcid":false,"given":"Vasileios","family":"Mezaris","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,29]]},"reference":[{"issue":"11","key":"15_CR1","doi-asserted-by":"publisher","first-page":"1838","DOI":"10.1109\/JPROC.2021.3117472","volume":"109","author":"E Apostolidis","year":"2021","unstructured":"Apostolidis, E., Adamantidou, E., Metsai, A.I., Mezaris, V., Patras, I.: Video summarization using deep neural networks: a survey. Proc. IEEE 109(11), 1838\u20131863 (2021). https:\/\/doi.org\/10.1109\/JPROC.2021.3117472","journal-title":"Proc. IEEE"},{"key":"15_CR2","doi-asserted-by":"publisher","unstructured":"Apostolidis, E., Balaouras, G., Mezaris, V., Patras, I.: Summarizing videos using concentrated attention and considering the uniqueness and diversity of the video frames. In: Proceedings of the 2022 International Conference on Multimedia Retrieval, ICMR 2022, pp. 407\u2013415. Association for Computing Machinery, New York, NY, USA (2022). https:\/\/doi.org\/10.1145\/3512527.3531404","DOI":"10.1145\/3512527.3531404"},{"key":"15_CR3","doi-asserted-by":"publisher","first-page":"200","DOI":"10.1016\/j.cag.2022.06.002","volume":"106","author":"E Bernal-Berdun","year":"2022","unstructured":"Bernal-Berdun, E., Martin, D., Gutierrez, D., Masia, B.: SST-Sal: a spherical spatio-temporal approach for saliency prediction in 360 videos. Comput. Graph. 106, 200\u2013209 (2022). https:\/\/doi.org\/10.1016\/j.cag.2022.06.002","journal-title":"Comput. Graph."},{"issue":"3","key":"15_CR4","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1109\/TPAMI.2018.2815601","volume":"41","author":"Z Bylinskii","year":"2019","unstructured":"Bylinskii, Z., Judd, T., Oliva, A., Torralba, A., Durand, F.: What do different evaluation metrics tell us about saliency models? IEEE Trans. Pattern Anal. Mach. Intell. 41(3), 740\u2013757 (2019). https:\/\/doi.org\/10.1109\/TPAMI.2018.2815601","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"15_CR5","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"305","DOI":"10.1007\/978-3-030-68796-0_22","volume-title":"Pattern Recognition. ICPR International Workshops and Challenges","author":"Y Dahou","year":"2021","unstructured":"Dahou, Y., Tliba, M., McGuinness, K., O\u2019Connor, N.: ATSal: an attention based architecture for saliency prediction in 360$$^\\circ $$ videos. In: Del Bimbo, A., et al. (eds.) ICPR 2021. LNCS, vol. 12663, pp. 305\u2013320. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-68796-0_22"},{"key":"15_CR6","unstructured":"Ester, M., Kriegel, H.P., Sander, J., Xu, X.: A density-based algorithm for discovering clusters in large spatial databases with noise. In: Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, pp. 226\u2013231. AAAI Press (1996)"},{"key":"15_CR7","doi-asserted-by":"publisher","unstructured":"Guti\u00e9rrez, J., David, E.J., Coutrot, A., Da Silva, M.P., Callet, P.L.: Introducing un salient360! Benchmark: a platform for evaluating visual attention models for 360$$^\\circ $$ contents. In: 2018 10th International Conference on Quality of Multimedia Experience (QoMEX), pp. 1\u20133 (2018). https:\/\/doi.org\/10.1109\/QoMEX.2018.8463369","DOI":"10.1109\/QoMEX.2018.8463369"},{"key":"15_CR8","doi-asserted-by":"crossref","unstructured":"Haidar Sharif, M., Martinet, J., Djeraba, C.: Motion saliency. Encycl. Multimedia, 442\u2013444 (2008)","DOI":"10.1007\/978-0-387-78414-4_39"},{"key":"15_CR9","doi-asserted-by":"crossref","unstructured":"Hu, H.N., Lin, Y.C., Liu, M.Y., Cheng, H.T., Chang, Y.J., Sun, M.: Deep 360 pilot: learning a deep agent for piloting through 360$$^\\circ $$ sports videos. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.153"},{"key":"15_CR10","doi-asserted-by":"publisher","unstructured":"Kang, K., Cho, S.: Interactive and automatic navigation for 360$$^\\circ $$ video playback. ACM Trans. Graph. 38(4) (2019). https:\/\/doi.org\/10.1145\/3306346.3323046","DOI":"10.1145\/3306346.3323046"},{"key":"15_CR11","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1016\/j.image.2018.03.006","volume":"69","author":"P Lebreton","year":"2018","unstructured":"Lebreton, P., Raake, A.: GBVS360, BMS360, ProSal: extending existing saliency prediction models from 2D to omnidirectional images. Sig. Process. Image Commun. 69, 69\u201378 (2018). https:\/\/doi.org\/10.1016\/j.image.2018.03.006","journal-title":"Sig. Process. Image Commun."},{"key":"15_CR12","doi-asserted-by":"crossref","unstructured":"Lee, S., Sung, J., Yu, Y., Kim, G.: A memory network approach for story-based temporal summarization of 360$$^\\circ $$ videos. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00153"},{"key":"15_CR13","doi-asserted-by":"publisher","unstructured":"Li, J., Han, L., Zhang, C., Li, Q., Liu, Z.: Spherical convolution empowered viewport prediction in 360 video multicast with limited FoV feedback. ACM Trans. Multimedia Comput. Commun. Appl. 19(1) (2023). https:\/\/doi.org\/10.1145\/3511603","DOI":"10.1145\/3511603"},{"key":"15_CR14","unstructured":"Linardos, P., Mohedano, E., Nieto, J.J., O\u2019Connor, N.E., Gir\u00f3-i-Nieto, X., McGuinness, K.: Simple vs complex temporal recurrences for video saliency prediction. CoRR abs\/1907.01869 (2019). https:\/\/arxiv.org\/abs\/1907.01869"},{"key":"15_CR15","unstructured":"Nichat, M.: Landmark based shortest path detection by using a* algorithm and haversine formula (2013)"},{"key":"15_CR16","doi-asserted-by":"publisher","first-page":"748","DOI":"10.1109\/TMM.2020.2987682","volume":"23","author":"M Qiao","year":"2021","unstructured":"Qiao, M., Xu, M., Wang, Z., Borji, A.: Viewport-dependent saliency prediction in 360$$^\\circ $$ video. IEEE Trans. Multimedia 23, 748\u2013760 (2021). https:\/\/doi.org\/10.1109\/TMM.2020.2987682","journal-title":"IEEE Trans. Multimedia"},{"key":"15_CR17","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. In: Cortes, C., Lawrence, N., Lee, D., Sugiyama, M., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 28. Curran Associates, Inc. (2015)"},{"key":"15_CR18","doi-asserted-by":"publisher","unstructured":"Setayesh, M., Wong, V.W.: A content-based viewport prediction framework for 360$$^\\circ $$ video using personalized federated learning and fusion techniques. In: 2023 IEEE International Conference on Multimedia and Expo (ICME), pp. 654\u2013659 (2023). https:\/\/doi.org\/10.1109\/ICME55011.2023.00118","DOI":"10.1109\/ICME55011.2023.00118"},{"key":"15_CR19","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: International Conference on Learning Representations (2015)"},{"issue":"4","key":"15_CR20","doi-asserted-by":"publisher","first-page":"1633","DOI":"10.1109\/TVCG.2018.2793599","volume":"24","author":"V Sitzmann","year":"2018","unstructured":"Sitzmann, V., et al.: Saliency in VR: how do people explore virtual environments? IEEE Trans. Visual Comput. Graphics 24(4), 1633\u20131642 (2018). https:\/\/doi.org\/10.1109\/TVCG.2018.2793599","journal-title":"IEEE Trans. Visual Comput. Graphics"},{"key":"15_CR21","doi-asserted-by":"publisher","unstructured":"Song, Y., Vallmitjana, J., Stent, A., Jaimes, A.: TVSum: summarizing web videos using titles. In: 2015 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5179\u20135187 (2015). https:\/\/doi.org\/10.1109\/CVPR.2015.7299154","DOI":"10.1109\/CVPR.2015.7299154"},{"key":"15_CR22","doi-asserted-by":"crossref","unstructured":"Su, Y.C., Grauman, K.: Making 360$$^\\circ $$ video watchable in 2D: learning videography for click free viewing. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.150"},{"key":"15_CR23","doi-asserted-by":"crossref","unstructured":"Su, Y.C., Jayaraman, D., Grauman, K.: Pano2Vid: automatic cinematography for watching 360 videos. In: Proceedings of the Asian Conference on Computer Vision (ACCV) (2016)","DOI":"10.1007\/978-3-319-54190-7_10"},{"key":"15_CR24","doi-asserted-by":"publisher","unstructured":"Szegedy, C., et al.: Going deeper with convolutions. In: 2015 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1\u20139 (2015). https:\/\/doi.org\/10.1109\/CVPR.2015.7298594","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"15_CR25","doi-asserted-by":"publisher","unstructured":"Wang, M., Li, Y.J., Zhang, W.X., Richardt, C., Hu, S.M.: Transitioning360: content-aware NFoV virtual camera paths for 360$$^\\circ $$ video playback. In: 2020 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), pp. 185\u2013194 (2020). https:\/\/doi.org\/10.1109\/ISMAR50242.2020.00040","DOI":"10.1109\/ISMAR50242.2020.00040"},{"key":"15_CR26","doi-asserted-by":"publisher","unstructured":"Xu, Y., et al.: Gaze prediction in dynamic 360$$^\\circ $$ immersive videos. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5333\u20135342 (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00559","DOI":"10.1109\/CVPR.2018.00559"},{"key":"15_CR27","doi-asserted-by":"crossref","unstructured":"Yu, Y., Lee, S., Na, J., Kang, J., Kim, G.: A deep ranking model for spatio-temporal highlight detection from a 360 video. In: Proceedings of the 2018 AAAI Conference on Artificial Intelligence (2018)","DOI":"10.1609\/aaai.v32i1.12335"},{"key":"15_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"504","DOI":"10.1007\/978-3-030-01234-2_30","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Z Zhang","year":"2018","unstructured":"Zhang, Z., Xu, Y., Yu, J., Gao, S.: Saliency detection in 360$$^\\circ $$ videos. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 504\u2013520. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01234-2_30"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-53302-0_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,9]],"date-time":"2024-11-09T10:10:32Z","timestamp":1731147032000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-53302-0_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031533013","9783031533020"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-53302-0_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"29 January 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Amsterdam","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"The Netherlands","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 January 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 February 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ConfTool Pro","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"297","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"112","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"38% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}