{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:14:05Z","timestamp":1775578445957,"version":"3.50.1"},"reference-count":343,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2025,4,10]],"date-time":"2025-04-10T00:00:00Z","timestamp":1744243200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,4,10]],"date-time":"2025-04-10T00:00:00Z","timestamp":1744243200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s11263-025-02391-w","type":"journal-article","created":{"date-parts":[[2025,4,9]],"date-time":"2025-04-09T19:05:28Z","timestamp":1744225528000},"page":"4973-5012","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["A Survey of Representation Learning, Optimization Strategies, and Applications for Omnidirectional Vision"],"prefix":"10.1007","volume":"133","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2104-3352","authenticated-orcid":false,"given":"Hao","family":"Ai","sequence":"first","affiliation":[]},{"given":"Zidong","family":"Cao","sequence":"additional","affiliation":[]},{"given":"Lin","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,4,10]]},"reference":[{"key":"2391_CR1","doi-asserted-by":"crossref","unstructured":"Ai, H., Cao, Z., Cao, Y. P., Shan, Y., & Wang, L. (2023). Hrdfuse: Monocular 360 depth estimation by collaboratively learning holistic-with-regional depth distributions. In CVPR.","DOI":"10.1109\/CVPR52729.2023.01275"},{"key":"2391_CR2","doi-asserted-by":"crossref","unstructured":"Ai H, Cao Z, Lu H, Chen C, Ma J, Zhou P, Kim TK, Hui P & Wang L. (2024). Dream360: Diverse and immersive outdoor virtual scene creation via transformer-based 360 image outpainting. In TVCG.","DOI":"10.1109\/TVCG.2024.3372085"},{"key":"2391_CR3","doi-asserted-by":"crossref","unstructured":"Akimoto, N., Kasai, S., Hayashi, M., & Aoki, Y. (2019). 360-degree image completion by two-stage conditional gans. In ICIP.","DOI":"10.1109\/ICIP.2019.8803435"},{"key":"2391_CR4","doi-asserted-by":"crossref","unstructured":"Akimoto, N., Matsuo, Y., & Aoki, Y. (2022). Diverse plausible 360-degree image outpainting for efficient 3dcg background creation. In CVPR.","DOI":"10.1109\/CVPR52688.2022.01115"},{"key":"2391_CR5","doi-asserted-by":"crossref","unstructured":"Albanis, G., Zioulis, N., Drakoulis, P., Gkitsas, V., Sterzentsenko, V., Alvarez, F., Zarpalas, D., & Daras, P., (2021). Pano3d: A holistic benchmark and a solid baseline for 360$$^\\circ $$ depth estimation. In CVPR Workshop.","DOI":"10.1109\/CVPRW53098.2021.00413"},{"key":"2391_CR6","unstructured":"Apitzsch, A., Seidel, R., & Hirtz, G. (2018). Cubes3d: Neural network based optical flow in omnidirectional image scenes. arXiv preprint arXiv:1804.09004."},{"key":"2391_CR7","unstructured":"Armeni, I., Sax, S., Zamir, A.\u00a0R, & Savarese, S. (2017). Joint 2d-3d-semantic data for indoor scene understanding."},{"key":"2391_CR8","doi-asserted-by":"crossref","unstructured":"Artizzu, C.-O., Zhang, H., Allibert, G., & Demonceaux, C. (2021). Omniflownet: a perspective neural network adaptation for optical flow estimation in omnidirectional images. In ICPR.","DOI":"10.1109\/ICPR48806.2021.9412745"},{"key":"2391_CR9","unstructured":"Bai, J., Lai, S., Qin, H., Guo, J., & Guo, Y. (2022). Global-to-local panoramic depth estimation. arXiv: Glpanodepth."},{"key":"2391_CR10","doi-asserted-by":"crossref","unstructured":"Ban, Y., Zhang, Y., Zhang, H., Zhang, X., & Guo, Z. (2020). Ma360: Multi-agent deep reinforcement learning based live 360-degree video streaming on edge. In ICME (pp. 1\u20136).","DOI":"10.1109\/ICME46284.2020.9102836"},{"issue":"8","key":"2391_CR11","doi-asserted-by":"publisher","first-page":"921","DOI":"10.1364\/OPTICA.6.000921","volume":"6","author":"G Barbastathis","year":"2019","unstructured":"Barbastathis, G., Ozcan, A., & Situ, G. (2019). On the use of deep learning for computational imaging. Optica, 6(8), 921\u2013943.","journal-title":"Optica"},{"key":"2391_CR12","unstructured":"Bar-Tal, O. Yariv, L., Lipman, Y., & Dekel, T. (2023). Fusing diffusion paths for controlled image generation: Multidiffusion. In ICML."},{"key":"2391_CR13","doi-asserted-by":"crossref","unstructured":"Beltr\u00e1n, J., Guindel, C., Cort\u00e9s, I., Barrera, A., Astudillo, A., Urdiales, J., \u00c1lvarez, M., Bekka, F., Milan\u00e9s, V., & Garc\u00eda, F. (2020). Towards autonomous driving: a multi-modal 360$$^\\circ $$ perception proposal. In ITSC.","DOI":"10.1109\/ITSC45102.2020.9294494"},{"key":"2391_CR14","doi-asserted-by":"crossref","unstructured":"Berenguel-Baeta, B., Bermudez-Cameo, J., & Guerrero, J.\u00a0J. (2021). Scaled 360 layouts: Revisiting non-central panoramas. In CVPR.","DOI":"10.1109\/CVPRW53098.2021.00410"},{"key":"2391_CR15","doi-asserted-by":"crossref","unstructured":"Berenguel-Baeta, B., Bermudez-Cameo, J., & Guerrero, J.\u00a0J. (2022). Atlanta scaled layouts from non-central panoramas. In PR.","DOI":"10.1016\/j.patcog.2022.108740"},{"key":"2391_CR16","doi-asserted-by":"crossref","unstructured":"Bhandari, K., Duan, B., Liu, G., Latapie, H., Zong, Z., & Yan, Y. (2022). Learning omnidirectional flow in 360$$^\\circ $$ video via siamese representation. In ECCV.","DOI":"10.1007\/978-3-031-20074-8_32"},{"key":"2391_CR17","doi-asserted-by":"crossref","unstructured":"Bhandari, K., Zong, Z., & Yan, Y. (2021). Revisiting optical flow estimation in 360 videos. In 2020 25th International Conference on Pattern Recognition (ICPR) (pp. 8196\u20138203). IEEE","DOI":"10.1109\/ICPR48806.2021.9412035"},{"key":"2391_CR18","doi-asserted-by":"crossref","unstructured":"Caesar, H., Bankiti, V., Lang, A.\u00a0H., Vora, S., Liong, V.\u00a0E., Xu, Q., Krishnan, A., Pan, Y., Baldan, G., & Beijbom, O. (2020). nuscenes: A multimodal dataset for autonomous driving. In CVPR.","DOI":"10.1109\/CVPR42600.2020.01164"},{"key":"2391_CR19","doi-asserted-by":"crossref","unstructured":"Cao, Z., Ai, H., Cao, Y. P., Shan, Y., Qie, X., & Wang, L. (2023). Omnizoomer: Learning to move and zoom in on sphere at high-resolution. In ICCV.","DOI":"10.1109\/ICCV51070.2023.01185"},{"key":"2391_CR20","unstructured":"Cao, M., Ikehata, S., & Aizawa, K. (2022). Field-of-view iou for object detection in 360$$^\\circ $$ images. In IEEE TIP."},{"key":"2391_CR21","unstructured":"Cao, M., Mou, C., Yu, F., Wang, X., Zheng, Y., Zhang, J., Dong, C., Li, G., Shan, Y., Timofte, R., & Sun, X. (2023). Ntire 2023 challenge on 360$$^\\circ $$ omnidirectional image and video super-resolution: Datasets, methods and results. In CVPR Workshop."},{"key":"2391_CR22","doi-asserted-by":"crossref","unstructured":"Caruso, D., Engel, J. J., & Cremers, D. (2015). Large-scale direct slam for omnidirectional cameras. In IROS.","DOI":"10.1109\/IROS.2015.7353366"},{"key":"2391_CR23","doi-asserted-by":"crossref","unstructured":"Chai, X., & Shao, F. (2021). Blind quality assessment of omnidirectional videos using spatio-temporal convolutional neural networks. In Optik.","DOI":"10.1016\/j.ijleo.2020.165887"},{"key":"2391_CR24","doi-asserted-by":"crossref","unstructured":"Chang, S. H., Chiu, C. Y., Chang, C. S., Chen, K. W., Yao, C. Y., Lee, R. R., & Chu, H. K. (2018). Generating 360 outdoor panorama dataset with reliable sun position estimation. In SIGGRAPH Asia 2018 Posters.","DOI":"10.1145\/3283289.3283348"},{"key":"2391_CR25","doi-asserted-by":"crossref","unstructured":"Chang, A., Dai, A., Funkhouser, T., Halber, M., Niessner, M., Savva, M., Song, S., Zeng, A., & Zhang, Y. (2017). Matterport3d: Learning from rgb-d data in indoor environments.","DOI":"10.1109\/3DV.2017.00081"},{"key":"2391_CR26","doi-asserted-by":"crossref","unstructured":"Chang, W., Zhang, Y., & Xiong, Z. (2023). Depth estimation from indoor panoramas with neural scene representation. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (pp. 899\u2013908).","DOI":"10.1109\/CVPR52729.2023.00093"},{"key":"2391_CR27","doi-asserted-by":"crossref","unstructured":"Chen, Y., Dai, X., Liu, M., Chen, D., Yuan, L., & Liu, Z. (2019). Dynamic convolution: Attention over convolution kernels. In CVPR (pp. 11027\u201311036).","DOI":"10.1109\/CVPR42600.2020.01104"},{"key":"2391_CR28","unstructured":"Chen, Y., Du, Y., Xiao, Z., Zhao, L., Zhang, L., Liu, D., Zhu, D., Zhang, T., Hu, X., Liu, T., & Jiang, X. (2022d). A unified and biologically-plausible relational graph representation of vision transformers. In IEEE TNNLS."},{"key":"2391_CR29","doi-asserted-by":"crossref","unstructured":"Chen, H., Hou, Y., Qu, C., Testini, I., Hong, X., & Jiao, J. (2024). 360+ x: A panoptic multi-modal scene understanding dataset. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (pp. 19373\u201319382).","DOI":"10.1109\/CVPR52733.2024.01833"},{"key":"2391_CR30","doi-asserted-by":"crossref","unstructured":"Chen, Z., Li, Y., & Zhang, Y. (2018). Projection and evaluation. Signal Process: Recent advances in omnidirectional video coding for virtual reality.","DOI":"10.1016\/j.sigpro.2018.01.004"},{"key":"2391_CR31","doi-asserted-by":"crossref","unstructured":"Chen, Z., Lin, C., Lang, N., Liao, K., & Zhao, Y. (2023). Unsupervised omnimvs: Efficient omnidirectional depth inference via establishing pseudo-stereo supervision. IROS.","DOI":"10.1109\/IROS55552.2023.10342332"},{"key":"2391_CR32","unstructured":"Chen, D., O\u2019Bray, L., & Borgwardt, K. M. (2022a). Structure-aware transformer for graph representation learning."},{"key":"2391_CR33","doi-asserted-by":"crossref","unstructured":"Chen, D., Qing, C., Lin, X., Ye, M., Xu, M., & Dickinson, P. (2022b). Intra- and inter-reasoning graph convolutional network for saliency prediction on 360$$^\\circ $$ images. In IEEE TCVST.","DOI":"10.1109\/TCSVT.2022.3197159"},{"key":"2391_CR34","doi-asserted-by":"crossref","unstructured":"Chen, D., Qing, C., Xu, X., & Zhu, H. (2020). Salbinet360: Saliency prediction on 360$$^\\circ $$ images with local-global bifurcated deep network. In IEEE VR.","DOI":"10.1109\/VR46266.2020.00027"},{"key":"2391_CR35","doi-asserted-by":"crossref","unstructured":"Chen, G., Shao, F., Chai, X., Jiang, Q., & Ho, Y. S. (2024). Multi-stage salient object detection in 360$$^\\circ $$ omnidirectional image using complementary object-level semantic information. In IEEE TETCI.","DOI":"10.1109\/TETCI.2023.3259433"},{"key":"2391_CR36","doi-asserted-by":"crossref","unstructured":"Chen, S., Sun, P., Song, Y., & Luo, P. (2022c). Diffusiondet: Diffusion model for object detection. ICCV","DOI":"10.1109\/ICCV51070.2023.01816"},{"key":"2391_CR37","doi-asserted-by":"crossref","unstructured":"Chen, Z., Wang, G., & Liu, Z. (2022e). Text2light: Zero-shot text-driven HDR panorama generation. In TOG.","DOI":"10.1145\/3550454.3555447"},{"key":"2391_CR38","doi-asserted-by":"crossref","unstructured":"Chen, S., Zhang, Y., Li, Y., Chen, Z., & Wang, Z. (2018). Spherical structural similarity index for objective omnidirectional video quality assessment. In ICME.","DOI":"10.1109\/ICME.2018.8486584"},{"key":"2391_CR39","doi-asserted-by":"crossref","unstructured":"Cheng, H. T., Chao, C. H., Dong, J. D., Wen, H. K., Liu, T. L., & Sun, M. (2018). Cube padding for weakly-supervised saliency prediction in 360 videos. In CVPR.","DOI":"10.1109\/CVPR.2018.00154"},{"key":"2391_CR40","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.121853","volume":"238","author":"S Cho","year":"2024","unstructured":"Cho, S., Jung, R., & Kwon, J. (2024). Sampling based spherical transformer for 360 degree image classification. Expert Systems with Applications, 238, 121853.","journal-title":"Expert Systems with Applications"},{"key":"2391_CR41","doi-asserted-by":"crossref","unstructured":"Chou, S. H., Sun, C., Chang, W. Y., Hsu, W. T., Sun, M., & Fu, J. (2020). 360-indoor: towards learning real-world objects in 360deg indoor equirectangular images. In WACV.","DOI":"10.1109\/WACV45572.2020.9093262"},{"key":"2391_CR42","unstructured":"Cohen, T., Weiler, M., Kicanaoglu, B., & Welling, M. (2019). Gauge equivariant convolutional networks and the icosahedral CNN. In ICML."},{"key":"2391_CR43","volume-title":"Group equivariant convolutional networks","author":"T Cohen","year":"2016","unstructured":"Cohen, T., & Welling, M. (2016). Group equivariant convolutional networks. In International conference on machine learning: PMLR."},{"key":"2391_CR44","doi-asserted-by":"crossref","unstructured":"Coors, B., Condurache, A.\u00a0P., & Geiger, A. (2018). Spherenet: Learning spherical representations for detection and classification in omnidirectional images. In ECCV.","DOI":"10.1007\/978-3-030-01240-3_32"},{"key":"2391_CR45","unstructured":"Coughlan, J.\u00a0M., & Yuille, A.\u00a0L. (2000). The manhattan world assumption: Regularities in scene statistics which enable Bayesian inference. In NIPS."},{"key":"2391_CR46","doi-asserted-by":"crossref","unstructured":"Cruz, S., Hutchcroft, W., Li, Y., Khosravan, N., Boyadzhiev, I., & Kang, S.\u00a0B. (2021). Zillow indoor dataset: Annotated floor plans with 360deg panoramas and 3d room layouts. In CVPR.","DOI":"10.1109\/CVPR46437.2021.00217"},{"key":"2391_CR47","doi-asserted-by":"crossref","unstructured":"da\u00a0Silveira, T. L., Pinto, P. G., Murrugarra-Llerena, J., & Jung, C.\u00a0R. (2022). 3d scene geometry estimation from 360$$^\\circ $$ imagery: A survey. In ACM CSUR.","DOI":"10.1145\/3519021"},{"key":"2391_CR48","doi-asserted-by":"crossref","unstructured":"Dai, F., Zhang, Y., Ma, Y., Li, H., & Zhao, Q. (2020). Dilated convolutional neural networks for panoramic image saliency prediction. In ICASSP.","DOI":"10.1109\/ICASSP40776.2020.9053888"},{"key":"2391_CR49","doi-asserted-by":"crossref","unstructured":"Dastjerdi, M. R. K., Hold-Geoffroy, Y., Eisenmann, J., & Lalonde, J. F. (2023). Everlight: Indoor-outdoor editable HDR lighting estimation. ICCV.","DOI":"10.1109\/ICCV51070.2023.00682"},{"key":"2391_CR50","doi-asserted-by":"crossref","unstructured":"Dastjerdi, M. R.\u00a0K., Hold-Geoffroy, Y., Eisenmann, J., Khodadadeh, S., & Lalonde, J. F. (2022). Guided co-modulated gan for 360$$^\\circ $$ field of view extrapolation. In 3DV.","DOI":"10.1109\/3DV57658.2022.00059"},{"key":"2391_CR51","unstructured":"de\u00a0Albuquerque\u00a0Azevedo, R.\u00a0G., Birkbeck, N., Janatra, I., Adsumilli, B., & Frossard, P. (2020). A viewport-driven multi-metric fusion approach for 360-degree video quality assessment. In ICME."},{"key":"2391_CR52","doi-asserted-by":"crossref","unstructured":"Dedhia, B., Chiang, J. C., & Char, Y. F. (2019). Saliency prediction for omnidirectional images considering optimization on sphere domain. In ICASSP.","DOI":"10.1109\/ICASSP.2019.8683125"},{"key":"2391_CR53","unstructured":"Defferrard, M., Milani, M., Gusset, F., & Perraudin, N., (2020). Deepsphere: a graph-based spherical cnn. In ICLR."},{"key":"2391_CR54","doi-asserted-by":"crossref","unstructured":"Deng, X., Wang, H., Xu, M., Guo, Y., Song, Y., & Yang, L. (2021). Lau-net: Latitude adaptive upscaling network for omnidirectional imag super-resolution. In CVPR.","DOI":"10.1109\/CVPR46437.2021.00907"},{"key":"2391_CR55","doi-asserted-by":"crossref","unstructured":"Deng, L., Yang, M., Li, H., Li, T., Hu, B., & Wang, C. (2020). Restricted deformable convolution-based road scene semantic segmentation using surround view cameras. In IEEE TITS.","DOI":"10.1109\/TITS.2019.2939832"},{"key":"2391_CR56","doi-asserted-by":"crossref","unstructured":"Deng, L., Yang, M., Qian, Y., Wang, C., & Wang, B. (2017). CNN based semantic segmentation for urban traffic scenes using fisheye camera. In IV.","DOI":"10.1109\/IVS.2017.7995725"},{"issue":"6","key":"2391_CR57","doi-asserted-by":"publisher","first-page":"141","DOI":"10.1109\/MSP.2012.2211477","volume":"29","author":"L Deng","year":"2012","unstructured":"Deng, L. (2012). The mnist database of handwritten digit images for machine learning research. IEEE Signal Processing Magazine, 29(6), 141\u2013142.","journal-title":"IEEE Signal Processing Magazine"},{"key":"2391_CR58","doi-asserted-by":"crossref","unstructured":"Djemai, I., Fezza, S.\u00a0A., Hamidouche, W., & D\u00e9forges, O. (2020). Extending 2d saliency models for head movement prediction in 360-degree images using CNN-based fusion. In ISCAS.","DOI":"10.1109\/ISCAS45731.2020.9181229"},{"key":"2391_CR59","doi-asserted-by":"crossref","unstructured":"Djilali, Y. A. D., Krishna, T., McGuinness, K., & O\u2019Connor, N. E. (2021). $$\\text{Rethinking}^\\circ $$ 360 image visual attention modelling with unsupervised learning. In ICCV.","DOI":"10.1109\/ICCV48922.2021.01513"},{"key":"2391_CR60","volume-title":"Panoramic total scene understanding with a transformer","author":"Y Dong","year":"2023","unstructured":"Dong, Y., Fang, C., Dong, Z., Bo, L., & Tan, P. (2023). Panoramic total scene understanding with a transformer. ArXiv: Panocontext-former."},{"key":"2391_CR61","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., & Uszkoreit, J. (2020). An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR."},{"key":"2391_CR62","doi-asserted-by":"crossref","unstructured":"Duan, H., Zhai, G., Min, X., Zhu, Y.,\u00a0Fang, Y., & Yang, X. (2018). Perceptual quality assessment of omnidirectional images. In ISCAS.","DOI":"10.1109\/ISCAS.2018.8351786"},{"key":"2391_CR63","doi-asserted-by":"crossref","unstructured":"Eder, M., Moulon, P., &\u00a0Guan, L. (2019). Pano popups: Indoor 3d reconstruction with a plane-aware network. In 3DV.","DOI":"10.1109\/3DV.2019.00018"},{"key":"2391_CR64","doi-asserted-by":"crossref","unstructured":"Eder, M., Shvets, M., Lim, J., & Frahm, J. M. (2020). Tangent images for mitigating spherical distortion. In CVPR.","DOI":"10.1109\/CVPR42600.2020.01244"},{"key":"2391_CR65","doi-asserted-by":"crossref","unstructured":"Engel, J. J., Sch\u00f6ps, T., & Cremers, D. (2014). Lsd-slam: Large-scale direct monocular slam. European Conference on Computer Vision (pp. 834\u2013849). Springer International Publishing.","DOI":"10.1007\/978-3-319-10605-2_54"},{"key":"2391_CR66","doi-asserted-by":"crossref","unstructured":"Esser, P., Rombach, R., & Ommer, B. (2020). Taming transformers for high-resolution image synthesis. In CVPR.","DOI":"10.1109\/CVPR46437.2021.01268"},{"key":"2391_CR67","doi-asserted-by":"crossref","unstructured":"Feng, Q., Shum, H.\u00a0P. H., & Morishima, S. (2022). 360 depth estimation in the wild-the depth360 dataset and the segfuse network. In VR.","DOI":"10.1109\/VR51125.2022.00087"},{"key":"2391_CR68","doi-asserted-by":"crossref","unstructured":"Feng, B.\u00a0Y., Yao, W., Liu, Z., & Varshney, A. (2020). Deep depth estimation on 360 images with a double quaternion loss. In 3DV.","DOI":"10.1109\/3DV50981.2020.00062"},{"key":"2391_CR69","doi-asserted-by":"crossref","unstructured":"Fernandez-Labrador, C., F\u00e1cil, J.\u00a0M., P\u00e9rez-Yus, A., Demonceaux, C., Civera, J., & Guerrero, J.\u00a0J. (2020). Corners for layout: End-to-end layout recovery from 360 images. In RAL.","DOI":"10.1109\/LRA.2020.2967274"},{"key":"2391_CR70","doi-asserted-by":"crossref","unstructured":"Fernandez-Labrador, C., Perez-Yus, A., Lopez-Nicolas, G., & Guerrero, J.\u00a0J. (2018). Layouts from panoramic images with geometry and deep learning. In RAL.","DOI":"10.1109\/LRA.2018.2850532"},{"key":"2391_CR71","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s00371-024-03293-7","volume":"40","author":"M Flores","year":"2024","unstructured":"Flores, M., Valiente, D., Peidr\u00f3, A., Reinoso, O., & Pay\u00e1, L. (2024). Generating a full spherical view by modeling the relation between two fisheye images. The Visual Computer, 40, 1\u201326.","journal-title":"The Visual Computer"},{"key":"2391_CR72","doi-asserted-by":"crossref","unstructured":"Fu, J., Hou, C., Zhou, W., Xu, J., & Chen, Z. (2021). Adaptive hypergraph convolutional network for no-reference 360-degree image quality assessment. In ACM MM.","DOI":"10.1145\/3503161.3548337"},{"key":"2391_CR73","doi-asserted-by":"crossref","unstructured":"Gao, P., Zhang, P., & Smolic, A. (2022a). Quality assessment for omnidirectional video: A spatio-temporal distortion modeling approach. TMM.","DOI":"10.1109\/TMM.2020.3044458"},{"key":"2391_CR74","first-page":"1","volume":"71","author":"S Gao","year":"2022","unstructured":"Gao, S., Yang, K., Shi, H., Wang, K., & Bai, J. (2022). Review on panoramic imaging and its applications in scene understanding. IEEE Transactions on Instrumentation and Measurement, 71, 1\u201334.","journal-title":"IEEE Transactions on Instrumentation and Measurement"},{"key":"2391_CR75","doi-asserted-by":"crossref","unstructured":"Gardner, M. A., Hold-Geoffroy, Y., Sunkavalli, K., Gagn\u00e9, C., & Lalonde, J. F. (2019). Deep parametric indoor lighting estimation. In ICCV.","DOI":"10.1109\/ICCV.2019.00727"},{"key":"2391_CR76","doi-asserted-by":"crossref","unstructured":"Gardner, M. A., Sunkavalli, K., Yumer, E., Shen, X., Gambaretto, E., Gagn\u00e9, C., & Lalonde, J. F. (2017). Learning to predict indoor illumination from a single image. In TOG.","DOI":"10.1145\/3130800.3130891"},{"key":"2391_CR77","doi-asserted-by":"crossref","unstructured":"Garon, M., Sunkavalli, K., Hadap, S., Carr, N.\u00a0A., & Lalonde, J. F. (2019). Fast spatially-varying indoor lighting estimation. In CVPR.","DOI":"10.1109\/CVPR.2019.00707"},{"key":"2391_CR78","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., & Bengio, Y. (2014). Generative adversarial nets. In NIPS."},{"issue":"2","key":"2391_CR79","doi-asserted-by":"publisher","first-page":"759","DOI":"10.1086\/427976","volume":"622","author":"KM Gorski","year":"2005","unstructured":"Gorski, K. M., Hivon, E., Banday, A. J., Wandelt, B. D., Hansen, F. K., Reinecke, M., & Bartelman, M. (2005). Healpix: A framework for high-resolution discretization and fast analysis of data distributed on the sphere. The Astrophysical Journal, 622(2), 759.","journal-title":"The Astrophysical Journal"},{"key":"2391_CR80","doi-asserted-by":"crossref","unstructured":"Han, B. J., & Sim, J. Y. (2022). Zero-shot learning for reflection removal of single 360-degree image. European Conference on Computer Vision (pp. 533\u2013548). Springer.","DOI":"10.1007\/978-3-031-19800-7_31"},{"key":"2391_CR81","unstructured":"Han, S.\u00a0W., & Suh, D.\u00a0Y. (2020). Piinet: A 360-degree panoramic image inpainting network using a cube map. In arXiv."},{"key":"2391_CR82","unstructured":"Hara, T., & Harada, T. (2022). Enhancement of novel view synthesis using omnidirectional image completion."},{"key":"2391_CR83","doi-asserted-by":"crossref","unstructured":"Hara, T., Mukuta, Y., & Harada, T. (2021). Spherical image generation from a single image by considering scene symmetry. In AAAI.","DOI":"10.1609\/aaai.v35i2.16242"},{"key":"2391_CR84","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In CVPR.","DOI":"10.1109\/CVPR.2016.90"},{"key":"2391_CR85","doi-asserted-by":"crossref","unstructured":"Heshmat, Y., Jones,B., Xiong, X., Neustaedter, C., Tang, A., Riecke, B. E., & Yang, L. (2018). Geocaching with a beam: Shared outdoor activities through a telepresence robot with 360 degree viewing. In CHI.","DOI":"10.1145\/3173574.3173933"},{"key":"2391_CR86","unstructured":"Ho, J., & Ermon, S. (2016). Generative adversarial imitation learning. In NIPS."},{"key":"2391_CR87","doi-asserted-by":"crossref","unstructured":"Hold-Geoffroy, Y., Athawale, A., & Lalonde, J. F. (2019). Deep sky modeling for single image outdoor lighting estimation. In CVPR.","DOI":"10.1109\/CVPR.2019.00709"},{"key":"2391_CR88","doi-asserted-by":"crossref","unstructured":"Hold-Geoffroy, Y., Sunkavalli, K., Hadap, S., Gambaretto, E., & Lalonde, J. F. (2017). Deep outdoor illumination estimation. In CVPR.","DOI":"10.1109\/CVPR.2017.255"},{"key":"2391_CR89","doi-asserted-by":"crossref","unstructured":"Hong, Y., Zheng, Q., Zhao, L., Jiang, X., Kot, A. C., & Shi, B. (2021). Panoramic image reflection removal. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (pp. 7762\u20137771).","DOI":"10.1109\/CVPR46437.2021.00767"},{"issue":"10","key":"2391_CR90","doi-asserted-by":"publisher","first-page":"12192","DOI":"10.1109\/TPAMI.2023.3286429","volume":"45","author":"Y Hong","year":"2023","unstructured":"Hong, Y., Zheng, Q., Zhao, L., Jiang, X., Kot, A. C., & Shi, B. (2023). Par 2 net: End-to-end panoramic image reflection removal. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(10), 12192\u201312205.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"2391_CR91","volume-title":"PSNR vs. SSIM","author":"A Hor\u00e9","year":"2010","unstructured":"Hor\u00e9, A., & Ziou, D. (2010). PSNR vs. SSIM. ICPR: Image quality metrics."},{"key":"2391_CR92","unstructured":"Hsu, C.\u00a0Y., Sun, C., & Chen, H. T. (2021). Moving in a 360 world: Synthesizing panoramic parallaxes from a single panorama."},{"key":"2391_CR93","doi-asserted-by":"crossref","unstructured":"Hu, Z., Liu, L., & Sang, Q. (2024). Omnidirectional video quality assessment with causal intervention. In IEEE T-OB.","DOI":"10.1109\/TBC.2023.3342707"},{"key":"2391_CR94","unstructured":"Hu, E. J., shen, y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., & Chen, W. (2022). LoRA: Low-rank adaptation of large language models. In ICLR."},{"key":"2391_CR95","doi-asserted-by":"crossref","unstructured":"Huang, H., & Yeung, S. K. (2022). 360vo: Visual odometry using a single 360 camera. In 2022 International Conference on Robotics and Automation (ICRA) (pp. 5594\u20135600),","DOI":"10.1109\/ICRA46639.2022.9812203"},{"key":"2391_CR96","doi-asserted-by":"crossref","unstructured":"Huang, H., Liu, C., Zhu, Y., Cheng, H., Braud, T., & Yeung, S. K. (2024). 360loc: A dataset and benchmark for omnidirectional visual localization with cross-device queries. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (pp. 22314\u201322324).","DOI":"10.1109\/CVPR52733.2024.02106"},{"key":"2391_CR97","doi-asserted-by":"crossref","unstructured":"Jakab, D., Deegan, B.M., Sharma, S., Grua, E.M., Horgan, J., Ward, E., Van De Ven, P., Scanlan, A., & Eising, C. (2024). Surround-view fisheye optics in computer vision and simulation: Survey and challenges. IEEE Transactions on Intelligent Transportation Systems.","DOI":"10.1109\/TITS.2024.3368136"},{"key":"2391_CR98","doi-asserted-by":"crossref","unstructured":"Jaus, A., Yang, K., & Stiefelhagen, R. (2021). Panoramic panoptic segmentation: Towards complete surrounding understanding via unsupervised contrastive learning. In 2021 IEEE Intelligent Vehicles Symposium (IV) (pp. 1421\u20131427). IEEE","DOI":"10.1109\/IV48863.2021.9575904"},{"issue":"4","key":"2391_CR99","doi-asserted-by":"publisher","first-page":"4438","DOI":"10.1109\/TITS.2022.3232897","volume":"24","author":"A Jaus","year":"2023","unstructured":"Jaus, A., Yang, K., & Stiefelhagen, R. (2023). Panoramic panoptic segmentation: Insights into surrounding parsing for mobile agents via unsupervised contrastive learning. IEEE Transactions on Intelligent Transportation Systems, 24(4), 4438\u20134453.","journal-title":"IEEE Transactions on Intelligent Transportation Systems"},{"key":"2391_CR100","doi-asserted-by":"crossref","unstructured":"Jayasuriya, M., Ranasinghe, R., & Dissanayake, G. (2020). Active perception for outdoor localisation with an omnidirectional camera. In 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp. 4567\u20134574).","DOI":"10.1109\/IROS45743.2020.9340974"},{"key":"2391_CR101","doi-asserted-by":"crossref","unstructured":"Jia, H., Yi, H., Fujiki, H., Zhang, H., Wang, W., & Odamaki, M. (2022). 3d room layout recovery generalizing across manhattan and non-manhattan worlds. In CVPR Workshop.","DOI":"10.1109\/CVPRW56347.2022.00567"},{"key":"2391_CR102","doi-asserted-by":"crossref","unstructured":"Jiang, Q., Gao, S., Gao, Y., Yang, K., Yi, Z., Shi, H., Sun, L., & Wang, K. (2024). Minimalist and high-quality panoramic imaging with psf-aware transformers. In IEEE Transactions on Image Processing.","DOI":"10.1109\/TIP.2024.3441370"},{"key":"2391_CR103","doi-asserted-by":"crossref","unstructured":"Jiang, S., Li, Y., Weng, D., You, K., & Chen, W. (2023). 3d reconstruction of spherical images: A review of techniques, applications, and prospects.","DOI":"10.1080\/10095020.2024.2313328"},{"key":"2391_CR104","unstructured":"Jiang, C., Max, H., Jingwei, K., Karthik, P., Marcus, P., & Niessner , M. (2019). Spherical cnns on unstructured grids. In ICLR."},{"key":"2391_CR105","doi-asserted-by":"crossref","unstructured":"Jiang, H., Sheng, Z., Zhu, S., Dong, Z., & Huang, R. (2021b). Unifuse: Unidirectional fusion for 360 panorama depth estimation. In RAL.","DOI":"10.1109\/LRA.2021.3058957"},{"key":"2391_CR106","doi-asserted-by":"crossref","unstructured":"Jiang, Z., Xiang, Z., Xu, J., & Zhao, M. (2022b). Indoor panoramic room layout estimation with geometry-aware transformer network: Lgt-net. In CVPR.","DOI":"10.1109\/CVPR52688.2022.00170"},{"key":"2391_CR107","doi-asserted-by":"crossref","unstructured":"Jiang, H., Yi\u00a0Jiang, G., Yu, M., Zhang, Y., Yang, Y., Peng, Z., Chen, F., & Zhang, Q. (2021a). Cubemap-based perception-driven blind quality assessment for 360-degree images. In TIP.","DOI":"10.1109\/TIP.2021.3052073"},{"key":"2391_CR108","doi-asserted-by":"crossref","unstructured":"Jiang, Z., Zhang, X., Xu, Y., Ma, Z., Sun, J., & Zhang, Y. (2020). Reinforcement learning based rate adaptation for 360-degree video streaming. In IEEE T-OB.","DOI":"10.1109\/TBC.2020.3028286"},{"key":"2391_CR109","doi-asserted-by":"publisher","first-page":"1250","DOI":"10.1109\/TCI.2022.3233467","volume":"8","author":"Q Jiang","year":"2022","unstructured":"Jiang, Q., Shi, H., Sun, L., Gao, S., Yang, K., & Wang, K. (2022). Annular computational imaging: Capture clear panoramic images through simple lens. IEEE Transactions on Computational Imaging, 8, 1250\u20131264.","journal-title":"IEEE Transactions on Computational Imaging"},{"key":"2391_CR110","doi-asserted-by":"crossref","unstructured":"Jin, L., Xu, Y., Zheng, J., Zhang, J., Tang, R., Xu, S., Yu, J., & Gao, S. (2020). Geometric structure based and regularized depth estimation from 360 indoor imagery. In CVPR.","DOI":"10.1109\/CVPR42600.2020.00097"},{"key":"2391_CR111","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1016\/j.isprsjprs.2019.11.014","volume":"159","author":"S Ji","year":"2020","unstructured":"Ji, S., Qin, Z., Shan, J., & Lu, M. (2020). Panoramic slam from a multiple fisheye camera rig. ISPRS Journal of Photogrammetry and Remote Sensing, 159, 169\u2013183.","journal-title":"ISPRS Journal of Photogrammetry and Remote Sensing"},{"key":"2391_CR112","doi-asserted-by":"crossref","unstructured":"Kan, N., Zou, J., Li, C., Dai, W., & Xiong, H. (2021). Rapt360: Reinforcement learning-based rate adaptation for 360-degree video streaming with adaptive prediction and tiling. In IEEE TCSVT.","DOI":"10.1109\/TCSVT.2021.3076585"},{"key":"2391_CR113","doi-asserted-by":"crossref","unstructured":"Keinert, B., Innmann, M., S\u00e4nger, M., & Stamminger, M. (2015). Spherical fibonacci mapping. In TOG.","DOI":"10.1145\/2816795.2818131"},{"key":"2391_CR114","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3592433","volume":"42","author":"B Kerbl","year":"2023","unstructured":"Kerbl, B., Kopanas, G., Leimkuehler, T., & Drettakis, G. (2023). 3d gaussian splatting for real-time radiance field rendering. ACM Transactions on Graphics (TOG), 42, 1\u201314.","journal-title":"ACM Transactions on Graphics (TOG)"},{"key":"2391_CR115","unstructured":"Khasanova, R., & Frossard, P. (2019). Geometry aware convolutional filters for omnidirectional images representation. In ICML."},{"key":"2391_CR116","doi-asserted-by":"crossref","unstructured":"Kim, H., & Hilton, A. (2013). 3d scene reconstruction from multiple spherical stereo pairs. In IJCV.","DOI":"10.1007\/s11263-013-0616-1"},{"key":"2391_CR117","doi-asserted-by":"crossref","unstructured":"Kim, H., Gu, L., Heoun, t., & Ro, Y. M. (2020). Deep virtual reality image quality assessment with human perception guider for omnidirectional image. In IEEE TCSVT.","DOI":"10.1109\/TCSVT.2019.2898732"},{"key":"2391_CR118","doi-asserted-by":"crossref","unstructured":"Kim, H., Hernaggi, L., Jackson, P. J.\u00a0B., & Hilton, A. (2019). Immersive spatial audio reproduction for VR\/AR using room acoustic modelling from 360$$^\\circ $$ images. In VR.","DOI":"10.1109\/VR.2019.8798247"},{"key":"2391_CR119","doi-asserted-by":"crossref","unstructured":"Kim, J., Lee, E., & Kim, Y. (2023). Calibrating panoramic depth estimation for practical localization and mapping. In 2023 IEEE\/CVF International Conference on Computer Vision (ICCV) (pp. 8796\u20138806).","DOI":"10.1109\/ICCV51070.2023.00811"},{"key":"2391_CR120","doi-asserted-by":"crossref","unstructured":"Kinzig, C., Miller, H., Lauer, M., & Stiller, C. (2024). Panoptic segmentation from stitched panoramic view for automated driving. In 2024 IEEE Intelligent Vehicles Symposium (IV) (pp. 3342\u20133347). IEEE","DOI":"10.1109\/IV55156.2024.10588453"},{"key":"2391_CR121","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A. C., Lo, W. Y., & Dollar, P. (2023). Segment anything. In Proceedings of the IEEE\/CVF International Conference on Computer Vision (pp. 4015\u20134026).","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"2391_CR122","doi-asserted-by":"publisher","DOI":"10.3389\/fpsyg.2020.563474","volume":"11","author":"A Kittel","year":"2020","unstructured":"Kittel, A., Larkin, P., Cunningham, I., & Spittle, M. (2020). 360 virtual reality: A swot analysis in comparison to virtual reality. Frontiers in Psychology, 11, 563474.","journal-title":"Frontiers in Psychology"},{"key":"2391_CR123","unstructured":"Krizhevsky, A. (2009). Learning multiple layers of features from tiny images."},{"key":"2391_CR124","doi-asserted-by":"crossref","unstructured":"Krolla, B., Diebold, M., Goldl\u00fccke, B., & Stricker, D. (2014). Spherical light fields. In BMVC.","DOI":"10.5244\/C.28.67"},{"key":"2391_CR125","doi-asserted-by":"crossref","unstructured":"Kulkarni, S., Yin, P., & Scherer, S. A. (2022). 360fusionnerf: Panoramic neural radiance fields with joint guidance. In 2023 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp. 7202\u20137209).","DOI":"10.1109\/IROS55552.2023.10341346"},{"key":"2391_CR126","doi-asserted-by":"crossref","unstructured":"Lai, P. K., Xie, S., Lang, J., & Lagani\u00e8re, R. (2019). Real-time panoramic depth maps from omni-directional stereo images for 6 dof videos in virtual reality. In 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR) (pp. 405\u2013412). IEEE","DOI":"10.1109\/VR.2019.8798016"},{"key":"2391_CR127","unstructured":"Laine, S., & Aila, T. (2017). Temporal ensembling for semi-supervised learning. In ICLR."},{"key":"2391_CR128","doi-asserted-by":"crossref","unstructured":"Lee, Y., Jeong, J., Yun, J., Cho, W., & Yoon, K. J. (2019). Spherephd: Applying cnns on a spherical polyhedron representation of 360$$^\\circ $$ images. In CVPR.","DOI":"10.1109\/CVPR.2019.00940"},{"key":"2391_CR129","doi-asserted-by":"crossref","unstructured":"Lee, J., Park, D., Hyuck, L., Dongwook, & Ji, D. (2022). Semi-supervised 360$$^\\circ $$ depth estimation from multiple fisheye cameras with pixel-level selective loss. In ICASSP.","DOI":"10.1109\/ICASSP43922.2022.9746232"},{"key":"2391_CR130","unstructured":"Li, S., & Fukamori, K. (2005). Spherical stereo for the construction of immersive vr environment. In IEEE Proceedings. VR 2005. Virtual Reality, 2005. (pp. 217\u2013222). IEEE"},{"key":"2391_CR131","doi-asserted-by":"crossref","unstructured":"Li, Y., Barnes, C., Huang, K., & Zhang, F. L. (2022b). Deep 360$$^\\circ $$ optical flow estimation based on multi-projection fusion. European Conference on Computer Vision (pp. 336\u2013352). Springer.","DOI":"10.1007\/978-3-031-19833-5_20"},{"key":"2391_CR132","doi-asserted-by":"crossref","unstructured":"Li, Y., Guo, Y., Yan, Z., Huang, X., Duan, Y., & Ren, L. (2022c). Omnifusion: 360 monocular depth estimation via geometry-aware fusion. In CVPR.","DOI":"10.1109\/CVPR52688.2022.00282"},{"key":"2391_CR133","doi-asserted-by":"crossref","unstructured":"Li, M., Jin, X., Hu, X., Dai, J., Du, S., & Li, Y. (2022a). Mode: Multi-view omnidirectional depth estimation with 360$$^\\circ $$ cameras. In ECCV.","DOI":"10.1007\/978-3-031-19827-4_12"},{"key":"2391_CR134","unstructured":"Li, J., Li, D., Savarese, S., & Hoi, S. (2023a). Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. In International conference on machine learning (pp. 19730\u201319742). PMLR."},{"key":"2391_CR135","doi-asserted-by":"crossref","unstructured":"Li, Z., Shafiei, M., Ramamoorthi, R., Sunkavalli, K., & Chandraker, M. (2020). Inverse rendering for complex indoor scenes: Shape, spatially-varying lighting and SVBRDF from a single image. In CVPR.","DOI":"10.1109\/CVPR42600.2020.00255"},{"key":"2391_CR136","doi-asserted-by":"crossref","unstructured":"Li, Z., Wang, Y., & Liu, Y. (2022d). Sad360: Spherical viewport-aware dynamic tiling for 360-degree video streaming. In VCIP.","DOI":"10.1109\/VCIP56404.2022.10008862"},{"key":"2391_CR137","doi-asserted-by":"crossref","unstructured":"Li, M., Wang, S., Yuan, W., Shen, W., Sheng, Z., & Dong, Z. (2023b). S2net: Accurate panorama depth estimation on spherical surface. In IEEE RAL.","DOI":"10.1109\/LRA.2023.3234820"},{"key":"2391_CR138","doi-asserted-by":"crossref","unstructured":"Li, X., Wu, T., Qi, Z., Wang, G., Shan, Y., & Li, X. (2023c). Sgat4pass: Spherical geometry-aware transformer for panoramic semantic segmentation. In IJCAI.","DOI":"10.24963\/ijcai.2023\/125"},{"key":"2391_CR139","doi-asserted-by":"crossref","unstructured":"Li, C., Xu, M., Jiang, L., Zhang, S., & Tao, X. (2019a). Viewport proposal cnn for 360$$^\\circ $$ video quality assessment. In CVPR.","DOI":"10.1109\/CVPR.2019.01042"},{"key":"2391_CR140","unstructured":"Liao, K., Xu, X., Lin, C., Ren, W., Wei, Y., & Zhao, Y. (2022). Cylin-painting: Seamless 360$$^\\circ $$ panoramic image outpainting and beyond. In IEEE TIP."},{"key":"2391_CR141","first-page":"21878","volume":"36","author":"J Li","year":"2023","unstructured":"Li, J., & Bansal, M. (2023). Panogen: Text-conditioned panoramic environment generation for vision-and-language navigation. Advances in Neural Information Processing Systems, 36, 21878\u201321894.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2391_CR142","doi-asserted-by":"crossref","unstructured":"Lim, H. T., K., Hak, G., & Ro, Y. M. (2018). Vr iqa net: Deep virtual reality image quality assessment using adversarial learning. In ICASSP.","DOI":"10.1109\/ICASSP.2018.8461317"},{"key":"2391_CR143","doi-asserted-by":"crossref","unstructured":"Ling, Z., Xing, Z., Zhou, X., Cao, M., & Zhou, G. (2023). A pano-style swin transformer for panorama understanding: Panoswin. In CVPR.","DOI":"10.1109\/CVPR52729.2023.01703"},{"key":"2391_CR144","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., & Guo, B. (2021b). Swin transformer: Hierarchical vision transformer using shifted windows. In ICCV.","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"2391_CR145","doi-asserted-by":"crossref","unstructured":"Liu, M., Wang, S., Guo, Y., He, Y., & Xue H. (2021a). Pano-sfmlearner: Self-supervised multi-task learning of depth and semantics in panoramic videos. In IEEE SPL.","DOI":"10.1109\/LSP.2021.3073627"},{"key":"2391_CR146","doi-asserted-by":"crossref","unstructured":"Liu, X., Xu, H., Chen, B., Zhao, Q., Ma, Y., Yan, C., & Dai, F. (2023). Sph2pob: Boosting object detection on spherical images with planar oriented boxes methods. In IJCAI.","DOI":"10.24963\/ijcai.2023\/137"},{"key":"2391_CR147","doi-asserted-by":"crossref","unstructured":"Liu, X., Yaann, J., Huang, L., Fang., Y., Wan, Z., & Liu, Y. (2024). Perceptual quality assessment of omnidirectional images: A benchmark and computational model. In ACM TOMM.","DOI":"10.1145\/3640344"},{"key":"2391_CR148","unstructured":"Liu, M., Yao, F., Choi, C., Sinha, A., & Ramani, K. (2018). Deep learning 3d shapes using alt-az anisotropic 2-sphere convolution. In ICLR."},{"key":"2391_CR149","doi-asserted-by":"crossref","unstructured":"Liu, R., Zhang, G., Wang, J., & Zhao, S. (2022). Cross-modal 360$$^\\circ $$ depth completion and reconstruction for large-scale indoor environment. In IEEE TITS.","DOI":"10.1109\/TITS.2022.3155925"},{"issue":"1","key":"2391_CR150","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1109\/JSTSP.2019.2953950","volume":"14","author":"J Li","year":"2019","unstructured":"Li, J., Zhao, Y., Ye, W., Yu, K., & Ge, S. (2019). Attentive deep stitching and quality assessment for 360$$^\\circ $$ omnidirectional images. IEEE Journal of Selected Topics in Signal Processing, 14(1), 209\u2013221.","journal-title":"IEEE Journal of Selected Topics in Signal Processing"},{"key":"2391_CR151","doi-asserted-by":"crossref","unstructured":"Lo, I. C., Shih, K. T., & Chen, H. H. (2018). Image stitching for dual fisheye cameras. In 2018 25th IEEE International Conference on Image Processing (ICIP) (pp. 3164\u20133168). IEEE","DOI":"10.1109\/ICIP.2018.8451333"},{"key":"2391_CR152","doi-asserted-by":"crossref","unstructured":"Lo, I. C., Shih, K. T.,, Ju, G. H., & Chen, H. H. (2020). Photometric consistency for dual fisheye cameras. In 2020 IEEE International Conference on Image Processing (ICIP) (pp. 261\u2013265). IEEE","DOI":"10.1109\/ICIP40778.2020.9190784"},{"key":"2391_CR153","doi-asserted-by":"publisher","first-page":"251","DOI":"10.1109\/TIP.2021.3130531","volume":"31","author":"IC Lo","year":"2021","unstructured":"Lo, I. C., Shih, K. T., & Chen, H. H. (2021). Efficient and accurate stitching for 360$$^\\circ $$ dual-fisheye images and videos. IEEE Transactions on Image Processing, 31, 251\u2013262.","journal-title":"IEEE Transactions on Image Processing"},{"key":"2391_CR154","doi-asserted-by":"crossref","unstructured":"Lu, Z., Hu, K., Wang, C., Bai, L., & Wang, Z. (2024). Autoregressive omni-aware outpainting for open-vocabulary 360-degree image generation. In AAAI.","DOI":"10.1609\/aaai.v38i13.29332"},{"key":"2391_CR155","unstructured":"Lucas, B. D., & Kanade, T. (1981). An iterative image registration technique with an application to stereo vision. In IJCAI\u201981: 7th international joint conference on Artificial intelligence, vol.\u00a02 (pp. 674\u2013679)."},{"key":"2391_CR156","doi-asserted-by":"crossref","unstructured":"Lv, H., Yang, Q., Li, C., Dai, W., Zou, J., & Xiong, H. (2020). Salgcn: Saliency prediction for 360-degree images based on spherical graph convolutional networks. In ACM MM.","DOI":"10.1145\/3394171.3413733"},{"key":"2391_CR157","doi-asserted-by":"crossref","unstructured":"Ma, C., Zhang, J., Yang, K., Roitberg, A., & Stiefelhagen, R. (2021). Densepass: Dense panoramic semantic segmentation via unsupervised domain adaptation with attention-augmented context exchange. In ITSC.","DOI":"10.1109\/ITSC48978.2021.9564920"},{"key":"2391_CR158","unstructured":"Maniotis, P., & Thomos, N. (2020). Viewport-aware deep reinforcement learning approach for 360$$^\\circ $$ video caching. In TMM."},{"key":"2391_CR159","doi-asserted-by":"crossref","unstructured":"Mansouri, S., Sharif, K., Petros. S., Kanellakis, C., Kominiak, D., & Nikolakopoulos, G. (2019). Vision-based mav navigation in underground mine using convolutional neural network. In IECON.","DOI":"10.1109\/IECON.2019.8927168"},{"key":"2391_CR160","doi-asserted-by":"publisher","first-page":"3693","DOI":"10.1109\/LRA.2018.2855443","volume":"3","author":"H Matsuki","year":"2018","unstructured":"Matsuki, H., von Stumberg, L., Usenko, V., St\u00fcckler, J., & Cremers, D. (2018). Omnidirectional dso: Direct sparse odometry with fisheye cameras. IEEE Robotics and Automation Letters, 3, 3693\u20133700.","journal-title":"IEEE Robotics and Automation Letters"},{"issue":"64\u201367","key":"2391_CR161","first-page":"2","volume":"5","author":"LR Medsker","year":"2001","unstructured":"Medsker, L. R., & Jain, L. C. (2001). Recurrent neural networks. Design and Applications, 5(64\u201367), 2.","journal-title":"Recurrent neural networks. Design and Applications"},{"key":"2391_CR162","doi-asserted-by":"crossref","unstructured":"Mei, J., Zhu, A. Z., Yan, X., Yan, H., Qiao, S., Chen, L. C., & Kretzschmar, H. (2022). Waymo open dataset: Panoramic video panoptic segmentation. European Conference on Computer Vision (pp. 53\u201372). Springer.","DOI":"10.1007\/978-3-031-19818-2_4"},{"key":"2391_CR163","doi-asserted-by":"crossref","unstructured":"Meuleman, A., Jang, H., Jeon, D. S., & Kim. M. H. (2021). Real-time sphere sweeping stereo from multiview fisheye images. In CVPR.","DOI":"10.1109\/CVPR46437.2021.01126"},{"issue":"1","key":"2391_CR164","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1145\/3503250","volume":"65","author":"B Mildenhall","year":"2021","unstructured":"Mildenhall, B., Srinivasan, P. P., Tancik, M., Barron, J. T., Ramamoorthi, R., & Ng, R. (2021). Nerf: Representing scenes as neural radiance fields for view synthesis. Communications of the ACM, 65(1), 99\u2013106.","journal-title":"Communications of the ACM"},{"key":"2391_CR165","unstructured":"Mirza, M., & Osindero, S. (2014). Conditional generative adversarial nets. abs\/1411.1784,"},{"key":"2391_CR166","first-page":"26","volume":"69","author":"R Monroy","year":"2017","unstructured":"Monroy, R., Lutz, S., Chalasani, T., & Smolic, A. (2017). Salnet360: Saliency maps for omni-directional images with CNN. Image Commun: Signal Process, 69, 26\u201334.","journal-title":"Image Commun: Signal Process"},{"key":"2391_CR167","doi-asserted-by":"crossref","unstructured":"Nguyen, A., Yan, Z., & Nahrstedt, K.. (2018). Your attention is unique: Detecting 360-degree video saliency in head-mounted display for head movement prediction. In ACM MM.","DOI":"10.1145\/3240508.3240669"},{"issue":"7","key":"2391_CR168","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1088\/0031-9112\/13\/7\/027","volume":"13","author":"TH O\u2019Beirne","year":"1962","unstructured":"O\u2019Beirne, T. H. (1962). Introduction to geometry. Physics Bulletin, 13(7), 211.","journal-title":"Physics Bulletin"},{"key":"2391_CR169","unstructured":"Ocampo, Jeremy, P., Matthew. A., & McEwen, J. D. (2023). Scalable and equivariant spherical cnns by discrete-continuous convolutions. In ICLR."},{"key":"2391_CR170","volume-title":"Bi-modal indoor panorama synthesis via residual depth-aided adversarial learning","author":"C Oh","year":"2021","unstructured":"Oh, C., Cho, W., Park, D., Chae, Y., Wang, L., & Yoon, K. J. (2021). Bi-modal indoor panorama synthesis via residual depth-aided adversarial learning. ArXiv: Bips."},{"key":"2391_CR171","doi-asserted-by":"crossref","unstructured":"Ozcinar, C., Cabrera, J., & Smolic, A. (2019a). Visual attention-aware omnidirectional video streaming using optimal tiles for virtual reality. In IEEE JETCAS.","DOI":"10.1109\/JETCAS.2019.2895096"},{"key":"2391_CR172","doi-asserted-by":"crossref","unstructured":"Ozcinar, C., Rana, A., & Smolic, A. (2019b). Super-resolution of omnidirectional images using adversarial learning. In MMSP.","DOI":"10.1109\/MMSP.2019.8901764"},{"key":"2391_CR173","doi-asserted-by":"crossref","unstructured":"Park, S., Hoai, M., Bhaacharya, A., & Das, S. R. (2021). Adaptive streaming of 360-degree videos with reinforcement learning. In WACV.","DOI":"10.1109\/WACV48630.2021.00188"},{"key":"2391_CR174","doi-asserted-by":"crossref","unstructured":"Park, J., Kim, H., Park, E., & Sim, J.-Y. (2024). Fully-automatic reflection removal for 360-degree images. In Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (pp. 1609\u20131617).","DOI":"10.1109\/WACV57701.2024.00163"},{"key":"2391_CR175","doi-asserted-by":"crossref","unstructured":"Peleg, S., & Ben-Ezra, M. (1999). Stereo panorama with a single camera. In CVPR.","DOI":"10.1109\/CVPR.1999.786969"},{"key":"2391_CR176","doi-asserted-by":"crossref","unstructured":"Peng, C. H., & Zhang, J. (2023). High-resolution depth estimation for 360deg panoramas through perspective and panoramic depth images registration. In Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (pp. 3116\u20133125).","DOI":"10.1109\/WACV56688.2023.00313"},{"key":"2391_CR177","doi-asserted-by":"crossref","unstructured":"Pintore, G., Agus, M., & Gobbetti, E. (2020). Atlantanet: Inferring the 3d indoor layout from a single $$360^\\circ $$ image beyond the manhattan world assumption. In ECCV.","DOI":"10.1007\/978-3-030-58598-3_26"},{"key":"2391_CR178","doi-asserted-by":"crossref","unstructured":"Pintore, G., Almansa, E., & Schneider, J. (2021b). Slicenet: deep dense depth estimation from a single indoor panorama using a slice-based representation. In CVPR.","DOI":"10.1109\/CVPR46437.2021.01137"},{"key":"2391_CR179","doi-asserted-by":"crossref","unstructured":"Pintore, G., Almansa, E., Agus, M., & Gobbetti, E. (2021a). Deep3dlayout: 3d reconstruction of an indoor layout from a spherical panoramic image. In TOG.","DOI":"10.1145\/3478513.3480480"},{"issue":"31","key":"2391_CR180","doi-asserted-by":"publisher","first-page":"7356","DOI":"10.1364\/AO.33.007356","volume":"33","author":"I Powell","year":"1994","unstructured":"Powell, I. (1994). Panoramic lens. Applied Optics, 33(31), 7356\u20137361.","journal-title":"Applied Optics"},{"key":"2391_CR181","doi-asserted-by":"crossref","unstructured":"Pudics, G., Szabo-Resch, M., Z., & V\u00e1mossy, Z. (2015). Safe robot navigation using an omnidirectional camera. In CINTI.","DOI":"10.1109\/CINTI.2015.7382928"},{"issue":"12","key":"2391_CR182","doi-asserted-by":"publisher","first-page":"22755","DOI":"10.1109\/TITS.2022.3210409","volume":"23","author":"Y Qian","year":"2022","unstructured":"Qian, Y., Yang, M., & Dolan, J. M. (2022). Survey on fish-eye cameras and their applications in intelligent vehicles. IEEE Transactions on Intelligent Transportation Systems, 23(12), 22755\u201322771.","journal-title":"IEEE Transactions on Intelligent Transportation Systems"},{"key":"2391_CR183","doi-asserted-by":"crossref","unstructured":"Qiao, M., Xu, M., Wang, Z., & Borji, A. (2021). Viewport-dependent saliency prediction in 360$$^\\circ $$ video. In TMM.","DOI":"10.1109\/TMM.2020.2987682"},{"key":"2391_CR184","unstructured":"Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., Krueger, G., & Sutskever, I. (2021). Learning transferable visual models from natural language supervision. In ICML."},{"key":"2391_CR185","doi-asserted-by":"crossref","unstructured":"Rai, Y., Guti\u00e9rrez, J., & Le\u00a0Callet, P. (2017). A dataset of head and eye movements for 360 degree images. In ACM MMSys.","DOI":"10.1145\/3192974"},{"key":"2391_CR186","doi-asserted-by":"crossref","unstructured":"Ranftl, R., B., Alexey, & Koltun, V. (2021). Vision transformers for dense prediction. In ICCV.","DOI":"10.1109\/ICCV48922.2021.01196"},{"issue":"6","key":"2391_CR187","doi-asserted-by":"publisher","first-page":"1341","DOI":"10.3390\/s17061341","volume":"17","author":"L Ran","year":"2017","unstructured":"Ran, L., Zhang, Y., Zhang, Q., & Yang, T. (2017). Convolutional neural network-based robot navigation using uncalibrated spherical images $$\\dagger $$. Sensors, 17(6), 1341.","journal-title":"Sensors"},{"key":"2391_CR188","unstructured":"Ren, S., He, K., Girshick, R.\u00a0B., & Sun, J. (2015). Faster r-cnn: Towards real-time object detection with region proposal networks. In IEEE T-PAMI."},{"key":"2391_CR189","doi-asserted-by":"crossref","unstructured":"Rey-Area, M., Yuan, M., & Richardt, C. (2022). 360monodepth: High-resolution 360$$^\\circ $$ monocular depth estimation. In CVPR.","DOI":"10.1109\/CVPR52688.2022.00374"},{"key":"2391_CR190","doi-asserted-by":"crossref","unstructured":"Roberto, R., Perazzo, D., Lima, J. P., Teichrieb, V., Quintino, J. P., da\u00a0Silva, F. Q, Santos, A. L., & Pinho, H. (2020). Using local refinements on 360 stitching from dual-fisheye cameras. In VISIGRAPP (5: VISAPP) (pp. 17\u201326).","DOI":"10.5220\/0008874100170026"},{"key":"2391_CR191","first-page":"1","volume":"2","author":"JP Rolland","year":"2005","unstructured":"Rolland, J. P., & Hua, H. (2005). Head-mounted display systems. Encyclopedia of Optical Engineering, 2, 1\u201314.","journal-title":"Encyclopedia of Optical Engineering"},{"key":"2391_CR192","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., & Ommer, B. (2021). High-resolution image synthesis with latent diffusion models. In CVPR.","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"2391_CR193","doi-asserted-by":"crossref","unstructured":"Schindler, G., & Dellaert, F. (2004). Atlanta world: an expectation maximization framework for simultaneous low-level edge grouping and camera calibration in complex man-made environments. In CVPR.","DOI":"10.1109\/CVPR.2004.1315033"},{"key":"2391_CR194","doi-asserted-by":"crossref","unstructured":"Seidel, R., Apitzsch, A., & Hirtz, G. (2021). Omniflow: Human omnidirectional optical flow. In CVPR Workshops.","DOI":"10.1109\/CVPRW53098.2021.00407"},{"key":"2391_CR195","doi-asserted-by":"crossref","unstructured":"Seok, H., & Lim, J. (2019). Rovo: Robust omnidirectional visual odometry for wide-baseline wide-fov camera systems. In 2019 International Conference on Robotics and Automation (ICRA) (pp. 6344\u20136350).","DOI":"10.1109\/ICRA.2019.8793758"},{"key":"2391_CR196","doi-asserted-by":"publisher","first-page":"6225","DOI":"10.1109\/LRA.2020.3010457","volume":"5","author":"H Seok","year":"2020","unstructured":"Seok, H., & Lim, J. (2020). Rovins: Robust omnidirectional visual inertial navigation system. IEEE Robotics and Automation Letters, 5, 6225\u20136232.","journal-title":"IEEE Robotics and Automation Letters"},{"key":"2391_CR197","unstructured":"Shakerinava, M., & Ravanbakhsh, S. (2021). Equivariant networks for pixelized spheres. In ICML."},{"key":"2391_CR198","doi-asserted-by":"crossref","unstructured":"Shang, Z., Liu, Y., Li, Guoyi, Z., Yunjian, M., Jingbo, L., Jinxia, & Wang, L. (2022). Viewport-oriented panoramic image inpainting. In ICIP.","DOI":"10.1109\/ICIP46576.2022.9897208"},{"key":"2391_CR199","doi-asserted-by":"crossref","unstructured":"Shen, Q. Y., Huang, T. G., Ding, P. X., & He, J. (2021a). Training real-time panoramic object detectors with virtual dataset. In ICASSP.","DOI":"10.1109\/ICASSP39728.2021.9414503"},{"key":"2391_CR200","doi-asserted-by":"crossref","unstructured":"Shen, Z., Lin, C., Liao, K., Nie, L., Zheng, Z., & Zhao, Y.(2022). Panoformer: Panorama transformer for indoor 360$$^\\circ $$ depth estimation. In ECCV.","DOI":"10.1007\/978-3-031-19769-7_12"},{"key":"2391_CR201","doi-asserted-by":"crossref","unstructured":"Shen, Z., Shen, T., Lin, Z., & Ma, J. (2021b). Pdo-es2cnns: Partial differential operator based equivariant spherical cnns. In AAAI.","DOI":"10.1609\/aaai.v35i11.17154"},{"key":"2391_CR202","doi-asserted-by":"crossref","unstructured":"Shen, Z., Zheng, Z., Lin, C., Nie, L., Liao, K., Zheng, S., & Zhao, Y. (2023). Disentangling orthogonal planes for indoor panoramic room layout estimation with cross-scale distortion awareness. In CVPR.","DOI":"10.1109\/CVPR52729.2023.01663"},{"key":"2391_CR203","unstructured":"Shi, X., Chen, Z., Wang, H., Yeung, D. Y., Wong, W. K., & Woo, W. C. (2015). Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In NIPS."},{"issue":"5","key":"2391_CR204","doi-asserted-by":"publisher","first-page":"5570","DOI":"10.1109\/TITS.2023.3241212","volume":"24","author":"H Shi","year":"2023","unstructured":"Shi, H., Zhou, Y., Yang, K., Yin, X., Wang, Z., Ye, Y., Yin, Z., Meng, S., Li, P., & Wang, K. (2023). Panoflow: Learning 360$$^\\circ $$ optical flow for surrounding temporal understanding. IEEE Transactions on Intelligent Transportation Systems, 24(5), 5570\u20135585.","journal-title":"IEEE Transactions on Intelligent Transportation Systems"},{"key":"2391_CR205","doi-asserted-by":"crossref","unstructured":"Shum, K., Pang, H. W., Hua, B. S., Nguyen, D. T., & Yeung, S. K. (2023). Conditional 360-degree image synthesis for immersive indoor scene decoration. In ICCV (pp. 4455\u20134465).","DOI":"10.1109\/ICCV51070.2023.00413"},{"key":"2391_CR206","unstructured":"Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition."},{"key":"2391_CR207","doi-asserted-by":"crossref","unstructured":"Singh, A. (2024). An overview of multi-view fisheye for vision-first autonomous driving. In Authorea Preprints.","DOI":"10.36227\/techrxiv.172253983.30594891\/v1"},{"key":"2391_CR208","doi-asserted-by":"crossref","unstructured":"Sitzmann, V., Serrano, A., Pavel, A., Agrawala, M., Gutierrez, D., Masia, B., & Wetzstein, G. (2018). Saliency in vr: How do people explore virtual environments? TVCG.","DOI":"10.1109\/TVCG.2018.2793599"},{"key":"2391_CR209","unstructured":"Sohn, K., Lee, H., & Yan, X. (2015). Learning structured output representation using deep conditional generative models. In NIPS."},{"key":"2391_CR210","doi-asserted-by":"crossref","unstructured":"Somanath, G., & Kurz, D. (2021). Hdr environment map estimation for real-time augmented reality. In CVPR.","DOI":"10.1109\/CVPR46437.2021.01114"},{"key":"2391_CR211","doi-asserted-by":"crossref","unstructured":"Song, S., & Funkhouser, T. A. (2019). Neural illumination: Lighting prediction for indoor environments. In CVPR (pp. 6911\u20136919).","DOI":"10.1109\/CVPR.2019.00708"},{"key":"2391_CR212","unstructured":"Straub, J., Whelan, T., Ma, L., Chen, Y., Wijmans, E., Green, S., Engel, J. J., Mur-Artal, R., Ren, C., Verma, S., & Clarkson, A. (2019). The replica dataset: A digital replica of indoor spaces."},{"key":"2391_CR213","unstructured":"Su, Y. C., & Grauman, K. (2017). Learning spherical convolution for fast features from 360$$^\\circ $$ imagery. In NIPS."},{"key":"2391_CR214","doi-asserted-by":"crossref","unstructured":"Su, Y. C., & Grauman, K. (2019). Kernel transformer networks for compact spherical convolution. In CVPR.","DOI":"10.1109\/CVPR.2019.00967"},{"key":"2391_CR215","doi-asserted-by":"crossref","unstructured":"Su, Y. C., & Grauman, K. (2021). Learning spherical convolution for 360 recognition. In TPAMI.","DOI":"10.1109\/TPAMI.2021.3113612"},{"key":"2391_CR216","doi-asserted-by":"crossref","unstructured":"Su, J. W. Peng, C. H., Wonka, P., & Chu, H. K. (2023). Multi-view layout estimation via a geometry-aware panorama registration network: Gpr-net. In CVPR.","DOI":"10.1109\/CVPRW59228.2023.00688"},{"key":"2391_CR217","unstructured":"Sui, X., Zhu, H., Liu, X., Fang, Y., Wang, S., & Wang, Z. (2023). Perceptual quality assessment of 360$$^\\circ $$ images based on generative scanpath representation."},{"key":"2391_CR218","doi-asserted-by":"crossref","unstructured":"Sun, C., Hsiao, C. W., Sun, M., & Chen, H. T. (2019). Horizonnet: Learning room layout with 1d representation and pano stretch data augmentation. In CVPR.","DOI":"10.1109\/CVPR.2019.00114"},{"key":"2391_CR219","doi-asserted-by":"crossref","unstructured":"Sun, Y., Lu, A., & Yu, L. (2017b). Weighted-to-spherically-uniform quality evaluation for omnidirectional video. In IEEE SPL.","DOI":"10.1109\/LSP.2017.2720693"},{"key":"2391_CR220","doi-asserted-by":"crossref","unstructured":"Sun, W., Min, X., Zhai, G., Gu, K., Duan, H., & Ma, S. (2020). Mc360iqa: A multi-channel CNN for blind 360-degree image quality assessment. In IEEE J-STSP.","DOI":"10.1109\/ISCAS.2019.8702664"},{"key":"2391_CR221","doi-asserted-by":"crossref","unstructured":"Sun, C., Sun, M., & Chen, H. T. (2021). Hohonet: 360 indoor holistic understanding with latent horizontal features. In CVPR.","DOI":"10.1109\/CVPR46437.2021.00260"},{"key":"2391_CR222","doi-asserted-by":"crossref","unstructured":"Sun, C., Tai, W. E., Shih, Y. L., Chen, K. W., Syu, Y. J., The Kent, S., Wang, Y. C., & Chen, H. T. (2023). Seg2reg: Differentiable 2d segmentation to 1d regression rendering for 360 room layout reconstruction.","DOI":"10.1109\/CVPR52733.2024.00993"},{"key":"2391_CR223","doi-asserted-by":"crossref","unstructured":"Sun, W. Gu, K., Zhai, G., Ma, S., Lin, W., & Callet, P., Le. (2017a). Cviqd: Subjective quality evaluation of compressed virtual reality images. In ICIP.","DOI":"10.1109\/ICIP.2017.8296923"},{"key":"2391_CR224","doi-asserted-by":"crossref","unstructured":"Suzuki, T., & Yamanaka, T. (2018). Saliency map estimation for omni-directional image considering prior distributions. In SMC.","DOI":"10.1109\/SMC.2018.00358"},{"key":"2391_CR225","doi-asserted-by":"crossref","unstructured":"Tateno, K., Navab, N., & Tombari, F. (2018). Distortion-aware convolutional filters for dense prediction in panoramic images. In ECCV.","DOI":"10.1007\/978-3-030-01270-0_43"},{"key":"2391_CR226","doi-asserted-by":"crossref","unstructured":"Tian, C., Shao, F., Chai, X., Jiang, Q., Xu, L., & Ho, Y. S. (2023). Viewport-sphere-branch network for blind quality assessment of stitched 360$$^\\circ $$ omnidirectional images. In IEEE TCSVT.","DOI":"10.1109\/TCSVT.2022.3225172"},{"key":"2391_CR227","doi-asserted-by":"crossref","unstructured":"Tofighi, N., Jabbari, E., Mohamed, H., Imamoglu, N., Ozcinar, C., Erdem, E., & Erdem, A. (2023). St360iq: No-reference omnidirectional image quality assessment with spherical vision transformers. In ICASSP.","DOI":"10.1109\/ICASSP49357.2023.10096750"},{"issue":"5","key":"2391_CR228","doi-asserted-by":"publisher","first-page":"500","DOI":"10.1049\/iet-cvi.2018.5304","volume":"13","author":"G Tong","year":"2019","unstructured":"Tong, G., Chen, H., Li, Y., Du, X., & Zhang, Q. (2019). In Object detection for panoramic images based on MS-RPN structure in traffic road scenes. IET Computer Vision, 13(5), 500\u2013506.","journal-title":"IET Computer Vision"},{"key":"2391_CR229","unstructured":"Tortorella, G., Luz, J., Brites, C., & Pereira, F. (2017). Saliency-driven omnidirectional imaging adaptive coding: Modeling and assessment. In MMSP."},{"key":"2391_CR230","doi-asserted-by":"crossref","unstructured":"Tran, P. V. (2021). Sslayout360: Semi-supervised indoor layout estimation from 360$$^\\circ $$ panorama. In CVPR.","DOI":"10.1109\/CVPR46437.2021.01510"},{"issue":"11","key":"2391_CR231","doi-asserted-by":"publisher","first-page":"3097","DOI":"10.3390\/s20113097","volume":"20","author":"H Ullah","year":"2020","unstructured":"Ullah, H., Zia, O., Kim, J. H., Han, K., & Lee, J. W. (2020). Automatic 360 mono-stereo panorama generation using a cost-effective multi-camera system. Sensors, 20(11), 3097.","journal-title":"Sensors"},{"key":"2391_CR232","doi-asserted-by":"crossref","unstructured":"Upenik, E., Rer\u00e1bek, M., & Ebrahimi, T. (2016). Testbed for subjective evaluation of omnidirectional visual content. In PCS.","DOI":"10.1109\/PCS.2016.7906378"},{"key":"2391_CR233","unstructured":"Uyen, T., Thi, H., Kwon, O. J., Choi, S., & Hussain, I. (2020). Subjective assessment of 360$$^\\circ $$ image projection formats. In IEEE Access."},{"key":"2391_CR234","unstructured":"Von Gioi, R. G., Jakubowicz, J., Morel, J. M., & Randall, G. (2008). Lsd: A fast line segment detector with a false detection control. PAMI."},{"key":"2391_CR235","doi-asserted-by":"crossref","unstructured":"Wang, K. H., & Lai, S. H. (2019). Object detection in curved space for 360-degree camera. In ICASSP.","DOI":"10.1109\/ICASSP.2019.8683093"},{"key":"2391_CR236","doi-asserted-by":"crossref","unstructured":"Wang, J. Q., Chen, Z., Ling, J., Xie, R., & Song, L. (2023). 360-degree panorama generation from few unregistered nfov images. In ACM MM.","DOI":"10.1145\/3581783.3612508"},{"key":"2391_CR237","doi-asserted-by":"crossref","unstructured":"Wang, F. E., Hu, H. N., Cheng, H. T., Lin, H. T., Yang, S. T., Shih, M. L., Chu, H. K., & Sun, M. (2018). Self-supervised learning of depth and camera motion from 360$$^\\circ $$ videos. In ACCV.","DOI":"10.1007\/978-3-030-20873-8_4"},{"key":"2391_CR238","doi-asserted-by":"crossref","unstructured":"Wang, Q., Li, W., Mou, C., Cheng, X., & Zhang, J. (2024b). 360dvd: Controllable panorama video generation with 360-degree video diffusion model.","DOI":"10.1109\/CVPR52733.2024.00660"},{"key":"2391_CR239","doi-asserted-by":"crossref","unstructured":"Wang, N. H., Solarte, B., Tsai, Y. H., Chiu, W. C., & Sun, M. (2020b). 360sd-net: 360 stereo depth estimation with learnable cost volume. In 2020 IEEE International Conference on Robotics and Automation (ICRA) (pp. 582\u2013588). IEEE","DOI":"10.1109\/ICRA40945.2020.9196975"},{"key":"2391_CR240","doi-asserted-by":"crossref","unstructured":"Wang, H., Xiang, X., Fan, Y., & Xue, J. H. (2024a). Customizing 360-degree panoramas through text-to-image diffusion models. In WACV.","DOI":"10.1109\/WACV57701.2024.00486"},{"key":"2391_CR241","unstructured":"Wang, P., Yang, A., Men, R., Lin, J., Bai, S., Li, Z., Ma, J., Zhou, C., Zhou, J., & Yang, H. (2022c). Ofa: Unifying architectures, tasks, and modalities through a simple sequence-to-sequence learning framework. In ICML."},{"key":"2391_CR242","doi-asserted-by":"crossref","unstructured":"Wang, Z., Yang, K., Shi, H., & Wang, K. (2022d). Lf-vio: A visual-inertial-odometry framework for large field-of-view cameras with negative plane. In 2022 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp. 4423\u20134430).","DOI":"10.1109\/IROS47612.2022.9981217"},{"key":"2391_CR243","doi-asserted-by":"crossref","unstructured":"Wang, F. E., Yeh, Y. H., Sun, M., Chiu, W. C., & Tsai, Y. H. (2020a). Bifuse: Monocular 360 depth estimation via bi-projection fusion. In CVPR.","DOI":"10.1109\/CVPR42600.2020.00054"},{"key":"2391_CR244","doi-asserted-by":"crossref","unstructured":"Wang, F. E., Yeh, Y. H., Sun, M., Chiu, W. C., & Tsai, Y. H. (2021). Led2-net: Monocular 360$$^\\circ $$ layout estimation via differentiable depth rendering. In CVPR.","DOI":"10.1109\/CVPR46437.2021.01276"},{"key":"2391_CR245","doi-asserted-by":"crossref","unstructured":"Wang, H. Hutchcroft, W., Li, Y., Wan, Z., Boyadzhiev, I., Tian, Y., & Kang, S. B. (2022b) Position-aware stereo merging network for room layout estimation: Psmnet. In CVPR.","DOI":"10.1109\/CVPR52688.2022.00842"},{"key":"2391_CR246","doi-asserted-by":"crossref","unstructured":"Wang, G. Yang, Yinuo, Loy, C. C., & Liu, Z. (2022a). HDR panorama generation for lighting estimation and editing: Stylelight. In ECCV.","DOI":"10.1007\/978-3-031-19784-0_28"},{"key":"2391_CR247","doi-asserted-by":"publisher","first-page":"6321","DOI":"10.1109\/TASE.2023.3324495","volume":"21","author":"ZY Wang","year":"2022","unstructured":"Wang, Z. Y., Kailun, S., Hao, M., Li, P., Gao, F., Bai, J., & Wang, K. (2022). Lf-vislam: A slam framework for large field-of-view cameras with negative imaging plane on mobile agents. IEEE Transactions on Automation Science and Engineering, 21, 6321\u20136335.","journal-title":"IEEE Transactions on Automation Science and Engineering"},{"key":"2391_CR248","doi-asserted-by":"crossref","unstructured":"Won, C., Seok, H., Cui, Z., Pollefeys, M., & Lim, J. (2020). Omnislam: Omnidirectional localization and dense mapping for wide-baseline multi-camera systems. In 2020 IEEE International Conference on Robotics and Automation (ICRA) (pp. 559\u2013566).","DOI":"10.1109\/ICRA40945.2020.9196695"},{"key":"2391_CR249","unstructured":"Wu, T., Shi, S., Cai, H., Cao, M., Xiao, J., Zheng, Y., & Yang, Y. (2023). Assessor360: Multi-sequence network for blind omnidirectional image quality assessment. In NIPS."},{"key":"2391_CR250","doi-asserted-by":"crossref","unstructured":"Wu, C., Zhang, R., Wang, Z., & Sun, L. (2020). A spherical convolution approach for learning long term viewport prediction in 360 immersive video. In AAAI.","DOI":"10.1609\/aaai.v34i01.7377"},{"key":"2391_CR251","unstructured":"Wu, T., Zheng, C., & Cham, T. J. (2024). Panodiffusion: 360-degree panorama outpainting via diffusion. In ICLR."},{"key":"2391_CR252","doi-asserted-by":"crossref","unstructured":"Xia, F., Zamir, A. R., He., Z., Sax, A., Malik, J., & Savarese, S. (2018). Gibson Env: real-world perception for embodied agents. In CVPR.","DOI":"10.1109\/CVPR.2018.00945"},{"key":"2391_CR253","unstructured":"Xiao, J., Ehinger, K. A., Oliva, A., & Torralba, A. (2012). Recognizing scene viewpoint using panoramic place representation. In CVPR."},{"key":"2391_CR254","doi-asserted-by":"crossref","unstructured":"Xie, W., Chu, G., Qian, Q., Yu, Y., Zhai, S., Chen, W., Nan, B., & Zhangv, G. (2024). Omnidirectional dense slam for back-to-back fisheye cameras. In 2024 IEEE International Conference on Robotics and Automation (ICRA) (pp. 1653\u20131660).","DOI":"10.1109\/ICRA57147.2024.10610351"},{"key":"2391_CR255","doi-asserted-by":"crossref","unstructured":"Xie, S., Lai, P. K., Laganiere, R., & Lang, J. (2019). Effective convolutional neural network layers in flow estimation for omnidirectional images. In 2019 International Conference on 3D Vision (3DV) (pp. 671\u2013680). IEEE","DOI":"10.1109\/3DV.2019.00079"},{"key":"2391_CR256","doi-asserted-by":"crossref","unstructured":"Xu, Mai, Song, Y., Wang, J., Qiao, M., Huo, L., & Wang, Z. (2019b). Predicting head movement in panoramic video: A deep reinforcement learning approach. TPAMI.","DOI":"10.1109\/TPAMI.2018.2858783"},{"key":"2391_CR257","doi-asserted-by":"crossref","unstructured":"Xu, Y., Dong, Y., Wu, J., Sun, Z., Shi, Z., Yu, J., & Gao, S. (2018b). Gaze prediction in dynamic 360$$^\\circ $$ immersive videos. In CVPR.","DOI":"10.1109\/CVPR.2018.00559"},{"key":"2391_CR258","doi-asserted-by":"crossref","unstructured":"Xu, M., Jiang, L., Li, C., Wang, Z., & Tao, X. (2020b). Viewport-based CNN: A multi-task approach for assessing 360$$^\\circ $$ video quality. In TPAMI.","DOI":"10.1109\/TPAMI.2020.3028509"},{"key":"2391_CR259","doi-asserted-by":"crossref","unstructured":"Xu, M., Li, C., Chen, Z., Wang, Z., & Guan, Z. (2019a). Assessing visual quality of omnidirectional videos. In IEEE TCSVT.","DOI":"10.1109\/TCSVT.2018.2886277"},{"key":"2391_CR260","doi-asserted-by":"crossref","unstructured":"Xu, M., Li, C., Zhang, S., & Callet, P.\u00a0L. (2020c). State-of-the-art in 360$$^\\circ $$ video\/image processing: Perception, assessment and compression. In IEEE J-STSP.","DOI":"10.1109\/JSTSP.2020.2966864"},{"key":"2391_CR261","doi-asserted-by":"crossref","unstructured":"Xu, H., Liu, X., Zhao, Q., Ma, Y., Yan, C., & Dai, F. (2023). Gaussian label distribution learning for spherical image object detection. In CVPR (pp. 1033\u20131042).","DOI":"10.1109\/CVPR52729.2023.00106"},{"key":"2391_CR262","doi-asserted-by":"crossref","unstructured":"Xu, M., Song, Y., Wang, J., Qiao, M., Huo, L., & Wang, Z. (2018a). Predicting head movement in panoramic video: A deep reinforcement learning approach. In IEEE T-PAMI.","DOI":"10.1109\/TPAMI.2018.2858783"},{"key":"2391_CR263","unstructured":"Xu, M., Yang, L., Tao, X., Duan, Y., & Wang, Z. (2019c). Saliency prediction on omnidirectional image with generative adversarial imitation learning. In IEEE TIP."},{"key":"2391_CR264","doi-asserted-by":"crossref","unstructured":"Xu, M., Yang, L., Tao, X., Duan, Y., & Wang, Z. (2021). Saliency prediction on omnidirectional image with generative adversarial imitation learning. In IEEE TIP.","DOI":"10.1109\/TIP.2021.3050861"},{"key":"2391_CR265","doi-asserted-by":"crossref","unstructured":"Xu, H., Zhao, Q., Ma, Y., Li, X. D., Yuan, P., Feng, B., Yan, C., & Dai, F. (2022). Pandora: A panoramic detection dataset for object with orientation. European Conference on Computer Vision (pp. 237\u2013252). Springer.","DOI":"10.1007\/978-3-031-20074-8_14"},{"key":"2391_CR266","doi-asserted-by":"crossref","unstructured":"Xu, J., Zhou, W., & Chen, Z. (2020a). Blind omnidirectional image quality assessment with viewport oriented graph convolutional networks. In IEEE TCSVT.","DOI":"10.1109\/TCSVT.2020.3015186"},{"key":"2391_CR267","unstructured":"Yan, Z., Li, X., Wang, K., Chen, S., Li, J., & Yang, J. (2023b). Distortion and uncertainty aware loss for panoramic depth completion. In International Conference on Machine Learning (pp. 39099\u201339109). PMLR."},{"key":"2391_CR268","doi-asserted-by":"crossref","unstructured":"Yan, Z., Li, X., Wang, K., Zhang, Z., Li, J., & Yang, J. (2022). Multi-modal masked pre-training for monocular panoramic depth completion. European Conference on Computer Vision (pp. 378\u2013395). Springer.","DOI":"10.1007\/978-3-031-19769-7_22"},{"key":"2391_CR269","doi-asserted-by":"crossref","unstructured":"Yang, K., Hu, X., & Stiefelhagen, R. (2021a). Is context-aware cnn ready for the surroundings? panoramic semantic segmentation in the wild. In IEEE TIP.","DOI":"10.1109\/TIP.2020.3048682"},{"key":"2391_CR270","doi-asserted-by":"crossref","unstructured":"Yang, K., Hu, X., Bergasa, L. M. , Eduardo, R., & Wang, K. (2020a). Pass: Panoramic annular semantic segmentation. IEEE TITS.","DOI":"10.1109\/TITS.2019.2938965"},{"key":"2391_CR271","doi-asserted-by":"crossref","unstructured":"Yang, K., Hu, X., Bergasa, L. M., Romera, E., Huang, X., Sun, D., & Wang, K. (2019a). Can we pass beyond the field of view? panoramic annular semantic segmentation for real-world surrounding perception. In IV.","DOI":"10.1109\/IVS.2019.8814042"},{"key":"2391_CR272","doi-asserted-by":"crossref","unstructured":"Yang, K., Hu, X., Chen, H., Xiang, K., Wang, K., & Stiefelhagen, R. (2019b). Ds-pass: Detail-sensitive panoramic annular semantic segmentation through swaftnet for surrounding sensing. In IV.","DOI":"10.1109\/IV47402.2020.9304706"},{"key":"2391_CR273","unstructured":"Yang, K., Hu, X., Fang, Y., Wang, K., & Stiefelhagen, R. (2020b). Omnisupervised omnidirectional semantic segmentation. In IEEE TITS."},{"key":"2391_CR274","doi-asserted-by":"crossref","unstructured":"Yang, S., Jiang, L., Liu, Z., & Loy, C. C. (2022b). Pastiche master: Exemplar-based high-resolution portrait style transfer. In CVPR.","DOI":"10.1109\/CVPR52688.2022.00754"},{"key":"2391_CR275","doi-asserted-by":"crossref","unstructured":"Yang, Q., Li, C., Dai, W., Zou, J., Qi, G. J., & Xiong, H. (2020c). Rotation equivariant graph convolutional network for spherical image classification. In CVPR.","DOI":"10.1109\/CVPR42600.2020.00436"},{"key":"2391_CR276","doi-asserted-by":"crossref","unstructured":"Yang, W., Qian, Y., K\u00e4m\u00e4r\u00e4inen, J. K., Cricri, F., & Fan, L. (2018). Object detection in equirectangular panorama. In ICPR.","DOI":"10.1109\/ICPR.2018.8546070"},{"key":"2391_CR277","doi-asserted-by":"crossref","unstructured":"Yang, S. T., Wang, F. E., Peng, C. H., Wonka, P., Sun, M., & Chu, H. K. (2019c). Dula-net: A dual-projection network for estimating room layouts from a single rgb panorama. In CVPR.","DOI":"10.1109\/CVPR.2019.00348"},{"key":"2391_CR278","doi-asserted-by":"crossref","unstructured":"Yang, L., Xu, M., Deng, X., & Feng, B. (2021c). Spatial attention-based non-reference perceptual quality prediction network for omnidirectional images. In ICME.","DOI":"10.1109\/ICME51207.2021.9428390"},{"key":"2391_CR279","doi-asserted-by":"crossref","unstructured":"Yang, L., Xu, M., Liu, T., Huo, L., & Gao, X. (2022a). Tvformer: Trajectory-guided visual quality assessment on 360$$^\\circ $$ images with transformers. In ACM MM.","DOI":"10.1145\/3503161.3547748"},{"key":"2391_CR280","doi-asserted-by":"crossref","unstructured":"Yang, K., Zhang, J., Rei\u00df, S., Hu, X., & Stiefelhagen, R. (2021b). Capturing omni-range context for omnidirectional segmentation. In CVPR.","DOI":"10.1109\/CVPR46437.2021.00143"},{"key":"2391_CR281","doi-asserted-by":"crossref","unstructured":"Yang, Y., Zhu, Y., Gao, Z., & Zhai, G. (2021). Salgfcn: Graph based fully convolutional network for panoramic saliency prediction. In VCIP.","DOI":"10.1109\/VCIP53242.2021.9675373"},{"key":"2391_CR282","doi-asserted-by":"publisher","first-page":"3556","DOI":"10.1109\/TMECH.2023.3348986","volume":"29","author":"Y Yang","year":"2024","unstructured":"Yang, Y., Pan, M., Tang, D., Wang, T., Yue, Y., Liu, T., & Fu, M. (2024). Mcov-slam: A multicamera omnidirectional visual slam system. IEEE\/ASME Transactions on Mechatronics, 29, 3556\u20133567.","journal-title":"IEEE\/ASME Transactions on Mechatronics"},{"key":"2391_CR283","doi-asserted-by":"publisher","first-page":"3863","DOI":"10.1007\/s11063-023-11226-z","volume":"55","author":"N Yan","year":"2023","unstructured":"Yan, N., Mei, Y., Xu, L., Yu, H., Sun, B., Wang, Z., & Chen, Y. (2023). Deep learning on image stitching with multi-viewpoint images: A survey. Neural Processing Letters, 55, 3863\u20133898.","journal-title":"Neural Processing Letters"},{"issue":"4","key":"2391_CR284","doi-asserted-by":"publisher","first-page":"2801","DOI":"10.1109\/COMST.2020.3006999","volume":"22","author":"A Yaqoob","year":"2020","unstructured":"Yaqoob, A., Bi, T., & Muntean, G. M. (2020). A survey on adaptive 360 video streaming: solutions, challenges & opportunities. IEEE Communications Surveys & Tutorials, 22(4), 2801\u20132838.","journal-title":"IEEE Communications Surveys & Tutorials"},{"key":"2391_CR285","unstructured":"Yoon, Y., Chung, I., Wang, L., & Yoon, K. J. (2022). Spheresr. In CVPR."},{"key":"2391_CR286","doi-asserted-by":"crossref","unstructured":"Yu, L., Gao, Y., Pakdaman, F., & Gabbouj, M. (2024). Panoramic image inpainting with gated convolution and contextual reconstruction loss.","DOI":"10.1109\/ICASSP48485.2024.10446469"},{"key":"2391_CR287","doi-asserted-by":"crossref","unstructured":"Yu, J., Grassi, A., & Hirtz, G. (2023c). Applications of deep learning for top-view omnidirectional imaging: A survey. In CVPR Workshop.","DOI":"10.1109\/CVPRW59228.2023.00683"},{"key":"2391_CR288","doi-asserted-by":"crossref","unstructured":"Yu, H., He, L., Jian, B., Feng, W., & Liu, S. (2023b). Panelnet: Understanding 360 indoor environment via panel representation. In CVPR.","DOI":"10.1109\/CVPR52729.2023.00091"},{"key":"2391_CR289","doi-asserted-by":"crossref","unstructured":"Yu, M.\u00a0C., Lakshman, H., & Girod, B. (2015). A framework to evaluate omnidirectional video coding schemes. In ISMAR.","DOI":"10.1109\/ISMAR.2015.12"},{"key":"2391_CR290","doi-asserted-by":"crossref","unstructured":"Yu, F., Wang, X., Cao, M., Li, G., Shan, Y., & Dong, C. (2023a). Omnidirectional image super-resolution with distortion-aware transformer: Osrt. In CVPR.","DOI":"10.1109\/CVPR52729.2023.01276"},{"key":"2391_CR291","doi-asserted-by":"crossref","unstructured":"Yuan, M., & Richardt, C. (2021). 360 optical flow using tangent images. In British Machine Vision Conference:(BMVC). Christian Richardt","DOI":"10.5244\/C.35.80"},{"issue":"11","key":"2391_CR292","doi-asserted-by":"publisher","first-page":"2622","DOI":"10.3390\/s19112622","volume":"19","author":"D Yu","year":"2019","unstructured":"Yu, D., & Ji, S. (2019). Grid based spherical CNN for object detection from panoramic images. Sensors, 19(11), 2622.","journal-title":"Sensors"},{"key":"2391_CR293","doi-asserted-by":"crossref","unstructured":"Yun, I.\u00a0D., Lee, H. J., & Rhee, C. E. (2021). Improving 360 monocular depth estimation via non-local dense prediction transformer and joint supervised and self-supervised learning. In AAAI.","DOI":"10.1609\/aaai.v36i3.20231"},{"key":"2391_CR294","doi-asserted-by":"crossref","unstructured":"Yun, I., Shin, C., Lee, H., Lee, H. J., & Rhee, C. E. (2023). Egformer: Equirectangular geometry-biased transformer for 360 depth estimation. In ICCV.","DOI":"10.1109\/ICCV51070.2023.00561"},{"key":"2391_CR295","doi-asserted-by":"crossref","unstructured":"Zakharchenko, V., Choi, K.\u00a0P., & Park, J. (2016). Quality metric for spherical panoramic video. In Optical Engineering + Applications.","DOI":"10.1117\/12.2235885"},{"key":"2391_CR296","doi-asserted-by":"crossref","unstructured":"Zhan, F., Yu, Y., Wu, R., Zhang, C., Lu, S., Shao, L., Ma, F., & Xie, X. (2021). Gmlight: Lighting estimation via geometric distribution approximation. In IEEE TIP.","DOI":"10.1109\/TIP.2022.3151997"},{"key":"2391_CR297","unstructured":"Zhang, J. (2018). A 360$$^\\circ $$ video-based robot platform for telepresent redirected walking. In VAM-HRI."},{"key":"2391_CR298","doi-asserted-by":"crossref","unstructured":"Zhang, C., Cui, Z., Chen, C., Liu, S., Zeng, B., Bao, H., & Zhang, Y. (2021a). Deeppanocontext: Panoramic 3d scene understanding with holistic scene context graph and relation-based optimization. In ICCV.","DOI":"10.1109\/ICCV48922.2021.01240"},{"key":"2391_CR299","doi-asserted-by":"crossref","unstructured":"Zhang, X., Li, Z., Gong, Y., Jin, D., Li, J., Wang, L., Zhu, Y., & Liu, H. (2022c). Openmpd: An open multimodal perception dataset for autonomous driving. In IEEE TVT.","DOI":"10.1007\/978-981-99-3280-1_7"},{"key":"2391_CR300","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Li, K., Li, K., Wang, L., Zhong, B., & Fu, Y. (2018a). Image super-resolution using very deep residual channel attention networks. In Proceedings of the European conference on computer vision (ECCV) (pp. 286\u2013301).","DOI":"10.1007\/978-3-030-01234-2_18"},{"key":"2391_CR301","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Liu, Y., Liu, J., Miao, J., Argyriou, A., Wang, L., & Xu, Z. (2022d). 360-attack: Distortion-aware perturbations from perspective-views. In CVPR.","DOI":"10.1109\/CVPR52688.2022.01461"},{"key":"2391_CR302","doi-asserted-by":"crossref","unstructured":"Zhang, W., Liu, Y.,\u00a0Zheng, X., & Wang, L. (2024). Goodsam: Bridging domain and capacity gaps via segment anything model for distortion-aware panoramic semantic segmentation. In 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 28264\u201328273).","DOI":"10.1109\/CVPR52733.2024.02670"},{"key":"2391_CR303","doi-asserted-by":"crossref","unstructured":"Zhang, C., Liwicki, S., He, S., Smith, W., & Cipolla, R. (2023a). Hexnet: An orientation-aware deep learning framework for omni-directional input. In TPAMI.","DOI":"10.1109\/TPAMI.2023.3307152"},{"key":"2391_CR304","doi-asserted-by":"crossref","unstructured":"Zhang, C., Liwicki, S., Smith, W., & Cipolla, R. (2019a). Orientation-aware semantic segmentation on icosahedron spheres. In ICCV.","DOI":"10.1109\/ICCV.2019.00363"},{"key":"2391_CR305","doi-asserted-by":"crossref","unstructured":"Zhang, J., Ma, C., Yang, K., Roitberg, A., Peng, K., & Stiefelhagen, R. (2021b). Transfer beyond the field of view: Dense panoramic semantic segmentation via unsupervised domain adaptation. In IEEE TITS.","DOI":"10.1109\/TITS.2021.3123070"},{"key":"2391_CR306","doi-asserted-by":"crossref","unstructured":"Zhang, L., Rao, A., & Agrawala, M. (2023b). Adding conditional control to text-to-image diffusion models. In ICCV.","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"2391_CR307","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Song, S., Tan, P., & Xiao, J. (2014). Panocontext: A whole-room 3d context model for panoramic scene understanding. In ECCV.","DOI":"10.1007\/978-3-319-10599-4_43"},{"key":"2391_CR308","doi-asserted-by":"crossref","unstructured":"Zhang, J., Sunkavalli, K., Hold-Geoffroy, Y., Hadap, S., Eisenmann, J., & Lalonde, J. F. (2019b). All-weather deep outdoor lighting estimation. In CVPR (pp. 10150\u201310158).","DOI":"10.1109\/CVPR.2019.01040"},{"key":"2391_CR309","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Xu, Y., Yu, J., & Gao, S. (2018b). Saliency detection in 360$$^\\circ $$ videos. In ECCV.","DOI":"10.1007\/978-3-030-01234-2_30"},{"key":"2391_CR310","doi-asserted-by":"crossref","unstructured":"Zhang, J., Yang, K., Ma, C., Rei\u00df, S., Peng, K., & Stiefelhagen, R. (2022a). Bending reality: Distortion-aware transformers for adapting to panoramic semantic segmentation. In CVPR.","DOI":"10.1109\/CVPR52688.2022.01641"},{"key":"2391_CR311","unstructured":"Zhang, J., Yang, K., Shi, H., Rei\u00df, S., Peng, K., Ma, C., Fu, H., Wang, K., & Stiefelhagen, R. (2022b). Behind every domain there is a shift: Adapting distortion-aware vision transformers for panoramic semantic segmentation. In Arxiv."},{"key":"2391_CR312","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Zhao, P., Bian, K., Liu, Y., Song, L., & Li, X. (2019c). Drl360: 360-degree video streaming with deep reinforcement learning. In INFOCOM (pp. 1252\u20131260).","DOI":"10.1109\/INFOCOM.2019.8737361"},{"key":"2391_CR313","doi-asserted-by":"crossref","unstructured":"Zhang, P.,\u00a0Zhang, B., Zhang, T., Chen, D., Wang, Y., & Wen, F. (2021c). Prototypical pseudo label denoising and target structure learning for domain adaptive semantic segmentation. In CVPR.","DOI":"10.1109\/CVPR46437.2021.01223"},{"key":"2391_CR314","doi-asserted-by":"crossref","unstructured":"Zhang, Y. Liu, Y., Liu, J., Zhan, P., Wang, L., & Xu, Z. (2022e). Single-perspective attack for generating adversarial omnidirectional images: Sp attack. In ICASSP.","DOI":"10.1109\/ICASSP43922.2022.9746052"},{"key":"2391_CR315","volume-title":"Lighting estimation via spherical distribution approximation","author":"F Zhan","year":"2020","unstructured":"Zhan, F., Zhang, C., Yu, Y., Chang, Y., Lu, S., Ma, F., & Xie, X. (2020). Lighting estimation via spherical distribution approximation. ArXiv: Emlight."},{"key":"2391_CR316","unstructured":"Zhao, Q., Chen, B., Xu, H., Ma, Y., Li, X. D., Feng, B., Yan, C., & Dai, F. (2021a). Unbiased iou for spherical image object detection. In AAAI."},{"key":"2391_CR317","unstructured":"Zhao, S., Cui, J., Sheng, Y., Dong, Y., Liang, X., , Eric I. C., & Xu, Y. (2021b). Large scale image completion via co-modulated generative adversarial networks. In ICLR."},{"key":"2391_CR318","doi-asserted-by":"crossref","unstructured":"Zhao, P., You, A., Zhang, Y., Liu, J., Bian, K., & Tong, Y. (2020). Spherical criteria for fast and accurate 360 object detection. In AAAI.","DOI":"10.1609\/aaai.v34i07.6995"},{"key":"2391_CR319","doi-asserted-by":"publisher","first-page":"143","DOI":"10.1007\/s11263-014-0787-4","volume":"113","author":"Q Zhao","year":"2014","unstructured":"Zhao, Q., Feng, W., Wan, L., & Zhang, J. (2014). Sphorb: A fast and robust binary feature on the sphere. International Journal of Computer Vision, 113, 143\u2013159.","journal-title":"International Journal of Computer Vision"},{"key":"2391_CR320","doi-asserted-by":"crossref","unstructured":"Zheng, X., Pan, T., Luo, Y., & Wang, L. (2023b). Look at the neighbor: Distortion-aware unsupervised domain adaptation for panoramic semantic segmentation. In ICCV.","DOI":"10.1109\/ICCV51070.2023.01713"},{"key":"2391_CR321","doi-asserted-by":"crossref","unstructured":"Zheng, J., Zhang, J., Li, J., Tang, R., Gao, S., & Zhou, Z. (2020). Structured3d: A large photo-realistic dataset for structured 3d modeling. In ECCV.","DOI":"10.1007\/978-3-030-58545-7_30"},{"key":"2391_CR322","doi-asserted-by":"crossref","unstructured":"Zheng, X., Zhou, P., Vasilakos, A.\u00a0V., & Wang, L. (2024). Semantics distortion and style matter: Towards source-free uda for panoramic segmentation. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (pp. 27885\u201327895).","DOI":"10.1109\/CVPR52733.2024.02634"},{"key":"2391_CR323","doi-asserted-by":"crossref","unstructured":"Zheng, X., Zhu, J., Liu, Y., Cao, Z., Fu, C., & Wang, L. (2023a). Dual-path unsupervised domain adaptation for panoramic semantic segmentation: Both style and distortion matter. In CVPR.","DOI":"10.1109\/CVPR52729.2023.00130"},{"key":"2391_CR324","doi-asserted-by":"crossref","unstructured":"Zhou, B., Khosla, A., Lapedriza, \u00c0., Oliva, A., & Torralba, A. (2016). Learning deep features for discriminative localization. In CVPR.","DOI":"10.1109\/CVPR.2016.319"},{"key":"2391_CR325","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Sun, Y., Li, L.,\u00a0Gu, K., & Fang, Y. (2022). Omnidirectional image quality assessment by distortion discrimination assisted multi-stream network. In IEEE TCSVT.","DOI":"10.1109\/TCSVT.2021.3081162"},{"key":"2391_CR326","doi-asserted-by":"crossref","unstructured":"Zhou, W., Xu, J., Jiang, Q., & Chen, Z. (2021). No-reference quality assessment for 360-degree images by analysis of multifrequency information and local-global naturalness. In IEEE TCSVT.","DOI":"10.1109\/TCSVT.2021.3081182"},{"key":"2391_CR327","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Yu, M., Ma, H., Shao, H., & Jiang, G. (2018). Weighted-to-spherically-uniform ssim objective quality evaluation for panoramic video. In ICSP.","DOI":"10.1109\/ICSP.2018.8652269"},{"key":"2391_CR328","doi-asserted-by":"crossref","unstructured":"Zhu, D., Chen, Y., Han, T., Zhao, D., Zhu, Y., Zhou, Q., Zhai, G., & Yang, X. (2020a). Ransp: Ranking attention network for saliency prediction on omnidirectional images. In ICME.","DOI":"10.1109\/ICME46284.2020.9102867"},{"key":"2391_CR329","doi-asserted-by":"crossref","unstructured":"Zhu, D., Chen, Y., Zhao, D., Min, X., Zhou, Q., Yu, S., Zhai, G., & Yang, X. (2021a). A lightweight saliency prediction model for omnidirectional images. In ICME.","DOI":"10.1109\/ICME51207.2021.9428420"},{"key":"2391_CR330","doi-asserted-by":"crossref","unstructured":"Zhu, W., Wen, T., Song, G., Ma, X., & Wang, L. (2023). Hierarchical transformer for scalable graph learning.","DOI":"10.24963\/ijcai.2023\/523"},{"key":"2391_CR331","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Zhai, G., Min, X., & Zhou, J. (2020b). The prediction of saliency map for head and eye movements in 360 degree images. In TMM.","DOI":"10.1109\/TMM.2019.2957986"},{"key":"2391_CR332","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Zhai, G., Yang, Y., Duan, H., Min, X., & Yang, X. (2021c). Viewing behavior supported visual saliency predictor for 360 degree videos. In IEEE TCSVT.","DOI":"10.1109\/TCSVT.2021.3126590"},{"key":"2391_CR333","doi-asserted-by":"crossref","unstructured":"Zhu, D., Zhang, K., Zhang, G., Zhou, Q., Min, X., Zhai, G., & Yang, X. (2022). Decoupled dynamic group equivariant filter for saliency prediction on omnidirectional image. In Neurocomputing.","DOI":"10.1016\/j.neucom.2022.09.107"},{"key":"2391_CR334","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Zhang, Y.,\u00a0Li, S., & Shi, B. (2021b). Spatially-varying outdoor lighting estimation from intrinsics. In CVPR.","DOI":"10.1109\/CVPR46437.2021.01264"},{"key":"2391_CR335","doi-asserted-by":"crossref","unstructured":"Zhuang, C., Lu, Z., Wang, Y., Xiao, J., & Wang, Y. (2022). Adaptively combined dilated convolution for monocular panorama depth estimation: Acdnet. In AAAI.","DOI":"10.1609\/aaai.v36i3.20278"},{"issue":"10","key":"2391_CR336","doi-asserted-by":"publisher","first-page":"12474","DOI":"10.1109\/TPAMI.2023.3272949","volume":"45","author":"C Zhuang","year":"2023","unstructured":"Zhuang, C., Lu, Z., Wang, Y., Xiao, J., & Wang, Y. (2023). Spdet: Edge-aware self-supervised panoramic depth estimation transformer with spherical geometry. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(10), 12474\u201312489.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"2391_CR337","doi-asserted-by":"crossref","unstructured":"Zia, O., Kim, J. H., Han, K., & Lee, J.\u00a0W. (2019). 360 panorama generation using drone mounted fisheye cameras. In 2019 IEEE International Conference on Consumer Electronics (ICCE) (pp. 1\u20133). IEEE","DOI":"10.1109\/ICCE.2019.8661954"},{"key":"2391_CR338","doi-asserted-by":"crossref","unstructured":"Zink, M., Sitaraman, R.\u00a0K., & Nahrstedt, K. (2019). Scalable 360$$^\\circ $$ video stream delivery: Challenges, solutions, and opportunities. In Proceedings of the IEEE.","DOI":"10.1109\/JPROC.2019.2894817"},{"key":"2391_CR339","doi-asserted-by":"crossref","unstructured":"Zioulis, N., Karakottas, A., Zarpalas, D., & Daras, P. (2018). Dense depth estimation for indoors spherical panoramas: Omnidepth. In ECCV.","DOI":"10.1007\/978-3-030-01231-1_28"},{"key":"2391_CR340","doi-asserted-by":"crossref","unstructured":"Zioulis, N., Karakottas, A., Zarpalas, D., Alvarez, F., & Daras, P. (2019). Spherical view synthesis for self-supervised 360 depth estimation. In 3DV.","DOI":"10.1109\/3DV.2019.00081"},{"key":"2391_CR341","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2021.104160","volume":"110","author":"N Zioulis","year":"2021","unstructured":"Zioulis, N., \u00c1lvarez, F., Zarpalas, D., & Daras, P. (2021). Single-shot cuboids: Geodesics-based end-to-end Manhattan aligned layout estimation from spherical panoramas. Image and Vision Computing, 110, 104160.","journal-title":"Image and Vision Computing"},{"key":"2391_CR342","doi-asserted-by":"crossref","unstructured":"Zou, C., Colburn, A.,\u00a0Shan, Q., & Hoiem, D. (2018). Layoutnet: Reconstructing the 3d room layout from a single rgb image. In CVPR.","DOI":"10.1109\/CVPR.2018.00219"},{"key":"2391_CR343","doi-asserted-by":"crossref","unstructured":"Zou, C., Su, J. W., Peng, C. H., Colburn, A., Shan, Q., Wonka, P., Chu, H. K., & Hoiem, D. (2021). Manhattan room layout reconstruction from a single 360$$^\\circ $$ image: A comparative study of state-of-the-art methods. In IJCV.","DOI":"10.1007\/s11263-020-01426-8"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02391-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-025-02391-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02391-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T12:24:29Z","timestamp":1757161469000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-025-02391-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,10]]},"references-count":343,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["2391"],"URL":"https:\/\/doi.org\/10.1007\/s11263-025-02391-w","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,4,10]]},"assertion":[{"value":"15 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 February 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 April 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}