{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T16:08:33Z","timestamp":1775146113716,"version":"3.50.1"},"publisher-location":"Cham","reference-count":56,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031198380","type":"print"},{"value":"9783031198397","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19839-7_14","type":"book-chapter","created":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T11:40:06Z","timestamp":1666438806000},"page":"233-250","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":35,"title":["Efficient Deep Visual and\u00a0Inertial Odometry with\u00a0Adaptive Visual Modality Selection"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1301-6493","authenticated-orcid":false,"given":"Mingyu","family":"Yang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3008-9208","authenticated-orcid":false,"given":"Yu","family":"Chen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6658-5502","authenticated-orcid":false,"given":"Hun-Seok","family":"Kim","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,23]]},"reference":[{"key":"14_CR1","doi-asserted-by":"crossref","unstructured":"Almalioglu, Y., et al.: SelfVio: self-supervised deep monocular visual-inertial odometry and depth estimation. arXiv preprint arXiv:1911.09968 (2019)","DOI":"10.1109\/ICRA.2019.8793512"},{"key":"14_CR2","unstructured":"Bolukbasi, T., Wang, J., Dekel, O., Saligrama, V.: Adaptive neural networks for efficient inference. In: International Conference on Machine Learning, pp. 527\u2013536. PMLR (2017)"},{"issue":"6","key":"14_CR3","doi-asserted-by":"publisher","first-page":"1309","DOI":"10.1109\/TRO.2016.2624754","volume":"32","author":"C Cadena","year":"2016","unstructured":"Cadena, C., et al.: Past, present, and future of simultaneous localization and mapping: toward the robust-perception age. IEEE Trans. Rob. 32(6), 1309\u20131332 (2016)","journal-title":"IEEE Trans. Rob."},{"key":"14_CR4","unstructured":"Campos, V., Jou, B., Gir\u00f3-i Nieto, X., Torres, J., Chang, S.F.: Skip RNN: learning to skip state updates in recurrent neural networks. arXiv preprint arXiv:1708.06834 (2017)"},{"key":"14_CR5","unstructured":"Chen, C., Rosa, S., Lu, C.X., Trigoni, N., Markham, A.: SelectFusion: a generic framework to selectively learn multisensory fusion. arXiv preprint arXiv:1912.13077 (2019)"},{"key":"14_CR6","doi-asserted-by":"crossref","unstructured":"Chen, C., et al.: Selective sensor fusion for neural visual-inertial odometry. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10542\u201310551 (2019)","DOI":"10.1109\/CVPR.2019.01079"},{"key":"14_CR7","doi-asserted-by":"crossref","unstructured":"Clark, R., Wang, S., Wen, H., Markham, A., Trigoni, N.: ViNet: visual-inertial odometry as a sequence-to-sequence learning problem. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31 (2017)","DOI":"10.1609\/aaai.v31i1.11215"},{"key":"14_CR8","doi-asserted-by":"crossref","unstructured":"Dosovitskiy, A., et al.: FlowNet: learning optical flow with convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2758\u20132766 (2015)","DOI":"10.1109\/ICCV.2015.316"},{"issue":"3","key":"14_CR9","doi-asserted-by":"publisher","first-page":"611","DOI":"10.1109\/TPAMI.2017.2658577","volume":"40","author":"J Engel","year":"2017","unstructured":"Engel, J., Koltun, V., Cremers, D.: Direct sparse odometry. IEEE Trans. Pattern Anal. Mach. Intell. 40(3), 611\u2013625 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"14_CR10","doi-asserted-by":"crossref","unstructured":"Forster, C., Pizzoli, M., Scaramuzza, D.: SVO: fast semi-direct monocular visual odometry. In: 2014 IEEE International Conference on Robotics and Automation (ICRA), pp. 15\u201322. IEEE (2014)","DOI":"10.1109\/ICRA.2014.6906584"},{"key":"14_CR11","doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? The Kitti vision benchmark suite. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 3354\u20133361. IEEE (2012)","DOI":"10.1109\/CVPR.2012.6248074"},{"issue":"10","key":"14_CR12","doi-asserted-by":"publisher","first-page":"75","DOI":"10.1145\/84537.84552","volume":"33","author":"PW Glynn","year":"1990","unstructured":"Glynn, P.W.: Likelihood ratio gradient estimation for stochastic systems. Commun. ACM 33(10), 75\u201384 (1990)","journal-title":"Commun. ACM"},{"key":"14_CR13","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac Aodha, O., Firman, M., Brostow, G.J.: Digging into self-supervised monocular depth estimation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3828\u20133838 (2019)","DOI":"10.1109\/ICCV.2019.00393"},{"key":"14_CR14","unstructured":"Graves, A.: Adaptive computation time for recurrent neural networks. arXiv preprint arXiv:1603.08983 (2016)"},{"key":"14_CR15","doi-asserted-by":"crossref","unstructured":"Han, L., Lin, Y., Du, G., Lian, S.: DeepVio: self-supervised deep learning of monocular visual inertial odometry using 3D geometric constraints. In: 2019 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 6906\u20136913. IEEE (2019)","DOI":"10.1109\/IROS40897.2019.8968467"},{"key":"14_CR16","unstructured":"Hansen, C., Hansen, C., Alstrup, S., Simonsen, J.G., Lioma, C.: Neural speed reading with structural-jump-LSTM. arXiv preprint arXiv:1904.00761 (2019)"},{"key":"14_CR17","doi-asserted-by":"crossref","unstructured":"Hong, E., Lim, J.: Visual inertial odometry using coupled nonlinear optimization. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 6879\u20136885. IEEE (2017)","DOI":"10.1109\/IROS.2017.8206610"},{"key":"14_CR18","unstructured":"Hua, W., Zhou, Y., De Sa, C.M., Zhang, Z., Suh, G.E.: Channel gating neural networks. Adv. Neural Inf. Process. Syst. 32 (2019)"},{"key":"14_CR19","unstructured":"Huang, G., Chen, D., Li, T., Wu, F., Van Der Maaten, L., Weinberger, K.Q.: Multi-scale dense networks for resource efficient image classification. arXiv preprint arXiv:1703.09844 (2017)"},{"key":"14_CR20","unstructured":"Jang, E., Gu, S., Poole, B.: Categorical reparameterization with gumbel-softmax. arXiv preprint arXiv:1611.01144 (2016)"},{"key":"14_CR21","unstructured":"Kingma, D.P., Welling, M.: Auto-encoding variational Bayes. arXiv preprint arXiv:1312.6114 (2013)"},{"key":"14_CR22","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 25 (2012)"},{"issue":"3","key":"14_CR23","doi-asserted-by":"publisher","first-page":"314","DOI":"10.1177\/0278364914554813","volume":"34","author":"S Leutenegger","year":"2015","unstructured":"Leutenegger, S., Lynen, S., Bosse, M., Siegwart, R., Furgale, P.: Keyframe-based visual-inertial odometry using nonlinear optimization. Int. J. Robot. Res. 34(3), 314\u2013334 (2015)","journal-title":"Int. J. Robot. Res."},{"issue":"6","key":"14_CR24","doi-asserted-by":"publisher","first-page":"690","DOI":"10.1177\/0278364913481251","volume":"32","author":"M Li","year":"2013","unstructured":"Li, M., Mourikis, A.I.: High-precision, consistent EKF-based visual-inertial odometry. Int. J. Robot. Res. 32(6), 690\u2013711 (2013)","journal-title":"Int. J. Robot. Res."},{"key":"14_CR25","doi-asserted-by":"crossref","unstructured":"Liu, L., Li, G., Li, T.H.: AtVio: attention guided visual-inertial odometry. In: ICASSP 2021\u20132021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4125\u20134129. IEEE (2021)","DOI":"10.1109\/ICASSP39728.2021.9413912"},{"key":"14_CR26","doi-asserted-by":"crossref","unstructured":"Lowe, D.G.: Object recognition from local scale-invariant features. In: Proceedings of the Seventh IEEE International Conference on Computer Vision, vol. 2, pp. 1150\u20131157. IEEE (1999)","DOI":"10.1109\/ICCV.1999.790410"},{"key":"14_CR27","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1007\/978-3-030-58571-6_6","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Meng","year":"2020","unstructured":"Meng, Y., et al.: AR-Net: adaptive frame resolution for efficient action recognition. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12352, pp. 86\u2013104. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58571-6_6"},{"key":"14_CR28","unstructured":"Meng, Y., et al.: AdaFuse: adaptive temporal fusion network for efficient action recognition. arXiv preprint arXiv:2102.05775 (2021)"},{"issue":"132","key":"14_CR29","first-page":"1","volume":"21","author":"S Mohamed","year":"2020","unstructured":"Mohamed, S., Rosca, M., Figurnov, M., Mnih, A.: Monte Carlo gradient estimation in machine learning. J. Mach. Learn. Res. 21(132), 1\u201362 (2020)","journal-title":"J. Mach. Learn. Res."},{"issue":"5","key":"14_CR30","doi-asserted-by":"publisher","first-page":"1147","DOI":"10.1109\/TRO.2015.2463671","volume":"31","author":"R Mur-Artal","year":"2015","unstructured":"Mur-Artal, R., Montiel, J.M.M., Tardos, J.D.: ORB-SLAM: a versatile and accurate monocular slam system. IEEE Trans. Rob. 31(5), 1147\u20131163 (2015)","journal-title":"IEEE Trans. Rob."},{"issue":"5","key":"14_CR31","doi-asserted-by":"publisher","first-page":"1255","DOI":"10.1109\/TRO.2017.2705103","volume":"33","author":"R Mur-Artal","year":"2017","unstructured":"Mur-Artal, R., Tard\u00f3s, J.D.: ORB-SLAM2: an open-source slam system for monocular, stereo, and RGB-D cameras. IEEE Trans. Rob. 33(5), 1255\u20131262 (2017)","journal-title":"IEEE Trans. Rob."},{"key":"14_CR32","doi-asserted-by":"crossref","unstructured":"Nist\u00e9r, D., Naroditsky, O., Bergen, J.: Visual odometry. In: Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2004, vol. 1, p. I-I. IEEE (2004)","DOI":"10.1109\/CVPR.2004.1315094"},{"key":"14_CR33","doi-asserted-by":"crossref","unstructured":"Panda, R., et al.: AdaMML: adaptive multi-modal learning for efficient video recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7576\u20137585 (2021)","DOI":"10.1109\/ICCV48922.2021.00748"},{"issue":"4","key":"14_CR34","doi-asserted-by":"publisher","first-page":"1004","DOI":"10.1109\/TRO.2018.2853729","volume":"34","author":"T Qin","year":"2018","unstructured":"Qin, T., Li, P., Shen, S.: VINS-MONO: a robust and versatile monocular visual-inertial state estimator. IEEE Trans. Rob. 34(4), 1004\u20131020 (2018)","journal-title":"IEEE Trans. Rob."},{"key":"14_CR35","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 28 (2015)"},{"key":"14_CR36","unstructured":"Rezende, D.J., Mohamed, S., Wierstra, D.: Stochastic backpropagation and approximate inference in deep generative models. In: International Conference on Machine Learning, pp. 1278\u20131286. PMLR (2014)"},{"key":"14_CR37","doi-asserted-by":"crossref","unstructured":"Rublee, E., Rabaud, V., Konolige, K., Bradski, G.: ORB: an efficient alternative to sift or surf. In: 2011 International Conference on Computer Vision, pp. 2564\u20132571. IEEE (2011)","DOI":"10.1109\/ICCV.2011.6126544"},{"issue":"4","key":"14_CR38","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1109\/MRA.2011.943233","volume":"18","author":"D Scaramuzza","year":"2011","unstructured":"Scaramuzza, D., Fraundorfer, F.: Visual odometry [tutorial]. IEEE Robot. Autom. Mag. 18(4), 80\u201392 (2011)","journal-title":"IEEE Robot. Autom. Mag."},{"key":"14_CR39","unstructured":"Seo, M., Min, S., Farhadi, A., Hajishirzi, H.: Neural speed reading via skim-RNN. arXiv preprint arXiv:1711.02085 (2017)"},{"key":"14_CR40","doi-asserted-by":"crossref","unstructured":"Shamwell, E.J., Leung, S., Nothwang, W.D.: Vision-aided absolute trajectory estimation using an unsupervised deep network with online error correction. In: 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2524\u20132531. IEEE (2018)","DOI":"10.1109\/IROS.2018.8593573"},{"key":"14_CR41","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"14_CR42","doi-asserted-by":"crossref","unstructured":"Teerapittayanon, S., McDanel, B., Kung, H.T.: BranchyNet: fast inference via early exiting from deep neural networks. In: 2016 23rd International Conference on Pattern Recognition (ICPR), pp. 2464\u20132469. IEEE (2016)","DOI":"10.1109\/ICPR.2016.7900006"},{"key":"14_CR43","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-01246-5_1","volume-title":"Computer Vision \u2013 ECCV 2018","author":"A Veit","year":"2018","unstructured":"Veit, A., Belongie, S.: Convolutional networks with adaptive inference graphs. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11205, pp. 3\u201318. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01246-5_1"},{"key":"14_CR44","doi-asserted-by":"crossref","unstructured":"Wang, S., Clark, R., Wen, H., Trigoni, N.: DeepVO: towards end-to-end visual odometry with deep recurrent convolutional neural networks. In: 2017 IEEE International Conference On Robotics and Automation (ICRA), pp. 2043\u20132050. IEEE (2017)","DOI":"10.1109\/ICRA.2017.7989236"},{"key":"14_CR45","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"420","DOI":"10.1007\/978-3-030-01261-8_25","volume-title":"Computer Vision \u2013 ECCV 2018","author":"X Wang","year":"2018","unstructured":"Wang, X., Yu, F., Dou, Z.-Y., Darrell, T., Gonzalez, J.E.: SkipNet: learning dynamic routing in convolutional networks. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11217, pp. 420\u2013436. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01261-8_25"},{"key":"14_CR46","doi-asserted-by":"crossref","unstructured":"Wang, Y., Chen, Z., Jiang, H., Song, S., Han, Y., Huang, G.: Adaptive focus for efficient video recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 16249\u201316258 (2021)","DOI":"10.1109\/ICCV48922.2021.01594"},{"issue":"3","key":"14_CR47","doi-asserted-by":"publisher","first-page":"229","DOI":"10.1007\/BF00992696","volume":"8","author":"RJ Williams","year":"1992","unstructured":"Williams, R.J.: Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach. Learn. 8(3), 229\u2013256 (1992)","journal-title":"Mach. Learn."},{"key":"14_CR48","doi-asserted-by":"crossref","unstructured":"Wu, Z., et al.: BlockDrop: dynamic inference paths in residual networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8817\u20138826 (2018)","DOI":"10.1109\/CVPR.2018.00919"},{"key":"14_CR49","doi-asserted-by":"crossref","unstructured":"Wu, Z., Xiong, C., Ma, C.Y., Socher, R., Davis, L.S.: AdaFrame: adaptive frame selection for fast video recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1278\u20131287 (2019)","DOI":"10.1109\/CVPR.2019.00137"},{"key":"14_CR50","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1007\/978-3-030-20876-9_19","volume-title":"Computer Vision \u2013 ACCV 2018","author":"F Xue","year":"2019","unstructured":"Xue, F., Wang, Q., Wang, X., Dong, W., Wang, J., Zha, H.: Guided feature selection for deep visual odometry. In: Jawahar, C.V., Li, H., Mori, G., Schindler, K. (eds.) ACCV 2018. LNCS, vol. 11366, pp. 293\u2013308. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-20876-9_19"},{"key":"14_CR51","doi-asserted-by":"crossref","unstructured":"Xue, F., Wang, X., Li, S., Wang, Q., Wang, J., Zha, H.: Beyond tracking: selecting memory and refining poses for deep visual odometry. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8575\u20138583 (2019)","DOI":"10.1109\/CVPR.2019.00877"},{"key":"14_CR52","doi-asserted-by":"crossref","unstructured":"Yang, M., Kim, H.S.: Deep joint source-channel coding for wireless image transmission with adaptive rate control. arXiv preprint arXiv:2110.04456 (2021)","DOI":"10.1109\/ICC42927.2021.9500996"},{"issue":"4","key":"14_CR53","doi-asserted-by":"publisher","first-page":"2878","DOI":"10.1109\/LRA.2018.2846813","volume":"3","author":"N Yang","year":"2018","unstructured":"Yang, N., Wang, R., Gao, X., Cremers, D.: Challenges in monocular visual odometry: photometric calibration, motion bias, and rolling shutter effect. IEEE Robot. Autom. Lett. 3(4), 2878\u20132885 (2018)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"14_CR54","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"175","DOI":"10.1007\/978-3-030-58536-5_11","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Z Yuan","year":"2020","unstructured":"Yuan, Z., Wu, B., Sun, G., Liang, Z., Zhao, S., Bi, W.: S2DNAS: transforming static CNN model for dynamic inference via neural architecture search. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12347, pp. 175\u2013192. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_11"},{"key":"14_CR55","doi-asserted-by":"crossref","unstructured":"Zhou, T., Brown, M., Snavely, N., Lowe, D.G.: Unsupervised learning of depth and ego-motion from video. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1851\u20131858 (2017)","DOI":"10.1109\/CVPR.2017.700"},{"key":"14_CR56","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"710","DOI":"10.1007\/978-3-030-58568-6_42","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Zou","year":"2020","unstructured":"Zou, Y., Ji, P., Tran, Q.-H., Huang, J.-B., Chandraker, M.: Learning monocular visual odometry via self-supervised long-term modeling. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12359, pp. 710\u2013727. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58568-6_42"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19839-7_14","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,6]],"date-time":"2024-10-06T10:10:33Z","timestamp":1728209433000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19839-7_14"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198380","9783031198397"],"references-count":56,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19839-7_14","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"23 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}