{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T16:55:56Z","timestamp":1767372956337,"version":"3.40.3"},"publisher-location":"Cham","reference-count":50,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031264306"},{"type":"electronic","value":"9783031264313"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-26431-3_15","type":"book-chapter","created":{"date-parts":[[2023,4,27]],"date-time":"2023-04-27T18:13:33Z","timestamp":1682619213000},"page":"186-196","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Depth Estimation of Traffic Scenes from Image Sequence Using Deep Learning"],"prefix":"10.1007","author":[{"given":"Xiaoxu","family":"Liu","sequence":"first","affiliation":[]},{"given":"Wei Qi","family":"Yan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,28]]},"reference":[{"key":"15_CR1","doi-asserted-by":"crossref","unstructured":"Li,Y., Tong, G., Yang, J., Zhang, L. Peng, H.: 3D point cloud scene data ac-quisition and its key technologies for scene understanding. Laser Optoelectron. Prog., 040002 (2019)","DOI":"10.3788\/LOP56.040002"},{"issue":"2","key":"15_CR2","doi-asserted-by":"publisher","first-page":"261","DOI":"10.1007\/s11263-019-01247-4","volume":"128","author":"L Liu","year":"2019","unstructured":"Liu, L., et al.: Deep learning for generic object detection: a survey. Int. J. Comput. Vis. 128(2), 261\u2013318 (2019). https:\/\/doi.org\/10.1007\/s11263-019-01247-4","journal-title":"Int. J. Comput. Vis."},{"key":"15_CR3","doi-asserted-by":"publisher","first-page":"1241","DOI":"10.1016\/j.drudis.2018.01.039","volume":"23","author":"H Chen","year":"2019","unstructured":"Chen, H., Engkvist, O., Wang, Y., Olivecrona, M., Blaschke, T.: The rise of deep learning in drug discovery. Drug Discovery Today 23, 1241\u20131250 (2019)","journal-title":"Drug Discovery Today"},{"key":"15_CR4","doi-asserted-by":"crossref","unstructured":"Husain, F., Dellen, B., Torras, C.: Scene understanding using deep learning, pp. 373\u2013382. Academic Press, Cambridge (2017)","DOI":"10.1016\/B978-0-12-811318-9.00020-X"},{"key":"15_CR5","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1109\/TSMC.2018.2868372","volume":"49","author":"S Yang","year":"2019","unstructured":"Yang, S., Wang, W., Liu, C., Deng, W.: Scene understanding in deep learning-based end-to-end controllers for autonomous vehicles. IEEE Trans. Syst. Man Cybernet. Syst. 49, 53\u201363 (2019)","journal-title":"IEEE Trans. Syst. Man Cybernet. Syst."},{"key":"15_CR6","unstructured":"Lecun, Y., Muller, U., Ben, J., Cosatto, E., Flepp, B.: Off-road obstacle avoidance through end-to-end learning. In: International Conference on Neural Information Processing Systems, pp. 739\u2013746 (2005)"},{"issue":"1","key":"15_CR7","doi-asserted-by":"publisher","first-page":"9425","DOI":"10.1038\/s41598-017-09891-x","volume":"7","author":"H Ohsugi","year":"2017","unstructured":"Ohsugi, H., Tabuchi, H., Enno, H., Ishitobi, N.: Accuracy of deep learning, a machine-learning technology, using ultra-wide-field fundus ophthalmoscopy for detecting hematogenous retinal detachment. Sci. Rep. 7(1), 9425 (2017)","journal-title":"Sci. Rep."},{"issue":"8","key":"15_CR8","first-page":"1037","volume":"9","author":"F Li","year":"2009","unstructured":"Li, F., Deng, J., Li, K.: ImageNet: constructing a largescale image database. J. Vis. 9(8), 1037\u20131038 (2009)","journal-title":"J. Vis."},{"key":"15_CR9","doi-asserted-by":"crossref","unstructured":"Laina, I., Rupprecht, C., Belagiannis, V., Tombari, F., Navab, N.: Deeper depth prediction with fully convolutional residual networks. In: International Conference on 3D Vision (3DV) (2016)","DOI":"10.1109\/3DV.2016.32"},{"key":"15_CR10","doi-asserted-by":"crossref","unstructured":"Eigen, D., Fergus, R.: Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture. In: IEEE International Conference on Computer Vision, pp. 2650\u20132658 (2014)","DOI":"10.1109\/ICCV.2015.304"},{"key":"15_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-46484-8_45","volume-title":"Computer Vision \u2013 ECCV 2016","author":"R Garg","year":"2016","unstructured":"Garg, R., B.G., V.K., Carneiro, G., Reid, I.: Unsupervised CNN for single view depth estimation: geometry to the rescue. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9912, pp. 740\u2013756. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46484-8_45"},{"key":"15_CR12","doi-asserted-by":"crossref","unstructured":"Godard, C., Aodha, O., Gabriel, J.: Unsupervised monocular depth estimation with left-right consistency. In: IEEE CVPR, pp. 270\u2013279 (2017)","DOI":"10.1109\/CVPR.2017.699"},{"key":"15_CR13","first-page":"1","volume":"01","author":"R Ranftl","year":"2020","unstructured":"Ranftl, R., Lasinger, K., Hafner, D., Schindler, K., Koltun, V.: Towards robust monocular depth estimation: Mixing datasets for zero-shot cross-dataset transfer. IEEE Trans. Pattern Anal. Mach. Intell. 01, 1 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"15_CR14","doi-asserted-by":"crossref","unstructured":"Miangoleh, S.M., Dille, S., Mai, L., Paris, S., Aksoy, Y.: Boosting monocular depth estimation models to high-resolution via content-adaptive multi-resolution merging. IEEE CVPR, pp. 9685\u20139694 (2021)","DOI":"10.1109\/CVPR46437.2021.00956"},{"issue":"9","key":"15_CR15","doi-asserted-by":"publisher","first-page":"1612","DOI":"10.1007\/s11431-020-1582-8","volume":"63","author":"C Zhao","year":"2020","unstructured":"Zhao, C., Sun, Q., Zhang, C., Tang, Y., Qian, F.: Monocular depth estimation based on deep learning: an overview. Sci. China Technol. Sci. 63(9), 1612\u20131627 (2020). https:\/\/doi.org\/10.1007\/s11431-020-1582-8","journal-title":"Sci. China Technol. Sci."},{"key":"15_CR16","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"288","DOI":"10.1007\/978-3-030-33676-9_20","volume-title":"Pattern Recognition","author":"M Ochs","year":"2019","unstructured":"Ochs, M., Kretz, A., Mester, R.: SDNet: semantically guided depth estimation network. In: Fink, G.A., Frintrop, S., Jiang, X. (eds.) DAGM GCPR 2019. LNCS, vol. 11824, pp. 288\u2013302. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-33676-9_20"},{"key":"15_CR17","doi-asserted-by":"publisher","first-page":"165","DOI":"10.1016\/S0963-8695(01)00041-X","volume":"35","author":"A Darabi","year":"2012","unstructured":"Darabi, A., Maldague, X.: Neural network based defect detection and depth es-timation in TNDE. NDT E Int. 35, 165\u2013175 (2012)","journal-title":"NDT E Int."},{"key":"15_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"298","DOI":"10.1007\/978-3-030-20893-6_19","volume-title":"Computer Vision \u2013 ACCV 2018","author":"P Zama Ramirez","year":"2019","unstructured":"Zama Ramirez, P., Poggi, M., Tosi, F., Mattoccia, S., Di Stefano, L.: Geometry meets semantics for semi-supervised monocular depth estimation. In: Jawahar, C.V., Li, Hongdong, Mori, Greg, Schindler, Konrad (eds.) ACCV 2018. LNCS, vol. 11363, pp. 298\u2013313. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-20893-6_19"},{"key":"15_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1007\/978-3-030-34869-4_23","volume-title":"Pattern Recognition and Machine Intelligence","author":"VK Repala","year":"2019","unstructured":"Repala, V.K., Dubey, S.R.: Dual CNN models for unsupervised monocular depth estimation. In: Deka, Bhabesh, Maji, Pradipta, Mitra, Sushmita, Bhattacharyya, Dhruba Kumar, Bora, Prabin Kumar, Pal, Sankar Kumar (eds.) PReMI 2019. LNCS, vol. 11941, pp. 209\u2013217. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-34869-4_23"},{"key":"15_CR20","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1007\/978-3-319-54187-7_2","volume-title":"Computer Vision \u2013 ACCV 2016","author":"K Honauer","year":"2017","unstructured":"Honauer, K., Johannsen, O., Kondermann, D., Goldluecke, B.: A dataset and evaluation methodology for depth estimation on 4D\u00a0light fields. In: Lai, S.H., Lepetit, V., Nishino, K., Sato, Y. (eds.) ACCV 2016. LNCS, vol. 10113, pp. 19\u201334. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-54187-7_2"},{"key":"15_CR21","doi-asserted-by":"crossref","unstructured":"Liu, F., Shen, C., Lin, G.: Deep convolutional neural fields for depth estimation from a single image. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 5162\u20135170 (2015)","DOI":"10.1109\/CVPR.2015.7299152"},{"key":"15_CR22","unstructured":"Dan, X. et al. Multiscale continuous CRFs as sequential deep networks for monocular depth estimation. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 5354\u20135362 (2017)"},{"key":"15_CR23","doi-asserted-by":"publisher","first-page":"255","DOI":"10.1016\/j.isprsjprs.2020.06.004","volume":"166","author":"J Liu","year":"2020","unstructured":"Liu, J., Li, Q., Cao, R., et al.: MiniNet: an extremely lightweight convolutional neural network for real-time unsupervised monocular depth estimation. ISPRS J. Photogrammetry Remote Sens. 166, 255\u2013267 (2020)","journal-title":"ISPRS J. Photogrammetry Remote Sens."},{"key":"15_CR24","doi-asserted-by":"crossref","unstructured":"Hu, J., Zhang, Y.Z., Takayuki, O.: Visualization of convolutional neural networks for monocular depth estimation. In: International Conference on Computer Vision, pp. 3869\u20133878 (2019)","DOI":"10.1109\/ICCV.2019.00397"},{"key":"15_CR25","doi-asserted-by":"crossref","unstructured":"Ding, X., Wang, Y., Zhang, J., et al.: Underwater image dehaze using scene depth estimation with adaptive color correction. In: OCEANS, pp.1\u20135 (2017)","DOI":"10.1109\/OCEANSE.2017.8084665"},{"key":"15_CR26","doi-asserted-by":"publisher","first-page":"1226","DOI":"10.1109\/TPAMI.2002.1033214","volume":"24","author":"A Torralba","year":"2002","unstructured":"Torralba, A., Aude, O.: Depth estimation from image structure. IEEE Trans. Pattern Anal. Mach. Intell. 24, 1226\u20131238 (2002)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"15_CR27","doi-asserted-by":"crossref","unstructured":"Song, W., et al.: A rapid scene depth estimation model based on underwater light attenuation prior for underwater image restoration. In: Pacific Rim Conference on Multimedia, pp.1\u20139 (2018)","DOI":"10.1007\/978-3-030-00776-8_62"},{"key":"15_CR28","doi-asserted-by":"publisher","first-page":"1521","DOI":"10.1109\/TPAMI.2004.102","volume":"26","author":"A Rajagopalan","year":"2014","unstructured":"Rajagopalan, A., Chaudhuri, S., Mudenagudi, U.: Depth estimation and image restoration using defocused stereo pairs. IEEE Trans. Pattern Anal. Mach. Intell. 26, 1521\u20131525 (2014)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"15_CR29","doi-asserted-by":"crossref","unstructured":"Chen, P., et al.: Towards scene understanding: unsupervised monocular depth estimation with semantic-aware representation. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2624\u20132632 (2019)","DOI":"10.1109\/CVPR.2019.00273"},{"key":"15_CR30","doi-asserted-by":"crossref","unstructured":"Watson, J., Firman, M., Brostow, G.J., Turmukhambetov, D.: Self-supervised monocular depth hints. In: IEEE International Conference on Computer Vision, pp. 2162\u20132171 (2019)","DOI":"10.1109\/ICCV.2019.00225"},{"key":"15_CR31","doi-asserted-by":"crossref","unstructured":"Godard, C., Aodha, O.M., Brostow, G.J.: Unsupervised monocular depth estimation with left-right consistency. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 270\u2013279 (2017)","DOI":"10.1109\/CVPR.2017.699"},{"issue":"2","key":"15_CR32","doi-asserted-by":"publisher","first-page":"328","DOI":"10.1109\/TPAMI.2007.1166","volume":"30","author":"H Hirschmuller","year":"2008","unstructured":"Hirschmuller, H.: Stereo processing by semiglobal matching and mutual information. IEEE Trans. Pattern Anal. Mach. Intell. 30(2), 328\u2013341 (2008)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"15_CR33","doi-asserted-by":"crossref","unstructured":"Godard, C., Aodha, O.M., Firman, M., Brostow, G.J.: Digging into self-supervised monocular depth estimation. In: IEEE International Conference on Computer Vision, pp. 3828\u20133838 (2019)","DOI":"10.1109\/ICCV.2019.00393"},{"issue":"10","key":"15_CR34","doi-asserted-by":"publisher","first-page":"15161","DOI":"10.1007\/s11042-020-10455-x","volume":"80","author":"X Liu","year":"2021","unstructured":"Liu, X., Yan, W.Q.: Traffic-light sign recognition using capsule network. Multimed. Tools Appl. 80(10), 15161\u201315171 (2021). https:\/\/doi.org\/10.1007\/s11042-020-10455-x","journal-title":"Multimed. Tools Appl."},{"key":"15_CR35","doi-asserted-by":"crossref","unstructured":"Liu, X., Yan, W.: Vehicle-related scene segmentation using CapsNets. In: IEEE IVCNZ (2020)","DOI":"10.1109\/IVCNZ51579.2020.9290664"},{"key":"15_CR36","series-title":"Communications in Computer and Information Science","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1007\/978-981-15-3651-9_7","volume-title":"Pattern Recognition","author":"X Liu","year":"2020","unstructured":"Liu, X., Neuyen, M., Yan, W.Q.: Vehicle-related scene understanding using deep learning. In: Cree, Michael, Huang, Fay, Yuan, Junsong, Yan, Wei Qi (eds.) ACPR 2019. CCIS, vol. 1180, pp. 61\u201373. Springer, Singapore (2020). https:\/\/doi.org\/10.1007\/978-981-15-3651-9_7"},{"key":"15_CR37","doi-asserted-by":"crossref","unstructured":"Liu, X.: Vehicle-related Scene Understanding Using Deep Learning. Master\u2019s Thesis, Auckland University of Technology, New Zealand (2019)","DOI":"10.1007\/978-981-15-3651-9_7"},{"key":"15_CR38","doi-asserted-by":"crossref","unstructured":"Mehtab, S., Yan, W.: FlexiNet: fast and accurate vehicle detection for autonomous vehicles-2D vehicle detection using deep neural network. In: ACM ICCCV (2021)","DOI":"10.1145\/3484274.3484282"},{"key":"15_CR39","doi-asserted-by":"publisher","first-page":"7169","DOI":"10.1007\/s11042-022-11933-0","volume":"81","author":"S Mehtab","year":"2021","unstructured":"Mehtab, S., Yan, W.: Flexible neural network for fast and accurate road scene perception. Multimed. Tools Appl. 81, 7169\u20137181 (2021). https:\/\/doi.org\/10.1007\/s11042-022-11933-0","journal-title":"Multimed. Tools Appl."},{"key":"15_CR40","doi-asserted-by":"crossref","unstructured":"Mehtab, S., Yan, W., Narayanan, A.: 3D vehicle detection using cheap LiDAR and camera sensors. In: IEEE IVCNZ (2021)","DOI":"10.1109\/IVCNZ54163.2021.9653358"},{"key":"15_CR41","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-61081-4","volume-title":"Computational Methods for Deep Learning: Theoretic Practice and Applications","author":"W Yan","year":"2021","unstructured":"Yan, W.: Computational Methods for Deep Learning: Theoretic Practice and Applications. Springer, Berlin (2021)"},{"key":"15_CR42","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-10713-0","volume-title":"Introduction to Intelligent Surveillance: Surveillance Data Capture, Transmission, and Analytics","author":"W Yan","year":"2019","unstructured":"Yan, W.: Introduction to Intelligent Surveillance: Surveillance Data Capture, Transmission, and Analytics. Springer, Berlin (2019)"},{"issue":"6","key":"15_CR43","doi-asserted-by":"publisher","first-page":"06310210","DOI":"10.1117\/1.OE.56.6.063102","volume":"56","author":"Q Gu","year":"2017","unstructured":"Gu, Q., Yang, J., Kong, L., Yan, W., Klette, R.: Embedded and real-time vehicle detection system for challenging on-road scenes. Opt. Eng. 56(6), 06310210 (2017)","journal-title":"Opt. Eng."},{"key":"15_CR44","first-page":"97","volume":"14","author":"Y Ming","year":"2021","unstructured":"Ming, Y., Li, Y., Zhang, Z., Yan, W.: A survey of path planning algorithms for autonomous vehicles. Int. J. Commercial Veh. 14, 97\u2013109 (2021)","journal-title":"Int. J. Commercial Veh."},{"key":"15_CR45","doi-asserted-by":"crossref","unstructured":"Shen, D., Xin, C., Nguyen, M., Yan, W.: Flame detection using deep learning. In: International Conference on Control, Automation and Robotics (2018)","DOI":"10.1109\/ICCAR.2018.8384711"},{"key":"15_CR46","doi-asserted-by":"crossref","unstructured":"Xin, C., Nguyen, M., Yan, W.: Multiple flames recognition using deep learning. In: Handbook of Research on Multimedia Cyber Security, pp. 296\u2013307 (2020)","DOI":"10.4018\/978-1-7998-2701-6.ch015"},{"key":"15_CR47","doi-asserted-by":"crossref","unstructured":"Luo, Z., Nguyen, M., Yan, W.: Kayak and sailboat detection based on the im-proved YOLO with transformer. In: ACM ICCCV (2022)","DOI":"10.1145\/3561613.3561619"},{"key":"15_CR48","doi-asserted-by":"crossref","unstructured":"Le, R., Nguyen, M., Yan, W.: Training a convolutional neural network for transportation sign detection using synthetic dataset. In: IEEE IVCNZ (2021)","DOI":"10.1109\/IVCNZ54163.2021.9653398"},{"issue":"27-28","key":"15_CR49","doi-asserted-by":"publisher","first-page":"19925","DOI":"10.1007\/s11042-020-08866-x","volume":"79","author":"C Pan","year":"2020","unstructured":"Pan, C., Yan, W.Q.: Object detection based on saturation of visual perception. Multimed. Tools Appl. 79(27\u201328), 19925\u201319944 (2020). https:\/\/doi.org\/10.1007\/s11042-020-08866-x","journal-title":"Multimed. Tools Appl."},{"key":"15_CR50","doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? The KITTI vision benchmark suite. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3354\u20133361 (2012)","DOI":"10.1109\/CVPR.2012.6248074"}],"container-title":["Lecture Notes in Computer Science","Image and Video Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-26431-3_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,27]],"date-time":"2023-04-27T18:15:55Z","timestamp":1682619355000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-26431-3_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031264306","9783031264313"],"references-count":50,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-26431-3_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"28 April 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PSIVT","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Pacific-Rim Symposium on Image and Video Technology","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 November 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14 November 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"psivt2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.cis-ram.org\/psivt2022\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}