{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T22:09:39Z","timestamp":1774476579898,"version":"3.50.1"},"publisher-location":"Cham","reference-count":70,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030585228","type":"print"},{"value":"9783030585235","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-58523-5_4","type":"book-chapter","created":{"date-parts":[[2020,12,3]],"date-time":"2020-12-03T20:13:16Z","timestamp":1607026396000},"page":"52-69","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":147,"title":["Accurate RGB-D Salient Object Detection via Collaborative Learning"],"prefix":"10.1007","author":[{"given":"Wei","family":"Ji","sequence":"first","affiliation":[]},{"given":"Jingjing","family":"Li","sequence":"additional","affiliation":[]},{"given":"Miao","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Yongri","family":"Piao","sequence":"additional","affiliation":[]},{"given":"Huchuan","family":"Lu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,12,4]]},"reference":[{"key":"4_CR1","doi-asserted-by":"crossref","unstructured":"Achanta, R., Hemami, S.S., Estrada, F.J., S\u00fcsstrunk, S.: Frequency-tuned salient region detection. In: CVPR, pp. 1597\u20131604 (2009)","DOI":"10.1109\/CVPR.2009.5206596"},{"issue":"12","key":"4_CR2","first-page":"5706","volume":"24","author":"A Borji","year":"2015","unstructured":"Borji, A., Cheng, M.M., Jiang, H., Li, J.: Salient object detection: a benchmark. TIP 24(12), 5706\u20135722 (2015)","journal-title":"TIP"},{"key":"4_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"414","DOI":"10.1007\/978-3-642-33709-3_30","volume-title":"Computer Vision \u2013 ECCV 2012","author":"A Borji","year":"2012","unstructured":"Borji, A., Sihite, D.N., Itti, L.: Salient object detection: a benchmark. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7573, pp. 414\u2013429. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33709-3_30"},{"issue":"6","key":"4_CR4","doi-asserted-by":"publisher","first-page":"679","DOI":"10.1109\/TPAMI.1986.4767851","volume":"8","author":"J Canny","year":"1986","unstructured":"Canny, J.: A computational approach to edge detection. TPAMI 8(6), 679\u2013698 (1986)","journal-title":"TPAMI"},{"key":"4_CR5","doi-asserted-by":"crossref","unstructured":"Chen, H., Li, Y.: Progressively complementarity-aware fusion network for RGB-D salient object detection. In: CVPR, pp. 3051\u20133060 (2018)","DOI":"10.1109\/CVPR.2018.00322"},{"issue":"6","key":"4_CR6","first-page":"2825","volume":"28","author":"H Chen","year":"2019","unstructured":"Chen, H., Li, Y.: Three-stream attention-aware network for RGB-D salient object detection. TIP 28(6), 2825\u20132835 (2019)","journal-title":"TIP"},{"key":"4_CR7","first-page":"376","volume":"86","author":"H Chen","year":"2019","unstructured":"Chen, H., Li, Y., Su, D.: Multi-modal fusion network with multi-scale multi-path and cross-modal interactions for RGB-D salient object detection. PR 86, 376\u2013385 (2019)","journal-title":"PR"},{"issue":"4","key":"4_CR8","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1109\/TPAMI.2017.2699184","volume":"40","author":"LC Chen","year":"2018","unstructured":"Chen, L.C., Papandreou, G., Kokkinos, I., Murphy, K., Yuille, A.L.: Deeplab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. TPAMI 40(4), 834\u2013848 (2018)","journal-title":"TPAMI"},{"key":"4_CR9","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"833","DOI":"10.1007\/978-3-030-01234-2_49","volume-title":"Computer Vision \u2013 ECCV 2018","author":"L-C Chen","year":"2018","unstructured":"Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 833\u2013851. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01234-2_49"},{"issue":"3","key":"4_CR10","first-page":"409","volume":"37","author":"MM Cheng","year":"2011","unstructured":"Cheng, M.M., Zhang, G.X., Mitra, N.J., Huang, X., Hu, S.M.: Global contrast based salient region detection. TPAMI 37(3), 409\u2013416 (2011)","journal-title":"TPAMI"},{"key":"4_CR11","doi-asserted-by":"crossref","unstructured":"Cheng, Y., Fu, H., Wei, X., Xiao, J., Cao, X.: Depth enhanced saliency detection method. In: ICIMCS, pp. 23\u201327 (2014)","DOI":"10.1145\/2632856.2632866"},{"issue":"6","key":"4_CR12","first-page":"819","volume":"23","author":"R Cong","year":"2016","unstructured":"Cong, R., Lei, J., Zhang, C., Huang, Q., Cao, X., Hou, C.: Saliency detection for stereoscopic images based on depth confidence analysis and multiple cues fusion. SPL 23(6), 819\u2013823 (2016)","journal-title":"SPL"},{"key":"4_CR13","doi-asserted-by":"crossref","unstructured":"Craye, C., Filliat, D., Goudou, J.F.: Environment exploration for object-based visual saliency learning. In: ICRA, pp. 2303\u20132309 (2016)","DOI":"10.1109\/ICRA.2016.7487379"},{"key":"4_CR14","unstructured":"Dai, J., Li, Y., He, K., Sun, J.: R-FCN: object detection via region-based fully convolutional networks. In: NIPS, pp. 379\u2013387 (2016)"},{"key":"4_CR15","doi-asserted-by":"crossref","unstructured":"Deng, Z., et al.: R$$^3$$net: recurrent residual refinement network for saliency detection. In: IJCAI, pp. 684\u2013690 (2018)","DOI":"10.24963\/ijcai.2018\/95"},{"key":"4_CR16","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"196","DOI":"10.1007\/978-3-030-01267-0_12","volume-title":"Computer Vision \u2013 ECCV 2018","author":"D-P Fan","year":"2018","unstructured":"Fan, D.-P., Cheng, M.-M., Liu, J.-J., Gao, S.-H., Hou, Q., Borji, A.: Salient objects in clutter: bringing salient object detection to the foreground. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11219, pp. 196\u2013212. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01267-0_12"},{"key":"4_CR17","doi-asserted-by":"crossref","unstructured":"Fan, D.P., Cheng, M.M., Liu, Y., Li, T., Borji, A.: Structure-measure: a new way to evaluate foreground maps. In: ICCV, pp. 4558\u20134567 (2017)","DOI":"10.1109\/ICCV.2017.487"},{"key":"4_CR18","doi-asserted-by":"crossref","unstructured":"Fan, D.P., Gong, C., Cao, Y., Ren, B., Cheng, M.M., Borji, A.: Enhanced-alignment measure for binary foreground map evaluation. In: IJCAI, pp. 698\u2013704 (2018)","DOI":"10.24963\/ijcai.2018\/97"},{"key":"4_CR19","unstructured":"Fan, D.P., et al.: Rethinking RGB-D salient object detection: models, datasets, and large-scale benchmarks. arXiv preprint arXiv:1907.06781 (2019)"},{"key":"4_CR20","doi-asserted-by":"crossref","unstructured":"Fan, D.P., Wang, W., Cheng, M.M., Shen, J.: Shifting more attention to video salient object detection. In: CVPR, pp. 8554\u20138564 (2019)","DOI":"10.1109\/CVPR.2019.00875"},{"key":"4_CR21","doi-asserted-by":"crossref","unstructured":"Feng, M., Lu, H., Ding, E.: Attentive feedback network for boundary-aware salient object detection. In: CVPR, pp. 1623\u20131632 (2019)","DOI":"10.1109\/CVPR.2019.00172"},{"key":"4_CR22","doi-asserted-by":"crossref","unstructured":"Girshick, R.: Fast R-CNN. In: ICCV, pp. 1440\u20131448 (2015)","DOI":"10.1109\/ICCV.2015.169"},{"issue":"11","key":"4_CR23","first-page":"3171","volume":"48","author":"J Han","year":"2018","unstructured":"Han, J., Chen, H., Liu, N., Yan, C., Li, X.: CNNs-based RGB-D saliency detection via cross-view transfer and multiview fusion. IEEE Trans. Syst. Man Cybern. 48(11), 3171\u20133183 (2018)","journal-title":"IEEE Trans. Syst. Man Cybern."},{"key":"4_CR24","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"4_CR25","unstructured":"Hong, S., You, T., Kwak, S., Han, B.: Online tracking by learning discriminative saliency map with convolutional neural network. In: ICML, pp. 597\u2013606 (2015)"},{"key":"4_CR26","doi-asserted-by":"crossref","unstructured":"Hou, Q., Cheng, M.M., Hu, X., Borji, A., Tu, Z., Torr, P.H.S.: Deeply supervised salient object detection with short connections. In: CVPR, pp. 815\u2013828 (2017)","DOI":"10.1109\/TPAMI.2018.2815688"},{"issue":"11","key":"4_CR27","doi-asserted-by":"publisher","first-page":"1254","DOI":"10.1109\/34.730558","volume":"20","author":"L Itti","year":"1998","unstructured":"Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. TPAMI 20(11), 1254\u20131259 (1998)","journal-title":"TPAMI"},{"key":"4_CR28","doi-asserted-by":"crossref","unstructured":"Ju, R., Ge, L., Geng, W., Ren, T., Wu, G.: Depth saliency based on anisotropic center-surround difference. In: ICIP, pp. 1115\u20131119 (2014)","DOI":"10.1109\/ICIP.2014.7025222"},{"key":"4_CR29","unstructured":"Kr\u00e4henb\u00fchl, P., Koltun, V.: Efficient inference in fully connected CRFs with gaussian edge potentials. In: NIPS, pp. 109\u2013117 (2011)"},{"key":"4_CR30","doi-asserted-by":"crossref","unstructured":"Lee, G., Tai, Y.W., Kim, J.: Deep saliency with encoded low level distance map and high level features. In: CVPR, pp. 660\u2013668 (2016)","DOI":"10.1109\/CVPR.2016.78"},{"key":"4_CR31","doi-asserted-by":"crossref","unstructured":"Li, G., Zhu, C.: A three-pathway psychobiological framework of salient object detection using stereoscopic technology. In: ICCVW, pp. 3008\u20133014 (2017)","DOI":"10.1109\/ICCVW.2017.355"},{"key":"4_CR32","doi-asserted-by":"crossref","unstructured":"Li, G., Yu, Y.: Visual saliency based on multiscale deep features. In: CVPR, pp. 5455\u20135463 (2015)","DOI":"10.1109\/CVPR.2015.7299184"},{"issue":"11","key":"4_CR33","first-page":"5012","volume":"25","author":"G Li","year":"2016","unstructured":"Li, G., Yu, Y.: Visual saliency detection based on multiscale deep CNN features. TIP 25(11), 5012\u20135024 (2016)","journal-title":"TIP"},{"issue":"8","key":"4_CR34","doi-asserted-by":"publisher","first-page":"1605","DOI":"10.1109\/TPAMI.2016.2610425","volume":"39","author":"N Li","year":"2017","unstructured":"Li, N., Ye, J., Ji, Y., Ling, H., Yu, J.: Saliency detection on light field. TPAMI 39(8), 1605\u20131616 (2017)","journal-title":"TPAMI"},{"issue":"8","key":"4_CR35","first-page":"3919","volume":"25","author":"X Li","year":"2016","unstructured":"Li, X., et al.: Deepsaliency: multi-task deep neural network model for salient object detection. TIP 25(8), 3919\u20133930 (2016)","journal-title":"TIP"},{"key":"4_CR36","doi-asserted-by":"crossref","unstructured":"Li, Y., Hou, X., Koch, C., Rehg, J.M., Yuille, A.L.: The secrets of salient object segmentation. In: CVPR, pp. 280\u2013287 (2014)","DOI":"10.1109\/CVPR.2014.43"},{"key":"4_CR37","doi-asserted-by":"crossref","unstructured":"Lin, G., Milan, A., Shen, C., Reid, I.D.: RefineNet: multi-path refinement networks for high-resolution semantic segmentation. In: CVPR, pp. 5168\u20135177 (2017)","DOI":"10.1109\/CVPR.2017.549"},{"key":"4_CR38","doi-asserted-by":"crossref","unstructured":"Liu, J.J., Hou, Q., Cheng, M.M., Feng, J., Jiang, J.: A simple pooling-based design for real-time salient object detection. In: CVPR, pp. 3917\u20133926 (2019)","DOI":"10.1109\/CVPR.2019.00404"},{"key":"4_CR39","doi-asserted-by":"crossref","unstructured":"Liu, N., Han, J., Yang, M.H.: PicaNet: learning pixel-wise contextual attention for saliency detection. In: CVPR, pp. 3089\u20133098 (2018)","DOI":"10.1109\/CVPR.2018.00326"},{"key":"4_CR40","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: CVPR, pp. 3431\u20133440 (2015)","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"4_CR41","doi-asserted-by":"crossref","unstructured":"Luo, Z., Mishra, A.K., Achkar, A., Eichel, J.A., Li, S., Jodoin, P.M.: Non-local deep features for salient object detection. In: CVPR, pp. 6593\u20136601 (2017)","DOI":"10.1109\/CVPR.2017.698"},{"key":"4_CR42","doi-asserted-by":"crossref","unstructured":"Margolin, R., Zelnik-Manor, L., Tal, A.: How to evaluate foreground maps. In: CVPR, pp. 248\u2013255 (2014)","DOI":"10.1109\/CVPR.2014.39"},{"key":"4_CR43","doi-asserted-by":"crossref","unstructured":"Niu, Y., Geng, Y., Li, X., Liu, F.: Leveraging stereopsis for saliency analysis. In: CVPR, pp. 454\u2013461 (2012)","DOI":"10.1109\/CVPR.2012.6247708"},{"key":"4_CR44","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"92","DOI":"10.1007\/978-3-319-10578-9_7","volume-title":"Computer Vision \u2013 ECCV 2014","author":"H Peng","year":"2014","unstructured":"Peng, H., Li, B., Xiong, W., Hu, W., Ji, R.: RGBD salient object detection: a benchmark and algorithms. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8691, pp. 92\u2013109. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10578-9_7"},{"key":"4_CR45","doi-asserted-by":"crossref","unstructured":"Perazzi, F., Kr\u00e4henb\u00fchl, P., Pritch, Y., Hornung, A.: Saliency filters: contrast based filtering for salient region detection. In: CVPR, pp. 733\u2013740 (2012)","DOI":"10.1109\/CVPR.2012.6247743"},{"key":"4_CR46","doi-asserted-by":"crossref","unstructured":"Piao, Y., Ji, W., Li, J., Zhang, M., Lu, H.: Depth-induced multi-scale recurrent attention network for saliency detection. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00735"},{"key":"4_CR47","doi-asserted-by":"crossref","unstructured":"Qin, X., Zhang, Z., Huang, C., Gao, C., Dehghan, M., Jagersand, M.: BasNet: boundary-aware salient object detection. In: CVPR, pp. 7479\u20137489 (2019)","DOI":"10.1109\/CVPR.2019.00766"},{"issue":"5","key":"4_CR48","first-page":"2274","volume":"26","author":"L Qu","year":"2017","unstructured":"Qu, L., He, S., Zhang, J., Tian, J., Tang, Y., Yang, Q.: RGBD salient object detection via deep fusion. TIP 26(5), 2274\u20132285 (2017)","journal-title":"TIP"},{"key":"4_CR49","doi-asserted-by":"crossref","unstructured":"Ren, J., Gong, X., Yu, L., Zhou, W., Yang, M.Y.: Exploiting global priors for RGB-D saliency detection. In: CVPRW, pp. 25\u201332 (2015)","DOI":"10.1109\/CVPRW.2015.7301391"},{"key":"4_CR50","unstructured":"Ren, S., He, K., Girshick, R.B., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. In: NIPS 2015, pp. 91\u201399 (2015)"},{"key":"4_CR51","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: ICLR (2015)"},{"issue":"7","key":"4_CR52","doi-asserted-by":"publisher","first-page":"1442","DOI":"10.1109\/TPAMI.2013.230","volume":"36","author":"AWM Smeulders","year":"2014","unstructured":"Smeulders, A.W.M., Chu, D.M., Cucchiara, R., Calderara, S., Dehghan, A., Shah, M.: Visual tracking: an experimental survey. TPAMI 36(7), 1442\u20131468 (2014)","journal-title":"TPAMI"},{"key":"4_CR53","doi-asserted-by":"crossref","unstructured":"Wang, L., Lu, H., Ruan, X., Yang, M.H.: Deep networks for saliency detection via local estimation and global search. In: CVPR, pp. 3183\u20133192 (2015)","DOI":"10.1109\/CVPR.2015.7298938"},{"key":"4_CR54","unstructured":"Wang, P., Shen, X., Lin, Z., Cohen, S., Price, B., Yuille, A.: Towards unified depth and semantic prediction from a single image. In: CVPR, pp. 2800\u20132809 (2015)"},{"key":"4_CR55","unstructured":"Wang, W., Lai, Q., Fu, H., Shen, J., Ling, H.: Salient object detection in the deep learning era: an in-depth survey. arXiv preprint arXiv:1904.09146 (2019)"},{"issue":"5","key":"4_CR56","first-page":"2368","volume":"27","author":"W Wang","year":"2018","unstructured":"Wang, W., Shen, J.: Deep visual attention prediction. TIP 27(5), 2368\u20132378 (2018)","journal-title":"TIP"},{"key":"4_CR57","doi-asserted-by":"crossref","unstructured":"Wang, W., Shen, J., Dong, X., Borji, A.: Salient object detection driven by fixation prediction. In: CVPR, pp. 1711\u20131720 (2018)","DOI":"10.1109\/CVPR.2018.00184"},{"key":"4_CR58","unstructured":"Wang, W., Shen, J., Xie, J., Cheng, M.M., Ling, H., Borji, A.: Revisiting video saliency prediction in the deep learning era. TPAMI 1 (2019)"},{"key":"4_CR59","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-01234-2_1","volume-title":"Computer Vision \u2013 ECCV 2018","author":"S Woo","year":"2018","unstructured":"Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3\u201319. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01234-2_1"},{"key":"4_CR60","doi-asserted-by":"crossref","unstructured":"Wu, Z., Su, L., Huang, Q.: Cascaded partial decoder for fast and accurate salient object detection. In: CVPR, pp. 3907\u20133916 (2019)","DOI":"10.1109\/CVPR.2019.00403"},{"key":"4_CR61","doi-asserted-by":"crossref","unstructured":"Yan, Q., Xu, L., Shi, J., Jia, J.: Hierarchical saliency detection. In: CVPR, pp. 1155\u20131162 (2013)","DOI":"10.1109\/CVPR.2013.153"},{"key":"4_CR62","unstructured":"Yu, F., Koltun, V.: Multi-scale context aggregation by dilated convolutions. In: ICLR (2016)"},{"key":"4_CR63","doi-asserted-by":"crossref","unstructured":"Zhang, P., Wang, D., Lu, H., Wang, H., Ruan, X.: Amulet: aggregating multi-level convolutional features for salient object detection. In: ICCV, pp. 202\u2013211 (2017)","DOI":"10.1109\/ICCV.2017.31"},{"key":"4_CR64","doi-asserted-by":"crossref","unstructured":"Zhang, X., Wang, T., Qi, J., Lu, H., Wang, G.: Progressive attention guided recurrent network for salient object detection. In: CVPR, pp. 714\u2013722 (2018)","DOI":"10.1109\/CVPR.2018.00081"},{"key":"4_CR65","doi-asserted-by":"crossref","unstructured":"Zhao, J., Cao, Y., Fan, D., Cheng, M., LI, X., Zhang, L.: Contrast prior and fluid pyramid integration for RGBD salient object detection. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00405"},{"key":"4_CR66","doi-asserted-by":"crossref","unstructured":"Zhao, J., Liu, J., Fan, D.P., Cao, Y., Yang, J., Cheng, M.M.: EGNet: edge guidance network for salient object detection. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00887"},{"key":"4_CR67","doi-asserted-by":"crossref","unstructured":"Zhao, R., Ouyang, W., Li, H., Wang, X.: Saliency detection by multi-context deep learning. In: CVPR, pp. 1265\u20131274 (2015)","DOI":"10.1109\/CVPR.2015.7298731"},{"key":"4_CR68","doi-asserted-by":"crossref","unstructured":"Zhu, C., Cai, X., Huang, K., Li, T.H., Li, G.: PDNet: prior-model guided depth-enhanced network for salient object detection. In: ICME, pp. 199\u2013204 (2019)","DOI":"10.1109\/ICME.2019.00042"},{"key":"4_CR69","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1007\/978-3-319-64698-5_2","volume-title":"Computer Analysis of Images and Patterns","author":"C Zhu","year":"2017","unstructured":"Zhu, C., Li, G., Guo, X., Wang, W., Wang, R.: A multilayer backpropagation saliency detection algorithm based on depth mining. In: Felsberg, M., Heyden, A., Kr\u00fcger, N. (eds.) CAIP 2017. LNCS, vol. 10425, pp. 14\u201323. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-64698-5_2"},{"key":"4_CR70","doi-asserted-by":"crossref","unstructured":"Zhu, C., Li, G., Wang, W., Wang, R.: An innovative salient object detection using center-dark channel prior. In: ICCVW, pp. 1509\u20131515 (2017)","DOI":"10.1109\/ICCVW.2017.178"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2020"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-58523-5_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,3]],"date-time":"2024-12-03T00:03:11Z","timestamp":1733184191000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-58523-5_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030585228","9783030585235"],"references-count":70,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-58523-5_4","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"4 December 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Glasgow","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 August 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2020.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"OpenReview","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5025","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1360","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"27% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually due to the COVID-19 pandemic. From the ECCV Workshops 249 full papers, 18 short papers, and 21 further contributions were published out of a total of 467 submissions.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}