{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T16:25:27Z","timestamp":1761582327950,"version":"3.37.3"},"reference-count":56,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T00:00:00Z","timestamp":1705881600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T00:00:00Z","timestamp":1705881600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2022YFB3706902"],"award-info":[{"award-number":["2022YFB3706902"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1007\/s00371-023-03243-9","type":"journal-article","created":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T16:02:05Z","timestamp":1705939325000},"page":"8377-8396","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["An efficient parallel fusion structure of distilled and transformer-enhanced modules for lightweight image super-resolution"],"prefix":"10.1007","volume":"40","author":[{"given":"Guanqiang","family":"Wang","sequence":"first","affiliation":[]},{"given":"Mingsong","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Yongcheng","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Xianhua","family":"Tan","sequence":"additional","affiliation":[]},{"given":"Chizhou","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Wenxin","family":"Yao","sequence":"additional","affiliation":[]},{"given":"Baihui","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Weidong","family":"Zeng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,22]]},"reference":[{"key":"3243_CR1","doi-asserted-by":"publisher","first-page":"105486","DOI":"10.1016\/j.engappai.2022.105486","volume":"116","author":"Z Ullah","year":"2022","unstructured":"Ullah, Z., Qi, L., Hasan, A., Asim, M.: Improved deep CNN-based two stream super resolution and hybrid deep model-based facial emotion recognition. Eng. Appl. Artif. Intell. 116, 105486 (2022)","journal-title":"Eng. Appl. Artif. Intell."},{"key":"3243_CR2","doi-asserted-by":"publisher","first-page":"116793","DOI":"10.1016\/j.eswa.2022.116793","volume":"197","author":"Y Wang","year":"2022","unstructured":"Wang, Y., Bashir, S.M.A., Khan, M., Ullah, Q., Wang, R., et al.: Remote sensing image super-resolution and object detection: benchmark and state of the art. Expert Syst. Appl. 197, 116793 (2022)","journal-title":"Expert Syst. Appl."},{"key":"3243_CR3","doi-asserted-by":"crossref","unstructured":"Niu, T., Chen, B., Lyu, Q., Li, B., Luo, W., Wang, Z., Li, B.: Scoring Bayesian Neural Networks for learning from inconsistent labels in surface defect segmentation. Measurement. 225, 113998 (2024)","DOI":"10.1016\/j.measurement.2023.113998"},{"key":"3243_CR4","doi-asserted-by":"publisher","first-page":"120159","DOI":"10.1016\/j.eswa.2023.120159","volume":"226","author":"H Yang","year":"2023","unstructured":"Yang, H., Yang, X., Liu, K., Jeon, G., Zhu, C.: SCN: self-calibration network for fast and accurate image super-resolution. Expert Syst. Appl. 226, 120159 (2023)","journal-title":"Expert Syst. Appl."},{"issue":"6","key":"3243_CR5","doi-asserted-by":"publisher","first-page":"1153","DOI":"10.1109\/TASSP.1981.1163711","volume":"29","author":"R Keys","year":"1981","unstructured":"Keys, R.: Cubic convolution interpolation for digital image processing. IEEE Trans. Acoust. Speech Signal Process. 29(6), 1153\u20131160 (1981)","journal-title":"IEEE Trans. Acoust. Speech Signal Process."},{"issue":"2","key":"3243_CR6","doi-asserted-by":"publisher","first-page":"295","DOI":"10.1109\/TPAMI.2015.2439281","volume":"38","author":"C Dong","year":"2015","unstructured":"Dong, C., Loy, C.C., He, K., Tang, X.: Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. 38(2), 295\u2013307 (2015)","journal-title":"IEEE Trans. Pattern Anal."},{"issue":"3","key":"3243_CR7","doi-asserted-by":"publisher","first-page":"1192","DOI":"10.1109\/TPAMI.2020.3021088","volume":"44","author":"S Anwar","year":"2020","unstructured":"Anwar, S., Barnes, N.: Densely residual laplacian super-resolution. IEEE T. Pattern Anal. 44(3), 1192\u20131204 (2020)","journal-title":"IEEE T. Pattern Anal."},{"key":"3243_CR8","doi-asserted-by":"crossref","unstructured":"Park, D., Kim, K., Young Chun, S.: Efficient module based single image super resolution for multiple problems. In: Proceedings of 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 995\u20139958. IEEE, Salt Lake City (2018)","DOI":"10.1109\/CVPRW.2018.00133"},{"key":"3243_CR9","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-023-03044-0","author":"Y Liu","year":"2023","unstructured":"Liu, Y., Yang, D., Zhang, F., Xie, Q., Zhang, C.: Deep recurrent residual channel attention network for single image super-resolution. Vis. Comput. (2023). https:\/\/doi.org\/10.1007\/s00371-023-03044-0","journal-title":"Vis. Comput."},{"key":"3243_CR10","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-023-02968-x","author":"J Wang","year":"2023","unstructured":"Wang, J., Zou, Y., Wu, H.: Image super-resolution method based on attention aggregation hierarchy feature. Vis. Comput. (2023). https:\/\/doi.org\/10.1007\/s00371-023-02968-x","journal-title":"Vis. Comput."},{"key":"3243_CR11","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-023-02795-0","author":"Y Chen","year":"2023","unstructured":"Chen, Y., Xia, R., Yang, K., Zou, K.: MFFN: image super-resolution via multi-level features fusion network. Vis. Comput. (2023). https:\/\/doi.org\/10.1007\/s00371-023-02795-0","journal-title":"Vis. Comput."},{"key":"3243_CR12","doi-asserted-by":"crossref","unstructured":"Agustsson, E., Timofte, R.: Ntire 2017 challenge on single image super-resolution: Dataset and study. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 126\u2013135. IEEE, Honolulu (2017)","DOI":"10.1109\/CVPRW.2017.150"},{"key":"3243_CR13","doi-asserted-by":"crossref","unstructured":"Ahn, N., Kang, B., Sohn, K.: Fast, accurate, and lightweight super-resolution with cascading residual network. In: Proceedings of the European Conference on Computer Vision, pp. 252\u2013268. IEEE, Munich (2018)","DOI":"10.1007\/978-3-030-01249-6_16"},{"key":"3243_CR14","first-page":"20343","volume":"33","author":"W Li","year":"2020","unstructured":"Li, W., Zhou, K., Qi, L., Jiang, N., Lu, J., Jia, J.L.: Linearly-assembled pixel-adaptive regression network for single image super-resolution and beyond. Adv. Neural Inf. Proc. Syst. 33, 20343\u201320355 (2020)","journal-title":"Adv. Neural Inf. Proc. Syst."},{"key":"3243_CR15","unstructured":"Gao, Q., Zhao, Y., Li, G., Tong, T.: Image super-resolution using knowledge distillation. In: Asian Conference on Computer Vision, pp. 1103\u20131112. Springer (2018)"},{"key":"3243_CR16","doi-asserted-by":"publisher","first-page":"106407","DOI":"10.1016\/j.engappai.2023.106407","volume":"123","author":"S Angarano","year":"2023","unstructured":"Angarano, S., Salvetti, F., Martini, M., Chiaberge, M.: Generative adversarial super-resolution at the edge with knowledge distillation. Eng. Appl. Artif. Intel. 123, 106407 (2023)","journal-title":"Eng. Appl. Artif. Intel."},{"key":"3243_CR17","doi-asserted-by":"crossref","unstructured":"Wang, H., Li, J., Wu, H., Hovy, E., Sun, Y.: Pre-trained language models and their applications. Engineering. 25, 51\u201365 (2023)","DOI":"10.1016\/j.eng.2022.04.024"},{"issue":"3","key":"3243_CR18","doi-asserted-by":"publisher","first-page":"275","DOI":"10.1016\/j.eng.2019.12.014","volume":"6","author":"M Zhou","year":"2020","unstructured":"Zhou, M., Duan, N., Liu, S., Shum, H.: Progress in neural NLP: modeling, learning, and reasoning. Engineering 6(3), 275\u2013290 (2020)","journal-title":"Engineering"},{"key":"3243_CR19","doi-asserted-by":"publisher","first-page":"102352","DOI":"10.1016\/j.displa.2022.102352","volume":"76","author":"Z Zhou","year":"2023","unstructured":"Zhou, Z., Li, G., Wang, G.: A hybrid of transformer and CNN for efficient single image super-resolution via multi-level distillation. Displays 76, 102352 (2023)","journal-title":"Displays"},{"key":"3243_CR20","doi-asserted-by":"crossref","unstructured":"Fang, J., Lin, H., Chen, X., Zeng, K.: A hybrid network of cnn and transformer for lightweight image super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition , pp. 1103\u20131112. IEEE, New Orleans (2022)","DOI":"10.1109\/CVPRW56347.2022.00119"},{"key":"3243_CR21","doi-asserted-by":"crossref","unstructured":"Lu, Z., Li, J., Liu, H., Huang, C., Zhang, L., Zeng, T.: Transformer for single image super-resolution. In: Proceedings of the 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 456\u2013465. IEEE, New Orleans (2022)","DOI":"10.1109\/CVPRW56347.2022.00061"},{"key":"3243_CR22","doi-asserted-by":"publisher","first-page":"143","DOI":"10.1016\/j.eng.2021.03.023","volume":"18","author":"H Wang","year":"2021","unstructured":"Wang, H., Wu, H., He, Z., Huang, L., Church, K.W.: Progress in machine translation. Engineering 18, 143\u2013153 (2021)","journal-title":"Engineering"},{"key":"3243_CR23","doi-asserted-by":"crossref","unstructured":"Hui, Z., Gao, X., Yang, Y., Wang, X.: Lightweight image super-resolution with information multi-distillation network. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 2024\u20132032. New York, NY, USA (2019)","DOI":"10.1145\/3343031.3351084"},{"key":"3243_CR24","doi-asserted-by":"crossref","unstructured":"Luo, X., Xie, Y., Zhang, Y., Qu, Y., Li, C., Fu, Y.: Latticenet: Towards lightweight image super-resolution with lattice block. In: Computer Vision\u2014ECCV 2020: 16th European Conference, pp. 272\u2013289. Glasgow, UK (2020)","DOI":"10.1007\/978-3-030-58542-6_17"},{"key":"3243_CR25","doi-asserted-by":"crossref","unstructured":"Liu, J., Tang, J., Wu, G.: Residual feature distillation network for lightweight image super-resolution. In: Computer Vision\u2014ECCV 2020 Workshops, pp. 41\u201355. Glasgow, UK (2020)","DOI":"10.1007\/978-3-030-67070-2_2"},{"key":"3243_CR26","doi-asserted-by":"publisher","first-page":"117594","DOI":"10.1016\/j.eswa.2022.117594","volume":"204","author":"X Yang","year":"2022","unstructured":"Yang, X., Guo, Y., Li, Z., Zhou, D., Li, T.: MRDN: a lightweight multi-stage residual distillation network for image super-resolution. Expert Syst. Appl. 204, 117594 (2022)","journal-title":"Expert Syst. Appl."},{"key":"3243_CR27","doi-asserted-by":"crossref","unstructured":"Zamir, S. W., Arora, A., Khan, S., Hayat, M., Khan, F. S., Yang, M.: Restormer: efficient transformer for high-resolution image restoration. In: Proceedings of 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5718\u20135729. IEEE, New Orleans (2022)","DOI":"10.1109\/CVPR52688.2022.00564"},{"key":"3243_CR28","doi-asserted-by":"crossref","unstructured":"Wang, Z., Cun, X., Bao, J., Zhou, W., Liu, J., Li, H.: Uformer: a general U-shaped transformer for image restoration. In: Proceedings of 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition. IEEE, New Orleans, LA, USA, pp. 17662\u201317672 (2022)","DOI":"10.1109\/CVPR52688.2022.01716"},{"key":"3243_CR29","doi-asserted-by":"crossref","unstructured":"Chen, H., Wang, Y., Guo, T., Xu, C., Deng, Y., Liu, Z., et al.: Pre-trained image processing transformer. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 2165\u20130608. IEEE Safranbolu (2021)","DOI":"10.1109\/CVPR46437.2021.01212"},{"key":"3243_CR30","doi-asserted-by":"crossref","unstructured":"Liang, J., Cao, J., Sun, G., Zhang, K., Van Gool, L., Timofte, R.: Swinir: Image restoration using swin transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1833\u20131844. IEEE, Montreal (2021)","DOI":"10.1109\/ICCVW54120.2021.00210"},{"key":"3243_CR31","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022. IEEE, Montreal (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"3243_CR32","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zeng, H., Guo, S., Zhang, L.: Efficient long-range attention network for image super-resolution. In: Proceedings of European Conference on Computer Vision, p. 13677. Springer, Cham (2022)","DOI":"10.1007\/978-3-031-19790-1_39"},{"key":"3243_CR33","doi-asserted-by":"crossref","unstructured":"Xu, B., Yin, H.: A slimmer and deeper approach to deep network structures for low-level vision tasks. Expert Syst. e13092, 1\u201316 (2022)","DOI":"10.1111\/exsy.13092"},{"key":"3243_CR34","doi-asserted-by":"publisher","first-page":"1569","DOI":"10.1007\/s00371-020-01903-8","volume":"37","author":"W Shi","year":"2021","unstructured":"Shi, W., Du, H., Mei, W., Ma, Z.: (SARN)spatial-wise attention residual network for image super-resolution. Vis. Comput. 37, 1569 (2021)","journal-title":"Vis. Comput."},{"key":"3243_CR35","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-023-03046-y","author":"A Yang","year":"2023","unstructured":"Yang, A., Wei, Z., Wang, J., Cao, J., Ji, Z., Pang, Y.: Multi-feature self-attention super-resolution network. Vis. Comput. (2023). https:\/\/doi.org\/10.1007\/s00371-023-03046-y","journal-title":"Vis. Comput."},{"key":"3243_CR36","doi-asserted-by":"crossref","unstructured":"Wang, G., Chen, M., Lin, Y. C., Tan, X., Zhang, C., Yao, W., Gao, B., Li, K., Li, Z., Zeng, W.: Efficient multi-branch dynamic fusion network for super-resolution of industrial component image. Displays. 82, 102633 (2024)","DOI":"10.1016\/j.displa.2023.102633"},{"key":"3243_CR37","unstructured":"Kim, J. H., Choi, J. H., Cheon, M., Lee, J. S. Ram.: residual attention module for single image super-resolution. arXiv preprint arXiv:1811.12043 (2018)"},{"key":"3243_CR38","doi-asserted-by":"crossref","unstructured":"Niu, B., Wen, W., Ren, W., Zhang, X., Yang, L., Wang, S., et al.: Single image super-resolution via a holistic attention network. In: Proceedings of European Conference on Computer Vision, pp. 191\u2013207. Springer, Cham (2020)","DOI":"10.1007\/978-3-030-58610-2_12"},{"issue":"3","key":"3243_CR39","doi-asserted-by":"publisher","first-page":"1443","DOI":"10.1109\/TCYB.2020.2970104","volume":"51","author":"R Lan","year":"2020","unstructured":"Lan, R., Sun, L., Liu, Z., Lu, H., Pang, C., Luo, X.: MADNet: a fast and lightweight network for single-image super resolution. IEEE Trans. Cybern. 51(3), 1443\u20131453 (2020)","journal-title":"IEEE Trans. Cybern."},{"key":"3243_CR40","doi-asserted-by":"crossref","unstructured":"Wang, L., Li, D., Zhu, Y., Tian, L., Shan, Y.: Dual super-resolution learning for semantic segmentation. In: Proceedings of 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3773\u20133782. IEEE, Seattle (2020)","DOI":"10.1109\/CVPR42600.2020.00383"},{"key":"3243_CR41","doi-asserted-by":"publisher","first-page":"114450","DOI":"10.1016\/j.eswa.2020.114450","volume":"169","author":"Z Huang","year":"2021","unstructured":"Huang, Z., Li, W., Li, J., Zhou, D.: Dual-path attention network for single image super-resolution. Expert Syst. Appl. 169, 114450 (2021)","journal-title":"Expert Syst. Appl."},{"issue":"7","key":"3243_CR42","doi-asserted-by":"publisher","first-page":"1397","DOI":"10.1007\/s11760-021-01870-0","volume":"15","author":"X Yang","year":"2021","unstructured":"Yang, X., Guo, Y., Li, Z., Zhou, D.: Image super-resolution network based on a multi-branch attention mechanism. Signal Image Video Process. 15(7), 1397\u20131405 (2021)","journal-title":"Signal Image Video Process."},{"issue":"8","key":"3243_CR43","doi-asserted-by":"publisher","first-page":"3647","DOI":"10.1007\/s00371-023-02938-3","volume":"39","author":"S Huang","year":"2023","unstructured":"Huang, S., Liu, X., Tan, T., et al.: TransMRSR: transformer-based self-distilled generative prior for brain MRI super-resolution. Visual Comput. 39(8), 3647\u20133659 (2023)","journal-title":"Visual Comput."},{"key":"3243_CR44","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2023","unstructured":"Lin, X., Sun, S., Huang, W., Sheng, B., Li, P., Feng, D.: EAPT: efficient attention pyramid transformer for image processing. IEEE Trans. Multimed. 25, 50\u201361 (2023)","journal-title":"IEEE Trans. Multimed."},{"key":"3243_CR45","doi-asserted-by":"crossref","unstructured":"Liu, J., Tang, J., Wu, G.: Residual feature distillation network for lightweight image super-resolution. Computer Vision\u2013ECCV 2020 Workshops, pp. 41\u201355. Glasgow, UK (2020)","DOI":"10.1007\/978-3-030-67070-2_2"},{"key":"3243_CR46","doi-asserted-by":"crossref","unstructured":"Fang, J., Lin, H., Chen, X., Zeng, K.: A hybrid network of cnn and transformer for lightweight image super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1103\u20131112. IEEE, New Orleans (2022)","DOI":"10.1109\/CVPRW56347.2022.00119"},{"key":"3243_CR47","doi-asserted-by":"crossref","unstructured":"Lin, T., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2117\u20132125. IEEE, Honolulu (2017)","DOI":"10.1109\/CVPR.2017.106"},{"key":"3243_CR48","doi-asserted-by":"crossref","unstructured":"Bevilacqua, M., Roumy, A., Guillemot, C., Alberi-morel, M. L.: Low-complexity single-image super-resolution based on nonnegative neighbor embedding. British Machine Vision Conference, Surrey, UK, pp. 1\u201310 (2012)","DOI":"10.5244\/C.26.135"},{"key":"3243_CR49","doi-asserted-by":"crossref","unstructured":"Zeyde, R., Elad, M., Protter, M.: On single image scale-up using sparse-representations. In: Proceedings of International Conference on Curves and Surfaces, pp. 711\u2013730. Springer, Berlin (2010)","DOI":"10.1007\/978-3-642-27413-8_47"},{"key":"3243_CR50","doi-asserted-by":"crossref","unstructured":"Martin, D., Fowlkes, C., Tal, D., Malik, J.: A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In: Proceedings of Eighth IEEE International Conference on Computer Vision, pp. 416\u2013423. IEEE, Vancouver (2001)","DOI":"10.1109\/ICCV.2001.937655"},{"key":"3243_CR51","doi-asserted-by":"crossref","unstructured":"Huang, J., Singh, A., Ahuja, N.: Single image super-resolution from transformed self-exemplars. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5197\u20135206. IEEE, Waknaghat (2015)","DOI":"10.1109\/CVPR.2015.7299156"},{"key":"3243_CR52","doi-asserted-by":"publisher","first-page":"21811","DOI":"10.1007\/s11042-016-4020-z","volume":"76","author":"Y Matsui","year":"2017","unstructured":"Matsui, Y., Ito, K., Aramaki, Y., Fujimoto, A., Ogawa, T., Yamasaki, T., et al.: Sketch-based manga retrieval using manga109 dataset. Multimed. Tools Appl. 76, 21811\u201321838 (2017)","journal-title":"Multimed. Tools Appl."},{"key":"3243_CR53","unstructured":"Kingma, D. P., Ba, J. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"3243_CR54","doi-asserted-by":"crossref","unstructured":"Fang, J., Lin, H., Chen, X., Zeng, K. A hybrid network of cnn and transformer for lightweight image super-resolution. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 1103\u20131112. IEEE, New Orleans (2022)","DOI":"10.1109\/CVPRW56347.2022.00119"},{"key":"3243_CR55","unstructured":"Paszke, A. Gross, S., Massa, F, Lerer, A., Bradbury, J., Chanan.G, et al.: Pytorch: An imperative style, high-performance deep learning library. In: Proceedings of the international conference on neural information processing systems, pp. 8024\u20138035. Curran Associates, Vancouver (2019)"},{"key":"3243_CR56","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7132\u20137141. IEEE, Salt Lake City (2018)","DOI":"10.1109\/CVPR.2018.00745"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03243-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-03243-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03243-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,12]],"date-time":"2024-11-12T09:04:23Z","timestamp":1731402263000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-03243-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,22]]},"references-count":56,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["3243"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-03243-9","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2024,1,22]]},"assertion":[{"value":"19 December 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 January 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}