{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T09:16:18Z","timestamp":1743066978019,"version":"3.40.3"},"publisher-location":"Cham","reference-count":36,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030678319"},{"type":"electronic","value":"9783030678326"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-67832-6_54","type":"book-chapter","created":{"date-parts":[[2021,1,22]],"date-time":"2021-01-22T17:44:48Z","timestamp":1611337488000},"page":"666-677","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["An Acceleration Framework for\u00a0Super-Resolution Network via\u00a0Region Difficulty Self-adaption"],"prefix":"10.1007","author":[{"given":"Zhenfang","family":"Guo","sequence":"first","affiliation":[]},{"given":"Yuyao","family":"Ye","sequence":"additional","affiliation":[]},{"given":"Yang","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Ronggang","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,1,21]]},"reference":[{"key":"54_CR1","doi-asserted-by":"crossref","unstructured":"Dong, C., Loy, C.C., He, K.M., Tang, X.O.: Learning a deep convolutional network for image super-resolution. In: ECCV, 184\u2013199 (2014)","DOI":"10.1007\/978-3-319-10593-2_13"},{"key":"54_CR2","doi-asserted-by":"crossref","unstructured":"Lim, B., Son, S., Kim, H., Nah, S., Lee, K.M.: Enhanced deep residual networks for single image super-resolution. In: CVPR (2017)","DOI":"10.1109\/CVPRW.2017.151"},{"key":"54_CR3","doi-asserted-by":"crossref","unstructured":"Zhang, Y.L., Tian, Y.P., Kong, Y., Zhong, B., Fu, Y.: Residual dense network for image super-resolution. In: CVPR, pp. 2472\u20132481 (2018)","DOI":"10.1109\/CVPR.2018.00262"},{"key":"54_CR4","doi-asserted-by":"crossref","unstructured":"Zhang, Y.L., Li, K.P., Li, K., Wang, L.C., Zhong, B.N., Fu, Y.: Image super-resolution using very deep residual channel attention networks. In: ECCV, pp. 1646\u20131654 (2018)","DOI":"10.1007\/978-3-030-01234-2_18"},{"key":"54_CR5","unstructured":"Han, S., Mao, H.Z., Dally, W.J.: Compressing deep neural networks with pruning, trained quantization and Huffman coding. In: ICLR (2016)"},{"key":"54_CR6","doi-asserted-by":"crossref","unstructured":"Luo, J.H., Wu, J.X., Lin, W.Y.: ThiNet: a filter level pruning method for deep neural network compression. In: ICCV, pp. 5058\u20135066 (2017)","DOI":"10.1109\/ICCV.2017.541"},{"key":"54_CR7","doi-asserted-by":"crossref","unstructured":"Jacob, B., et al.: Quantization and training of neural networks for efficient integer-arithmetic-only inference. In: CVPR, pp. 2704\u20132713 (2018)","DOI":"10.1109\/CVPR.2018.00286"},{"key":"54_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"525","DOI":"10.1007\/978-3-319-46493-0_32","volume-title":"Computer Vision \u2013 ECCV 2016","author":"M Rastegari","year":"2016","unstructured":"Rastegari, M., Ordonez, V., Redmon, J., Farhadi, A.: XNOR-Net: ImageNet classification using binary convolutional neural networks. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9908, pp. 525\u2013542. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46493-0_32"},{"key":"54_CR9","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. In: arXiv preprint arXiv:1503.02531 (2015)"},{"key":"54_CR10","unstructured":"Li, H., Kadav, A., Durdanovic, I., Samet, H., Graf, H.P.: Pruning filters for efficient convnets. In: ICLR (2017)"},{"key":"54_CR11","unstructured":"Hinton, G., Oriol, V., Dean, J.: Distilling the knowledge in a neural network. In: arXiv preprint arXiv:1503.02531 (2015)"},{"key":"54_CR12","unstructured":"Howard, A.G., et al.: MobileNets: efficient convolutional neural networks for mobile vision applications. In: arXiv preprint arXiv:1704.04861 (2017)"},{"key":"54_CR13","doi-asserted-by":"crossref","unstructured":"Howard, A., et al.: Searching for mobilenetv3. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"key":"54_CR14","doi-asserted-by":"crossref","unstructured":"Zhang, X.Y., Zhou, X.Y., Lin, M.X., Sun, J.: ShuffleNet: an extremely efficient convolutional neural network for mobile devices. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00716"},{"key":"54_CR15","unstructured":"Baker, B., Gupta, O., Naik, N., Raskar, R.: Designing neural network architectures using reinforement learning. In: ICLR (2017)"},{"key":"54_CR16","doi-asserted-by":"crossref","unstructured":"Liu, C., et al.: Progressive neural architecture search. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01246-5_2"},{"key":"54_CR17","unstructured":"Pham, H., Guan, M.Y., Zoph, B., Le, Q.V., Dean, J.: Efficient neural architecture search via parameter sharing. In: ICML (2018)"},{"key":"54_CR18","doi-asserted-by":"crossref","unstructured":"Kim, J., Lee, J.K., Lee, K.M.: Accurate image super-resolution using very deep convolutional networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1646\u20131654 (2016)","DOI":"10.1109\/CVPR.2016.182"},{"key":"54_CR19","doi-asserted-by":"crossref","unstructured":"Ahn, N., Kang, B., Sohn, K.A.: Fast, accurate, and lightweight super-resolution with cascading residual network. In: ECCV, pp. 252\u2013268 (2018)","DOI":"10.1007\/978-3-030-01249-6_16"},{"key":"54_CR20","doi-asserted-by":"crossref","unstructured":"Dong, C., Loy, C.C., Tang, X.: Accelerating the super-resolution convolutional neural network. In: ECCV, pp. 391\u2013407 (2016)","DOI":"10.1007\/978-3-319-46475-6_25"},{"key":"54_CR21","doi-asserted-by":"crossref","unstructured":"Ledig, C., et al.: Photo-realistic single image super-resolution using a generative adversarial network. In: arXiv preprint arXiv:1609.04802 (2016)","DOI":"10.1109\/CVPR.2017.19"},{"key":"54_CR22","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: NIPS (2014)"},{"key":"54_CR23","unstructured":"Dosovitskiy, A., Brox, T.: Generating images with perceptural similarity metrics based on deep networks. In: arXiv preprint arXiv:1602.02644 (2016)"},{"key":"54_CR24","doi-asserted-by":"crossref","unstructured":"Justin, J., Alexandre, A., Li, F.F.: Perceptual losses for real-time style transfer and super-resolution. In: ECCV, pp. 694\u2013711 (2016)","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"54_CR25","doi-asserted-by":"crossref","unstructured":"Wang, X.T., Chan, K.C., Yu, K., Dong, C., Loy, C.C.: EDVR: video restoration with enhanced deformable convolutional networks. In: CVPR Workshops (2019)","DOI":"10.1109\/CVPRW.2019.00247"},{"key":"54_CR26","doi-asserted-by":"crossref","unstructured":"Liu, Z.C., Wu, B.Y., Luo, W.H., Yang, X., Liu, W., Cheng, K.T.: Bi-Real Net: enhancing the performance of 1-bit CNNs with improved representational capability and advanced training algorithm. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01267-0_44"},{"key":"54_CR27","unstructured":"Iandola, F.N., Moskewicz, M.W., Ashraf, K., Han, S., Dally, W.J., Keutzer, K.: SqueezeNet: Alexnet-level accuracy with 50x fewer parameters and 1 Mb model size. In: arXiv preprint arXiv:1602.07360 (2016)"},{"key":"54_CR28","unstructured":"Tan, M., Le, Q.V.: EfficientNet: rethinking model scaling for convolutional neural networks. In: ICML (2019)"},{"issue":"12","key":"54_CR29","doi-asserted-by":"publisher","first-page":"9828","DOI":"10.1109\/TIE.2018.2886792","volume":"66","author":"W Jia","year":"2018","unstructured":"Jia, W., Zhao, Y., Wang, R.G., Li, S.J., Min, H., Liu, X.P.: Are recent SISR techniques suitable for industrial applications at low magnification? IEEE Trans. Ind. Electron. 66(12), 9828\u20139836 (2018)","journal-title":"IEEE Trans. Ind. Electron."},{"key":"54_CR30","unstructured":"Timofte, R., Gu, S.H., Wu, J.Q., Gool, L.V., Zhang, L., Yang, M.H., et al.: NTIRE 2018 challenge on single image super-resolution: methods and results. In: CVPR Workshop (2018)"},{"key":"54_CR31","doi-asserted-by":"crossref","unstructured":"Abdelhamed, A., Timofte, R., Brown, M.S., Yu, S., Park, B.J., et al.: NTIRE 2019 challenge on real image denoising: methods and results. In: CVPR workshop (2019)","DOI":"10.1109\/CVPRW.2019.00273"},{"key":"54_CR32","doi-asserted-by":"publisher","first-page":"23767","DOI":"10.1109\/ACCESS.2018.2829908","volume":"6","author":"W Wang","year":"2018","unstructured":"Wang, W., Ren, C., He, X.H., Chen, H.G., Qing, L.B.: Video super-resolution via residual learning. IEEE Access 6, 23767\u201323777 (2018)","journal-title":"IEEE Access"},{"issue":"5","key":"54_CR33","doi-asserted-by":"publisher","first-page":"2530","DOI":"10.1109\/TIP.2018.2887017","volume":"28","author":"ZY Wang","year":"2019","unstructured":"Wang, Z.Y., Yi, P., Jiang, K., Jiang, J., Han, Z., Lu, T., Ma, J.Y.: Multi-memory convolutional neural network for video super-resolution. IEEE Trans. Image Process. 28(5), 2530\u20132544 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"54_CR34","unstructured":"Tianchi Homepage: https:\/\/tianchi.aliyun.com\/dataset\/dataDetail?datald=39568. Accessed 4 May 2020"},{"key":"54_CR35","unstructured":"Seungjun Nah Homepage: https:\/\/seungjunnah.github.io\/Datasets\/reds.html. Accessed 6 July 2020"},{"key":"54_CR36","unstructured":"Youtube Homepage: https:\/\/www.youtube.com\/watch?v=LXb3EKWsInQ. Accessed 15 Aug 2020"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-67832-6_54","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,6,14]],"date-time":"2021-06-14T08:26:15Z","timestamp":1623659175000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-67832-6_54"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030678319","9783030678326"],"references-count":36,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-67832-6_54","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"21 January 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Prague","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Czech Republic","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 June 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 June 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/mmm2021.cz\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"211","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"73","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"35% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2,63","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2,5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}