{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,31]],"date-time":"2025-12-31T00:14:00Z","timestamp":1767140040621,"version":"build-2238731810"},"reference-count":42,"publisher":"Springer Science and Business Media LLC","issue":"1-2","license":[{"start":{"date-parts":[[2019,12,17]],"date-time":"2019-12-17T00:00:00Z","timestamp":1576540800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2019,12,17]],"date-time":"2019-12-17T00:00:00Z","timestamp":1576540800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100000268","name":"Biotechnology and Biological Sciences Research Council","doi-asserted-by":"publisher","award":["BB\/P022790\/1"],"award-info":[{"award-number":["BB\/P022790\/1"]}],"id":[{"id":"10.13039\/501100000268","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Machine Vision and Applications"],"published-print":{"date-parts":[[2020,2]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>There is an increase in consumption of agricultural produce as a result of the rapidly growing human population, particularly in developing nations. This has triggered high-quality plant phenotyping research to help with the breeding of high-yielding plants that can adapt to our continuously changing climate. Novel, low-cost, fully automated plant phenotyping systems, capable of infield deployment, are required to help identify quantitative plant phenotypes. The identification of quantitative plant phenotypes is a key challenge which relies heavily on the precise segmentation of plant images. Recently, the plant phenotyping community has started to use very deep convolutional neural networks (CNNs) to help tackle this fundamental problem. However, these very deep CNNs rely on some millions of model parameters and generate very large weight matrices, thus making them difficult to deploy infield on low-cost, resource-limited devices. We explore how to compress existing very deep CNNs for plant image segmentation, thus making them easily deployable infield and on mobile devices. In particular, we focus on applying these models to the pixel-wise segmentation of plants into multiple classes including background, a challenging problem in the plant phenotyping community. We combined two approaches (separable convolutions and SVD) to reduce model parameter numbers and weight matrices of these very deep CNN-based models. Using our combined method (separable convolution and SVD) reduced the weight matrix by up to 95% without affecting pixel-wise accuracy. These methods have been evaluated on two public plant datasets and one non-plant dataset to illustrate generality. We have successfully tested our models on a mobile device.<\/jats:p>","DOI":"10.1007\/s00138-019-01051-7","type":"journal-article","created":{"date-parts":[[2019,12,17]],"date-time":"2019-12-17T05:23:17Z","timestamp":1576560197000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":16,"title":["Towards infield, live plant phenotyping using a reduced-parameter CNN"],"prefix":"10.1007","volume":"31","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2307-2720","authenticated-orcid":false,"given":"John","family":"Atanbori","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8313-2898","authenticated-orcid":false,"given":"Andrew P.","family":"French","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9485-1978","authenticated-orcid":false,"given":"Tony P.","family":"Pridmore","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,12,17]]},"reference":[{"key":"1051_CR1","unstructured":"Aich, S., Stavness, I.: Leaf counting with deep convolutional and deconvolutional networks. (2017) arXiv preprint arXiv:1708.07570"},{"key":"1051_CR2","unstructured":"Aich, S., Stavness, I.: Object counting with small datasets of large images. (2018) arXiv preprint arXiv:1805.11123"},{"key":"1051_CR3","unstructured":"Aich, S., Josuttes, A., Ovsyannikov, I., Strueby, K., Ahmed, I., Duddu, H.S., Pozniak, C., Shirtliffe, S., Stavness, I.: Deepwheat: Estimating phenotypic traits from crop images with deep learning. In: IEEE Winter Conference on Applications of Computer Vision (WACV), 2018, IEEE, pp 323\u2013332 (2018)"},{"key":"1051_CR4","doi-asserted-by":"crossref","unstructured":"Aich, S., van der Kamp, W., Stavness, I.: Semantic binary segmentation using convolutional networks without decoders. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), IEEE, pp. 182\u20131824 (2018)","DOI":"10.1109\/CVPRW.2018.00032"},{"key":"1051_CR5","unstructured":"Alexandratos, N., Bruinsma, J. et al.: World agriculture towards 2030\/2050: the 2012 revision. Tech. rep., ESA Working paper FAO, Rome (2012)"},{"key":"1051_CR6","unstructured":"Atanbori, J., Chen, F., French, A.P., Pridmore, T.: Towards low-cost image-based plant phenotyping using reduced-parameter cnn. In: S A Tsaftaris HS, Pridmore T (eds) Proceedings of the Computer Vision Problems in Plant Phenotyping (CVPPP), BMVA Press, (2018) http:\/\/bmvc2018.org\/contents\/workshops\/cvppp2018\/0023.pdf"},{"issue":"12","key":"1051_CR7","doi-asserted-by":"publisher","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","volume":"39","author":"V Badrinarayanan","year":"2017","unstructured":"Badrinarayanan, V., Kendall, A., Cipolla, R.: Segnet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481\u20132495 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1051_CR8","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: Deep learning with depthwise separable convolutions. (2016) arXiv preprint","DOI":"10.1109\/CVPR.2017.195"},{"key":"1051_CR9","unstructured":"Denton, E.L., Zaremba, W., Bruna, J., LeCun, Y., Fergus, R.: Exploiting linear structure within convolutional networks for efficient evaluation. In: Advances in neural information processing systems, pp 1269\u20131277 (2014)"},{"key":"1051_CR10","unstructured":"Girshick, R.: Fast r-cnn. (2015) arXiv preprint arXiv:1504.08083"},{"key":"1051_CR11","doi-asserted-by":"crossref","unstructured":"Giuffrida, M.V., Minervini, M., Tsaftaris, S.A.: Learning to count leaves in rosette plants (2016)","DOI":"10.5244\/C.29.CVPPP.1"},{"key":"1051_CR12","unstructured":"Han, S., Mao, H., Dally, W.J.: Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. (2015) arXiv preprint arXiv:1510.00149"},{"key":"1051_CR13","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., Adam, H.: Mobilenets: Efficient convolutional neural networks for mobile vision applications. (2017) arXiv preprint arXiv:1704.04861"},{"key":"1051_CR14","unstructured":"Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., Keutzer, K.: Squeezenet: Alexnet-level accuracy with 50x fewer parameters and $$< 0.5~\\text{mb}$$ model size. (2016) arXiv preprint arXiv:1602.07360"},{"key":"1051_CR15","unstructured":"Jin, J., Dundar, A., Culurciello, E.: Flattened convolutional neural networks for feedforward acceleration. (2014) arXiv preprint arXiv:1412.5474"},{"key":"1051_CR16","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1016\/j.rse.2017.06.007","volume":"198","author":"X Jin","year":"2017","unstructured":"Jin, X., Liu, S., Baret, F., Hemerl\u00e9, M., Comar, A.: Estimates of plant density of wheat crops at emergence from very low altitude uav imagery. Remote Sens. Environ. 198, 105\u2013114 (2017)","journal-title":"Remote Sens. Environ."},{"key":"1051_CR17","doi-asserted-by":"crossref","unstructured":"Lin, G., Milan, A., Shen, C., Reid, I.: Refinenet: Multi-path refinement networks for high-resolution semantic segmentation. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.549"},{"key":"1051_CR18","doi-asserted-by":"publisher","first-page":"739","DOI":"10.3389\/fpls.2017.00739","volume":"8","author":"S Liu","year":"2017","unstructured":"Liu, S., Baret, F., Andrieu, B., Burger, P., Hemmerle, M.: Estimation of wheat plant density at early stages using high resolution imagery. Front. Plant Sci. 8, 739 (2017)","journal-title":"Front. Plant Sci."},{"key":"1051_CR19","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3431\u20133440 (2015)","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"1051_CR20","unstructured":"Minervini, M., Fischbach, A., Scharr, H., Tsaftaris, S.: Plant phenotyping datasets. (2015) http:\/\/www.plant-phenotyping.org\/datasets"},{"key":"1051_CR21","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1016\/j.patrec.2015.10.013","volume":"81","author":"M Minervini","year":"2016","unstructured":"Minervini, M., Fischbach, A., Scharr, H., Tsaftaris, S.A.: Finely-grained annotated datasets for image-based plant phenotyping. Pattern Recogn. Lett. 81, 80\u201389 (2016)","journal-title":"Pattern Recogn. Lett."},{"key":"1051_CR22","doi-asserted-by":"crossref","unstructured":"Minervini, M., Giuffrida, M.V., Tsaftaris, S.A.: An interactive tool for semi-automated leaf annotation (2016)","DOI":"10.5244\/C.29.CVPPP.6"},{"issue":"6","key":"1051_CR23","doi-asserted-by":"publisher","first-page":"1049","DOI":"10.1016\/j.imavis.2009.10.001","volume":"28","author":"ME Nilsback","year":"2010","unstructured":"Nilsback, M.E., Zisserman, A.: Delving deeper into the whorl of flower segmentation. Image Vis. Comput. 28(6), 1049\u20131062 (2010)","journal-title":"Image Vis. Comput."},{"key":"1051_CR24","doi-asserted-by":"crossref","unstructured":"Pound, M.P., Atkinson, J.A., Townsend, A.J., Wilson, M.H., Griffiths, M., Jackson, A.S., Bulat, A., Tzimiropoulos, G., Wells, D.M., Murchie, E.H., et al.: Deep machine learning provides state-of-the-art performance in image-based plant phenotyping. GigaScience (2017)","DOI":"10.1093\/gigascience\/gix083"},{"key":"1051_CR25","doi-asserted-by":"crossref","unstructured":"Pound, M.P., Atkinson, J.A., Wells, D.M., Pridmore, T.P., French, A.P.: Deep learning for multi-task plant phenotyping. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2055\u20132063 (2017)","DOI":"10.1101\/204552"},{"key":"1051_CR26","unstructured":"Razafindradina, H.B., Randriamitantsoa, P.A., Razafindrakoto, N.R.: Image compression with svd: A new quality metric based on energy ratio. (2017) arXiv preprint arXiv:1701.06183"},{"key":"1051_CR27","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.C.: Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"issue":"4","key":"1051_CR28","doi-asserted-by":"publisher","first-page":"585","DOI":"10.1007\/s00138-015-0737-3","volume":"27","author":"H Scharr","year":"2016","unstructured":"Scharr, H., Minervini, M., French, A.P., Klukas, C., Kramer, D.M., Liu, X., Luengo, I., Pape, J.M., Polder, G., Vukadinovic, D., et al.: Leaf segmentation in plant phenotyping: a collation study. Mach. Vis. Appl. 27(4), 585\u2013606 (2016)","journal-title":"Mach. Vis. Appl."},{"key":"1051_CR29","doi-asserted-by":"crossref","unstructured":"Shi, W., Caballero, J., Husz\u00e1r, F., Totz, J., Aitken, A.P., Bishop, R., Rueckert, D., Wang, Z.: Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1874\u20131883 (2016)","DOI":"10.1109\/CVPR.2016.207"},{"issue":"2","key":"1051_CR30","doi-asserted-by":"publisher","first-page":"559","DOI":"10.13031\/2013.12945","volume":"46","author":"DS Shrestha","year":"2003","unstructured":"Shrestha, D.S., Steward, B.L.: Automatic corn plant population measurement using machine vision. Trans. ASAE 46(2), 559 (2003)","journal-title":"Trans. ASAE"},{"key":"1051_CR31","doi-asserted-by":"crossref","unstructured":"Sun, Y., Zheng, L., Deng, W., Wang, S.: Svdnet for pedestrian retrieval. (2017) arXiv preprint","DOI":"10.1109\/ICCV.2017.410"},{"key":"1051_CR32","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich, A., et al.: Going deeper with convolutions. Cvpr (2015)","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"1051_CR33","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818\u20132826 (2016)","DOI":"10.1109\/CVPR.2016.308"},{"issue":"1","key":"1051_CR34","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1186\/s13007-015-0056-8","volume":"11","author":"A Walter","year":"2015","unstructured":"Walter, A., Liebisch, F., Hund, A.: Plant phenotyping: from bean weighing to image analysis. Plant Methods 11(1), 14 (2015)","journal-title":"Plant Methods"},{"key":"1051_CR35","unstructured":"Wang, M., Liu, B., Foroosh, H.: Factorized convolutional neural networks. (2016) CoRR, arXiv:1608.04337"},{"key":"1051_CR36","doi-asserted-by":"crossref","unstructured":"Wu, J., Leng, C., Wang, Y., Hu, Q., Cheng, J.: Quantized convolutional neural networks for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4820\u20134828 (2016)","DOI":"10.1109\/CVPR.2016.521"},{"key":"1051_CR37","doi-asserted-by":"crossref","unstructured":"Xue, J., Li, J., Gong, Y.: Restructuring of deep neural network acoustic models with singular value decomposition. In: Interspeech, pp. 2365\u20132369 (2013)","DOI":"10.21437\/Interspeech.2013-552"},{"key":"1051_CR38","unstructured":"Yu, D., Seide, F., Li, G., Deng, L.: Exploiting sparseness in deep neural networks for large vocabulary speech recognition. In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2012, IEEE, pp. 4409\u20134412 (2012)"},{"key":"1051_CR39","unstructured":"Yu, F., Koltun, V.: Multi-scale context aggregation by dilated convolutions. (2015) arXiv preprint arXiv:1511.07122"},{"key":"1051_CR40","doi-asserted-by":"crossref","unstructured":"Yu, F., Koltun, V., Funkhouser, T.: Dilated residual networks. In: Computer Vision and Pattern Recognition, vol 1 (2017)","DOI":"10.1109\/CVPR.2017.75"},{"key":"1051_CR41","unstructured":"Zhang, X., Zhou, X., Lin, M., Sun, J.: Shufflenet: An extremely efficient convolutional neural network for mobile devices. (2017) arXiv preprint arXiv:1707.01083"},{"key":"1051_CR42","doi-asserted-by":"crossref","unstructured":"Zhao, H., Shi, J., Qi, X., Wang, X., Jia, J.: Pyramid scene parsing network. In: IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), pp. 2881\u20132890 (2017)","DOI":"10.1109\/CVPR.2017.660"}],"updated-by":[{"DOI":"10.1007\/s00138-020-01066-5","type":"correction","label":"Correction","source":"publisher","updated":{"date-parts":[[2020,2,27]],"date-time":"2020-02-27T00:00:00Z","timestamp":1582761600000}}],"container-title":["Machine Vision and Applications"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-019-01051-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s00138-019-01051-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-019-01051-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,8]],"date-time":"2022-10-08T13:49:09Z","timestamp":1665236949000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s00138-019-01051-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,12,17]]},"references-count":42,"journal-issue":{"issue":"1-2","published-print":{"date-parts":[[2020,2]]}},"alternative-id":["1051"],"URL":"https:\/\/doi.org\/10.1007\/s00138-019-01051-7","relation":{},"ISSN":["0932-8092","1432-1769"],"issn-type":[{"value":"0932-8092","type":"print"},{"value":"1432-1769","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019,12,17]]},"assertion":[{"value":"2 October 2018","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 April 2019","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 November 2019","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 December 2019","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 February 2020","order":5,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Correction","order":6,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The articles listed below were published in Issue January 2020, Issue 1, instead of Issue February 2020, Issues 1\u20132.","order":7,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"2"}}