{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,18]],"date-time":"2026-01-18T07:08:07Z","timestamp":1768720087828,"version":"3.49.0"},"reference-count":34,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,2,12]],"date-time":"2024-02-12T00:00:00Z","timestamp":1707696000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,2,12]],"date-time":"2024-02-12T00:00:00Z","timestamp":1707696000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Cogn Comput"],"published-print":{"date-parts":[[2024,5]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>The applications of deep learning algorithms with images to various scenarios have attracted significant research attention. However, application scenarios in animal breeding managements are still limited. In this paper we propose a new deep learning framework to estimate the chest circumference of domestic animals from images. This parameter is a key metric for breeding and monitoring the quality of animal in animal husbandry. We design a set of feature extraction methods based on a multi-task learning framework to address the challenging issues in the main estimation task. The multiple tasks in our proposed framework include object segmentation, keypoint estimation, and depth estimation of cow from monocular images. The domain-specific features extracted from these tasks improve upon our main estimation task. In addition, we also attempt to reduce unnecessary computations during the framework design to reduce the cost of subsequent practical implementation of the developed system. Our proposed framework is tested on our own collected dataset to evaluate its performance.<\/jats:p>","DOI":"10.1007\/s12559-024-10250-y","type":"journal-article","created":{"date-parts":[[2024,2,12]],"date-time":"2024-02-12T09:02:22Z","timestamp":1707728542000},"page":"1092-1102","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Deep Multi-task Learning for Animal Chest Circumference Estimation from Monocular Images"],"prefix":"10.1007","volume":"16","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-6037-670X","authenticated-orcid":false,"given":"Hongtao","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Dongbing","family":"Gu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,12]]},"reference":[{"key":"10250_CR1","doi-asserted-by":"crossref","unstructured":"Liu W, Anguelov D, Erhan D, Szegedy C, Reed SE, Fu C-Y, Berg AC. SSD: Single shot multibox detector. In: European Conference on Computer Vision. 2015.","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"10250_CR2","doi-asserted-by":"crossref","unstructured":"He K, Gkioxari G, Doll\u00e1r P, Girshick RB. Mask R-CNN. In: 2017 IEEE International Conference on Computer Vision (ICCV). 2017. p. 2980\u20138.","DOI":"10.1109\/ICCV.2017.322"},{"key":"10250_CR3","doi-asserted-by":"crossref","unstructured":"Cao Z, Simon T, Wei S-E, Sheikh Y. Realtime multi-person 2D pose estimation using part affinity fields. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2016. p. 1302\u201310.","DOI":"10.1109\/CVPR.2017.143"},{"key":"10250_CR4","doi-asserted-by":"crossref","unstructured":"Deng R, Zeng G, Rui G, Zha H. Image-based building reconstruction with Manhattan-world assumption. In: Pattern Recognition. 2011.","DOI":"10.1109\/ACPR.2011.6892193"},{"key":"10250_CR5","unstructured":"Chen L-C, Papandreou G, Schroff F, Adam H. Rethinking atrous convolution for semantic image segmentation. arXiv:1706.05587 [Preprint]. 2017. Available from: http:\/\/arxiv.org\/abs\/1706.05587."},{"key":"10250_CR6","unstructured":"Zhou X, Wang D, Kr\u00e4henb\u00fchl P. Objects as points. arXiv:1904.07850 [Preprint]. 2019. Available from: http:\/\/arxiv.org\/abs\/1904.07850."},{"key":"10250_CR7","doi-asserted-by":"crossref","unstructured":"Soviany P, Ionescu RT. Optimizing the trade-off between single-stage and two-stage deep object detectors using image difficulty prediction. In: 2018 20th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC). 2018. p. 209\u201314.","DOI":"10.1109\/SYNASC.2018.00041"},{"key":"10250_CR8","unstructured":"Ge Z, Liu S, Wang F, Li Z, Sun J. YOLOX: Exceeding YOLO series in 2021. arXiv:2107.08430 [Preprint]. 2021. Available from: http:\/\/arxiv.org\/abs\/2107.08430."},{"key":"10250_CR9","doi-asserted-by":"crossref","unstructured":"Tian Z, Shen C, Chen H, He T. FCOS: Fully convolutional one-stage object detection. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV). 2019. p. 9626\u201335.","DOI":"10.1109\/ICCV.2019.00972"},{"key":"10250_CR10","doi-asserted-by":"crossref","unstructured":"G\u00fcler RA, Trigeorgis G, Antonakos E, Snape P, Zafeiriou S, Kokkinos I. DenseReg: Fully convolutional dense shape regression in-the-wild. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2016. p. 2614\u201323.","DOI":"10.1109\/CVPR.2017.280"},{"key":"10250_CR11","unstructured":"Wang X, Zhang R, Kong T, Li L, Shen C. SOLOv2: Dynamic, faster and stronger. arXiv:2003.10152 [Preprint]. 2020. Available from: http:\/\/arxiv.org\/abs\/2003.10152."},{"key":"10250_CR12","doi-asserted-by":"crossref","unstructured":"Qiao S, Shen W, Zhang Z, Wang B, Yuille AL. Deep co-training for semi-supervised image recognition. arXiv:1803.05984 [Preprint]. 2018. Available from: http:\/\/arxiv.org\/abs\/1803.05984.","DOI":"10.1007\/978-3-030-01267-0_9"},{"issue":"8","key":"10250_CR13","doi-asserted-by":"publisher","first-page":"800","DOI":"10.1109\/34.946985","volume":"23","author":"Y Deng","year":"2001","unstructured":"Deng Y, Manjunath BS. Unsupervised segmentation of color-texture regions in images and video. IEEE Trans Patt Anal Mach Intell. 2001;23(8):800\u201310.","journal-title":"IEEE Trans Patt Anal Mach Intell"},{"key":"10250_CR14","doi-asserted-by":"crossref","unstructured":"Sun K, Xiao B, Liu D, Wang J. Deep high-resolution representation learning for human pose estimation. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2019. p. 5686\u201396.","DOI":"10.1109\/CVPR.2019.00584"},{"key":"10250_CR15","doi-asserted-by":"crossref","unstructured":"Deng J, Guo J, Zhou Y, Yu J, Kotsia I, Zafeiriou S. RetinaFace: Single-stage dense face localisation in the wild. arXiv:1905.00641 [Preprint]. 2019. Available from: http:\/\/arxiv.org\/abs\/1905.00641.","DOI":"10.1109\/CVPR42600.2020.00525"},{"key":"10250_CR16","doi-asserted-by":"crossref","unstructured":"Fang H, Xie S, Tai Y-W, Lu C. RMPE: Regional multi-person pose estimation. In: 2017 IEEE International Conference on Computer Vision (ICCV). 2016. p. 2353\u201362.","DOI":"10.1109\/ICCV.2017.256"},{"key":"10250_CR17","doi-asserted-by":"crossref","unstructured":"Godard C, Aodha OM, Brostow GJ. Digging into self-supervised monocular depth estimation. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV). 2018. p. 3827\u201337.","DOI":"10.1109\/ICCV.2019.00393"},{"issue":"4","key":"10250_CR18","first-page":"3242","volume":"34","author":"S Bhatia","year":"2020","unstructured":"Bhatia S, Hooi B, Yoon M, Shin K, Faloutsos C. Midas: Microcluster-based detector of anomalies in edge streams. Proc AAAI Conf Artif Intell. 2020;34(4):3242\u20139.","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"10250_CR19","doi-asserted-by":"publisher","first-page":"1623","DOI":"10.1109\/TPAMI.2020.3019967","volume":"44","author":"R Ranftl","year":"2019","unstructured":"Ranftl R, Lasinger K, Hafner D, Schindler K, Koltun V. Towards robust monocular depth estimation: Mixing datasets for zero-shot cross-dataset transfer. IEEE Trans Pattern Anal Mach Intell. 2019;44:1623\u201337.","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10250_CR20","doi-asserted-by":"crossref","unstructured":"Ranftl R, Bochkovskiy A, Koltun V. Vision transformers for dense prediction. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV). 2021. p. 12159\u201368.","DOI":"10.1109\/ICCV48922.2021.01196"},{"issue":"11","key":"10250_CR21","doi-asserted-by":"publisher","first-page":"1231","DOI":"10.1177\/0278364913491297","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger A, Lenz P, Stiller C, Urtasun R. Vision meets robotics: The KITTI dataset. Int J Robot Res. 2013;32(11):1231\u20137.","journal-title":"Int. J. Robot. Res."},{"key":"10250_CR22","unstructured":"Cordts M, Omran M, Ramos S, Scharw\u00e4chter T, Enzweiler M, Benenson R, Franke U, Roth S, Schiele B. The cityscapes dataset. In: CVPR Workshop on The Future of Datasets in Vision. 2015."},{"key":"10250_CR23","doi-asserted-by":"crossref","unstructured":"Gordon A, Li H, Jonschkowski R, Angelova A. Depth from videos in the wild: Unsupervised monocular depth learning from unknown cameras. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV). p. 8976\u201385.","DOI":"10.1109\/ICCV.2019.00907"},{"key":"10250_CR24","doi-asserted-by":"crossref","unstructured":"Yin Z, Shi J. GeoNet: Unsupervised learning of dense depth, optical flow and camera pose. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 2018. p. 1983\u201392.","DOI":"10.1109\/CVPR.2018.00212"},{"key":"10250_CR25","doi-asserted-by":"crossref","unstructured":"Chen T, Gu D. CSA6D: Channel-spatial attention networks for 6D object pose estimation. Cognit Comput. 2021.","DOI":"10.1007\/s12559-021-09966-y"},{"key":"10250_CR26","doi-asserted-by":"crossref","unstructured":"Wang F, Xie Z. An adversarial multi-task learning method for Chinese text correction with semantic detection. 2023.","DOI":"10.1007\/978-3-031-15931-2_14"},{"key":"10250_CR27","unstructured":"Xu P, Cai J, Gao Y, Rong Z, Xin H. MIRACLE: Multi-task Learning based Interpretable Regulation of Autoimmune Diseases through Common Latent Epigenetics. 2023."},{"key":"10250_CR28","doi-asserted-by":"crossref","unstructured":"Lin T-Y, Doll\u00e1r P, Girshick RB, He K, Hariharan B, Belongie SJ. Feature pyramid networks for object detection. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2016. p. 936\u201344.","DOI":"10.1109\/CVPR.2017.106"},{"key":"10250_CR29","doi-asserted-by":"crossref","unstructured":"Ding X, Zhang X, Ma N, Han J, Ding G, Sun J. RepVGG: Making VGG-style convnets great again. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2021. p. 13728\u201337.","DOI":"10.1109\/CVPR46437.2021.01352"},{"key":"10250_CR30","doi-asserted-by":"crossref","unstructured":"Woo S, Park J, Lee J-Y, Kweon I-S. CBAM: Convolutional block attention module. In: European Conference on Computer Vision. 2018.","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"10250_CR31","unstructured":"Loshchilov I, Hutter F. Fixing weight decay regularization in adam. arXiv:1711.05101 [Preprint]. 2017. Availalble from: http:\/\/arxiv.org\/abs\/1711.05101."},{"key":"10250_CR32","doi-asserted-by":"crossref","unstructured":"Cheng B, Girshick R, Doll\u00e1r P, Berg AC, Kirillov A. Boundary IoU: Improving object-centric image segmentation evaluation. 2021.","DOI":"10.1109\/CVPR46437.2021.01508"},{"key":"10250_CR33","doi-asserted-by":"crossref","unstructured":"Han K, Wang Y, Tian Q, Guo J, Xu C, Xu C. GhostNet: More features from cheap operations. 2020.","DOI":"10.1109\/CVPR42600.2020.00165"},{"key":"10250_CR34","doi-asserted-by":"crossref","unstructured":"Chen J, Kao S-H, He H, Zhuo W, Wen S, Lee C-H, Chan S-HG. Run, don\u2019t walk: Chasing higher FLOPS for faster neural networks. 2023.","DOI":"10.1109\/CVPR52729.2023.01157"}],"container-title":["Cognitive Computation"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12559-024-10250-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s12559-024-10250-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12559-024-10250-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,11]],"date-time":"2024-11-11T02:08:25Z","timestamp":1731290905000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s12559-024-10250-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,2,12]]},"references-count":34,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,5]]}},"alternative-id":["10250"],"URL":"https:\/\/doi.org\/10.1007\/s12559-024-10250-y","relation":{},"ISSN":["1866-9956","1866-9964"],"issn-type":[{"value":"1866-9956","type":"print"},{"value":"1866-9964","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,2,12]]},"assertion":[{"value":"20 September 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 January 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 February 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"This article does not contain any studies with human participants or live animals performed by any of the authors.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics Approval"}},{"value":"The authors declare no competing.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}}]}}