{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,22]],"date-time":"2026-03-22T05:52:22Z","timestamp":1774158742387,"version":"3.50.1"},"reference-count":27,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2022,4,30]],"date-time":"2022-04-30T00:00:00Z","timestamp":1651276800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,4,30]],"date-time":"2022-04-30T00:00:00Z","timestamp":1651276800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100020194","name":"Wellcome\/EPSRC Centre for Interventional and Surgical Sciences","doi-asserted-by":"crossref","award":["203145\/Z\/16\/Z"],"award-info":[{"award-number":["203145\/Z\/16\/Z"]}],"id":[{"id":"10.13039\/501100020194","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["EP\/P027938\/1, EP\/R004080\/1, EP\/P012841\/1"],"award-info":[{"award-number":["EP\/P027938\/1, EP\/R004080\/1, EP\/P012841\/1"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J CARS"],"published-print":{"date-parts":[[2022,5]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:sec>\n                <jats:title>Purpose<\/jats:title>\n                <jats:p>In obstetric ultrasound (US) scanning, the learner\u2019s ability to mentally build a three-dimensional (3D) map of the fetus from a two-dimensional (2D) US image represents a major challenge in skill acquisition. We aim to build a US plane localisation system for 3D visualisation, training, and guidance without integrating additional sensors.<\/jats:p>\n              <\/jats:sec><jats:sec>\n                <jats:title>Methods<\/jats:title>\n                <jats:p>We propose a regression convolutional neural network (CNN) using image features to estimate the six-dimensional pose of arbitrarily oriented US planes relative to the fetal brain centre. The network was trained on synthetic images acquired from phantom 3D US volumes and fine-tuned on real scans. Training data was generated by slicing US volumes into imaging planes in Unity at random coordinates and more densely around the standard transventricular (TV) plane.<\/jats:p>\n              <\/jats:sec><jats:sec>\n                <jats:title>Results<\/jats:title>\n                <jats:p>With phantom data, the median errors are 0.90 mm\/1.17<jats:inline-formula><jats:alternatives><jats:tex-math>$$^\\circ $$<\/jats:tex-math><mml:math xmlns:mml=\"http:\/\/www.w3.org\/1998\/Math\/MathML\">\n                    <mml:msup>\n                      <mml:mrow\/>\n                      <mml:mo>\u2218<\/mml:mo>\n                    <\/mml:msup>\n                  <\/mml:math><\/jats:alternatives><\/jats:inline-formula>\u00a0and 0.44 mm\/1.21<jats:inline-formula><jats:alternatives><jats:tex-math>$$^\\circ $$<\/jats:tex-math><mml:math xmlns:mml=\"http:\/\/www.w3.org\/1998\/Math\/MathML\">\n                    <mml:msup>\n                      <mml:mrow\/>\n                      <mml:mo>\u2218<\/mml:mo>\n                    <\/mml:msup>\n                  <\/mml:math><\/jats:alternatives><\/jats:inline-formula>\u00a0for random planes and planes close to the TV one, respectively. With real data, using a different fetus with the same gestational age (GA), these errors are 11.84 mm\/25.17<jats:inline-formula><jats:alternatives><jats:tex-math>$$^\\circ $$<\/jats:tex-math><mml:math xmlns:mml=\"http:\/\/www.w3.org\/1998\/Math\/MathML\">\n                    <mml:msup>\n                      <mml:mrow\/>\n                      <mml:mo>\u2218<\/mml:mo>\n                    <\/mml:msup>\n                  <\/mml:math><\/jats:alternatives><\/jats:inline-formula>. The average inference time is 2.97 ms per plane.<\/jats:p>\n              <\/jats:sec><jats:sec>\n                <jats:title>Conclusion<\/jats:title>\n                <jats:p>The proposed network reliably localises US planes within the fetal brain in phantom data and successfully generalises pose regression for an unseen fetal brain from a similar GA as in training. Future development will expand the prediction to volumes of the whole fetus and assess its potential for vision-based, freehand US-assisted navigation when acquiring standard fetal planes.<\/jats:p>\n              <\/jats:sec>","DOI":"10.1007\/s11548-022-02609-z","type":"journal-article","created":{"date-parts":[[2022,4,30]],"date-time":"2022-04-30T05:22:23Z","timestamp":1651296143000},"page":"833-839","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":16,"title":["Deep learning-based plane pose regression in obstetric ultrasound"],"prefix":"10.1007","volume":"17","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2853-9791","authenticated-orcid":false,"given":"Chiara","family":"Di Vece","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9494-9724","authenticated-orcid":false,"given":"Brian","family":"Dromey","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4609-1177","authenticated-orcid":false,"given":"Francisco","family":"Vasconcelos","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0199-6140","authenticated-orcid":false,"given":"Anna L.","family":"David","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8562-1970","authenticated-orcid":false,"given":"Donald","family":"Peebles","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0980-3227","authenticated-orcid":false,"given":"Danail","family":"Stoyanov","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,4,30]]},"reference":[{"key":"2609_CR1","doi-asserted-by":"publisher","first-page":"333","DOI":"10.1016\/0002-9378(85)90298-4","volume":"151","author":"FP Hadlock","year":"1985","unstructured":"Hadlock FP, Harrist RB, Sharman RS, Deter RL, Park SK (1985) Estimation of fetal weight with the use of head body and femur measurements: a prospective study. Am J Obstet Gynecol 151:333","journal-title":"Am J Obstet Gynecol"},{"key":"2609_CR2","doi-asserted-by":"publisher","first-page":"116","DOI":"10.1002\/uog.8831","volume":"37","author":"LJ Salomon","year":"2011","unstructured":"Salomon LJ, Alfirevic Z, Berghella V, Bilardo CM, Hernandez-Andrade E, Johnsen SL, Kalache K, Leung KY, Malinger G, Munoz H, Prefumo F, Toi A, Lee W (2011) Practice guidelines for performance of the routine mid-trimester fetal ultrasound scan. Ultrasound Obstet Gynecol 37:116\u2013126","journal-title":"Ultrasound Obstet Gynecol"},{"key":"2609_CR3","doi-asserted-by":"publisher","first-page":"681","DOI":"10.1002\/uog.8997","volume":"38","author":"I Sarris","year":"2011","unstructured":"Sarris I, Ioannou C, Dighe M, Mitidieri A, Oberto M, Qingqing W, Shah J, Sohoni S, Zidjali WA, Hoch L, Altman DG, Papageorghiou AT (2011) Standardization of fetal ultrasound biometry measurements: improving the quality and consistency of measurements. Ultrasound Obstet Gynecol 38:681\u2013687","journal-title":"Ultrasound Obstet Gynecol"},{"key":"2609_CR4","doi-asserted-by":"publisher","first-page":"183","DOI":"10.7863\/ultra.15.02036","volume":"35","author":"DP Bahner","year":"2016","unstructured":"Bahner DP, Blickendorf JM, Bockbrader M, Adkins E, Vira A, Boulger C, Panchal AR (2016) Language of transducer manipulation: codifying terms for effective teaching. J Ultrasound Med 35:183\u2013188","journal-title":"J Ultrasound Med"},{"key":"2609_CR5","unstructured":"Chandrasekaran D, Patel H, Myriokefalitaki E, Woodhead N, Jones K, Gebeh AK, Jeve Y (2016) Simulation training - trainees want it but don\u2019t use it: A study by midlands research collaborative in obstetrics and gynaecology. In: Royal College of Obstetricians and Gynaecologists World Congress"},{"key":"2609_CR6","doi-asserted-by":"crossref","unstructured":"Zhou Y, Barnes C, Lu J, Yang J, Li H (2019) On the continuity of rotation representations in neural networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 5738\u20135746","DOI":"10.1109\/CVPR.2019.00589"},{"key":"2609_CR7","doi-asserted-by":"publisher","first-page":"391","DOI":"10.1007\/s11548-009-0316-z","volume":"4","author":"D Ruijters","year":"2009","unstructured":"Ruijters D, ter Haar Romeny BM, Suetens P (2009) Vesselness-based 2d\u20133d registration of the coronary arteries. Int J Comput Assist Radiol Surg 4:391\u2013397","journal-title":"Int J Comput Assist Radiol Surg"},{"key":"2609_CR8","doi-asserted-by":"publisher","first-page":"1749","DOI":"10.1118\/1.597276","volume":"21","author":"ND Kitchen","year":"1994","unstructured":"Kitchen ND, Thomas DGT (1994) A patient-to-computed-tomography image registration method based on digitally reconstructed radiographs. Med Phys 21:1749\u20131760","journal-title":"Med Phys"},{"key":"2609_CR9","doi-asserted-by":"crossref","unstructured":"Guo H, Xu X, Xu S, Wood BJ, Yan P (2021) End-to-end ultrasound frame to volume registration. In: Medical Image Computing and Computer-Assisted Intervention Conference, pp. 56\u201365","DOI":"10.1007\/978-3-030-87202-1_6"},{"key":"2609_CR10","doi-asserted-by":"crossref","unstructured":"Jaganathan S, Wang J, Borsdorf A, Shetty K, Maier A Deep iterative 2d\/3d registration. In: Medical Image Computing and Computer-Assisted Intervention Conference, pp. 383\u2013392 (2021)","DOI":"10.1007\/978-3-030-87202-1_37"},{"issue":"7","key":"2609_CR11","doi-asserted-by":"publisher","first-page":"1505","DOI":"10.1109\/TBME.2015.2496253","volume":"63","author":"G Wu","year":"2016","unstructured":"Wu G, Kim M, Wang Q, Munsell BC, Shen D (2016) Scalable high-performance image registration framework by unsupervised deep feature representations learning. IEEE Trans Biomed Eng 63(7):1505\u20131516","journal-title":"IEEE Trans Biomed Eng"},{"key":"2609_CR12","doi-asserted-by":"crossref","unstructured":"Pavlakos G, Zhou X, Chan A, Derpanis KG, Daniilidis K (2017) 6-dof object pose from semantic keypoints. In: IEEE International Conference on Robotics and Automation, pp. 2011\u20132018","DOI":"10.1109\/ICRA.2017.7989233"},{"key":"2609_CR13","doi-asserted-by":"crossref","unstructured":"Tulsiani S, Malik J (2015) Viewpoints and keypoints. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1510\u20131519","DOI":"10.1109\/CVPR.2015.7298758"},{"key":"2609_CR14","doi-asserted-by":"crossref","unstructured":"Su H, Qi CR, Li Y, Guibas LJ (2015) Render for cnn: Viewpoint estimation in images using cnns trained with rendered 3d model views. In: IEEE International Conference on Computer Vision, pp. 2686\u20132694","DOI":"10.1109\/ICCV.2015.308"},{"key":"2609_CR15","doi-asserted-by":"crossref","unstructured":"Simonovsky M, Guti\u00e9rrez-Becker B, Mateus D, Navab N, Komodakis N (2016) A deep metric for multimodal registration. In: Medical Image Computing and Computer-Assisted Intervention, pp. 10\u201318","DOI":"10.1007\/978-3-319-46726-9_2"},{"key":"2609_CR16","doi-asserted-by":"crossref","unstructured":"Mahendran S, Ali H, Vidal R (2017) 3d pose regression using convolutional neural networks. In: IEEE International Conference on Computer Vision Workshops, pp. 2174\u20132182","DOI":"10.1109\/ICCVW.2017.254"},{"key":"2609_CR17","doi-asserted-by":"publisher","first-page":"1737","DOI":"10.1109\/TMI.2018.2798801","volume":"37","author":"B Hou","year":"2018","unstructured":"Hou B, Khanal B, Alansary A, McDonagh S, Davidson A, Rutherford M, Hajnal JV, Rueckert D, Glocker B, Kainz B (2018) 3d reconstruction in canonical co-ordinate space from arbitrarily oriented 2d images. IEEE Trans Med Imag 37:1737\u20131750","journal-title":"IEEE Trans Med Imag"},{"key":"2609_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.media.2018.02.006","volume":"46","author":"AIL Namburete","year":"2018","unstructured":"Namburete AIL, Xie W, Yaqub M, Zisserman A, Noble JA (2018) Fully-automated alignment of 3d fetal brain ultrasound to a canonical reference space using multi-task learning. Med Image Anal 46:1\u201314","journal-title":"Med Image Anal"},{"key":"2609_CR19","doi-asserted-by":"publisher","first-page":"1352","DOI":"10.1109\/TMI.2016.2521800","volume":"35","author":"S Miao","year":"2016","unstructured":"Miao S, Wang ZJ, Liao R (2016) A cnn regression approach for real-time 2d\/3d registration. IEEE Trans Med Imag 35:1352\u20131363","journal-title":"IEEE Trans Med Imag"},{"key":"2609_CR20","doi-asserted-by":"crossref","unstructured":"Li Y, Khanal B, Hou B, Alansary A, Cerrolaza JJ, Sinclair M, Matthew J, Gupta C, Knight C, Kainz B, Rueckert D (2018) Standard plane detection in 3d fetal ultrasound using an iterative transformation network. In: Medical Image Computing and Computer Assisted Intervention Conference, pp. 392\u2013400","DOI":"10.1007\/978-3-030-00928-1_45"},{"key":"2609_CR21","doi-asserted-by":"publisher","first-page":"470","DOI":"10.1109\/TMI.2018.2866442","volume":"38","author":"SSM Salehi","year":"2019","unstructured":"Salehi SSM, Khan S, Erdogmus D, Gholipour A (2019) Real-time deep pose estimation with geodesic loss for image-to-template rigid registration. IEEE Trans Med Imag 38:470\u2013481","journal-title":"IEEE Trans Med Imag"},{"key":"2609_CR22","doi-asserted-by":"crossref","unstructured":"Kendall A, Cipolla R (2017) Geometric loss functions for camera pose regression with deep learning. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 6555\u20136564","DOI":"10.1109\/CVPR.2017.694"},{"key":"2609_CR23","doi-asserted-by":"crossref","unstructured":"Xiang Y, Schmidt T, Narayanan V, Fox D (2018) Posecnn: A convolutional neural network for 6d object pose estimation in cluttered scenes. In: Robotics: Science and Systems Foundation Conference","DOI":"10.15607\/RSS.2018.XIV.019"},{"key":"2609_CR24","doi-asserted-by":"crossref","unstructured":"Kanazawa A, Black MJ, Jacobs DW, Malik J (2018) End-to-end recovery of human shape and pose. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 7122\u20137131","DOI":"10.1109\/CVPR.2018.00744"},{"issue":"6","key":"2609_CR25","doi-asserted-by":"publisher","first-page":"700","DOI":"10.1002\/uog.7705","volume":"36","author":"LR Pistorius","year":"2010","unstructured":"Pistorius LR, Stoutenbeek P, Groenendaal F, de Vries L, Manten G, Mulder E, Visser G (2010) Grade and symmetry of normal fetal cortical development: a longitudinal two- and three-dimensional ultrasound study. Ultrasound Obstet Gynecol 36(6):700\u2013708","journal-title":"Ultrasound Obstet Gynecol"},{"key":"2609_CR26","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"2609_CR27","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2010) Imagenet: A large-scale hierarchical image database. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"}],"container-title":["International Journal of Computer Assisted Radiology and Surgery"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-022-02609-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11548-022-02609-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-022-02609-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,16]],"date-time":"2022-05-16T11:36:08Z","timestamp":1652700968000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11548-022-02609-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,30]]},"references-count":27,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2022,5]]}},"alternative-id":["2609"],"URL":"https:\/\/doi.org\/10.1007\/s11548-022-02609-z","relation":{},"ISSN":["1861-6429"],"issn-type":[{"value":"1861-6429","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4,30]]},"assertion":[{"value":"8 March 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 March 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 April 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This article does not contain any studies with human participants performed by any of the authors.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"This article does not contain patient data.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}}]}}