{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T22:30:41Z","timestamp":1768343441623,"version":"3.49.0"},"reference-count":27,"publisher":"Springer Science and Business Media LLC","issue":"11","license":[{"start":{"date-parts":[[2020,8,17]],"date-time":"2020-08-17T00:00:00Z","timestamp":1597622400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2020,8,17]],"date-time":"2020-08-17T00:00:00Z","timestamp":1597622400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"Wellcome\/EPSRC","award":["203145Z\/16\/Z"],"award-info":[{"award-number":["203145Z\/16\/Z"]}]},{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["EP\/P027938\/1"],"award-info":[{"award-number":["EP\/P027938\/1"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["EP\/R004080\/1"],"award-info":[{"award-number":["EP\/R004080\/1"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["NS\/A000027\/1"],"award-info":[{"award-number":["NS\/A000027\/1"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100010664","name":"H2020 Future and Emerging Technologies","doi-asserted-by":"publisher","award":["GA 863146"],"award-info":[{"award-number":["GA 863146"]}],"id":[{"id":"10.13039\/100010664","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Royal Academy of Engineering Chair in Emerging Technologies","award":["CiET1819\/2\/36"],"award-info":[{"award-number":["CiET1819\/2\/36"]}]},{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["EP\/P012841\/1"],"award-info":[{"award-number":["EP\/P012841\/1"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Medtronic\/Royal Academy of Engineering Research Chair","award":["RCSRF1819\/7\/34"],"award-info":[{"award-number":["RCSRF1819\/7\/34"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J CARS"],"published-print":{"date-parts":[[2020,11]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:sec>\n                <jats:title>Purpose<\/jats:title>\n                <jats:p>Fetoscopic laser photocoagulation is a minimally invasive surgical procedure used to treat twin-to-twin transfusion syndrome (TTTS), which involves localization and ablation of abnormal vascular connections on the placenta to regulate the blood flow in both fetuses. This procedure is particularly challenging due to the limited field of view, poor visibility, occasional bleeding, and poor image quality. Fetoscopic mosaicking can help in creating an image with the expanded field of view which could facilitate the clinicians during the TTTS procedure.<\/jats:p>\n              <\/jats:sec><jats:sec>\n                <jats:title>Methods<\/jats:title>\n                <jats:p>We propose a deep learning-based mosaicking framework for diverse fetoscopic videos captured from different settings such as simulation, phantoms, ex vivo, and in vivo environments. The proposed mosaicking framework extends an existing deep image homography model to handle video data by introducing the controlled data generation and consistent homography estimation modules. Training is performed on a small subset of fetoscopic images which are independent of the testing videos.<\/jats:p>\n              <\/jats:sec><jats:sec>\n                <jats:title>Results<\/jats:title>\n                <jats:p>We perform both quantitative and qualitative evaluations on 5 diverse fetoscopic videos (2400 frames) that captured different environments. To demonstrate the robustness of the proposed framework, a comparison is performed with the existing feature-based and deep image homography methods.<\/jats:p>\n              <\/jats:sec><jats:sec>\n                <jats:title>Conclusion<\/jats:title>\n                <jats:p>The proposed mosaicking framework outperformed existing methods and generated meaningful mosaic, while reducing the accumulated drift, even in the presence of visual challenges such as specular highlights, reflection, texture paucity, and low video resolution.<\/jats:p>\n              <\/jats:sec>","DOI":"10.1007\/s11548-020-02242-8","type":"journal-article","created":{"date-parts":[[2020,8,17]],"date-time":"2020-08-17T19:02:32Z","timestamp":1597690952000},"page":"1807-1816","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":24,"title":["Deep learning-based fetoscopic mosaicking for field-of-view expansion"],"prefix":"10.1007","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1329-4565","authenticated-orcid":false,"given":"Sophia","family":"Bano","sequence":"first","affiliation":[]},{"given":"Francisco","family":"Vasconcelos","sequence":"additional","affiliation":[]},{"given":"Marcel","family":"Tella-Amo","sequence":"additional","affiliation":[]},{"given":"George","family":"Dwyer","sequence":"additional","affiliation":[]},{"given":"Caspar","family":"Gruijthuijsen","sequence":"additional","affiliation":[]},{"given":"Emmanuel","family":"Vander Poorten","sequence":"additional","affiliation":[]},{"given":"Tom","family":"Vercauteren","sequence":"additional","affiliation":[]},{"given":"Sebastien","family":"Ourselin","sequence":"additional","affiliation":[]},{"given":"Jan","family":"Deprest","sequence":"additional","affiliation":[]},{"given":"Danail","family":"Stoyanov","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,8,17]]},"reference":[{"key":"2242_CR1","unstructured":"Baker S, Datta A, Kanade T (2006) Parameterizing homographies. Technical Report CMU-RI-TR-06-11"},{"key":"2242_CR2","doi-asserted-by":"crossref","unstructured":"Bano S, Vasconcelos F, Amo MT, Dwyer G, Gruijthuijsen C, Deprest J, Ourselin S, Vander\u00a0Poorten E, Vercauteren T, Stoyanov D (2019) Deep sequential mosaicking of fetoscopic videos. In: International conference on medical image computing and computer-assisted intervention, Springer, pp 311\u2013319","DOI":"10.1007\/978-3-030-32239-7_35"},{"key":"2242_CR3","doi-asserted-by":"crossref","unstructured":"Bano S, Vasconcelos F, Shepherd LM, Vander Poorten E, Vercauteren T, Ourselin S, David A L, Deprest J, Stoyanov D (2020) Deep placental vessel segmentation for fetoscopic mosaicking. In: International conference on medical image computing and computer-assisted intervention. Springer. arXiv:2007.04349","DOI":"10.1007\/978-3-030-59716-0_73"},{"issue":"5","key":"2242_CR4","doi-asserted-by":"publisher","first-page":"791","DOI":"10.1007\/s11548-020-02169-0","volume":"15","author":"S Bano","year":"2020","unstructured":"Bano S, Vasconcelos F, Vander Poorten E, Vercauteren T, Ourselin S, Deprest J, Stoyanov D (2020) FetNet: a recurrent convolutional network for occlusion identification in fetoscopic videos. Int J Comput Assist Radiol Surg 15(5):791\u2013801","journal-title":"Int J Comput Assist Radiol Surg"},{"issue":"2","key":"2242_CR5","first-page":"107","volume":"39","author":"A Baschat","year":"2011","unstructured":"Baschat A, Chmait RH, Deprest J, Gratac\u00f3s E, Hecher K, Kontopoulos E, Quintero R, Skupski DW, Valsky DV, Ville Y (2011) Twin-to-twin transfusion syndrome (TTTS). J Perinat Med 39(2):107\u2013112","journal-title":"J Perinat Med"},{"key":"2242_CR6","doi-asserted-by":"crossref","unstructured":"Bay H, Tuytelaars T, Van\u00a0Gool L (2006) SURF: Speeded up robust features. In: Proceedings of the European conference on computer vision, Springer, pp 404\u2013417","DOI":"10.1007\/11744023_32"},{"key":"2242_CR7","doi-asserted-by":"crossref","unstructured":"Blinn J (2002) Consider the lowly 2$$\\times $$2 matrix, chap 5. In: Jim Blinn\u2019s corner: notation, notation, notation. Morgan Kaufmann","DOI":"10.1016\/B978-155860860-3\/50006-6"},{"issue":"1","key":"2242_CR8","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1007\/s11263-006-0002-3","volume":"74","author":"M Brown","year":"2007","unstructured":"Brown M, Lowe DG (2007) Automatic panoramic image stitching using invariant features. Int J Comput Vis 74(1):59\u201373","journal-title":"Int J Comput Vis"},{"key":"2242_CR9","first-page":"1","volume":"128","author":"F Chadebecq","year":"2019","unstructured":"Chadebecq F, Vasconcelos F, Lacher R, Maneas E, Desjardins A, Ourselin S, Vercauteren T, Stoyanov D (2019) Refractive two-view reconstruction for underwater 3d vision. Int J Comput Vis 128:1\u201317","journal-title":"Int J Comput Vis"},{"key":"2242_CR10","doi-asserted-by":"crossref","unstructured":"Daga P, Chadebecq F, Shakir DI, Herrera LCGP, Tella M, Dwyer G, David AL, Deprest J, Stoyanov D, Vercauteren T (2016) Real-time mosaicing of fetoscopic videos using SIFT. In: Medical Imaging: Image-Guided Procedures","DOI":"10.1117\/12.2217172"},{"issue":"5","key":"2242_CR11","doi-asserted-by":"publisher","first-page":"347","DOI":"10.1046\/j.1469-0705.1998.11050347.x","volume":"11","author":"J Deprest","year":"1998","unstructured":"Deprest J, Van Schoubroeck D, Van Ballaer P, Flageole H, Van Assche FA, Vandenberghe K (1998) Alternative technique for Nd: Yag laser coagulation in twin-to-twin transfusion syndrome with anterior placenta. Ultrasound Obstet Gynecol J 11(5):347\u2013352","journal-title":"Ultrasound Obstet Gynecol J"},{"key":"2242_CR12","unstructured":"DeTone D, Malisiewicz T, Rabinovich A (2016) Deep image homography estimation. RSS Workshop on Limits and Potentials of Deep Learning in Robotics"},{"key":"2242_CR13","doi-asserted-by":"crossref","unstructured":"Dusmanu M, Rocco I, Pajdla T, Pollefeys M, Sivic J, Torii A, Sattler T (2019) D2-Net: A trainable CNN for joint description and detection of local features. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8092\u20138101","DOI":"10.1109\/CVPR.2019.00828"},{"issue":"3","key":"2242_CR14","doi-asserted-by":"publisher","first-page":"1656","DOI":"10.1109\/LRA.2017.2679902","volume":"2","author":"G Dwyer","year":"2017","unstructured":"Dwyer G, Chadebecq F, Amo MT, Bergeles C, Maneas E, Pawar V, Vander Poorten E, Deprest J, Ourselin S, De Coppi P (2017) A continuum robot and control interface for surgical assist in fetoscopic interventions. IEEE Robot Autom Lett 2(3):1656\u20131663","journal-title":"IEEE Robot Autom Lett"},{"issue":"1","key":"2242_CR15","doi-asserted-by":"publisher","first-page":"24","DOI":"10.3390\/jimaging4010024","volume":"4","author":"F Gaisser","year":"2018","unstructured":"Gaisser F, Peeters SH, Lenseigne BA, Jonker PP, Oepkes D (2018) Stable image registration for in-vivo fetoscopic panorama reconstruction. J Imaging 4(1):24","journal-title":"J Imaging"},{"key":"2242_CR16","volume-title":"Multiple view geometry in computer vision, Chapter 4: Estimation - 2D Projective Transformations","author":"R Hartley","year":"2003","unstructured":"Hartley R, Zisserman A (2003) Multiple view geometry in computer vision, Chapter 4: Estimation - 2D Projective Transformations. Cambridge University Press, Cambridge"},{"issue":"12","key":"2242_CR17","doi-asserted-by":"publisher","first-page":"1949","DOI":"10.1007\/s11548-018-1822-7","volume":"13","author":"A Javaux","year":"2018","unstructured":"Javaux A, Bouget D, Gruijthuijsen C, Stoyanov D, Vercauteren T, Ourselin S, Deprest J, Denis K, Vander Poorten E (2018) A mixed-reality surgical trainer with comprehensive sensing for fetal laser minimally invasive surgery. Int J Comput Assist Radiol Surg 13(12):1949\u20131957","journal-title":"Int J Comput Assist Radiol Surg"},{"issue":"1","key":"2242_CR18","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1016\/j.ajog.2012.09.025","volume":"208","author":"L Lewi","year":"2013","unstructured":"Lewi L, Deprest J, Hecher K (2013) The vascular anastomoses in monochorionic twin pregnancies and their clinical consequences. Am J Obstet Gynecol 208(1):19\u201330","journal-title":"Am J Obstet Gynecol"},{"key":"2242_CR19","unstructured":"Malis E, Vargas M (2007) Deeper understanding of the homography decomposition for vision-based control. Ph.D. Thesis, INRIA"},{"issue":"3","key":"2242_CR20","doi-asserted-by":"publisher","first-page":"2346","DOI":"10.1109\/LRA.2018.2809549","volume":"3","author":"T Nguyen","year":"2018","unstructured":"Nguyen T, Chen SW, Shivakumar SS, Taylor CJ, Kumar V (2018) Unsupervised deep homography: a fast and robust homography estimation model. IEEE Robot Autom Lett 3(3):2346\u20132353","journal-title":"IEEE Robot Autom Lett"},{"issue":"5","key":"2242_CR21","doi-asserted-by":"publisher","first-page":"713","DOI":"10.1007\/s11548-018-1728-4","volume":"13","author":"L Peter","year":"2018","unstructured":"Peter L, Tella-Amo M, Shakir DI, Attilakos G, Wimalasundera R, Deprest J, Ourselin S, Vercauteren T (2018) Retrieval and registration of long-range overlapping frames for scalable mosaicking of in vivo fetoscopy. Int J Comput Assist Radiol Surg 13(5):713\u2013720","journal-title":"Int J Comput Assist Radiol Surg"},{"key":"2242_CR22","unstructured":"Reeff M, Gerhard F, Cattin P, G\u00e1bor S (2006) Mosaicing of endoscopic placenta images. INFORMATIK 2006\u2013Informatik f\u00fcr Menschen, Band 1"},{"key":"2242_CR23","doi-asserted-by":"crossref","unstructured":"Shen X, Darmon F, Efros AA, Aubry M (2020) RANSAC-Flow: generic two-stage image alignment. arXiv preprint arXiv:2004.01526","DOI":"10.1007\/978-3-030-58548-8_36"},{"key":"2242_CR24","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: International conference on learning representations"},{"key":"2242_CR25","doi-asserted-by":"crossref","unstructured":"Srivastava S, Chopra A, Kumar AC, Bhandarkar S, Sharma D (2019) Matching disparate image pairs using shape-aware convnets. In: IEEE winter conference on applications of computer vision, IEEE, pp 531\u2013540","DOI":"10.1109\/WACV.2019.00062"},{"issue":"2","key":"2242_CR26","doi-asserted-by":"publisher","first-page":"021217","DOI":"10.1117\/1.JMI.5.2.021217","volume":"5","author":"M Tella-Amo","year":"2018","unstructured":"Tella-Amo M, Peter L, Shakir DI, Deprest J, Stoyanov D, Iglesias JE, Vercauteren T, Ourselin S (2018) Probabilistic visual and electromagnetic data fusion for robust drift-free sequential mosaicking: application to fetoscopy. J Med Imaging 5(2):021217","journal-title":"J Med Imaging"},{"issue":"3","key":"2242_CR27","doi-asserted-by":"publisher","first-page":"035001","DOI":"10.1117\/1.JMI.6.3.035001","volume":"6","author":"M Tella-Amo","year":"2019","unstructured":"Tella-Amo M, Peter L, Shakir DI, Deprest J, Stoyanov D, Vercauteren T, Ourselin S (2019) Pruning strategies for efficient online globally-consistent mosaicking in fetoscopy. J Med Imaging 6(3):035001","journal-title":"J Med Imaging"}],"container-title":["International Journal of Computer Assisted Radiology and Surgery"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-020-02242-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11548-020-02242-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-020-02242-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,8,16]],"date-time":"2021-08-16T23:34:23Z","timestamp":1629156863000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11548-020-02242-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,8,17]]},"references-count":27,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2020,11]]}},"alternative-id":["2242"],"URL":"https:\/\/doi.org\/10.1007\/s11548-020-02242-8","relation":{},"ISSN":["1861-6410","1861-6429"],"issn-type":[{"value":"1861-6410","type":"print"},{"value":"1861-6429","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,8,17]]},"assertion":[{"value":"13 April 2020","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 July 2020","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 August 2020","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Compliance with ethical standards"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"For this type of study, formal consent is not required.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"No animals or humans were involved in this research. All videos were anonymized before delivery to the researchers.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}}]}}