{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T01:56:28Z","timestamp":1772848588991,"version":"3.50.1"},"publisher-location":"Cham","reference-count":24,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032051264","type":"print"},{"value":"9783032051271","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,9,20]],"date-time":"2025-09-20T00:00:00Z","timestamp":1758326400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,20]],"date-time":"2025-09-20T00:00:00Z","timestamp":1758326400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-05127-1_30","type":"book-chapter","created":{"date-parts":[[2025,9,19]],"date-time":"2025-09-19T21:16:11Z","timestamp":1758316571000},"page":"308-317","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["LiteTracker: Leveraging Temporal Causality for\u00a0Accurate Low-Latency Tissue Tracking"],"prefix":"10.1007","author":[{"given":"Mert Asim","family":"Karaoglu","sequence":"first","affiliation":[]},{"given":"Wenbo","family":"Ji","sequence":"additional","affiliation":[]},{"given":"Ahmed","family":"Abbas","sequence":"additional","affiliation":[]},{"given":"Nassir","family":"Navab","sequence":"additional","affiliation":[]},{"given":"Benjamin","family":"Busam","sequence":"additional","affiliation":[]},{"given":"Alexander","family":"Ladikos","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,20]]},"reference":[{"key":"30_CR1","unstructured":"Aydemir, G., Cai, X., Xie, W., G\u00fcney, F.: Track-On: transformer-based online point tracking with memory. arXiv preprint arXiv:2501.18487 (2025)"},{"key":"30_CR2","doi-asserted-by":"crossref","unstructured":"Chen, Y., Wu, Z., Schmidt, A., Salcudean, S.E.: A-MFST: adaptive multi-flow sparse tracker for real-time tissue tracking under occlusion. arXiv preprint arXiv:2410.19996 (2024)","DOI":"10.1007\/s11548-025-03414-0"},{"key":"30_CR3","doi-asserted-by":"crossref","unstructured":"Cho, S., Huang, J., Nam, J., An, H., Kim, S., Lee, J.Y.: Local all-pair correspondence for point tracking. In: European Conference on Computer Vision, pp. 306\u2013325. Springer (2024)","DOI":"10.1007\/978-3-031-72684-2_18"},{"key":"30_CR4","first-page":"13610","volume":"35","author":"C Doersch","year":"2022","unstructured":"Doersch, C., et al.: TAP-VID: a benchmark for tracking any point in a video. Adv. Neural. Inf. Process. Syst. 35, 13610\u201313626 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"30_CR5","doi-asserted-by":"crossref","unstructured":"Doersch, C., et al.: Bootstrapped training for tracking-any-point. In: Proceedings of the Asian Conference on Computer Vision, pp. 3257\u20133274 (2024)","DOI":"10.1007\/978-981-96-0901-7_28"},{"key":"30_CR6","doi-asserted-by":"crossref","unstructured":"Harley, A.W., Fang, Z., Fragkiadaki, K.: Particle Video Revisited: tracking through occlusions using point trajectories. In: European Conference on Computer Vision, pp. 59\u201375. Springer (2022)","DOI":"10.1007\/978-3-031-20047-2_4"},{"key":"30_CR7","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11548-023-02919-w","volume":"18","author":"M Hayoz","year":"2023","unstructured":"Hayoz, M., et al.: Learning how to robustly estimate camera pose in endoscopic videos. Int. J. Comput. Assist. Radiol. Surg. 18, 1\u20138 (2023). https:\/\/doi.org\/10.1007\/s11548-023-02919-w","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"30_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"54","DOI":"10.1007\/978-3-030-59716-0_6","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2020","author":"S Ihler","year":"2020","unstructured":"Ihler, S., Kuhnke, F., Laves, M.-H., Ortmaier, T.: Self-supervised domain adaptation for patient-specific, real-time tissue tracking. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12263, pp. 54\u201364. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-59716-0_6"},{"key":"30_CR9","unstructured":"Karaev, N., Makarov, I., Wang, J., Neverova, N., Vedaldi, A., Rupprecht, C.: CoTracker3: simpler and better point tracking by pseudo-labelling real videos. arXiv preprint arXiv:2410.11831 (2024)"},{"key":"30_CR10","doi-asserted-by":"crossref","unstructured":"Karaoglu, M.A., Markova, V., Navab, N., Busam, B., Ladikos, A.: RIDE: self-supervised learning of rotation-equivariant keypoint detection and invariant description for endoscopy. In: 2024 IEEE International Conference on Robotics and Automation (ICRA), pp. 10764\u201310771. IEEE (2024)","DOI":"10.1109\/ICRA57147.2024.10611381"},{"issue":"2","key":"30_CR11","doi-asserted-by":"publisher","first-page":"2294","DOI":"10.1109\/LRA.2020.2970659","volume":"5","author":"Y Li","year":"2020","unstructured":"Li, Y., et al.: Super: a surgical perception framework for endoscopic tissue manipulation with surgical robotics. IEEE Rob. Autom. Lett. 5(2), 2294\u20132301 (2020)","journal-title":"IEEE Rob. Autom. Lett."},{"key":"30_CR12","doi-asserted-by":"crossref","unstructured":"Liang, X., Liu, F., Zhang, Y., Li, Y., Lin, S., Yip, M.: Real-to-sim deformable object manipulation: optimizing physics models with residual mappings for robotic surgery. In: 2024 IEEE International Conference on Robotics and Automation (ICRA), pp. 15471\u201315477. IEEE (2024)","DOI":"10.1109\/ICRA57147.2024.10610263"},{"key":"30_CR13","doi-asserted-by":"crossref","unstructured":"Neoral, M., \u0160er\u1ef3ch, J., Matas, J.: MFT: long-term tracking of every pixel. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 6837\u20136847 (2024)","DOI":"10.1109\/WACV57701.2024.00669"},{"key":"30_CR14","unstructured":"Paszke, A., et\u00a0al.: PyTorch: an imperative style, high-performance deep learning library. In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"30_CR15","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1007\/s11263-008-0136-6","volume":"80","author":"P Sand","year":"2008","unstructured":"Sand, P., Teller, S.: Particle video: long-range motion estimation using point trajectories. Int. J. Comput. Vis. 80, 72\u201391 (2008)","journal-title":"Int. J. Comput. Vis."},{"key":"30_CR16","unstructured":"Schmidt, A., et\u00a0al.: Point tracking in surgery\u2013the 2024 surgical tattoos in infrared (STIR) challenge. arXiv preprint arXiv:2503.24306 (2025)"},{"key":"30_CR17","doi-asserted-by":"crossref","unstructured":"Schmidt, A., Mohareri, O., DiMaio, S., Salcudean, S.E.: Fast graph refinement and implicit neural representation for tissue tracking. In: 2022 International Conference on Robotics and Automation (ICRA), pp. 1281\u20131288. IEEE (2022)","DOI":"10.1109\/ICRA46639.2022.9811742"},{"key":"30_CR18","doi-asserted-by":"crossref","unstructured":"Schmidt, A., Mohareri, O., DiMaio, S., Salcudean, S.E.: SENDD: sparse efficient neural depth and deformation for tissue tracking. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 238\u2013248. Springer (2023)","DOI":"10.1007\/978-3-031-43996-4_23"},{"key":"30_CR19","doi-asserted-by":"crossref","unstructured":"Schmidt, A., Mohareri, O., DiMaio, S., Salcudean, S.E.: Surgical tattoos in infrared: a dataset for quantifying tissue tracking and mapping. IEEE Trans. Med. Imaging (2024)","DOI":"10.1109\/TMI.2024.3372828"},{"key":"30_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103131","volume":"94","author":"A Schmidt","year":"2024","unstructured":"Schmidt, A., Mohareri, O., DiMaio, S., Yip, M.C., Salcudean, S.E.: Tracking and mapping in medical computer vision: a review. Med. Image Anal. 94, 103131 (2024)","journal-title":"Med. Image Anal."},{"key":"30_CR21","doi-asserted-by":"crossref","unstructured":"Serych, J., Neoral, M., Matas, J.: MFTIQ: multi-flow tracker with independent matching quality estimation. arXiv preprint arXiv:2411.09551 (2024)","DOI":"10.1109\/WACV61041.2025.00784"},{"key":"30_CR22","doi-asserted-by":"crossref","unstructured":"Shinde, N.U., et al.: Jiggle: an active sensing framework for boundary parameters estimation in deformable surgical environments. arXiv preprint arXiv:2405.09743 (2024)","DOI":"10.15607\/RSS.2024.XX.007"},{"key":"30_CR23","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11548-024-03171-6","volume":"19","author":"T Teufel","year":"2024","unstructured":"Teufel, T., et al.: OneSLAM to map them all: a generalized approach to SLAM for monocular endoscopic imaging based on tracking any point. Int. J. Comput. Assist. Radiol. Surg. 19, 1\u20138 (2024)","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"30_CR24","doi-asserted-by":"crossref","unstructured":"Wang, Y., Lipson, L., Deng, J.: SEA-RAFT: simple, efficient, accurate raft for optical flow. In: European Conference on Computer Vision, pp. 36\u201354. Springer (2024)","DOI":"10.1007\/978-3-031-72667-5_3"}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2025"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-05127-1_30","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,19]],"date-time":"2025-09-19T21:16:17Z","timestamp":1758316577000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-05127-1_30"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,20]]},"ISBN":["9783032051264","9783032051271"],"references-count":24,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-05127-1_30","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,9,20]]},"assertion":[{"value":"20 September 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Daejeon","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Korea (Republic of)","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 September 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 September 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/conferences.miccai.org\/2025\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}