{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,16]],"date-time":"2026-04-16T06:34:32Z","timestamp":1776321272329,"version":"3.50.1"},"publisher-location":"Cham","reference-count":22,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783031301100","type":"print"},{"value":"9783031301117","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-30111-7_13","type":"book-chapter","created":{"date-parts":[[2023,4,12]],"date-time":"2023-04-12T05:02:51Z","timestamp":1681275771000},"page":"145-157","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Transformer Based High-Frequency Predictive Model for\u00a0Visual-Haptic Feedback of\u00a0Virtual Surgery Navigation"],"prefix":"10.1007","author":[{"given":"Jianyong","family":"Huang","sequence":"first","affiliation":[]},{"given":"Jianhui","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Zhekang","family":"Qiu","sequence":"additional","affiliation":[]},{"given":"Zhiyong","family":"Yuan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,13]]},"reference":[{"key":"13_CR1","unstructured":"Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: ICLR (2015)"},{"key":"13_CR2","unstructured":"Bai, S., Kolter, J.Z., Koltun, V.: An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint arXiv:1803.01271 (2018)"},{"key":"13_CR3","unstructured":"Berkelman, P., Bozlee, S., Miyasaka, M.: Interactive rigid-body dynamics and deformable surface simulations with co-located maglev haptic and 3d graphic display (2013)"},{"key":"13_CR4","doi-asserted-by":"crossref","unstructured":"Berkelman, P., Miyasaka, M., Anderson, J.: Co-located 3d graphic and haptic display using electromagnetic levitation. In: 2012 IEEE Haptics Symposium (HAPTICS), pp. 77\u201381. IEEE (2012)","DOI":"10.1109\/HAPTIC.2012.6183773"},{"key":"13_CR5","doi-asserted-by":"crossref","unstructured":"Boulanger, P., Wu, G., Bischof, W., Yang, X.: Hapto-audio-visual environments for collaborative training of ophthalmic surgery over optical network. In: 2006 IEEE International Workshop on Haptic Audio Visual Environments and their Applications (HAVE 2006), pp. 21\u201326. IEEE (2006)","DOI":"10.1109\/HAVE.2006.283801"},{"key":"13_CR6","doi-asserted-by":"crossref","unstructured":"Clark, R., Wang, S., Wen, H., Markham, A., Trigoni, N.: Vinet: visual-inertial odometry as a sequence-to-sequence learning problem. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31 (2017)","DOI":"10.1609\/aaai.v31i1.11215"},{"key":"13_CR7","unstructured":"Hamza-Lup, F.G., Bogdan, C.M., Popovici, D.M., Costea, O.D.: A survey of visuo-haptic simulation in surgical training. arXiv preprint arXiv:1903.03272 (2019)"},{"key":"13_CR8","unstructured":"Kitaev, N., Kaiser, L., Levskaya, A.: Reformer: the efficient transformer. In: ICLR (2020). https:\/\/openreview.net\/forum?id=rkgNKkHtvB"},{"key":"13_CR9","doi-asserted-by":"crossref","unstructured":"Lai, G., Chang, W.C., Yang, Y., Liu, H.: Modeling long-and short-term temporal patterns with deep neural networks. In: SIGIR (2018)","DOI":"10.1145\/3209978.3210006"},{"issue":"3","key":"13_CR10","doi-asserted-by":"publisher","first-page":"252","DOI":"10.1016\/j.mechatronics.2014.02.001","volume":"24","author":"M Miyasaka","year":"2014","unstructured":"Miyasaka, M., Berkelman, P.: Magnetic levitation with unlimited omnidirectional rotation range. Mechatronics 24(3), 252\u2013264 (2014)","journal-title":"Mechatronics"},{"key":"13_CR11","unstructured":"Moody, L., Baber, C., Arvanitis, T.N.: The role of haptic feedback in the training and assessment of surgeons using a virtual environment. In: proceedings of Eurohaptics, pp. 170\u2013173 (2001)"},{"key":"13_CR12","doi-asserted-by":"crossref","unstructured":"Mourikis, A.I., Roumeliotis, S.I.: A multi-state constraint kalman filter for vision-aided inertial navigation. In: Proceedings 2007 IEEE International Conference on Robotics and Automation, pp. 3565\u20133572. IEEE (2007)","DOI":"10.1109\/ROBOT.2007.364024"},{"issue":"4","key":"13_CR13","doi-asserted-by":"publisher","first-page":"567","DOI":"10.1109\/TOH.2017.2679000","volume":"10","author":"SA Pedram","year":"2017","unstructured":"Pedram, S.A., Klatzky, R.L., Berkelman, P.: Torque contribution to haptic rendering of virtual textures. IEEE Trans. Haptics 10(4), 567\u2013579 (2017)","journal-title":"IEEE Trans. Haptics"},{"issue":"4","key":"13_CR14","doi-asserted-by":"publisher","first-page":"1004","DOI":"10.1109\/TRO.2018.2853729","volume":"34","author":"T Qin","year":"2018","unstructured":"Qin, T., Li, P., Shen, S.: Vins-mono: a robust and versatile monocular visual-inertial state estimator. IEEE Trans. Rob. 34(4), 1004\u20131020 (2018)","journal-title":"IEEE Trans. Rob."},{"key":"13_CR15","doi-asserted-by":"crossref","unstructured":"Satava, R.M., Jones, S.B.: Current and future applications of virtual reality for medicine. In: Proceedings of the IEEE, vol. 86, no. 3, pp. 484\u2013489 (1998)","DOI":"10.1109\/5.662873"},{"key":"13_CR16","doi-asserted-by":"crossref","unstructured":"Teixeira, B., Silva, H., Matos, A., Silva, E.: Deep learning approaches assessment for underwater scene understanding and egomotion estimation. In: OCEANS 2019 MTS\/IEEE SEATTLE, pp. 1\u20139. IEEE (2019)","DOI":"10.23919\/OCEANS40490.2019.8962872"},{"issue":"12","key":"13_CR17","doi-asserted-by":"publisher","first-page":"3123","DOI":"10.1109\/TVCG.2017.2772236","volume":"24","author":"Q Tong","year":"2017","unstructured":"Tong, Q., Yuan, Z., Liao, X., Zheng, M., Yuan, T., Zhao, J.: Magnetic levitation haptic augmentation for virtual tissue stiffness perception. IEEE Trans. Visual Comput. Graph. 24(12), 3123\u20133136 (2017)","journal-title":"IEEE Trans. Visual Comput. Graph."},{"key":"13_CR18","doi-asserted-by":"crossref","unstructured":"Tong, Q., Yuan, Z., Zheng, M., Zhu, W., Zhang, G., Liao, X.: A novel magnetic levitation haptic device for augmentation of tissue stiffness perception. In: Proceedings of the 22nd ACM Conference on Virtual Reality Software and Technology, pp. 143\u2013152 (2016)","DOI":"10.1145\/2993369.2993377"},{"key":"13_CR19","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017)"},{"key":"13_CR20","doi-asserted-by":"crossref","unstructured":"Viitanen, M., Vanne, J., H\u00e4m\u00e4l\u00e4inen, T.D., Kulmala, A.: Low latency edge rendering scheme for interactive 360 degree virtual reality gaming. In: 2018 IEEE 38th International Conference on Distributed Computing Systems (ICDCS), pp. 1557\u20131560. IEEE (2018)","DOI":"10.1109\/ICDCS.2018.00168"},{"key":"13_CR21","unstructured":"Wu, C., Wu, F., Qi, T., Huang, Y., Xie, X.: Fastformer: additive attention can be all you need. arXiv preprint arXiv:2108.09084 (2021)"},{"key":"13_CR22","doi-asserted-by":"crossref","unstructured":"Zhou, H., et al.: Informer: beyond efficient transformer for long sequence time-series forecasting. In: Proceedings of AAAI (2021)","DOI":"10.1609\/aaai.v35i12.17325"}],"container-title":["Lecture Notes in Computer Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-30111-7_13","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,12]],"date-time":"2023-04-12T05:06:10Z","timestamp":1681275970000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-30111-7_13"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031301100","9783031301117"],"references-count":22,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-30111-7_13","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"13 April 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Delhi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 November 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 November 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iconip2022.apnns.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easy Chair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"810","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"359","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"44% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.65","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ICONIP 2022 consists of a two-volume set, LNCS & CCIS, which includes 146 and 213 papers","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}