{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,19]],"date-time":"2025-11-19T17:19:37Z","timestamp":1763572777009,"version":"3.37.3"},"reference-count":46,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2024,5,27]],"date-time":"2024-05-27T00:00:00Z","timestamp":1716768000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,5,27]],"date-time":"2024-05-27T00:00:00Z","timestamp":1716768000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100010426","name":"UGC-DAE\u00a0Consortium for Scientific Research, University Grants Commission","doi-asserted-by":"publisher","award":["F.\/2017-18\/NFO-2017-18-OBC-KER-60500"],"award-info":[{"award-number":["F.\/2017-18\/NFO-2017-18-OBC-KER-60500"]}],"id":[{"id":"10.13039\/501100010426","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Intell Robot Syst"],"abstract":"<jats:title>Abstract<\/jats:title><jats:p>The surge of haptic technology has greatly impacted Robotic-assisted surgery in recent years due to its inspirational advancement in the field. Delivering tactile feedback to the surgeon has a significant role in improving the user experience in RAMIS. This work proposes a Modified inception ResNet network along with dimensionality reduction to regenerate the variable force produced during the surgical intervention. This work collects the relevant dataset from two ex vivo porcine skins and one ex vivo artificial skin for the validation of the results. The proposed framework is used to model both spatial and temporal data collected from the sensors, tissue, manipulators, and surgical tools. The evaluations are based on three distinct datasets with modest variations in tissue properties. The results of the proposed framework show an improvement of force prediction accuracy by 10.81% over RNN, 6.02% over RNN\u2009+\u2009LSTM, and 3.81% over the CNN\u2009+\u2009LSTM framework, and torque prediction accuracy by 12.41% over RNN, 5.75% over RNN\u2009+\u2009LSTM, and 3.75% over CNN\u2009+\u2009LSTM. The sensitivity study demonstrates that features such as torque (96.93%), deformation (94.02%), position (93.98%), vision (92.12%), stiffness (87.95%), tool diameter (89.24%), rotation (65.10%), and orientation (62.51%) have respective influences on the anticipated force. It was observed that the quality of the predicted force improved by 2.18% when performing feature selection and dimensionality reduction on features collected from tool, manipulator, tissue, and vision data and processing them simultaneously in all four architectures. The method has potential applications for online surgical tasks and surgeon training.<\/jats:p>","DOI":"10.1007\/s10846-024-02100-8","type":"journal-article","created":{"date-parts":[[2024,5,27]],"date-time":"2024-05-27T14:04:26Z","timestamp":1716818666000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["A Stereovision-based Approach for Retrieving Variable Force Feedback in Robotic-Assisted Surgery Using Modified Inception ResNet V2 Networks"],"prefix":"10.1007","volume":"110","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0952-5394","authenticated-orcid":false,"given":"P. V.","family":"Sabique","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5662-3694","authenticated-orcid":false,"given":"Ganesh","family":"Pasupathy","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0638-0302","authenticated-orcid":false,"given":"S.","family":"Kalaimagal","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5608-7737","authenticated-orcid":false,"given":"G.","family":"Shanmugasundar","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5517-2462","authenticated-orcid":false,"given":"V. K.","family":"Muneer","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,27]]},"reference":[{"key":"2100_CR1","doi-asserted-by":"crossref","unstructured":"Kroh M., Chalikonda, S.: Essentials of robotic surgery. (2015)","DOI":"10.1007\/978-3-319-09564-6"},{"key":"2100_CR2","doi-asserted-by":"crossref","unstructured":"Spinoglio, G., Marano, A., Formisano, G.: Robotic surgery: current applications and new trends. (2015)","DOI":"10.1007\/978-88-470-5714-2_1"},{"key":"2100_CR3","doi-asserted-by":"publisher","first-page":"88","DOI":"10.1109\/M-RA.2007.907921","volume":"14","author":"V Hayward","year":"2007","unstructured":"Hayward, V., MacLean, K.E.: Do it yourself haptics: part I. Robot. Autom. Mag. IEEE 14, 88\u2013104 (2007)","journal-title":"Robot. Autom. Mag. IEEE"},{"key":"2100_CR4","doi-asserted-by":"publisher","first-page":"1180","DOI":"10.1007\/s00464-008-0298-x","volume":"23","author":"OAJ van der Meijden","year":"2009","unstructured":"van der Meijden, O.A.J., Schijven, M.P.: The value of haptic feedback in conventional and robot-assisted minimal invasive surgery and virtual reality training: a current review. Surg. Endosc. 23, 1180\u20131190 (2009)","journal-title":"Surg. Endosc."},{"key":"2100_CR5","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-25457-9","volume-title":"Cutaneous Haptic Feedback in Robotic Teleoperation","author":"C Pacchierotti","year":"2015","unstructured":"Pacchierotti, C.: Cutaneous Haptic Feedback in Robotic Teleoperation. Springer, Berlin, Germany (2015)"},{"key":"2100_CR6","doi-asserted-by":"publisher","first-page":"401","DOI":"10.3182\/20090812-3-DK-2006.0035","volume":"42","author":"T Haidegger","year":"2009","unstructured":"Haidegger, T., Benyo, B., Kovacs, L., Benyo, Z.: Force Sensing and Force Control for Surgical Robots. IFAC Proc. Vol. 42, 401\u2013406 (2009)","journal-title":"IFAC Proc. Vol."},{"key":"2100_CR7","doi-asserted-by":"publisher","first-page":"2160","DOI":"10.1109\/LRA.2018.2810948","volume":"3","author":"N Haouchine","year":"2018","unstructured":"Haouchine, N., Kuang, W., Cotin, S., Yip, M.: Vision-Based Force Feedback Estimation for Robot-Assisted Surgery Using Instrument-Constrained Biomechanical Three-Dimensional Maps. IEEE Robot. Autom. Lett. 3, 2160\u20132165 (2018)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"2100_CR8","doi-asserted-by":"crossref","unstructured":"Gessert N., Beringhoff J., Otte C., Schlaefer A.: Force estimation from OCT volumes using 3D CNNs. Int. J. Comp. Assisted Radiol. Surg. 13, 1073\u20131082,\u00a02018\/07\/01\u00a0(2018)","DOI":"10.1007\/s11548-018-1777-8"},{"key":"2100_CR9","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1097\/MOU.0b013e32831a478c","volume":"19","author":"AM Okamura","year":"2009","unstructured":"Okamura, A.M.: Haptic feedback in robot-assisted minimally invasive surgery. Curr. Opin. Urol. 19, 102\u2013107 (2009)","journal-title":"Curr. Opin. Urol."},{"key":"2100_CR10","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1142\/S0219843615500139","volume":"12","author":"SM Yoon","year":"2015","unstructured":"Yoon, S.M., Lee, M.-C., Kim, C.Y.: Sliding Perturbation Observer Based Reaction Force Estimation Method of Surgical Robot Instrument for Haptic Realization. Int. J. Humanoid Robotics 12, 13\u201319 (2015)","journal-title":"Int. J. Humanoid Robotics"},{"key":"2100_CR11","doi-asserted-by":"crossref","unstructured":"Li, Y., Miyasaka, M., Haghighipanah, M., Lei, C., Hannaford B.: Dynamic modeling of cable driven elongated surgical instruments for sensorless grip force estimation. In: 2016 IEEE International Conference on Robotics and Automation (ICRA), 4128\u20134134 (2016)","DOI":"10.1109\/ICRA.2016.7487605"},{"key":"2100_CR12","first-page":"130","volume":"21","author":"D Lee","year":"2016","unstructured":"Lee, D., Kim, U., Gulrez, T., Yoon, W.J., Hannaford, B., Choi, H.R.: A Laparoscopic Grasping Tool With Force Sensing Capability. IEEE\/ASME Trans. Mechatron. 21, 130\u2013141 (2016)","journal-title":"IEEE\/ASME Trans. Mechatron."},{"key":"2100_CR13","doi-asserted-by":"crossref","unstructured":"Aviles, A. I., Marban, A., Sobrevilla, P., Fernandez, J., Casals. A.: A recurrent neural network approach for 3D vision-based force estimation. In: 2014 4th International Conference on Image Processing Theory, Tools and Applications (IPTA), pp. 1\u20136 (2014)","DOI":"10.1109\/IPTA.2014.7001941"},{"key":"2100_CR14","doi-asserted-by":"crossref","unstructured":"Aviles, A. I., Alsaleh, S., Sobrevilla, P., Casals, A.: Sensorless force estimation using a neuro-vision-based approach for robotic-assisted surgery. In: 2015 7th International IEEE\/EMBS Conference on Neural Engineering (NER), pp. 86\u201389 (2015)","DOI":"10.1109\/NER.2015.7146566"},{"key":"2100_CR15","doi-asserted-by":"crossref","unstructured":"Aviles, A. I., Alsaleh, S. M., Sobrevilla, P., Casals, A.: Force-feedback sensory substitution using supervised recurrent learning for robotic-assisted surgery. In: 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 1\u20134 (2015)","DOI":"10.1109\/EMBC.2015.7318246"},{"key":"2100_CR16","doi-asserted-by":"publisher","first-page":"431","DOI":"10.1109\/TOH.2016.2640289","volume":"10","author":"AI Aviles","year":"2017","unstructured":"Aviles, A.I., Alsaleh, S.M., Hahn, J.K., Casals, A.: Towards Retrieving Force Feedback in Robotic-Assisted Surgery: A Supervised Neuro-Recurrent-Vision Approach. IEEE Trans. Haptics 10, 431\u2013443 (2017)","journal-title":"IEEE Trans. Haptics"},{"key":"2100_CR17","doi-asserted-by":"crossref","unstructured":"Aviles, A. I., Alsaleh, S. M., Montseny, E., Sobrevilla, P., Casals, A.: A Deep-Neuro-Fuzzy approach for estimating the interaction forces in Robotic surgery. In: 2016 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), pp. 1113\u20131119 (2016)","DOI":"10.1109\/FUZZ-IEEE.2016.7737812"},{"key":"2100_CR18","doi-asserted-by":"publisher","first-page":"134","DOI":"10.1016\/j.bspc.2019.01.011","volume":"50","author":"A Marban","year":"2019","unstructured":"Marban, A., Srinivasan, V., Samek, W., Fern\u00e1ndez, J., Casals, A.: A recurrent convolutional neural network approach for sensorless force estimation in robotic surgery. Biomed. Signal Process. Control. 50, 134\u2013150 (2019)","journal-title":"Biomed. Signal Process. Control."},{"key":"2100_CR19","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun J.: Deep residual learning for image recognition. In: 2016 IEEE Conf. Comput. Vis. Pattern Recognit.\u00a0(CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"2100_CR20","doi-asserted-by":"crossref","unstructured":"Greminger, M. A., Nelson, B. J.: Modeling elastic objects with neural networks for vision-based force measurement.\u00a0In: Proceedings 2003 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453), pp. 1278\u20131283 (2003)","DOI":"10.1109\/IROS.2003.1248821"},{"key":"2100_CR21","unstructured":"Angelica, I. A., Samar, M. A., Eduard, M., Alicia, C.: V-ANFIS for Dealing with visual uncertainty for force estimation in robotic surgery. In: Proceedings of the 2015 Conference of the International Fuzzy Systems Association and the European Society for Fuzzy Logic and Technology, pp. 1465\u20131472 (2015)"},{"key":"2100_CR22","doi-asserted-by":"crossref","unstructured":"Marban, A., Srinivasan, V., Samek, W., Fern\u00e1ndez, J., Casals, A.: Estimation of interaction forces in robotic surgery using a semi-supervised deep neural network model. In: 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 761\u2013768 (2018)","DOI":"10.1109\/IROS.2018.8593701"},{"key":"2100_CR23","doi-asserted-by":"publisher","first-page":"143","DOI":"10.3233\/JIFS-211835","volume":"43","author":"SD Jeyabalan","year":"2022","unstructured":"Jeyabalan, S.D., Yesudhas, N.J., Harichandran, K.N., Sridharan, G.: Multivariate temporal data classification framework for ozone level prediction. Journal of Intelligent & Fuzzy Systems 43, 143\u2013157 (2022)","journal-title":"Journal of Intelligent & Fuzzy Systems"},{"key":"2100_CR24","doi-asserted-by":"publisher","first-page":"101730","DOI":"10.1016\/j.media.2020.101730","volume":"64","author":"N Gessert","year":"2020","unstructured":"Gessert, N., Bengs, M., Schluter, M., Schlaefer, A.: Deep learning with 4D spatio-temporal data representations for OCT-based force estimation. Med. Image Anal. 64, 101730 (2020)","journal-title":"Med. Image Anal."},{"key":"2100_CR25","doi-asserted-by":"crossref","unstructured":"Gao, C., Liu, X., Peven, M., Unberath, M., Reiter, A.: Learning to see forces: surgical force prediction with RGB-Point cloud temporal convolutional networks. In: OR 2.0 Context-Aware Operating Theaters, Computer Assisted Robotic Endoscopy, Clinical Image-Based Procedures, and Skin Image Analysis, Cham, pp. 118\u2013127 (2018)","DOI":"10.1007\/978-3-030-01201-4_14"},{"key":"2100_CR26","doi-asserted-by":"publisher","first-page":"1601","DOI":"10.1007\/s11548-019-02048-3","volume":"14","author":"A Mendizabal","year":"2019","unstructured":"Mendizabal, A., Sznitman, R., Cotin, S.: Force classification during robotic interventions through simulation-trained neural networks. Int. J. Comp. Assisted Radiol. Surg. 14, 1601\u20131610 (2019)","journal-title":"Int. J. Comp. Assisted Radiol. Surg."},{"key":"2100_CR27","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3389\/frobt.2019.00056","volume":"6","author":"S Abeywardena","year":"2019","unstructured":"Abeywardena, S., Yuan, Q., Tzemanaki, A., Psomopoulou, E., Droukas, L., Melhuish, C., et al.: Estimation of Tool-Tissue Forces in Robot-Assisted Minimally Invasive Surgery Using Neural Networks. Front. Robot. AI 6, 1\u201310 (2019)","journal-title":"Front. Robot. AI"},{"key":"2100_CR28","first-page":"414","volume":"9","author":"PJE Edwards","year":"2021","unstructured":"Edwards, P.J.E., Colleoni, E., Sridhar, A., Kelly, J.D., Stoyanov, D.: Visual kinematic force estimation in robot-assisted surgery \u2013 application to knot tying. Comput. methods Biomech. Biomed. Eng. Imaging Vis. 9, 414\u2013420 (2021)","journal-title":"Biomed. Eng. Imaging Vis."},{"key":"2100_CR29","doi-asserted-by":"publisher","first-page":"110","DOI":"10.3390\/s21010110","volume":"21","author":"W-J Jung","year":"2021","unstructured":"Jung, W.-J., Kwak, K.-S., Lim, S.-C.: Vision-Based Suture Tensile Force Estimation in Robotic Surgery. Sensors 21, 110 (2021)","journal-title":"Sensors"},{"key":"2100_CR30","doi-asserted-by":"publisher","first-page":"14648","DOI":"10.1007\/s11227-022-04432-4","volume":"78","author":"PV Sabique","year":"2022","unstructured":"Sabique, P.V., Ganesh, P., Sivaramakrishnan, R.: Stereovision based force estimation with stiffness mapping in surgical tool insertion using recurrent neural network. J. Supercomput. 78, 14648\u201314679 (2022)","journal-title":"J. Supercomput."},{"key":"2100_CR31","doi-asserted-by":"crossref","unstructured":"Chua, Z., Jarc, A. M., Okamura, A. M.: Toward force estimation in robot-assisted surgery using deep learning with vision and robot state. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 12335\u201312341 (2021)","DOI":"10.1109\/ICRA48506.2021.9560945"},{"key":"2100_CR32","doi-asserted-by":"crossref","unstructured":"Ko, D.-K., Lee, K.-W., Lee, D. H., Lim, S.-C.: Vision-based interaction force estimation for robot grip motion without tactile\/force sensor.\u00a0Expert Syst. with Appl.\u00a0vol. 211, p. 118441 (2022)","DOI":"10.1016\/j.eswa.2022.118441"},{"key":"2100_CR33","doi-asserted-by":"publisher","first-page":"290","DOI":"10.1109\/TPAMI.2004.1262305","volume":"26","author":"MA Greminger","year":"2004","unstructured":"Greminger, M.A., Nelson, B.J.: Vision-based force measurement. IEEE Trans. Pattern Anal. Mach. Intell. 26, 290\u2013298 (2004)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2100_CR34","doi-asserted-by":"publisher","first-page":"636","DOI":"10.1109\/TMECH.2009.2031240","volume":"15","author":"J Kim","year":"2010","unstructured":"Kim, J., Janabi-Sharifi, F., Kim, J.: A Haptic Interaction Method Using Visual Information and Physically Based Modeling. IEEE\/ASME Trans. Mechatron. 15, 636\u2013645 (2010)","journal-title":"IEEE\/ASME Trans. Mechatron."},{"key":"2100_CR35","doi-asserted-by":"crossref","unstructured":"Noohi, E., Parastegari, S., \u017defran, M.: Using monocular images to estimate interaction forces during minimally invasive surgery. In: 2014 IEEE\/RSJ International Conference on Intelligent Robots and Systems, pp. 4297\u20134302 (2014)","DOI":"10.1109\/IROS.2014.6943169"},{"key":"2100_CR36","doi-asserted-by":"crossref","unstructured":"Sabique, P. V., Pasupathy, G., Ramachandran, D., Shanmugasundar, G.: Investigating the influence of dimensionality reduction on force estimation in robotic-assisted surgery using recurrent and convolutional networks.\u00a0Eng. Appl. Art. Intel. 126, p. 107045 (2023)","DOI":"10.1016\/j.engappai.2023.107045"},{"key":"2100_CR37","doi-asserted-by":"crossref","unstructured":"Sabique, P.V., Pasupathy, G., Ramachandran, S.: A data driven recurrent neural network approach for reproduction of variable visuo-haptic force feedback in surgical tool insertion. Expert Sys. Appl. 238, p. 122221\u00a02024\/03\/15\/ (2024)","DOI":"10.1016\/j.eswa.2023.122221"},{"key":"2100_CR38","doi-asserted-by":"publisher","first-page":"143","DOI":"10.1109\/RBME.2017.2773521","volume":"11","author":"J Zhang","year":"2018","unstructured":"Zhang, J., Zhong, Y., Gu, C.: Deformable Models for Surgical Simulation: A Survey. IEEE Rev. Biomed. Eng. 11, 143\u2013164 (2018)","journal-title":"IEEE Rev. Biomed. Eng."},{"key":"2100_CR39","first-page":"49","volume":"5","author":"A Campeau-Lecours","year":"2017","unstructured":"Campeau-Lecours, A., Lamontagne, H., Latour, S., Fauteux, P., Maheu, V., Boucher, F., et al.: Kinova Modular Robot Arms for Service Robotics Applications. Int. J. Robot. Appl. Technol. 5, 49\u201371 (2017)","journal-title":"Int. J. Robot. Appl. Technol."},{"key":"2100_CR40","doi-asserted-by":"publisher","first-page":"452","DOI":"10.1016\/j.procs.2017.01.159","volume":"104","author":"O Krutikova","year":"2017","unstructured":"Krutikova, O., Sisojevs, A., Kovalovs, M.: Creation of a Depth Map from Stereo Images of Faces for 3D Model Reconstruction. Procedia Comput. Sci. 104, 452\u2013459 (2017)","journal-title":"Procedia Comput. Sci."},{"key":"2100_CR41","doi-asserted-by":"crossref","unstructured":"Pfister, T., Simonyan, K., Charles, J., Zisserman, A.: Deep convolutional neural networks for efficient pose estimation in gesture videos. (2014)","DOI":"10.1007\/978-3-319-16865-4_35"},{"key":"2100_CR42","doi-asserted-by":"publisher","first-page":"2385","DOI":"10.1162\/089976600300014980","volume":"12","author":"G Baudat","year":"2000","unstructured":"Baudat, G., Anouar, F.: Generalized discriminant analysis using a kernel approach. Neural Comput. 12, 2385\u20132404 (2000)","journal-title":"Neural Comput."},{"key":"2100_CR43","doi-asserted-by":"publisher","first-page":"785","DOI":"10.1016\/j.biomaterials.2005.06.019","volume":"27","author":"KS Wu","year":"2006","unstructured":"Wu, K.S., van Osdol, W.W., Dauskardt, R.H.: Mechanical properties of human stratum corneum: Effects of temperature, hydration, and chemical treatment. Biomaterials 27, 785\u2013795 (2006)","journal-title":"Biomaterials"},{"key":"2100_CR44","doi-asserted-by":"publisher","first-page":"2647","DOI":"10.1016\/j.bbamem.2007.05.028","volume":"1768","author":"CL Silva","year":"2007","unstructured":"Silva, C.L., Topgaard, D., Kocherbitov, V., Sousa, J.J.S., Pais, A.A.C.C., Sparr, E.: Stratum corneum hydration Phase transformations and mobility in stratum corneum, extracted lipids and isolated corneocytes. Biochim. Biophys. Acta (BBA) Biomembranes 1768, 2647\u20132659 (2007)","journal-title":"Biochim. Biophys. Acta (BBA) Biomembranes"},{"key":"2100_CR45","doi-asserted-by":"publisher","first-page":"6","DOI":"10.1016\/j.colsurfb.2005.12.013","volume":"48","author":"Y Yuan","year":"2006","unstructured":"Yuan, Y., Verma, R.: Measuring microelastic properties of stratum corneum. Colloids Surf B: Biointerfaces 48, 6\u201312 (2006)","journal-title":"Colloids Surf B: Biointerfaces"},{"key":"2100_CR46","unstructured":"Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., et al.: TensorFlow: large-scale machine learning on heterogeneous distributed systems. arXiv 1603.04467 (2015)"}],"container-title":["Journal of Intelligent &amp; Robotic Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10846-024-02100-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10846-024-02100-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10846-024-02100-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,20]],"date-time":"2024-11-20T03:06:42Z","timestamp":1732072002000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10846-024-02100-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,27]]},"references-count":46,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2024,6]]}},"alternative-id":["2100"],"URL":"https:\/\/doi.org\/10.1007\/s10846-024-02100-8","relation":{},"ISSN":["1573-0409"],"issn-type":[{"type":"electronic","value":"1573-0409"}],"subject":[],"published":{"date-parts":[[2024,5,27]]},"assertion":[{"value":"21 March 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 April 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 May 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declaration"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}},{"value":"All applicable international, national, and Anna University guidelines for the care and use of animals were followed, and all the experiments were conducted under the authority of strategies developed by the Animal Welfare Board of India. No formal consent is required for this type of study.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical Approval"}},{"value":"All individuals expressed their informed consent to participate in the study. No formal consent was necessary for this type of study.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to Participate"}},{"value":"All individuals expressed their consent to publish the work.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to Publish"}}],"article-number":"81"}}