{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,19]],"date-time":"2025-05-19T15:46:00Z","timestamp":1747669560902,"version":"3.37.3"},"reference-count":66,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2022,4,2]],"date-time":"2022-04-02T00:00:00Z","timestamp":1648857600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,4,2]],"date-time":"2022-04-02T00:00:00Z","timestamp":1648857600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2023,8]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Although state-of-the-art deep neural network models are known to be robust to random perturbations, it was verified that these architectures are indeed quite vulnerable to deliberately crafted perturbations, albeit being quasi-imperceptible. These vulnerabilities make it challenging to deploy deep neural network models in the areas where security is a critical concern. In recent years, many research studies have been conducted to develop new attack methods and come up with new defense techniques that enable more robust and reliable models. In this study, we use the quantified epistemic uncertainty obtained from the model\u2019s final probability outputs, along with the model\u2019s own loss function, to generate more effective adversarial samples. And we propose a novel defense approach against attacks like Deepfool which result in adversarial samples located near the model\u2019s decision boundary. We have verified the effectiveness of our attack method on MNIST (Digit), MNIST (Fashion) and CIFAR-10 datasets. In our experiments, we showed that our proposed uncertainty-based reversal method achieved a worst case success rate of around 95% without compromising clean accuracy.<\/jats:p>","DOI":"10.1007\/s40747-022-00701-0","type":"journal-article","created":{"date-parts":[[2022,4,2]],"date-time":"2022-04-02T09:02:31Z","timestamp":1648890151000},"page":"3739-3757","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Uncertainty as a Swiss army knife: new adversarial attack and defense ideas based on epistemic uncertainty"],"prefix":"10.1007","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6214-6262","authenticated-orcid":false,"given":"Omer Faruk","family":"Tuna","sequence":"first","affiliation":[]},{"given":"Ferhat Ozgur","family":"Catak","sequence":"additional","affiliation":[]},{"given":"M. Taner","family":"Eskil","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,4,2]]},"reference":[{"key":"701_CR1","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2015) Deep residual learning for image recognition. arXiv:1512.03385","DOI":"10.1109\/CVPR.2016.90"},{"key":"701_CR2","unstructured":"Goodfellow IJ, Bulatov Y, Ibarz J, Arnoud S, Shet V (2014) Multi-digit number recognition from street view imagery using deep convolutional neural networks. arXiv:1312.6082"},{"key":"701_CR3","doi-asserted-by":"publisher","DOI":"10.1038\/nature.2016.19575","author":"T Chouard","year":"2016","unstructured":"Chouard T (2016) The go files: Ai computer wraps up 4-1 victory against human champion. Nature. https:\/\/doi.org\/10.1038\/nature.2016.19575","journal-title":"Nature"},{"issue":"1","key":"701_CR4","doi-asserted-by":"publisher","first-page":"12495","DOI":"10.1038\/s41598-019-48995-4","volume":"9","author":"L Shen","year":"2019","unstructured":"Shen L, Margolies LR, Rothstein JH, Fluder E, McBride R, Sieh W (2019) Deep learning to improve breast cancer detection on screening mammography. Sci Rep 9(1):12495. https:\/\/doi.org\/10.1038\/s41598-019-48995-4","journal-title":"Sci Rep"},{"issue":"1","key":"701_CR5","doi-asserted-by":"publisher","first-page":"9286","DOI":"10.1038\/s41598-018-27569-w","volume":"8","author":"JL Causey","year":"2018","unstructured":"Causey JL, Zhang J, Ma S, Jiang B, Qualls JA, Politte DG, Prior F, Zhang S, Huang X (2018) Highly accurate model for prediction of lung nodule malignancy with ct scans. Sci Rep 8(1):9286. https:\/\/doi.org\/10.1038\/s41598-018-27569-w","journal-title":"Sci Rep"},{"key":"701_CR6","doi-asserted-by":"publisher","unstructured":"Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: unified, real-time object detection. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 779\u2013788. https:\/\/doi.org\/10.1109\/CVPR.2016.91","DOI":"10.1109\/CVPR.2016.91"},{"key":"701_CR7","unstructured":"Ren S, He K, Girshick R, Sun J (2015) Faster r-cnn: towards real-time object detection with region proposal networks. In: Cortes C, Lawrence N, Lee D, Sugiyama M, Garnett R (eds) Advances in neural information processing systems, vol 28. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2015\/file\/14bfa6bb14875e45bba028a21ed38046-Paper.pdf. Accessed 20 Dec 2021"},{"key":"701_CR8","doi-asserted-by":"crossref","unstructured":"Justesen N, Bontrager P, Togelius J, Risi S (2019) Deep learning for video game playing (2019). arXiv:1708.07902","DOI":"10.1109\/TG.2019.2896986"},{"key":"701_CR9","unstructured":"Bahdanau D, Cho K, Bengio Y (2016) Neural machine translation by jointly learning to align and translate. arXiv:1409.0473"},{"key":"701_CR10","doi-asserted-by":"crossref","unstructured":"Luong M-T, Pham H, Manning CD (2015) Effective approaches to attention-based neural machine translation. arXiv:1508.04025","DOI":"10.18653\/v1\/D15-1166"},{"key":"701_CR11","unstructured":"Szegedy C, Zaremba W, Sutskever I, Bruna J, Erhan D, Goodfellow I, Fergus R (2014) Intriguing properties of neural networks. arXiv:1312.6199"},{"key":"701_CR12","doi-asserted-by":"crossref","unstructured":"Sato M, Suzuki J, Shindo H, Matsumoto Y (2018) Interpretable adversarial perturbation in input embedding space for text. arXiv:1805.02917","DOI":"10.24963\/ijcai.2018\/601"},{"key":"701_CR13","doi-asserted-by":"crossref","unstructured":"Carlini N, Wagner D (2018) Audio adversarial examples: targeted attacks on speech-to-text. arXiv:1801.01944","DOI":"10.1109\/SPW.2018.00009"},{"key":"701_CR14","doi-asserted-by":"crossref","unstructured":"Finlayson SG, Chung HW, Kohane IS, Beam AL (2019) Adversarial attacks against medical deep learning systems. arXiv:1804.05296","DOI":"10.1126\/science.aaw4399"},{"key":"701_CR15","unstructured":"Sitawarin C, Bhagoji AN, Mosenia A, Chiang M, Mittal P (2018) Darts: deceiving autonomous cars with toxic signs. arXiv:1802.06430"},{"key":"701_CR16","unstructured":"Morgulis N, Kreines A, Mendelowitz S, Weisglass Y (2019) Fooling a real car with adversarial traffic signs. arXiv:1907.00374"},{"key":"701_CR17","doi-asserted-by":"crossref","unstructured":"Tuna OF, Catak FO, Eskil MT (2022) Exploiting epistemic uncertainty of the deep learning models to generate adversarial samples. Multimedia Tools App year. https:\/\/doi.org\/10.1007\/s11042-022-12132-7 ISBN: 1573\u20137721","DOI":"10.1007\/s11042-022-12132-7"},{"key":"701_CR18","doi-asserted-by":"publisher","first-page":"100270","DOI":"10.1016\/j.cosrev.2020.100270","volume":"37","author":"X Huang","year":"2020","unstructured":"Huang X, Kroening D, Ruan W, Sharp J, Sun Y, Thamo E, Wu M, Yi X (2020) A survey of safety and trustworthiness of deep neural networks: verification, testing, adversarial attack and defence, and interpretability. Comput Sci Rev 37:100270. https:\/\/doi.org\/10.1016\/j.cosrev.2020.100270","journal-title":"Comput Sci Rev"},{"key":"701_CR19","doi-asserted-by":"crossref","unstructured":"Catak FO, Sivaslioglu S, Sahinbas K (2020) A generative model based adversarial security of deep learning and linear classifier models. arXiv:2010.08546","DOI":"10.31449\/inf.v45i1.3234"},{"issue":"2","key":"701_CR20","doi-asserted-by":"publisher","first-page":"998","DOI":"10.1109\/COMST.2020.2975048","volume":"22","author":"A Qayyum","year":"2020","unstructured":"Qayyum A, Usama M, Qadir J, Al-Fuqaha A (2020) Securing connected autonomous vehicles: challenges posed by adversarial machine learning and the way forward. IEEE Commun Surv Tutor 22(2):998\u20131026. https:\/\/doi.org\/10.1109\/COMST.2020.2975048","journal-title":"IEEE Commun Surv Tutor"},{"issue":"4","key":"701_CR21","doi-asserted-by":"publisher","first-page":"450","DOI":"10.1109\/TETCI.2020.2968933","volume":"4","author":"K Sadeghi","year":"2020","unstructured":"Sadeghi K, Banerjee A, Gupta SKS (2020) A system-driven taxonomy of attacks and defenses in adversarial machine learning. IEEE Trans Emerg Top Comput Intell 4(4):450\u2013467. https:\/\/doi.org\/10.1109\/TETCI.2020.2968933","journal-title":"IEEE Trans Emerg Top Comput Intell"},{"key":"701_CR22","unstructured":"Goodfellow IJ, Shlens J, Szegedy C (2015) Explaining and harnessing adversarial examples. arXiv:1412.6572"},{"key":"701_CR23","doi-asserted-by":"crossref","unstructured":"Kurakin A, Goodfellow I, Bengio S (2017) Adversarial examples in the physical world. arXiv:1607.02533","DOI":"10.1201\/9781351251389-8"},{"key":"701_CR24","unstructured":"Kurakin A, Goodfellow IJ, Bengio S Adversarial machine learning at scale, CoRR abs\/1611.01236. arXiv:1611.01236"},{"key":"701_CR25","unstructured":"Madry A, Makelov A, Schmidt L, Tsipras D, Vladu A (2019) Towards deep learning models resistant to adversarial attacks. arXiv:1706.06083"},{"key":"701_CR26","doi-asserted-by":"crossref","unstructured":"Papernot N, McDaniel P, Goodfellow I, Jha S, Celik ZB, Swami A (2017) Practical black-box attacks against machine learning. arXiv:1602.02697","DOI":"10.1145\/3052973.3053009"},{"key":"701_CR27","doi-asserted-by":"crossref","unstructured":"Carlini N, Wagner D (2017) Towards evaluating the robustness of neural networks. arXiv:1608.04644","DOI":"10.1109\/SP.2017.49"},{"key":"701_CR28","doi-asserted-by":"crossref","unstructured":"Moosavi-Dezfooli S-M, Fawzi A, Frossard P (2016) Deepfool: a simple and accurate method to fool deep neural networks. arXiv:1511.04599","DOI":"10.1109\/CVPR.2016.282"},{"key":"701_CR29","doi-asserted-by":"publisher","unstructured":"Liu H, Ji R, Li J, Zhang B, Gao Y, Wu Y, Huang F (2019) Universal adversarial perturbation via prior driven uncertainty approximation. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp 2941\u20132949. https:\/\/doi.org\/10.1109\/ICCV.2019.00303","DOI":"10.1109\/ICCV.2019.00303"},{"key":"701_CR30","doi-asserted-by":"publisher","DOI":"10.1109\/TCDS.2020.2974509","author":"X Qu","year":"2020","unstructured":"Qu X, Sun Z, Ong YS, Gupta A, Wei P (2020) Minimalistic attacks: How little it takes to fool deep reinforcement learning policies. IEEE Trans Cogn Dev Syst. https:\/\/doi.org\/10.1109\/TCDS.2020.2974509","journal-title":"IEEE Trans Cogn Dev Syst"},{"key":"701_CR31","unstructured":"Guo C, Pleiss G, Sun Y, Weinberger KQ (2017) On calibration of modern neural networks. In: Precup D, Teh YW (eds) Proceedings of the 34th international conference on machine learning, vol 70 Proceedings of machine learning research, PMLR, pp 1321\u20131330. https:\/\/proceedings.mlr.press\/v70\/guo17a.html. Accessed 20 Dec 2021"},{"key":"701_CR32","unstructured":"Gawlikowski J, Tassi CRN, Ali M, Lee J, Humt M, Feng J, Kruspe A, Triebel R, Jung P, Roscher R, Shahzad M, Yang W, Bamler R, Zhu XX (2021) A survey of uncertainty in deep neural networks. arXiv:2107.03342"},{"key":"701_CR33","doi-asserted-by":"crossref","unstructured":"Papernot N, McDaniel P, Wu X, Jha S, Swami A (2016) Distillation as a defense to adversarial perturbations against deep neural networks. arXiv:1511.04508","DOI":"10.1109\/SP.2016.41"},{"key":"701_CR34","doi-asserted-by":"crossref","unstructured":"Xie C, Wu Y, van\u00a0der Maaten L, Yuille A, He K (2019) Feature denoising for improving adversarial robustness. arXiv:1812.03411","DOI":"10.1109\/CVPR.2019.00059"},{"key":"701_CR35","unstructured":"Carlini N, Katz G, Barrett C, Dill DL (2018) Ground-truth adversarial examples. https:\/\/openreview.net\/forum?id=Hki-ZlbA-. Accessed 20 Dec 2021"},{"key":"701_CR36","doi-asserted-by":"crossref","unstructured":"Meng D, Chen H (2017) Magnet: a two-pronged defense against adversarial examples. arXiv:1705.09064","DOI":"10.1145\/3133956.3134057"},{"key":"701_CR37","unstructured":"Carlini N, Wagner D (2017) Magnet and \u201cefficient defenses against adversarial attacks\u201d are not robust to adversarial examples. arXiv:1711.08478"},{"key":"701_CR38","doi-asserted-by":"crossref","unstructured":"Liao F, Liang M, Dong Y, Pang T, Hu X, Zhu J (2018) Defense against adversarial attacks using high-level representation guided denoiser. arXiv:1712.02976","DOI":"10.1109\/CVPR.2018.00191"},{"key":"701_CR39","unstructured":"Shen S, Jin G, Gao K, Zhang Y (2017) Ape-gan: adversarial perturbation elimination with gan. arXiv:1707.05474"},{"key":"701_CR40","unstructured":"Raghunathan A, Steinhardt J, Liang P (2020) Certified defenses against adversarial examples. arXiv:1801.09344"},{"key":"701_CR41","unstructured":"Laves M-H, Ihler S, Ortmaier T (2019) Uncertainty quantification in computer-aided diagnosis: Make your model say \u201ci don\u2019t know\u201d for ambiguous cases. arXiv:1908.00792"},{"key":"701_CR42","unstructured":"Tuna OF, Catak FO, Eskil MT (2020) Closeness and uncertainty aware adversarial examples detection in adversarial machine learning. arXiv:2012.06390"},{"key":"701_CR43","doi-asserted-by":"crossref","unstructured":"H\u00fcllermeier E, Waegeman W (2020) Aleatoric and epistemic uncertainty in machine learning: an introduction to concepts and methods. arXiv:1910.09457","DOI":"10.1007\/s10994-021-05946-3"},{"key":"701_CR44","doi-asserted-by":"publisher","first-page":"110617","DOI":"10.1016\/j.jss.2020.110617","volume":"167","author":"D An","year":"2020","unstructured":"An D, Liu J, Zhang M, Chen X, Chen M, Sun H (2020) Uncertainty modeling and runtime verification for autonomous vehicles driving control: a machine learning-based approach. J Syst Softw 167:110617. https:\/\/doi.org\/10.1016\/j.jss.2020.110617","journal-title":"J Syst Softw"},{"key":"701_CR45","doi-asserted-by":"publisher","first-page":"107046","DOI":"10.1016\/j.asoc.2020.107046","volume":"101","author":"R Zheng","year":"2021","unstructured":"Zheng R, Zhang S, Liu L, Luo Y, Sun M (2021) Uncertainty in bayesian deep label distribution learning. Appl Soft Comput 101:107046. https:\/\/doi.org\/10.1016\/j.asoc.2020.107046","journal-title":"Appl Soft Comput"},{"key":"701_CR46","doi-asserted-by":"publisher","first-page":"746","DOI":"10.1016\/j.future.2019.09.006","volume":"102","author":"F Antonelli","year":"2020","unstructured":"Antonelli F, Cortellessa V, Gribaudo M, Pinciroli R, Trivedi KS, Trubiani C (2020) Analytical modeling of performance indices under epistemic uncertainty applied to cloud computing systems. Future Gener Comput Syst 102:746\u2013761. https:\/\/doi.org\/10.1016\/j.future.2019.09.006","journal-title":"Future Gener Comput Syst"},{"key":"701_CR47","unstructured":"Zhou D-X (2018) Universality of deep convolutional neural networks. arXiv:1805.10769"},{"issue":"4","key":"701_CR48","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/BF02551274","volume":"2","author":"G Cybenko","year":"1989","unstructured":"Cybenko G (1989) Approximation by superpositions of a sigmoidal function. Math Control Signals Syst (MCSS) 2(4):303\u2013314. https:\/\/doi.org\/10.1007\/BF02551274","journal-title":"Math Control Signals Syst (MCSS)"},{"issue":"2","key":"701_CR49","doi-asserted-by":"publisher","first-page":"3153","DOI":"10.1109\/LRA.2020.2974682","volume":"5","author":"A Loquercio","year":"2020","unstructured":"Loquercio A, Segu M, Scaramuzza D (2020) A general framework for uncertainty estimation in deep learning. IEEE Robot Autom Lett 5(2):3153\u20133160. https:\/\/doi.org\/10.1109\/LRA.2020.2974682","journal-title":"IEEE Robot Autom Lett"},{"key":"701_CR50","doi-asserted-by":"publisher","unstructured":"Gurevich P, Stuke H (2019) Pairing an arbitrary regressor with an artificial neural network estimating aleatoric uncertainty. Neurocomputing 350:291\u2013306. https:\/\/doi.org\/10.1016\/j.neucom.2019.03.031","DOI":"10.1016\/j.neucom.2019.03.031"},{"key":"701_CR51","doi-asserted-by":"publisher","first-page":"16","DOI":"10.1016\/j.ins.2013.07.030","volume":"255","author":"R Senge","year":"2014","unstructured":"Senge R, B\u00f6sner S, Dembczy\u0144ski K, Haasenritter J, Hirsch O, Donner-Banzhoff N, H\u00fcllermeier E (2014) Reliable classification: learning classifiers that distinguish aleatoric and epistemic uncertainty. Inf Sci 255:16\u201329. https:\/\/doi.org\/10.1016\/j.ins.2013.07.030","journal-title":"Inf Sci"},{"key":"701_CR52","unstructured":"Hinton GE, Neal R (1995) Bayesian Learning for Neural Networks. University of Toronto, CAN ISBN: 0612026760. https:\/\/dl.acm.org\/doi\/book\/10.5555\/922680"},{"key":"701_CR53","unstructured":"Graves A (2011) Practical variational inference for neural networks. In: Shawe-Taylor J, Zemel R, Bartlett P, Pereira F, Weinberger KQ (eds) Advances in neural information processing systems, vol 24. Curran Associates, Inc., pp 2348\u20132356. https:\/\/proceedings.neurips.cc\/paper\/2011\/file\/7eb3c8be3d411e8ebfab08eba5f49632-Paper.pdf. Accessed 20 Dec 2021"},{"key":"701_CR54","unstructured":"Paisley J, Blei D, Jordan M (2012) Variational bayesian inference with stochastic search. arXiv:1206.6430"},{"key":"701_CR55","unstructured":"Hoffman M, Blei DM, Wang C, Paisley J (2013) Stochastic variational inference. arXiv:1206.7051"},{"key":"701_CR56","unstructured":"Blundell C, Cornebise J, Kavukcuoglu K, Wierstra D (2015) Weight uncertainty in neural networks. arXiv:1505.05424"},{"key":"701_CR57","unstructured":"Lakshminarayanan B, Pritzel A, Blundell C (2017) Simple and scalable predictive uncertainty estimation using deep ensembles. arXiv:1612.01474"},{"key":"701_CR58","unstructured":"Gal Y, Ghahramani Z (2016) Dropout as a bayesian approximation: representing model uncertainty in deep learning. arXiv:1506.02142"},{"key":"701_CR59","doi-asserted-by":"publisher","unstructured":"Combalia M, Hueto F, Puig S, Malvehy J, Vilaplana V (2020) Uncertainty estimation in deep neural networks for dermoscopic image classification. In: 2020 IEEE\/CVF conference on computer vision and pattern recognition workshops (CVPRW), pp 3211\u20133220. https:\/\/doi.org\/10.1109\/CVPRW50498.2020.00380","DOI":"10.1109\/CVPRW50498.2020.00380"},{"key":"701_CR60","doi-asserted-by":"publisher","unstructured":"Aladag M, Catak FO, Gul E (2019) Preventing data poisoning attacks by using generative models, in: 2019 1st international informatics and software engineering conference (UBMYK), pp 1\u20135. https:\/\/doi.org\/10.1109\/UBMYK48245.2019.8965459","DOI":"10.1109\/UBMYK48245.2019.8965459"},{"key":"701_CR61","unstructured":"Carlini N, Athalye A, Papernot N, Brendel W, Rauber J, Tsipras D, Goodfellow I, Madry A, Kurakin A (2019) On evaluating adversarial robustness. arXiv:1902.06705"},{"key":"701_CR62","unstructured":"LeCun Y, Cortes C MNIST handwritten digit database. http:\/\/yann.lecun.com\/exdb\/mnist\/. Accessed 20 Dec 2021"},{"key":"701_CR63","unstructured":"Xiao H, Rasul K, Vollgraf R (2017) Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv:1708.07747"},{"key":"701_CR64","unstructured":"Krizhevsky A, Nair V, Hinton G Cifar-10 (canadian institute for advanced research). http:\/\/www.cs.toronto.edu\/~kriz\/cifar.html. Accessed 20 Dec 2021"},{"key":"701_CR65","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556"},{"key":"701_CR66","unstructured":"Rauber J, Brendel W, Bethge M (2018) Foolbox: a python toolbox to benchmark the robustness of machine learning models. arXiv:1707.04131"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-022-00701-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-022-00701-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-022-00701-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,27]],"date-time":"2023-07-27T13:10:50Z","timestamp":1690463450000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-022-00701-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,2]]},"references-count":66,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2023,8]]}},"alternative-id":["701"],"URL":"https:\/\/doi.org\/10.1007\/s40747-022-00701-0","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"type":"print","value":"2199-4536"},{"type":"electronic","value":"2198-6053"}],"subject":[],"published":{"date-parts":[[2022,4,2]]},"assertion":[{"value":"25 March 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 February 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 April 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"On behalf of all authors, the corresponding author states that they do not have any conflicts of interest to declare.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}