{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T17:45:17Z","timestamp":1772300717331,"version":"3.50.1"},"reference-count":60,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2023,2,7]],"date-time":"2023-02-07T00:00:00Z","timestamp":1675728000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,2,7]],"date-time":"2023-02-07T00:00:00Z","timestamp":1675728000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Sci Comput"],"published-print":{"date-parts":[[2023,3]]},"DOI":"10.1007\/s10915-023-02125-5","type":"journal-article","created":{"date-parts":[[2023,2,7]],"date-time":"2023-02-07T11:23:08Z","timestamp":1675768988000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":28,"title":["DCT-Former: Efficient Self-Attention with Discrete Cosine Transform"],"prefix":"10.1007","volume":"94","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1006-7826","authenticated-orcid":false,"given":"Carmelo","family":"Scribano","sequence":"first","affiliation":[]},{"given":"Giorgia","family":"Franchini","sequence":"additional","affiliation":[]},{"given":"Marco","family":"Prato","sequence":"additional","affiliation":[]},{"given":"Marko","family":"Bertogna","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,7]]},"reference":[{"key":"2125_CR1","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017)"},{"key":"2125_CR2","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"issue":"8","key":"2125_CR3","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I., et al.: Language models are unsupervised multitask learners. OpenAI blog 1(8), 9 (2019)","journal-title":"OpenAI blog"},{"key":"2125_CR4","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877\u20131901 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2125_CR5","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"2125_CR6","doi-asserted-by":"crossref","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: European Conference on Computer Vision, pp. 213\u2013229 (2020). Springer","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"2125_CR7","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., Guo, B.: Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"issue":"1","key":"2125_CR8","doi-asserted-by":"publisher","first-page":"90","DOI":"10.1109\/T-C.1974.223784","volume":"100","author":"N Ahmed","year":"1974","unstructured":"Ahmed, N., Natarajan, T., Rao, K.R.: Discrete cosine transform. IEEE Trans. Comput. 100(1), 90\u201393 (1974)","journal-title":"IEEE Trans. Comput."},{"key":"2125_CR9","unstructured":"Maas, A.L., Daly, R.E., Pham, P.T., Huang, D., Ng, A.Y., Potts, C.: Learning word vectors for sentiment analysis. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pp. 142\u2013150. Association for Computational Linguistics, Portland, Oregon, USA (2011). https:\/\/aclanthology.org\/P11-1015"},{"key":"2125_CR10","unstructured":"Goodfellow, I.J., Bengio, Y., Courville, A.: Deep Learning. MIT Press, Cambridge, MA, USA (2016). http:\/\/www.deeplearningbook.org"},{"issue":"8","key":"2125_CR11","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735\u20131780 (1997)","journal-title":"Neural Comput."},{"key":"2125_CR12","doi-asserted-by":"crossref","unstructured":"Cho, K., Van\u00a0Merri\u00ebnboer, B., Bahdanau, D., Bengio, Y.: On the properties of neural machine translation: Encoder-decoder approaches. arXiv preprint arXiv:1409.1259 (2014)","DOI":"10.3115\/v1\/W14-4012"},{"key":"2125_CR13","unstructured":"Ba, J.L., Kiros, J.R., Hinton, G.E.: Layer normalization. arXiv preprint arXiv:1607.06450 (2016)"},{"key":"2125_CR14","unstructured":"Bridle, J.: Training stochastic model recognition algorithms as networks can lead to maximum mutual information estimation of parameters. Advances in neural information processing systems 2 (1989)"},{"key":"2125_CR15","doi-asserted-by":"crossref","unstructured":"Pennington, J., Socher, R., Manning, C.D.: Glove: Global vectors for word representation. In: Empirical Methods in Natural Language Processing (EMNLP), pp. 1532\u20131543 (2014). http:\/\/www.aclweb.org\/anthology\/D14-1162","DOI":"10.3115\/v1\/D14-1162"},{"key":"2125_CR16","doi-asserted-by":"crossref","unstructured":"Peters, M.E., Ammar, W., Bhagavatula, C., Power, R.: Semi-supervised sequence tagging with bidirectional language models. arXiv preprint arXiv:1705.00108 (2017)","DOI":"10.18653\/v1\/P17-1161"},{"key":"2125_CR17","unstructured":"Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., Soricut, R.: Albert: A lite bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942 (2019)"},{"key":"2125_CR18","unstructured":"Clark, K., Luong, M.-T., Le, Q.V., Manning, C.D.: Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555 (2020)"},{"key":"2125_CR19","unstructured":"He, P., Liu, X., Gao, J., Chen, W.: Deberta: Decoding-enhanced bert with disentangled attention. arXiv preprint arXiv:2006.03654 (2020)"},{"key":"2125_CR20","unstructured":"Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., Stoyanov, V.: Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)"},{"key":"2125_CR21","doi-asserted-by":"crossref","unstructured":"Iandola, F.N., Shaw, A.E., Krishna, R., Keutzer, K.W.: Squeezebert: What can computer vision teach nlp about efficient neural networks? arXiv preprint arXiv:2006.11316 (2020)","DOI":"10.18653\/v1\/2020.sustainlp-1.17"},{"key":"2125_CR22","unstructured":"Wu, Y., Schuster, M., Chen, Z., Le, Q.V., Norouzi, M., Macherey, W., Krikun, M., Cao, Y., Gao, Q., Macherey, K., et al.: Google\u2019s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144 (2016)"},{"key":"2125_CR23","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Kiros, R., Zemel, R., Salakhutdinov, R., Urtasun, R., Torralba, A., Fidler, S.: Aligning books and movies: Towards story-like visual explanations by watching movies and reading books. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 19\u201327 (2015)","DOI":"10.1109\/ICCV.2015.11"},{"key":"2125_CR24","unstructured":"Sharir, O., Peleg, B., Shoham, Y.: The cost of training nlp models: A concise overview. arXiv preprint arXiv:2004.08900 (2020)"},{"key":"2125_CR25","doi-asserted-by":"crossref","unstructured":"Izsak, P., Berchansky, M., Levy, O.: How to train bert with an academic budget. arXiv preprint arXiv:2104.07705 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.831"},{"key":"2125_CR26","doi-asserted-by":"crossref","unstructured":"Raid, A.M., Khedr, W.M., El-dosuky, M.A., Ahmed, W.: Jpeg image compression using discrete cosine transform a survey. arXiv preprint arXiv:1405.6147 (2014)","DOI":"10.5121\/ijcses.2014.5204"},{"issue":"6","key":"2125_CR27","doi-asserted-by":"publisher","first-page":"1553","DOI":"10.1016\/j.sigpro.2008.01.004","volume":"88","author":"X Shao","year":"2008","unstructured":"Shao, X., Johnson, S.G.: Type-II\/III DCT\/DST algorithms with reduced number of arithmetic operations. Signal Process. 88(6), 1553\u20131564 (2008)","journal-title":"Signal Process."},{"key":"2125_CR28","unstructured":"Wang, S., Li, B.Z., Khabsa, M., Fang, H., Ma, H.: Linformer: Self-attention with linear complexity. arXiv preprint arXiv:2006.04768 (2020)"},{"key":"2125_CR29","unstructured":"Choromanski, K.M., Likhosherstov, V., Dohan, D., Song, X., Gane, A., Sarlos, T., Hawkins, P., Davis, J.Q., Mohiuddin, A., Kaiser, L., et\u00a0al.: Rethinking attention with performers. In: International Conference on Learning Representations, p. 636 (2021)"},{"key":"2125_CR30","unstructured":"Kitaev, N., Kaiser, L., Levskaya, A.: Reformer: The efficient transformer. In: International Conference on Learning Representations, p. 1838 (2020)"},{"key":"2125_CR31","doi-asserted-by":"crossref","unstructured":"Xiong, Y., Zeng, Z., Chakraborty, R., Tan, M., Fung, G., Li, Y., Singh, V.: Nystr\u00f6mformer: A Nystr\u00f6m-based algorithm for approximating self-attention. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 14138\u201314148 (2021)","DOI":"10.1609\/aaai.v35i16.17664"},{"key":"2125_CR32","first-page":"21297","volume":"34","author":"J Lu","year":"2021","unstructured":"Lu, J., Yao, J., Zhang, J., Zhu, X., Xu, H., Gao, W., Xu, C., Xiang, T., Zhang, L.: Soft: Softmax-free transformer with linear complexity. Adv. Neural. Inf. Process. Syst. 34, 21297\u201321309 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2125_CR33","first-page":"22470","volume":"34","author":"H Ren","year":"2021","unstructured":"Ren, H., Dai, H., Dai, Z., Yang, M., Leskovec, J., Schuurmans, D., Dai, B.: Combiner: Full attention transformer with sparse computation cost. Adv. Neural. Inf. Process. Syst. 34, 22470\u201322482 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2125_CR34","first-page":"29449","volume":"34","author":"T Nguyen","year":"2021","unstructured":"Nguyen, T., Suliafu, V., Osher, S., Chen, L., Wang, B.: Fmmformer: Efficient and flexible transformer via decomposed near-field and far-field attention. Adv. Neural. Inf. Process. Syst. 34, 29449\u201329463 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2125_CR35","unstructured":"Wu, C., Wu, F., Qi, T., Huang, Y., Xie, X.: Fastformer: Additive attention can be all you need. arXiv preprint arXiv:2108.09084 (2021)"},{"key":"2125_CR36","first-page":"9895","volume":"34","author":"S Jaszczur","year":"2021","unstructured":"Jaszczur, S., Chowdhery, A., Mohiuddin, A., Kaiser, \u0141, Gajewski, W., Michalewski, H., Kanerva, J.: Sparse is enough in scaling transformers. Adv. Neural. Inf. Process. Syst. 34, 9895\u20139907 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2125_CR37","unstructured":"Beltagy, I., Peters, M.E., Cohan, A.: Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150 (2020)"},{"key":"2125_CR38","unstructured":"Tay, Y., Bahri, D., Metzler, D., Juan, D.-C., Zhao, Z., Zheng, C.: Synthesizer: Rethinking self-attention for transformer models. In: International Conference on Machine Learning, pp. 10183\u201310192 (2021)"},{"key":"2125_CR39","first-page":"17723","volume":"34","author":"C Zhu","year":"2021","unstructured":"Zhu, C., Ping, W., Xiao, C., Shoeybi, M., Goldstein, T., Anandkumar, A., Catanzaro, B.: Long-short transformer: Efficient transformers for language and vision. Adv. Neural. Inf. Process. Syst. 34, 17723\u201317736 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2125_CR40","first-page":"17413","volume":"34","author":"B Chen","year":"2021","unstructured":"Chen, B., Dao, T., Winsor, E., Song, Z., Rudra, A., R\u00e9, C.: Scatterbrain: Unifying sparse and low-rank attention. Adv. Neural. Inf. Process. Syst. 34, 17413\u201317426 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2125_CR41","first-page":"24261","volume":"34","author":"IO Tolstikhin","year":"2021","unstructured":"Tolstikhin, I.O., Houlsby, N., Kolesnikov, A., Beyer, L., Zhai, X., Unterthiner, T., Yung, J., Steiner, A., Keysers, D., Uszkoreit, J., et al.: Mlp-mixer: An all-mlp architecture for vision. Adv. Neural. Inf. Process. Syst. 34, 24261\u201324272 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2125_CR42","doi-asserted-by":"crossref","unstructured":"Lee-Thorp, J., Ainslie, J., Eckstein, I., Ontanon, S.: Fnet: Mixing tokens with fourier transforms. arXiv preprint arXiv:2105.03824 (2021)","DOI":"10.18653\/v1\/2022.naacl-main.319"},{"key":"2125_CR43","doi-asserted-by":"crossref","unstructured":"You, W., Sun, S., Iyyer, M.: Hard-coded gaussian attention for neural machine translation. arXiv preprint arXiv:2005.00742 (2020)","DOI":"10.18653\/v1\/2020.acl-main.687"},{"key":"2125_CR44","doi-asserted-by":"crossref","unstructured":"Jacob, B., Kligys, S., Chen, B., Zhu, M., Tang, M., Howard, A., Adam, H., Kalenichenko, D.: Quantization and training of neural networks for efficient integer-arithmetic-only inference. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2704\u20132713 (2018)","DOI":"10.1109\/CVPR.2018.00286"},{"key":"2125_CR45","unstructured":"Han, S., Pool, J., Tran, J., Dally, W.: Learning both weights and connections for efficient neural network. In: Cortes, C., Lawrence, N., Lee, D., Sugiyama, M., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 28, pp. 1135\u20131143 (2015)"},{"key":"2125_CR46","unstructured":"Hinton, G., Vinyals, O., Dean, J., et al.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.025312(7) (2015)"},{"key":"2125_CR47","unstructured":"Vanhoucke, V., Senior, A., Mao, M.Z.: Improving the speed of neural networks on CPUs. In: Deep Learning and Unsupervised Feature Learning Workshop, NIPS 2011 (2011)"},{"issue":"6","key":"2125_CR48","doi-asserted-by":"publisher","first-page":"1402","DOI":"10.1016\/j.sigpro.2006.12.004","volume":"87","author":"Y Zhou","year":"2007","unstructured":"Zhou, Y., Noras, J.M., Shepherd, S.J.: Novel design of multiplier-less FFT processors. Signal Process. 87(6), 1402\u20131407 (2007)","journal-title":"Signal Process."},{"issue":"16","key":"2125_CR49","doi-asserted-by":"publisher","first-page":"4367","DOI":"10.1109\/TSP.2017.2712120","volume":"65","author":"C Rusu","year":"2017","unstructured":"Rusu, C., Thompson, J.: Learning fast sparsifying transforms. IEEE Trans. Signal Process. 65(16), 4367\u20134378 (2017)","journal-title":"IEEE Trans. Signal Process."},{"key":"2125_CR50","doi-asserted-by":"crossref","unstructured":"Le\u00a0Magoarou, L., Gribonval, R.: Chasing butterflies: In search of efficient dictionaries. In: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 3287\u20133291 (2015)","DOI":"10.1109\/ICASSP.2015.7178579"},{"key":"2125_CR51","first-page":"3933","volume":"31","author":"L Gueguen","year":"2018","unstructured":"Gueguen, L., Sergeev, A., Kadlec, B., Liu, R., Yosinski, J.: Faster neural networks straight from jpeg. Adv. Neural. Inf. Process. Syst. 31, 3933\u20133944 (2018)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2125_CR52","unstructured":"Dziedzic, A., Paparrizos, J., Krishnan, S., Elmore, A., Franklin, M.: Band-limited training and inference for convolutional neural networks. In: International Conference on Machine Learning, pp. 1745\u20131754 (2019)"},{"key":"2125_CR53","doi-asserted-by":"crossref","unstructured":"Rajesh, B., Javed, M., Srivastava, S., et\u00a0al.: Dct-compcnn: A novel image classification network using jpeg compressed dct coefficients. In: 2019 IEEE Conference on Information and Communication Technology, pp. 1\u20136 (2019)","DOI":"10.1109\/CICT48419.2019.9066242"},{"key":"2125_CR54","doi-asserted-by":"crossref","unstructured":"Xu, K., Qin, M., Sun, F., Wang, Y., Chen, Y.-K., Ren, F.: Learning in the frequency domain. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1740\u20131749 (2020)","DOI":"10.1109\/CVPR42600.2020.00181"},{"key":"2125_CR55","doi-asserted-by":"crossref","unstructured":"dos Santos, S.F., Sebe, N., Almeida, J.: The good, the bad, and the ugly: Neural networks straight from jpeg. In: 2020 IEEE International Conference on Image Processing (ICIP), pp. 1896\u20131900 (2020)","DOI":"10.1109\/ICIP40778.2020.9190741"},{"issue":"1","key":"2125_CR56","doi-asserted-by":"publisher","first-page":"27","DOI":"10.1109\/TASSP.1980.1163351","volume":"28","author":"J Makhoul","year":"1980","unstructured":"Makhoul, J.: A fast cosine transform in one and two dimensions. IEEE Trans. Acoust. Speech Signal Process. 28(1), 27\u201334 (1980)","journal-title":"IEEE Trans. Acoust. Speech Signal Process."},{"key":"2125_CR57","unstructured":"Turc, I., Chang, M.-W., Lee, K., Toutanova, K.: Well-read students learn better: On the importance of pre-training compact models. arXiv preprint arXiv:1908.08962 (2019)"},{"key":"2125_CR58","doi-asserted-by":"crossref","unstructured":"Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C., Moi, A., Cistac, P., Rault, T., Louf, R., Funtowicz, M., Davison, J., Shleifer, S., von Platen, P., Ma, C., Jernite, Y., Plu, J., Xu, C., Scao, T.L., Gugger, S., Drame, M., Lhoest, Q., Rush, A.M.: Transformers: State-of-the-art natural language processing. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pp. 38\u201345. Association for Computational Linguistics, Online (2020)","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"2125_CR59","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"key":"2125_CR60","doi-asserted-by":"crossref","unstructured":"Rasley, J., Rajbhandari, S., Ruwase, O., He, Y.: Deepspeed: System optimizations enable training deep learning models with over 100 billion parameters. In: Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 3505\u20133506 (2020)","DOI":"10.1145\/3394486.3406703"}],"container-title":["Journal of Scientific Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10915-023-02125-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10915-023-02125-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10915-023-02125-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,22]],"date-time":"2023-02-22T21:13:14Z","timestamp":1677100394000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10915-023-02125-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,2,7]]},"references-count":60,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2023,3]]}},"alternative-id":["2125"],"URL":"https:\/\/doi.org\/10.1007\/s10915-023-02125-5","relation":{},"ISSN":["0885-7474","1573-7691"],"issn-type":[{"value":"0885-7474","type":"print"},{"value":"1573-7691","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,2,7]]},"assertion":[{"value":"28 February 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 October 2022","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 January 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 February 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no conflicts of interest to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"67"}}