{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,23]],"date-time":"2026-03-23T21:20:59Z","timestamp":1774300859240,"version":"3.50.1"},"reference-count":71,"publisher":"Springer Science and Business Media LLC","issue":"7","license":[{"start":{"date-parts":[[2024,11,22]],"date-time":"2024-11-22T00:00:00Z","timestamp":1732233600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,22]],"date-time":"2024-11-22T00:00:00Z","timestamp":1732233600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,5]]},"DOI":"10.1007\/s00371-024-03682-y","type":"journal-article","created":{"date-parts":[[2024,11,22]],"date-time":"2024-11-22T11:01:10Z","timestamp":1732273270000},"page":"4639-4660","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["DGL-GAN: discriminator-guided GAN compression"],"prefix":"10.1007","volume":"41","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6287-4174","authenticated-orcid":false,"given":"Yuesong","family":"Tian","sequence":"first","affiliation":[]},{"given":"Li","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Xiang","family":"Tian","sequence":"additional","affiliation":[]},{"given":"Dacheng","family":"Tao","sequence":"additional","affiliation":[]},{"given":"Zhifeng","family":"Li","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Yaowu","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,22]]},"reference":[{"key":"3682_CR1","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. In: Advances in Neural Information Processing Systems, pp. 2672\u20132680 (2014)"},{"key":"3682_CR2","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4401\u20134410 (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"3682_CR3","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8110\u20138119 (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"3682_CR4","unstructured":"Brock, A., Donahue, J., Simonyan, K.: Large scale gan training for high fidelity natural image synthesis. arXiv preprint arXiv:1809.11096 (2018)"},{"key":"3682_CR5","doi-asserted-by":"crossref","unstructured":"Sauer, A., Schwarz, K., Geiger, A.: Stylegan-xl: Scaling stylegan to large diverse datasets. In: ACM SIGGRAPH 2022 Conference Proceedings, pp. 1\u201310 (2022)","DOI":"10.1145\/3528233.3530738"},{"key":"3682_CR6","unstructured":"Sauer, A., Karras, T., Laine, S., Geiger, A., Aila, T.: Stylegan-t: Unlocking the power of gans for fast large-scale text-to-image synthesis. arXiv preprint arXiv:2301.09515 (2023)"},{"key":"3682_CR7","doi-asserted-by":"crossref","unstructured":"Kang, M., Zhu, J.-Y., Zhang, R., Park, J., Shechtman, E., Paris, S., Park, T.: Scaling up gans for text-to-image synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10124\u201310134 (2023)","DOI":"10.1109\/CVPR52729.2023.00976"},{"key":"3682_CR8","doi-asserted-by":"crossref","unstructured":"Shen, Y., Gu, J., Tang, X., Zhou, B.: Interpreting the latent space of gans for semantic face editing. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9243\u20139252 (2020)","DOI":"10.1109\/CVPR42600.2020.00926"},{"key":"3682_CR9","doi-asserted-by":"crossref","unstructured":"Kim, H., Choi, Y., Kim, J., Yoo, S., Uh, Y.: Stylemapgan: Exploiting spatial dimensions of latent in gan for real-time image editing. arXiv preprint arXiv:2104.14754 (2021)","DOI":"10.1109\/CVPR46437.2021.00091"},{"key":"3682_CR10","doi-asserted-by":"crossref","unstructured":"Li, M., Jin, Y., Zhu, H.: Surrogate gradient field for latent space manipulation. arXiv preprint arXiv:2104.09065 (2021)","DOI":"10.1109\/CVPR46437.2021.00646"},{"key":"3682_CR11","doi-asserted-by":"crossref","unstructured":"Lin, J., Zhang, R., Ganz, F., Han, S., Zhu, J.-Y.: Anycost gans for interactive image synthesis and editing. arXiv preprint arXiv:2103.03243 (2021)","DOI":"10.1109\/CVPR46437.2021.01474"},{"key":"3682_CR12","unstructured":"Zhuang, P., Koyejo, O., Schwing, A.G.: Enjoy your editing: Controllable gans for image editing via latent space navigation. arXiv preprint arXiv:2102.01187 (2021)"},{"key":"3682_CR13","doi-asserted-by":"crossref","unstructured":"Kaneko, T., Harada, T.: Noise robust generative adversarial networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8404\u20138414 (2020)","DOI":"10.1109\/CVPR42600.2020.00843"},{"key":"3682_CR14","unstructured":"Dutta, P., Power, B., Halpert, A., Ezequiel, C., Subramanian, A., Chatterjee, C., Hari, S., Prindle, K., Vaddina, V., Leach, A., et al.: 3d conditional generative adversarial networks to enable large-scale seismic image enhancement. arXiv preprint arXiv:1911.06932 (2019)"},{"key":"3682_CR15","unstructured":"Karras, T., Aittala, M., Laine, S., H\u00e4rk\u00f6nen, E., Hellsten, J., Lehtinen, J., Aila, T.: Alias-free generative adversarial networks. arXiv preprint arXiv:2106.12423 (2021)"},{"key":"3682_CR16","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., Fei-Fei, L.: Imagenet: A large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255 (2009). Ieee","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"3682_CR17","unstructured":"Chen, X., Zhang, Z., Sui, Y., Chen, T.: {GAN}s can play lottery tickets too. In: International Conference on Learning Representations (2021). https:\/\/openreview.net\/forum?id=1AoMhc_9jER"},{"key":"3682_CR18","unstructured":"Wang, P., Wang, D., Ji, Y., Xie, X., Song, H., Liu, X., Lyu, Y., Xie, Y.: Qgan: Quantized generative adversarial networks. arXiv preprint arXiv:1901.08263 (2019)"},{"key":"3682_CR19","doi-asserted-by":"crossref","unstructured":"Wang, H., Gui, S., Yang, H., Liu, J., Wang, Z.: Gan slimming: All-in-one gan compression by a unified optimization framework. In: European Conference on Computer Vision, pp. 54\u201373 (2020). Springer","DOI":"10.1007\/978-3-030-58548-8_4"},{"key":"3682_CR20","doi-asserted-by":"crossref","unstructured":"Liu, Y., Shu, Z., Li, Y., Lin, Z., Perazzi, F., Kung, S.-Y.: Content-aware gan compression. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12156\u201312166 (2021)","DOI":"10.1109\/CVPR46437.2021.01198"},{"key":"3682_CR21","unstructured":"Hou, L., Yuan, Z., Huang, L., Shen, H., Cheng, X., Wang, C.: Slimmable generative adversarial networks. arXiv preprint arXiv:2012.05660 (2020)"},{"key":"3682_CR22","unstructured":"Shen, Y., Shen, L., Huang, H.-Z., Wang, X., Liu, W.: Cpot: Channel pruning via optimal transport. arXiv preprint arXiv:2005.10451 (2020)"},{"key":"3682_CR23","doi-asserted-by":"publisher","unstructured":"Tian, Y., Shen, L., Shen, L., Su, G., Li, Z., Liu, W.: Alphagan: Fully differentiable architecture search for generative adversarial networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1\u20131 (2021). https:\/\/doi.org\/10.1109\/TPAMI.2021.3099829","DOI":"10.1109\/TPAMI.2021.3099829"},{"key":"3682_CR24","unstructured":"Liu, B., Zhu, Y., Song, K., Elgammal, A.: Towards faster and stabilized {gan} training for high-fidelity few-shot image synthesis. In: International Conference on Learning Representations (2021). https:\/\/openreview.net\/forum?id=1Fqg133qRaI"},{"key":"3682_CR25","unstructured":"Belousov, S.: Mobilestylegan: A lightweight convolutional neural network for high-fidelity image synthesis. arXiv preprint arXiv:2104.04767 (2021)"},{"key":"3682_CR26","unstructured":"Wang, J., Shu, H., Xia, W., Yang, Y., Wang, Y.: Coarse-to-fine searching for efficient generative adversarial networks. arXiv preprint arXiv:2104.09223 (2021)"},{"key":"3682_CR27","doi-asserted-by":"crossref","unstructured":"Li, J., Zhang, J., Gong, X., Lu, S.: Evolutionary generative adversarial networks with crossover based knowledge distillation. arXiv preprint arXiv:2101.11186 (2021)","DOI":"10.1109\/IJCNN52387.2021.9533612"},{"issue":"6","key":"3682_CR28","doi-asserted-by":"publisher","first-page":"921","DOI":"10.1109\/TEVC.2019.2895748","volume":"23","author":"C Wang","year":"2019","unstructured":"Wang, C., Xu, C., Yao, X., Tao, D.: Evolutionary generative adversarial networks. IEEE Trans. Evol. Comput. 23(6), 921\u2013934 (2019)","journal-title":"IEEE Trans. Evol. Comput."},{"key":"3682_CR29","unstructured":"Aguinaldo, A., Chiang, P.-Y., Gain, A., Patil, A., Pearson, K., Feizi, S.: Compressing gans using knowledge distillation. arXiv preprint arXiv:1902.00159 (2019)"},{"key":"3682_CR30","unstructured":"Yu, C., Pool, J.: Self-supervised generative adversarial compression. Advances in Neural Information Processing Systems 33 (2020)"},{"key":"3682_CR31","first-page":"3585","volume":"34","author":"H Chen","year":"2020","unstructured":"Chen, H., Wang, Y., Shu, H., Wen, C., Xu, C., Shi, B., Xu, C., Xu, C.: Distilling portable generative adversarial networks for image translation. Proc. AAAI Conf. Art. Intel. 34, 3585\u20133592 (2020)","journal-title":"Proc. AAAI Conf. Art. Intel."},{"key":"3682_CR32","doi-asserted-by":"crossref","unstructured":"Li, M., Lin, J., Ding, Y., Liu, Z., Zhu, J.-Y., Han, S.: Gan compression: Efficient architectures for interactive conditional gans. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5284\u20135294 (2020)","DOI":"10.1109\/CVPR42600.2020.00533"},{"key":"3682_CR33","doi-asserted-by":"crossref","unstructured":"Jin, Q., Ren, J., Woodford, O.J., Wang, J., Yuan, G., Wang, Y., Tulyakov, S.: Teachers do more than teach: Compressing image-to-image models. arXiv preprint arXiv:2103.03467 (2021)","DOI":"10.1109\/CVPR46437.2021.01339"},{"key":"3682_CR34","unstructured":"Li, S., Lin, M., Wang, Y., Xu, M., Huang, F., Wu, Y., Shao, L., Ji, R.: Learning efficient gans via differentiable masks and co-attention distillation. arXiv preprint arXiv:2011.08382 (2020)"},{"key":"3682_CR35","unstructured":"Fu, Y., Chen, W., Wang, H., Li, H., Lin, Y., Wang, Z.: Autogan-distiller: Searching to compress generative adversarial networks. arXiv preprint arXiv:2006.08198 (2020)"},{"key":"3682_CR36","unstructured":"Zhang, H., Goodfellow, I., Metaxas, D., Odena, A.: Self-attention generative adversarial networks. arXiv preprint arXiv:1805.08318 (2018)"},{"key":"3682_CR37","unstructured":"Li, S., Wu, J., Xiao, X., Chao, F., Mao, X., Ji, R.: Revisiting discriminator in gan compression: A generator-discriminator cooperative compression scheme. In: Thirty-Fifth Conference on Neural Information Processing Systems (2021)"},{"key":"3682_CR38","first-page":"18241","volume":"35","author":"M Kang","year":"2022","unstructured":"Kang, M., Yoo, H., Kang, E., Ki, S., Lee, H.E., Han, B.: Information-theoretic gan compression with variational energy-based model. Adv. Neural Inf. Process. Syst. 35, 18241\u201318255 (2022)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3682_CR39","unstructured":"Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., Chen, X.: Improved techniques for training gans. In: Advances in Neural Information Processing Systems, pp. 2234\u20132242 (2016)"},{"key":"3682_CR40","unstructured":"Sun, R., Fang, T., Schwing, A.: Towards a better global loss landscape of gans. Advances in Neural Information Processing Systems 33 (2020)"},{"key":"3682_CR41","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. In: Advances in Neural Information Processing Systems, pp. 6626\u20136637 (2017)"},{"key":"3682_CR42","unstructured":"Han, S., Mao, H., Dally, W.J.: Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149 (2015)"},{"key":"3682_CR43","unstructured":"Han, S., Pool, J., Tran, J., Dally, W.J.: Learning both weights and connections for efficient neural networks. arXiv preprint arXiv:1506.02626 (2015)"},{"key":"3682_CR44","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)"},{"key":"3682_CR45","unstructured":"Courbariaux, M., Bengio, Y., David, J.-P.: Binaryconnect: Training deep neural networks with binary weights during propagations. arXiv preprint arXiv:1511.00363 (2015)"},{"key":"3682_CR46","doi-asserted-by":"crossref","unstructured":"Wu, J., Leng, C., Wang, Y., Hu, Q., Cheng, J.: Quantized convolutional neural networks for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4820\u20134828 (2016)","DOI":"10.1109\/CVPR.2016.521"},{"key":"3682_CR47","unstructured":"Hu, H., Peng, R., Tai, Y.-W., Tang, C.-K.: Network trimming: A data-driven neuron pruning approach towards efficient deep architectures. arXiv preprint arXiv:1607.03250 (2016)"},{"key":"3682_CR48","unstructured":"Li, H., Kadav, A., Durdanovic, I., Samet, H., Graf, H.P.: Pruning filters for efficient convnets. arXiv preprint arXiv:1608.08710 (2016)"},{"key":"3682_CR49","doi-asserted-by":"crossref","unstructured":"Yim, J., Joo, D., Bae, J., Kim, J.: A gift from knowledge distillation: Fast optimization, network minimization and transfer learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4133\u20134141 (2017)","DOI":"10.1109\/CVPR.2017.754"},{"key":"3682_CR50","unstructured":"Tan, M., Le, Q.V.: Efficientnet: Rethinking model scaling for convolutional neural networks. arXiv preprint arXiv:1905.11946 (2019)"},{"key":"3682_CR51","unstructured":"Liu, H., Simonyan, K., Yang, Y.: Darts: Differentiable architecture search. arXiv preprint arXiv:1806.09055 (2018)"},{"key":"3682_CR52","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., Adam, H.: Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)"},{"key":"3682_CR53","doi-asserted-by":"crossref","unstructured":"Ren, Y., Wu, J., Xiao, X., Yang, J.: Online multi-granularity distillation for gan compression. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6793\u20136803 (2021)","DOI":"10.1109\/ICCV48922.2021.00672"},{"key":"3682_CR54","unstructured":"Liu, S., Tian, Y., Chen, T., Shen, L.: Sparse Unbalanced GAN Training with In-Time Over-Parameterization (2022). https:\/\/openreview.net\/forum?id=WLZ_2JjCz2a"},{"key":"3682_CR55","unstructured":"Radford, A., Metz, L., Chintala, S.: Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434 (2015)"},{"key":"3682_CR56","unstructured":"Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein generative adversarial networks. In: International Conference on Machine Learning, pp. 214\u2013223 (2017). PMLR"},{"key":"3682_CR57","unstructured":"Frankle, J., Carbin, M.: The lottery ticket hypothesis: Finding sparse, trainable neural networks. arXiv preprint arXiv:1803.03635 (2018)"},{"key":"3682_CR58","unstructured":"Miyato, T., Kataoka, T., Koyama, M., Yoshida, Y.: Spectral normalization for generative adversarial networks. arXiv preprint arXiv:1802.05957 (2018)"},{"issue":"2","key":"3682_CR59","doi-asserted-by":"publisher","first-page":"251","DOI":"10.1016\/0893-6080(91)90009-T","volume":"4","author":"K Hornik","year":"1991","unstructured":"Hornik, K.: Approximation capabilities of multilayer feedforward networks. Neural Netw. 4(2), 251\u2013257 (1991)","journal-title":"Neural Netw."},{"key":"3682_CR60","unstructured":"Allen-Zhu, Z., Li, Y., Song, Z.: A convergence theory for deep learning via over-parameterization. In: International Conference on Machine Learning, pp. 242\u2013252 (2019). PMLR"},{"key":"3682_CR61","unstructured":"Goodfellow, I.J., Shlens, J., Szegedy, C.: Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572 (2014)"},{"key":"3682_CR62","doi-asserted-by":"crossref","unstructured":"Bai, T., Luo, J., Zhao, J., Wen, B., Wang, Q.: Recent advances in adversarial training for adversarial robustness. arXiv preprint arXiv:2102.01356 (2021)","DOI":"10.24963\/ijcai.2021\/591"},{"key":"3682_CR63","doi-asserted-by":"crossref","unstructured":"Bai, J., Yuan, L., Xia, S.-T., Yan, S., Li, Z., Liu, W.: Improving vision transformers by revisiting high-frequency components. In: European Conference on Computer Vision, pp. 1\u201318 (2022). Springer","DOI":"10.1007\/978-3-031-20053-3_1"},{"key":"3682_CR64","unstructured":"Karras, T., Aittala, M., Hellsten, J., Laine, S., Lehtinen, J., Aila, T.: Training generative adversarial networks with limited data. arXiv preprint arXiv:2006.06676 (2020)"},{"key":"3682_CR65","unstructured":"Smilkov, D., Thorat, N., Kim, B., Vi\u00e9gas, F., Wattenberg, M.: Smoothgrad: removing noise by adding noise. arXiv preprint arXiv:1706.03825 (2017)"},{"key":"3682_CR66","unstructured":"Yu, F., Seff, A., Zhang, Y., Song, S., Funkhouser, T., Xiao, J.: Lsun: Construction of a large-scale image dataset using deep learning with humans in the loop. arXiv preprint arXiv:1506.03365 (2015)"},{"key":"3682_CR67","unstructured":"Zhao, S., Liu, Z., Lin, J., Zhu, J.-Y., Han, S.: Differentiable augmentation for data-efficient gan training. arXiv preprint arXiv:2006.10738 (2020)"},{"key":"3682_CR68","doi-asserted-by":"crossref","unstructured":"Zhou, P., Xie, L., Ni, B., Geng, C., Tian, Q.: Omni-gan: On the secrets of cgans and beyond. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14061\u201314071 (2021)","DOI":"10.1109\/ICCV48922.2021.01380"},{"key":"3682_CR69","unstructured":"Furlanello, T., Lipton, Z., Tschannen, M., Itti, L., Anandkumar, A.: Born again neural networks. In: International Conference on Machine Learning, pp. 1607\u20131616 (2018). PMLR"},{"key":"3682_CR70","unstructured":"Romero, A., Ballas, N., Kahou, S.E., Chassang, A., Gatta, C., Bengio, Y.: Fitnets: Hints for thin deep nets. arXiv preprint arXiv:1412.6550 (2014)"},{"key":"3682_CR71","unstructured":"Shen, Y., Yang, C., Tang, X., Zhou, B.: Interfacegan: Interpreting the disentangled face representation learned by gans. IEEE transactions on pattern analysis and machine intelligence (2020)"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03682-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-024-03682-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03682-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,24]],"date-time":"2025-04-24T10:04:24Z","timestamp":1745489064000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-024-03682-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,22]]},"references-count":71,"journal-issue":{"issue":"7","published-print":{"date-parts":[[2025,5]]}},"alternative-id":["3682"],"URL":"https:\/\/doi.org\/10.1007\/s00371-024-03682-y","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,22]]},"assertion":[{"value":"1 September 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 November 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The research does not involve human participants or animals. There is no conflict of interest to report.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}