{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T17:00:35Z","timestamp":1777654835964,"version":"3.51.4"},"publisher-location":"Cham","reference-count":64,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031200700","type":"print"},{"value":"9783031200717","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-20071-7_7","type":"book-chapter","created":{"date-parts":[[2022,11,12]],"date-time":"2022-11-12T05:15:09Z","timestamp":1668230109000},"page":"111-128","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":37,"title":["Learning Graph Neural Networks for\u00a0Image Style Transfer"],"prefix":"10.1007","author":[{"given":"Yongcheng","family":"Jing","sequence":"first","affiliation":[]},{"given":"Yining","family":"Mao","sequence":"additional","affiliation":[]},{"given":"Yiding","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Yibing","family":"Zhan","sequence":"additional","affiliation":[]},{"given":"Mingli","family":"Song","sequence":"additional","affiliation":[]},{"given":"Xinchao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Dacheng","family":"Tao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,13]]},"reference":[{"key":"7_CR1","doi-asserted-by":"crossref","unstructured":"An, J., Huang, S., Song, Y., Dou, D., Liu, W., Luo, J.: ArtFlow: unbiased image style transfer via reversible neural flows. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00092"},{"key":"7_CR2","unstructured":"Champandard, A.J.: Semantic style transfer and turning two-bit doodles into fine artworks. arXiv preprint arXiv:1603.01768 (2016)"},{"key":"7_CR3","doi-asserted-by":"crossref","unstructured":"Chen, D., Yuan, L., Liao, J., Yu, N., Hua, G.: StyleBank: an explicit representation for neural image style transfer. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.296"},{"key":"7_CR4","doi-asserted-by":"crossref","unstructured":"Chen, D., Yuan, L., Liao, J., Yu, N., Hua, G.: Explicit filterbank learning for neural image style transfer and image processing. TPAMI 43, 2373\u20132387 (2020)","DOI":"10.1109\/TPAMI.2020.2964205"},{"key":"7_CR5","doi-asserted-by":"crossref","unstructured":"Chen, H., et al.: Diverse image style transfer via invertible cross-space mapping. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01461"},{"key":"7_CR6","unstructured":"Chen, T.Q., Schmidt, M.: Fast patch-based style transfer of arbitrary style. In: NeurIPS Workshop on Constructive Machine Learning (2016)"},{"key":"7_CR7","doi-asserted-by":"crossref","unstructured":"Chen, Z., et al.: DPT: deformable patch-based transformer for visual recognition. In: ACM MM (2021)","DOI":"10.1145\/3474085.3475467"},{"key":"7_CR8","unstructured":"Ding, L., Wang, L., Liu, X., Wong, D.F., Tao, D., Tu, Z.: Understanding and improving lexical choice in non-autoregressive translation. In: ICLR (2021)"},{"key":"7_CR9","doi-asserted-by":"crossref","unstructured":"Ding, L., Wang, L., Tao, D.: Self-attention with cross-lingual position representation. In: ACL (2020)","DOI":"10.18653\/v1\/2020.acl-main.153"},{"key":"7_CR10","doi-asserted-by":"crossref","unstructured":"Ding, L., Wang, L., Wu, D., Tao, D., Tu, Z.: Context-aware cross-attention for non-autoregressive translation. In: COLING (2020)","DOI":"10.18653\/v1\/2020.coling-main.389"},{"key":"7_CR11","doi-asserted-by":"crossref","unstructured":"Gatys, L.A., Ecker, A.S., Bethge, M.: Image style transfer using convolutional neural networks. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.265"},{"key":"7_CR12","unstructured":"Hamilton, W.L., Ying, R., Leskovec, J.: Inductive representation learning on large graphs. In: NeurIPS (2017)"},{"key":"7_CR13","doi-asserted-by":"crossref","unstructured":"Hong, K., Jeon, S., Yang, H., Fu, J., Byun, H.: Domain-aware universal style transfer. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01434"},{"key":"7_CR14","doi-asserted-by":"crossref","unstructured":"Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.167"},{"key":"7_CR15","doi-asserted-by":"crossref","unstructured":"Huo, J., et al.: Manifold alignment for semantically aligned style transfer. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01459"},{"key":"7_CR16","doi-asserted-by":"crossref","unstructured":"Jing, Y., et al.: Dynamic instance normalization for arbitrary style transfer. In: AAAI (2020)","DOI":"10.1609\/aaai.v34i04.5862"},{"key":"7_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"244","DOI":"10.1007\/978-3-030-01261-8_15","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Jing","year":"2018","unstructured":"Jing, Y., et al.: Stroke controllable fast style transfer with adaptive receptive fields. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11217, pp. 244\u2013260. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01261-8_15"},{"key":"7_CR18","doi-asserted-by":"crossref","unstructured":"Jing, Y., Yang, Y., Feng, Z., Ye, J., Yu, Y., Song, M.: Neural style transfer: a review. TVCG 26, 3365\u20133385 (2019)","DOI":"10.1109\/TVCG.2019.2921336"},{"key":"7_CR19","doi-asserted-by":"crossref","unstructured":"Jing, Y., Yang, Y., Wang, X., Song, M., Tao, D.: Amalgamating knowledge from heterogeneous graph neural networks. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01545"},{"key":"7_CR20","doi-asserted-by":"crossref","unstructured":"Jing, Y., Yang, Y., Wang, X., Song, M., Tao, D.: Meta-aggregator: learning to aggregate for 1-bit graph neural networks. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00525"},{"key":"7_CR21","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"694","DOI":"10.1007\/978-3-319-46475-6_43","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J Johnson","year":"2016","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694\u2013711. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_43"},{"key":"7_CR22","doi-asserted-by":"crossref","unstructured":"Kalischek, N., Wegner, J.D., Schindler, K.: In the light of feature distributions: moment matching for neural style transfer. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00926"},{"key":"7_CR23","unstructured":"Kingma, D., Ba, J.: Adam: a method for stochastic optimization. In: ICLR (2015)"},{"key":"7_CR24","unstructured":"Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. In: ICLR (2017)"},{"key":"7_CR25","doi-asserted-by":"crossref","unstructured":"Kolkin, N., Salavon, J., Shakhnarovich, G.: Style transfer by relaxed optimal transport and self-similarity. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01029"},{"key":"7_CR26","doi-asserted-by":"crossref","unstructured":"Kong, Y., Liu, L., Wang, J., Tao, D.: Adaptive curriculum learning. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00502"},{"key":"7_CR27","doi-asserted-by":"crossref","unstructured":"Li, C., Wand, M.: Combining Markov random fields and convolutional neural networks for image synthesis. In: CVPR, pp. 2479\u20132486 (2016)","DOI":"10.1109\/CVPR.2016.272"},{"key":"7_CR28","doi-asserted-by":"crossref","unstructured":"Li, Y., Wang, N., Liu, J., Hou, X.: Demystifying neural style transfer. In: IJCAI (2017)","DOI":"10.24963\/ijcai.2017\/310"},{"key":"7_CR29","doi-asserted-by":"crossref","unstructured":"Li, Y., Chen, F., Yang, J., Wang, Z., Lu, X., Yang, M.H.: Diversified texture synthesis with feed-forward networks. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.36"},{"key":"7_CR30","unstructured":"Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., Yang, M.H.: Universal style transfer via feature transforms. In: NeurIPS (2017)"},{"key":"7_CR31","doi-asserted-by":"crossref","unstructured":"Liao, J., Yao, Y., Yuan, L., Hua, G., Kang, S.B.: Visual attribute transfer through deep image analogy. TOG 36, 1\u201315 (2017)","DOI":"10.1145\/3072959.3073683"},{"key":"7_CR32","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"7_CR33","doi-asserted-by":"crossref","unstructured":"Liu, H., Yang, Y., Wang, X.: Overcoming catastrophic forgetting in graph neural networks. In: AAAI (2021)","DOI":"10.1609\/aaai.v35i10.17049"},{"key":"7_CR34","doi-asserted-by":"crossref","unstructured":"Liu, S., et al.: Paint transformer: feed forward neural painting with stroke prediction. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00653"},{"key":"7_CR35","doi-asserted-by":"crossref","unstructured":"Liu, S., et al.: AdaAttN: revisit attention mechanism in arbitrary neural style transfer. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00658"},{"key":"7_CR36","doi-asserted-by":"crossref","unstructured":"Liu, X.C., Yang, Y.L., Hall, P.: Learning to warp for style transfer. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00370"},{"key":"7_CR37","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"800","DOI":"10.1007\/978-3-030-01264-9_47","volume-title":"Computer Vision \u2013 ECCV 2018","author":"R Mechrez","year":"2018","unstructured":"Mechrez, R., Talmi, I., Zelnik-Manor, L.: The contextual loss for image transformation with non-aligned data. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision \u2013 ECCV 2018. LNCS, vol. 11218, pp. 800\u2013815. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01264-9_47"},{"key":"7_CR38","unstructured":"Nichol, K.: Painter by numbers (2016). https:\/\/www.kaggle.com\/c\/painter-by-numbers"},{"key":"7_CR39","doi-asserted-by":"crossref","unstructured":"Ren, S., Zhou, D., He, S., Feng, J., Wang, X.: Shunted self-attention via multi-scale token aggregation. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01058"},{"key":"7_CR40","unstructured":"Risser, E., Wilmot, P., Barnes, C.: Stable and controllable neural texture synthesis and style transfer using histogram losses. arXiv preprint arXiv:1701.08893 (2017)"},{"key":"7_CR41","doi-asserted-by":"crossref","unstructured":"Shen, C., Yin, Y., Wang, X., Li, X., Song, J., Song, M.: Training generative adversarial networks in one stage. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00336"},{"key":"7_CR42","doi-asserted-by":"crossref","unstructured":"Sheng, L., Shao, J., Lin, Z., Warfield, S., Wang, X.: Avatar-Net: multi-scale zero-shot style transfer by feature decoration. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00860"},{"key":"7_CR43","unstructured":"Veli\u010dkovi\u0107, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., Bengio, Y.: Graph attention networks. In: ICLR (2018)"},{"key":"7_CR44","unstructured":"Wang, M., et al.: Deep graph library: towards efficient and scalable deep learning on graphs. In: ICLR Workshop (2019)"},{"key":"7_CR45","doi-asserted-by":"crossref","unstructured":"Wang, P., Li, Y., Vasconcelos, N.: Rethinking and improving the robustness of image style transfer. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00019"},{"key":"7_CR46","doi-asserted-by":"crossref","unstructured":"Wang, Y., Sun, Y., Liu, Z., Sarma, S.E., Bronstein, M.M., Solomon, J.M.: Dynamic graph CNN for learning on point clouds. TOG 38, 1\u201312 (2019)","DOI":"10.1145\/3326362"},{"key":"7_CR47","doi-asserted-by":"crossref","unstructured":"Wu, X., Hu, Z., Sheng, L., Xu, D.: StyleFormer: real-time arbitrary style transfer via parametric style composition. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01435"},{"key":"7_CR48","unstructured":"Xu, K., Hu, W., Leskovec, J., Jegelka, S.: How powerful are graph neural networks? In: ICLR (2019)"},{"key":"7_CR49","doi-asserted-by":"crossref","unstructured":"Xu, W., Long, C., Wang, R., Wang, G.: DRB-GAN: a dynamic ResBlock generative adversarial network for artistic style transfer. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00632"},{"key":"7_CR50","unstructured":"Xu, Y., Zhang, Q., Zhang, J., Tao, D.: ViTAE: vision transformer advanced by exploring intrinsic inductive bias. In: NeurIPS (2021)"},{"key":"7_CR51","unstructured":"Yang, Y., Feng, Z., Song, M., Wang, X.: Factorizable graph convolutional networks. In: NeurIPS (2020)"},{"key":"7_CR52","doi-asserted-by":"crossref","unstructured":"Yang, Y., Qiu, J., Song, M., Tao, D., Wang, X.: Distilling knowledge from graph convolutional networks. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00710"},{"key":"7_CR53","doi-asserted-by":"crossref","unstructured":"Yang, Y., Ren, Z., Li, H., Zhou, C., Wang, X., Hua, G.: Learning dynamics via graph neural networks for human pose estimation and tracking. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00798"},{"key":"7_CR54","doi-asserted-by":"crossref","unstructured":"Yang, Y., Wang, X., Song, M., Yuan, J., Tao, D.: SPAGAN: shortest path graph attention network. In: IJCAI (2019)","DOI":"10.24963\/ijcai.2019\/569"},{"key":"7_CR55","doi-asserted-by":"crossref","unstructured":"Ye, J., Jing, Y., Wang, X., Ou, K., Tao, D., Song, M.: Edge-sensitive human cutout with hierarchical granularity and loopy matting guidance. TIP 29, 1177\u20131191 (2019)","DOI":"10.1109\/TIP.2019.2930146"},{"key":"7_CR56","doi-asserted-by":"crossref","unstructured":"Yu, W., et al.: MetaFormer is actually what you need for vision. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01055"},{"key":"7_CR57","doi-asserted-by":"crossref","unstructured":"Zhan, Y., Yu, J., Yu, T., Tao, D.: On exploring undetermined relationships for visual relationship detection. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00527"},{"key":"7_CR58","doi-asserted-by":"crossref","unstructured":"Zhan, Y., Yu, J., Yu, T., Tao, D.: Multi-task compositional network for visual relationship detection. IJCV 128, 2146\u20132165 (2020)","DOI":"10.1007\/s11263-020-01353-8"},{"key":"7_CR59","unstructured":"Zhang, H., Dana, K.: Multi-style generative network for real-time transfer. arXiv preprint arXiv:1703.06953 (2017)"},{"key":"7_CR60","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Xu, Y., Zhang, J., Tao, D.: ViTAEv2: vision transformer advanced by exploring inductive bias for image recognition and beyond. arXiv preprint arXiv:2202.10108 (2022)","DOI":"10.1007\/s11263-022-01739-w"},{"key":"7_CR61","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Xu, Y., Zhang, J., Tao, D.: VSA: learning varied-size window attention in vision transformers. arXiv preprint arXiv:2204.08446 (2022)","DOI":"10.1007\/978-3-031-19806-9_27"},{"key":"7_CR62","doi-asserted-by":"crossref","unstructured":"Zhao, H., Bian, W., Yuan, B., Tao, D.: Collaborative learning of depth estimation, visual odometry and camera relocalization from monocular videos. In: IJCAI (2020)","DOI":"10.24963\/ijcai.2020\/68"},{"key":"7_CR63","unstructured":"Zhou, J., et al.: Graph neural networks: a review of methods and applications. arXiv preprint arXiv:1812.08434 (2018)"},{"key":"7_CR64","unstructured":"Zhou, S., Zhang, J., Zuo, W., Loy, C.C.: Cross-scale internal graph neural network for image super-resolution. In: NeurIPS (2020)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-20071-7_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,3,11]],"date-time":"2023-03-11T22:24:02Z","timestamp":1678573442000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-20071-7_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031200700","9783031200717"],"references-count":64,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-20071-7_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"13 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}