{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,13]],"date-time":"2025-12-13T09:46:43Z","timestamp":1765619203652,"version":"3.48.0"},"reference-count":63,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2025,8,8]],"date-time":"2025-08-08T00:00:00Z","timestamp":1754611200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,8,8]],"date-time":"2025-08-08T00:00:00Z","timestamp":1754611200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int. J. Mach. Learn. &amp; Cyber."],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s13042-025-02691-0","type":"journal-article","created":{"date-parts":[[2025,8,8]],"date-time":"2025-08-08T09:27:23Z","timestamp":1754645243000},"page":"10277-10293","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Cross-domain prompt-tuning for domain adaptation"],"prefix":"10.1007","volume":"16","author":[{"given":"Shuqin","family":"Wang","sequence":"first","affiliation":[]},{"given":"Yun","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yi","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Jipeng","family":"Qiang","sequence":"additional","affiliation":[]},{"given":"Yunhao","family":"Yuan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,8]]},"reference":[{"key":"2691_CR1","doi-asserted-by":"publisher","first-page":"151","DOI":"10.1007\/s10994-009-5152-4","volume":"79","author":"S Ben-David","year":"2010","unstructured":"Ben-David S, Blitzer J, Crammer K, Kulesza A, Pereira F, Vaughan JW (2010) A theory of learning from different domains. Mach Learn 79:151\u2013175","journal-title":"Mach Learn"},{"issue":"10","key":"2691_CR2","doi-asserted-by":"publisher","first-page":"1345","DOI":"10.1109\/TKDE.2009.191","volume":"22","author":"SJ Pan","year":"2010","unstructured":"Pan SJ, Yang Q (2010) A survey on transfer learning. IEEE Trans Knowl Data Eng 22(10):1345\u20131359","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"2691_CR3","unstructured":"Mansour Y, Mohri M, Rostamizadeh A (2009) Domain adaptation: Learning bounds and algorithms. arXiv preprint arXiv:0902.3430"},{"issue":"2","key":"2691_CR4","doi-asserted-by":"publisher","first-page":"199","DOI":"10.1109\/TNN.2010.2091281","volume":"22","author":"SJ Pan","year":"2011","unstructured":"Pan SJ, Tsang IW, Kwok JT, Yang Q (2011) Domain adaptation via transfer component analysis. IEEE Trans Neural Networks 22(2):199\u2013210","journal-title":"IEEE Trans Neural Networks"},{"issue":"1\u20132","key":"2691_CR5","first-page":"28","volume":"109","author":"H Judy","year":"2014","unstructured":"Judy H, Erik R, Jeff D, Trevor D, Kate S (2014) Asymmetric and category invariant feature transformations for domain adaptation. Int J Comput Vision 109(1\u20132):28\u201341","journal-title":"Int J Comput Vision"},{"issue":"07","key":"2691_CR6","first-page":"3897","volume":"10","author":"K Sharifani","year":"2023","unstructured":"Sharifani K, Amini M (2023) Machine learning and deep learning: A review of methods and applications. World Information Technology and Engineering Journal 10(07):3897\u20133904","journal-title":"World Information Technology and Engineering Journal"},{"issue":"3","key":"2691_CR7","doi-asserted-by":"publisher","first-page":"199","DOI":"10.1049\/iet-bmt.2017.0209","volume":"7","author":"FI Eyiokur","year":"2018","unstructured":"Eyiokur FI, Yaman D, Ekenel HK (2018) Domain adaptation for ear recognition using deep convolutional neural networks. iet Biometrics 7(3):199\u2013206","journal-title":"iet Biometrics"},{"key":"2691_CR8","doi-asserted-by":"publisher","first-page":"295","DOI":"10.1016\/j.neucom.2018.06.009","volume":"313","author":"Z Gao","year":"2018","unstructured":"Gao Z, Shen C, Xie C (2018) Stacked convolutional auto-encoders for single space target image blind deconvolution. Neurocomputing 313:295\u2013305","journal-title":"Neurocomputing"},{"issue":"3","key":"2691_CR9","doi-asserted-by":"publisher","first-page":"599","DOI":"10.1111\/coin.12225","volume":"35","author":"R Jin","year":"2019","unstructured":"Jin R, Lu L, Lee J, Usman A (2019) Multi-representational convolutional neural networks for text classification. Comput Intell 35(3):599\u2013609","journal-title":"Comput Intell"},{"key":"2691_CR10","doi-asserted-by":"crossref","unstructured":"Jaech A, Heck L, Ostendorf M (2016) Domain adaptation of recurrent neural networks for natural language understanding. arXiv preprint arXiv:1604.00117","DOI":"10.21437\/Interspeech.2016-1598"},{"key":"2691_CR11","doi-asserted-by":"crossref","unstructured":"Ding Y, Yu J, Jiang J (2017) Recurrent neural networks with auxiliary labels for cross-domain opinion target extraction. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31","DOI":"10.1609\/aaai.v31i1.11014"},{"key":"2691_CR12","doi-asserted-by":"crossref","unstructured":"Tzeng E, Hoffman J, Saenko K, Darrell T (2017) Adversarial discriminative domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7167\u20137176","DOI":"10.1109\/CVPR.2017.316"},{"key":"2691_CR13","doi-asserted-by":"crossref","unstructured":"Cao Z, Ma L, Long M, Wang J (2018) Partial adversarial domain adaptation. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 135\u2013150","DOI":"10.1007\/978-3-030-01237-3_9"},{"key":"2691_CR14","unstructured":"Long M, Cao Z, Wang J, Jordan MI (2018) Conditional adversarial domain adaptation. Advances in neural information processing systems 31"},{"key":"2691_CR15","unstructured":"Glorot X, Bordes A, Bengio Y (2011) Domain adaptation for large-scale sentiment classification: A deep learning approach. In: Proceedings of the 28th International Conference on Machine Learning (ICML-11), pp. 513\u2013520"},{"key":"2691_CR16","unstructured":"Chen M, Xu Z, Weinberger K, Sha F (2012) Marginalized denoising autoencoders for domain adaptation. arXiv preprint arXiv:1206.4683"},{"key":"2691_CR17","doi-asserted-by":"crossref","unstructured":"Kim Y (2014) Convolutional neural networks for sentence classification. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing, EMNLP, pp. 1746\u20131751","DOI":"10.3115\/v1\/D14-1181"},{"key":"2691_CR18","doi-asserted-by":"publisher","first-page":"182","DOI":"10.1016\/j.aiopen.2022.11.003","volume":"3","author":"X Han","year":"2022","unstructured":"Han X, Zhao W, Ding N, Liu Z, Sun M (2022) Ptr: Prompt tuning with rules for text classification. AI Open 3:182\u2013192","journal-title":"AI Open"},{"key":"2691_CR19","unstructured":"Devlin J, Chang M-W, Lee K, Toutanova K (2018) Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805"},{"key":"2691_CR20","unstructured":"Ding N, Wang X, Fu Y, Xu G, Wang R, Xie P, Shen Y, Huang F, Zheng H-T, Zhang R (2021) Prototypical representation learning for relation extraction. arXiv preprint arXiv:2103.11647"},{"issue":"1","key":"2691_CR21","first-page":"5485","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel C, Shazeer N, Roberts A, Lee K, Narang S, Matena M, Zhou Y, Li W, Liu PJ (2020) Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research 21(1):5485\u20135551","journal-title":"The Journal of Machine Learning Research"},{"key":"2691_CR22","unstructured":"Radford A, Narasimhan K, Salimans T, Sutskever I, et al (2018) Improving language understanding by generative pre-training"},{"key":"2691_CR23","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.118534","volume":"213","author":"I Ameer","year":"2023","unstructured":"Ameer I, B\u00f6l\u00fcc\u00fc N, Siddiqui MHF, Can B, Sidorov G, Gelbukh A (2023) Multi-label emotion classification in texts using transfer learning. Expert Syst Appl 213:118534","journal-title":"Expert Syst Appl"},{"key":"2691_CR24","first-page":"1029","volume":"2021","author":"P Yu","year":"2021","unstructured":"Yu P, Fei H, Li P (2021) Cross-lingual language model pretraining for retrieval. Proc Web Conf 2021:1029\u20131039","journal-title":"Proc Web Conf"},{"key":"2691_CR25","unstructured":"Song Y, Wang J, Liang Z, Liu Z, Jiang T (2020) Utilizing bert intermediate layers for aspect based sentiment analysis and natural language inference. arXiv preprint arXiv:2002.04815"},{"key":"2691_CR26","first-page":"8649","volume":"34","author":"J Qiang","year":"2020","unstructured":"Qiang J, Li Y, Zhu Y, Yuan Y, Wu X (2020) Lexical simplification with pretrained encoders. Proc AAAI Conf Artif Intell 34:8649\u20138656","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"2691_CR27","doi-asserted-by":"crossref","unstructured":"Zhang T, Wang D, Chen H, Zeng Z, Guo W, Miao C, Cui L (2020) Bdann: Bert-based domain adaptation neural network for multi-modal fake news detection. In: Proceedings of the International Joint Conference on Neural Networks (IJCNN), pp. 1\u20138","DOI":"10.1109\/IJCNN48605.2020.9206973"},{"key":"2691_CR28","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.121669","volume":"238","author":"S Xie","year":"2024","unstructured":"Xie S, Pan Q, Wang X, Luo X, Sugumaran V (2024) Combining prompt learning with contextual semantics for inductive relation prediction. Expert Syst Appl 238:121669","journal-title":"Expert Syst Appl"},{"key":"2691_CR29","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A et al (2020) Language models are few-shot learners. Adv Neural Inf Process Syst 33:1877\u20131901","journal-title":"Adv Neural Inf Process Syst"},{"issue":"9","key":"2691_CR30","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3560815","volume":"55","author":"P Liu","year":"2023","unstructured":"Liu P, Yuan W, Fu J, Jiang Z, Hayashi H, Neubig G (2023) Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Comput Surv 55(9):1\u201335","journal-title":"ACM Comput Surv"},{"key":"2691_CR31","doi-asserted-by":"crossref","unstructured":"Schick T, Schmid H, Sch\u00fctze H (2020) Automatically identifying words that can serve as labels for few-shot text classification. arXiv preprint arXiv:2010.13641","DOI":"10.18653\/v1\/2020.coling-main.488"},{"key":"2691_CR32","doi-asserted-by":"crossref","unstructured":"Hu S, Ding N, Wang H, Liu Z, Wang J, Li J, Wu W, Sun M (2022) Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 2225\u20132240. Association for Computational Linguistics, ???","DOI":"10.18653\/v1\/2022.acl-long.158"},{"key":"2691_CR33","doi-asserted-by":"crossref","unstructured":"Vu T, Lester B, Constant N, Al-Rfou R, Cer D (2021) Spot: Better frozen model adaptation through soft prompt transfer. arXiv preprint arXiv:2110.07904","DOI":"10.18653\/v1\/2022.acl-long.346"},{"key":"2691_CR34","doi-asserted-by":"crossref","unstructured":"Blitzer J, McDonald R, Pereira F (2006) Domain adaptation with structural correspondence learning. In: Proceedings of the 2006 Conference on Empirical Methods in Natural Language Processing, pp. 120\u2013128","DOI":"10.3115\/1610075.1610094"},{"issue":"2","key":"2691_CR35","doi-asserted-by":"publisher","first-page":"199","DOI":"10.1109\/TNN.2010.2091281","volume":"22","author":"SJ Pan","year":"2010","unstructured":"Pan SJ, Tsang IW, Kwok JT, Yang Q (2010) Domain adaptation via transfer component analysis. IEEE Trans Neural Networks 22(2):199\u2013210","journal-title":"IEEE Trans Neural Networks"},{"issue":"8","key":"2691_CR36","doi-asserted-by":"publisher","first-page":"1700","DOI":"10.1109\/TCSVT.2016.2539541","volume":"27","author":"J Li","year":"2016","unstructured":"Li J, Wu Y, Lu K (2016) Structured domain adaptation. IEEE Trans Circuits Syst Video Technol 27(8):1700\u20131713","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"2691_CR37","doi-asserted-by":"crossref","unstructured":"Sun B, Feng J, Saenko K (2016) Return of frustratingly easy domain adaptation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 30","DOI":"10.1609\/aaai.v30i1.10306"},{"key":"2691_CR38","unstructured":"Tzeng E, Hoffman J, Zhang N, Saenko K, Darrell T (2014) Deep domain confusion: Maximizing for domain invariance. arXiv preprint arXiv:1412.3474"},{"key":"2691_CR39","unstructured":"Long M, Cao Y, Wang J, Jordan M (2015) Learning transferable features with deep adaptation networks. In: International Conference on Machine Learning, pp. 97\u2013105. PMLR"},{"key":"2691_CR40","unstructured":"Long M, Zhu H, Wang J, Jordan MI (2016) Unsupervised domain adaptation with residual transfer networks. Advances in neural information processing systems 29"},{"key":"2691_CR41","unstructured":"Long M, Zhu H, Wang J, Jordan MI (2017) Deep transfer learning with joint adaptation networks. In: International Conference on Machine Learning, pp. 2208\u20132217. PMLR"},{"issue":"11","key":"2691_CR42","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2020) Generative adversarial networks. Commun ACM 63(11):139\u2013144","journal-title":"Commun ACM"},{"key":"2691_CR43","doi-asserted-by":"crossref","unstructured":"Zhang Y, Qiu Z, Yao T, Liu D, Mei T (2018) Fully convolutional adaptation networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6810\u20136818","DOI":"10.1109\/CVPR.2018.00712"},{"key":"2691_CR44","doi-asserted-by":"crossref","unstructured":"Pei Z, Cao Z, Long M, Wang J (2018) Multi-adversarial domain adaptation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32","DOI":"10.1609\/aaai.v32i1.11767"},{"key":"2691_CR45","unstructured":"Wei P, Ke Y, Goh CK (2016) Deep nonlinear feature coding for unsupervised domain adaptation. In: IJCAI, pp. 2189\u20132195"},{"issue":"5","key":"2691_CR46","doi-asserted-by":"publisher","first-page":"1321","DOI":"10.1109\/TNNLS.2018.2868709","volume":"30","author":"Pengfei Wei","year":"2018","unstructured":"Wei Pengfei, Ke Yiping, Goh Chi Keong (2018) Feature analysis of marginalized stacked denoising autoenconder for unsupervised domain adaptation. IEEE Trans Neural Netw Learn Syst 30(5):1321\u20131334","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"2691_CR47","first-page":"9073","volume":"34","author":"M Tu","year":"2020","unstructured":"Tu M, Huang K, Wang G, Huang J, He X, Zhou B (2020) Select, answer and explain: Interpretable multi-hop reading comprehension over multiple documents. Proc AAAI Conf Artif Intell 34:9073\u20139080","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"2691_CR48","doi-asserted-by":"crossref","unstructured":"Ramponi A, Plank B (2020) Neural unsupervised domain adaptation in nlp\u2014a survey. arXiv preprint arXiv:2006.00632","DOI":"10.18653\/v1\/2020.coling-main.603"},{"key":"2691_CR49","doi-asserted-by":"crossref","unstructured":"Gururangan S, Marasovi\u0107 A, Swayamdipta S, Lo K, Beltagy I, Downey D, Smith NA (2020) Don\u2019t stop pretraining: Adapt language models to domains and tasks. arXiv preprint arXiv:2004.10964","DOI":"10.18653\/v1\/2020.acl-main.740"},{"key":"2691_CR50","doi-asserted-by":"crossref","unstructured":"Basabain S, Cambria E, Alomar K, Hussain A (2023) Enhancing arabic-text feature extraction utilizing label-semantic augmentation in few\/zero-shot learning. Expert Systems, 13329","DOI":"10.1111\/exsy.13329"},{"key":"2691_CR51","unstructured":"Guo X, Yu H (2022) On the domain adaptation and generalization of pretrained language models: A survey. arXiv preprint arXiv:2211.03154"},{"key":"2691_CR52","doi-asserted-by":"crossref","unstructured":"Petroni F, Rockt\u00e4schel T, Lewis P, Bakhtin A, Wu Y, Miller AH, Riedel S (2019) Language models as knowledge bases? arXiv preprint arXiv:1909.01066","DOI":"10.18653\/v1\/D19-1250"},{"key":"2691_CR53","doi-asserted-by":"crossref","unstructured":"Schick T, Sch\u00fctze H (2021) Few-shot text generation with natural language instructions. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 390\u2013402","DOI":"10.18653\/v1\/2021.emnlp-main.32"},{"key":"2691_CR54","doi-asserted-by":"crossref","unstructured":"Shin T, Razeghi Y, Logan\u00a0IV RL, Wallace E, Singh S (2020) Autoprompt: Eliciting knowledge from language models with automatically generated prompts. arXiv preprint arXiv:2010.15980","DOI":"10.18653\/v1\/2020.emnlp-main.346"},{"key":"2691_CR55","unstructured":"Liu X, Zheng Y, Du Z, Ding M, Qian Y, Yang Z, Tang J (2021) Gpt understands, too. arXiv preprint arXiv:2103.10385"},{"key":"2691_CR56","unstructured":"Gao T, Fisch A, Chen D (2020) Making pre-trained language models better few-shot learners. arXiv preprint arXiv:2012.15723"},{"key":"2691_CR57","doi-asserted-by":"crossref","unstructured":"Wang Y, Wang W, Chen Q, Huang K, Nguyen A, De S (2023) Prompt-based zero-shot text classification with conceptual knowledge. In: Padmakumar, V., Vallejo, G., Fu, Y. (eds.) Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop), pp. 30\u201338. Association for Computational Linguistics, Toronto, Canada","DOI":"10.18653\/v1\/2023.acl-srw.4"},{"key":"2691_CR58","unstructured":"Zhu Y, Zhou X, Qiang J, Li Y, Yuan Y, Wu X (2022) Prompt-learning for short text classification. arXiv preprint arXiv:2202.11345"},{"key":"2691_CR59","doi-asserted-by":"crossref","unstructured":"Meng Y, Zhang Y, Huang J, Xiong C, Ji H, Zhang C, Han J (2020) Text classification using label names only: A language model self-training approach. In: Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 9006\u20139017","DOI":"10.18653\/v1\/2020.emnlp-main.724"},{"key":"2691_CR60","doi-asserted-by":"crossref","unstructured":"Cai R, Li Z, Wei P, Qiao J, Zhang K, Hao Z (2019) Learning disentangled semantic representation for domain adaptation. IJCAI\u201919, pp. 2060\u20132066. AAAI Press","DOI":"10.24963\/ijcai.2019\/285"},{"issue":"5","key":"2691_CR61","doi-asserted-by":"publisher","DOI":"10.1007\/s11704-022-1349-5","volume":"17","author":"Y Zhu","year":"2023","unstructured":"Zhu Y, Wu X, Qiang J, Yuan Y, Li Y (2023) Representation learning via an integrated autoencoder for unsupervised domain adaptation. Front Comp Sci 17(5):175334","journal-title":"Front Comp Sci"},{"key":"2691_CR62","doi-asserted-by":"crossref","unstructured":"Ding N, Hu S, Zhao W, Chen Y, Liu Z, Zheng H-T, Sun M (2021) Openprompt: An open-source framework for prompt-learning. arXiv preprint arXiv:2111.01998","DOI":"10.18653\/v1\/2022.acl-demo.10"},{"key":"2691_CR63","unstructured":"Jiang AQ, Sablayrolles A, Mensch A, Bamford C, Chaplot DS, Casas Ddl, Bressand F, Lengyel G, Lample G, Saulnier L, et al (2023) Mistral 7b. arXiv preprint arXiv:2310.06825"}],"container-title":["International Journal of Machine Learning and Cybernetics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-025-02691-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13042-025-02691-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-025-02691-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,13]],"date-time":"2025-12-13T09:41:50Z","timestamp":1765618910000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13042-025-02691-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,8]]},"references-count":63,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["2691"],"URL":"https:\/\/doi.org\/10.1007\/s13042-025-02691-0","relation":{},"ISSN":["1868-8071","1868-808X"],"issn-type":[{"type":"print","value":"1868-8071"},{"type":"electronic","value":"1868-808X"}],"subject":[],"published":{"date-parts":[[2025,8,8]]},"assertion":[{"value":"3 February 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 May 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 August 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}