{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T14:05:14Z","timestamp":1774447514186,"version":"3.50.1"},"reference-count":65,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2020AAA0108004"],"award-info":[{"award-number":["2020AAA0108004"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100013114","name":"Yunnan Key Research and Development Program","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100013114","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100018637","name":"Yunnan Key Research and Development Program","doi-asserted-by":"publisher","award":["202203AA080004"],"award-info":[{"award-number":["202203AA080004"]}],"id":[{"id":"10.13039\/501100018637","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Expert Systems with Applications"],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1016\/j.eswa.2025.130943","type":"journal-article","created":{"date-parts":[[2025,12,26]],"date-time":"2025-12-26T07:37:17Z","timestamp":1766734637000},"page":"130943","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["AlignExperts: Knowledge alignment experts for summarization faithfulness evaluation"],"prefix":"10.1016","volume":"307","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-5012-3706","authenticated-orcid":false,"given":"Jiuyi","family":"Li","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8860-7805","authenticated-orcid":false,"given":"Degen","family":"Huang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0000-4615-3318","authenticated-orcid":false,"given":"Junpeng","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-9436-8611","authenticated-orcid":false,"given":"Jinshuang","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.eswa.2025.130943_bib0001","unstructured":"Cao, J., Lin, T., He, H., Yan, R., Zhang, W., Li, J., Zhang, D., Tang, S., & Zhuang, Y. (2025). Moa: Heterogeneous mixture of adapters for parameter-efficient fine-tuning of LLMs. arXiv preprint arXiv: 2506.05928,. https:\/\/arxiv.org\/pdf\/2506.05928."},{"key":"10.1016\/j.eswa.2025.130943_bib0002","series-title":"Proceedings of the 60th annual meeting of the association for computational linguistics (volume 1: Long papers)","first-page":"3340","article-title":"Hallucinated but factual! inspecting the factuality of hallucinations in abstractive summarization","author":"Cao","year":"2022"},{"key":"10.1016\/j.eswa.2025.130943_bib0003","unstructured":"Chung, H. W., Hou, L., Longpre, S., Zoph, B., Tay, Y., Fedus, W., Li, Y., Wang, X., Dehghani, M., Brahma, S., Webson, A., Gu, S. S., Dai, Z., Suzgun, M., Chen, X., Chowdhery, A., Castro-Ros, A., Pellat, M., Robinson, K., Valter, D., Narang, S., Mishra, G., Yu, A., Zhao, V., Huang, Y., Dai, A., Yu, H., Petrov, S., Chi, E. H., Dean, J., Devlin, J., Roberts, A., Zhou, D., Le, Q. V., & Wei, J. (2022). Scaling instruction-finetuned language models. 10.48550\/arXiv.2210.11416."},{"key":"10.1016\/j.eswa.2025.130943_bib0004","unstructured":"Cong, P., Liu, W., Yu, W., Zhao, H., & Yang, T. (2025). Rank also matters: Hierarchical configuration for mixture of adapter experts in LLM fine-tuning. https:\/\/arxiv.org\/pdf\/2502.03884."},{"key":"10.1016\/j.eswa.2025.130943_bib0005","unstructured":"Demszky, D., Guu, K., & Liang, P. (2018). Transforming question answering datasets into natural language inference datasets. arXiv preprint arXiv: 1809.02922,."},{"key":"10.1016\/j.eswa.2025.130943_bib0006","series-title":"Proceedings of the 2021 conference on empirical methods in natural language processing","first-page":"7580","article-title":"Compression, transduction, and creation: A unified framework for evaluating natural language generation","author":"Deng","year":"2021"},{"key":"10.1016\/j.eswa.2025.130943_bib0007","series-title":"Findings of the association for computational linguistics: ACL 2022","first-page":"3759","article-title":"Benchmarking answer verification methods for question answering-based summarization evaluation metrics","author":"Deutsch","year":"2022"},{"key":"10.1016\/j.eswa.2025.130943_bib0008","series-title":"Proceedings of the 2019 conference of the north American chapter of the association for computational linguistics: human language technologies, volume 1 (long and short papers)","first-page":"4171","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2019"},{"key":"10.1016\/j.eswa.2025.130943_bib0009","series-title":"Proceedings of the 61st annual meeting of the association for computational linguistics (volume 1: Long papers), Toronto, Canada","first-page":"5113","article-title":"Mixture-of-domain-adapters: Decoupling and injecting domain knowledge to pre-trained language models\u2019 memories","author":"Diao","year":"2023"},{"key":"10.1016\/j.eswa.2025.130943_bib0010","series-title":"Proceedings of the 58th annual meeting of the association for computational linguistics","first-page":"5055","article-title":"FEQA: A question answering evaluation framework for faithfulness assessment in abstractive summarization","author":"Durmus","year":"2020"},{"key":"10.1016\/j.eswa.2025.130943_bib0011","series-title":"Proceedings of the 2022 conference of the north american chapter of the association for computational linguistics: human language technologies","first-page":"2587","article-title":"QAFActeval: Improved QA-based factual consistency evaluation for summarization","author":"Fabbri","year":"2022"},{"key":"10.1016\/j.eswa.2025.130943_bib0012","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2024.126031","article-title":"Your data is not perfect: Towards cross-domain out-of-distribution detection in class-imbalanced data","volume":"267","author":"Fang","year":"2025","journal-title":"Expert Systems with Applications"},{"issue":"02","key":"10.1016\/j.eswa.2025.130943_bib0013","doi-asserted-by":"crossref","first-page":"192","DOI":"10.1109\/TAI.2021.3116546","article-title":"ANIMC: A soft approach for autoweighted noisy and incomplete multiview clustering","volume":"3","author":"Fang","year":"2022","journal-title":"IEEE Transactions on Artificial Intelligence"},{"key":"10.1016\/j.eswa.2025.130943_bib0014","doi-asserted-by":"crossref","first-page":"7517","DOI":"10.1109\/TMM.2022.3222965","article-title":"Multi-modal cross-domain alignment network for video moment retrieval","volume":"25","author":"Fang","year":"2023","journal-title":"IEEE Transactions on Multimedia"},{"key":"10.1016\/j.eswa.2025.130943_bib0015","doi-asserted-by":"crossref","first-page":"3263","DOI":"10.1109\/TMM.2023.3309551","article-title":"Hierarchical local-global transformer for temporal sentence grounding","volume":"26","author":"Fang","year":"2024","journal-title":"IEEE Transactions on Multimedia"},{"key":"10.1016\/j.eswa.2025.130943_bib0016","series-title":"Proceedings of the 2023 conference on empirical methods in natural language processing","first-page":"2053","article-title":"Trueteacher: Learning factual consistency evaluation with large language models","author":"Gekhman","year":"2023"},{"key":"10.1016\/j.eswa.2025.130943_bib0017","series-title":"Findings of the association for computational linguistics: EMNLP 2020","first-page":"3592","article-title":"Evaluating factuality in generation with dependency-level entailment","author":"Goyal","year":"2020"},{"key":"10.1016\/j.eswa.2025.130943_bib0018","series-title":"Proceedings of the 2022 conference on empirical methods in natural language processing","first-page":"444","article-title":"SNaC: Coherence error detection for narrative summarization","author":"Goyal","year":"2022"},{"key":"10.1016\/j.eswa.2025.130943_bib0019","series-title":"Proceedings of the NAACL","article-title":"Annotation artifacts in natural language inference datasets","author":"Gururangan","year":"2018"},{"key":"10.1016\/j.eswa.2025.130943_bib0020","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2025.126712","article-title":"Halucheck: Explainable and verifiable automation for detecting hallucinations in LLM responses","volume":"272","author":"Heo","year":"2025","journal-title":"Expert Systems with Applications"},{"key":"10.1016\/j.eswa.2025.130943_bib0021","series-title":"Advances in neural information processing systems","first-page":"1693","article-title":"Teaching machines to read and comprehend","volume":"vol. 28","author":"Hermann","year":"2015"},{"key":"10.1016\/j.eswa.2025.130943_bib0022","unstructured":"Houlsby, N., Giurgiu, A., Jastrzebski, S., Morrone, B., de Laroussilhe, Q., Gesmundo, A., Attariyan, M., & Gelly, S. (2019a). Parameter-efficient transfer learning for NLP. arXiv preprint arXiv: 1902.00751,. https:\/\/arxiv.org\/abs\/1902.00751."},{"key":"10.1016\/j.eswa.2025.130943_bib0023","article-title":"Parameter-efficient transfer learning for NLP","volume":"abs\/1902.00751","author":"Houlsby","year":"2019","journal-title":"CoRR"},{"key":"10.1016\/j.eswa.2025.130943_bib0024","series-title":"Proceedings of the 2023 conference on empirical methods in natural language processing","first-page":"11017","article-title":"Zero-shot faithfulness evaluation for text summarization with foundation language model","author":"Jia","year":"2023"},{"key":"10.1016\/j.eswa.2025.130943_bib0025","series-title":"Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP)","first-page":"9332","article-title":"Evaluating the factual consistency of abstractive text summarization","author":"Kryscinski","year":"2020"},{"key":"10.1016\/j.eswa.2025.130943_bib0026","series-title":"Proceedings of the 2023 conference on empirical methods in natural language processing","first-page":"9662","article-title":"SummEdits: Measuring LLM ability at factual reasoning through the lens of summarization","author":"Laban","year":"2023"},{"key":"10.1016\/j.eswa.2025.130943_bib0027","doi-asserted-by":"crossref","first-page":"163","DOI":"10.1162\/tacl_a_00453","article-title":"SummaC: Re-visiting NLI-based models for inconsistency detection in summarization","volume":"10","author":"Laban","year":"2022","journal-title":"Transactions of the Association for Computational Linguistics"},{"key":"10.1016\/j.eswa.2025.130943_bib0028","series-title":"Proceedings of the 62nd annual meeting of the association for computational linguistics (volume 1: Long papers)","first-page":"10879","article-title":"The dawn after the dark: An empirical study on factuality hallucination in large language models","author":"Li","year":"2024"},{"key":"10.1016\/j.eswa.2025.130943_bib0029","series-title":"Text summarization branches out","first-page":"74","article-title":"ROUGE: A package for automatic evaluation of summaries","author":"Lin","year":"2004"},{"key":"10.1016\/j.eswa.2025.130943_bib0030","series-title":"Proceedings of the 60th annual meeting of the association for computational linguistics (volume 1: Long papers)","first-page":"6723","article-title":"A token-level reference-free hallucination detection benchmark for free-form text generation","author":"Liu","year":"2022"},{"key":"10.1016\/j.eswa.2025.130943_bib0031","series-title":"Proceedings of the ICCV","article-title":"Conflict-averse gradient descent for multi-task learning","author":"Liu","year":"2021"},{"key":"10.1016\/j.eswa.2025.130943_bib0032","series-title":"Proceedings of the EMNLP","article-title":"On the incompatibility of cross-task features in multi-task learning","author":"Liu","year":"2022"},{"key":"10.1016\/j.eswa.2025.130943_bib0033","unstructured":"Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., & Stoyanov, V. (2019). RoBERTa: A robustly optimized BERT pretraining approach. 10.48550\/arXiv.1907.11692."},{"key":"10.1016\/j.eswa.2025.130943_bib0034","unstructured":"Luo, Z., Xie, Q., & Ananiadou, S. (2023). ChatGPT as a factual inconsistency evaluator for text summarization. 10.48550\/arXiv.2303.15621."},{"key":"10.1016\/j.eswa.2025.130943_bib0035","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2024.124456","article-title":"Factual consistency evaluation of summarization in the era of large language models","volume":"254","author":"Luo","year":"2024","journal-title":"Expert Systems with Applications"},{"key":"10.1016\/j.eswa.2025.130943_bib0036","series-title":"Proceedings of the 58th annual meeting of the association for computational linguistics","first-page":"1906","article-title":"On faithfulness and factuality in abstractive summarization","author":"Maynez","year":"2020"},{"key":"10.1016\/j.eswa.2025.130943_bib0037","series-title":"Proceedings of the ACL","article-title":"Right for the wrong reasons: Diagnosing syntactic heuristics in natural language inference","author":"McCoy","year":"2019"},{"key":"10.1016\/j.eswa.2025.130943_bib0038","series-title":"Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (volume 1: Long papers)","first-page":"6881","article-title":"Improving factual consistency of abstractive summarization via question answering","author":"Nan","year":"2021"},{"key":"10.1016\/j.eswa.2025.130943_bib0039","series-title":"Proceedings of the 2018 conference on empirical methods in natural language processing","first-page":"1797","article-title":"Don\u2019t give me the details, just the summary! topic-aware convolutional neural networks for extreme summarization","author":"Narayan","year":"2018"},{"key":"10.1016\/j.eswa.2025.130943_bib0040","unstructured":"Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C. L., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., Schulman, J., Hilton, J., Kelton, F., Miller, L., Simens, M., Askell, A., Welinder, P., Christiano, P., Leike, J., & Lowe, R. (2022). Training language models to follow instructions with human feedback. 10.48550\/arXiv.2203.02155."},{"key":"10.1016\/j.eswa.2025.130943_bib0041","unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., & Liu, P. J. (2023). Exploring the limits of transfer learning with a unified text-to-text transformer. 10.48550\/arXiv.1910.10683."},{"key":"10.1016\/j.eswa.2025.130943_bib0042","series-title":"Proceedings of the 62nd annual meeting of the association for computational linguistics (volume 1: Long papers)","first-page":"3052","article-title":"MELoRA: Mini-ensemble low-rank adapters for parameter-efficient fine-tuning","author":"Ren","year":"2024"},{"key":"10.1016\/j.eswa.2025.130943_bib0043","series-title":"Proceedings of the 2022 conference of the north american chapter of the association for computational linguistics: Human language technologies","first-page":"3238","article-title":"FactGraph: Evaluating factuality in summarization with semantic graph representations","author":"Ribeiro","year":"2022"},{"key":"10.1016\/j.eswa.2025.130943_bib0044","unstructured":"Sansford, H., Richardson, N., Maretic, H. P., & Saada, J. N. (2024). Grapheval: A knowledge-graph based LLM hallucination evaluation framework. 10.48550\/arXiv.2407.10793."},{"key":"10.1016\/j.eswa.2025.130943_bib0045","series-title":"Proceedings of the 2021 conference on empirical methods in natural language processing","first-page":"6594","article-title":"QuestEval: Summarization asks for fact-based evaluation","author":"Scialom","year":"2021"},{"key":"10.1016\/j.eswa.2025.130943_bib0046","series-title":"Proceedings of the neurIPS","article-title":"Multi-task learning as multi-objective optimization","author":"Sener","year":"2018"},{"key":"10.1016\/j.eswa.2025.130943_bib0047","series-title":"Proceedings of the 62nd annual meeting of the association for computational linguistics (volume 2: Short papers)","first-page":"530","article-title":"The probabilities also matter: A more faithful metric for faithfulness of free-text explanations in large language models","author":"Siegel","year":"2024"},{"key":"10.1016\/j.eswa.2025.130943_bib0048","series-title":"Proceedings of the 62nd annual meeting of the association for computational linguistics (volume 1: Long papers)","first-page":"906","article-title":"FineSurE: Fine-grained summarization evaluation using LLMs","author":"Song","year":"2024"},{"key":"10.1016\/j.eswa.2025.130943_bib0049","series-title":"Proceedings of the 2024 conference of the north American chapter of the association for computational linguistics: human language technologies (volume 1: Long papers)","first-page":"4455","article-title":"Tofueval: Evaluating hallucinations of LLMs on topic-focused dialogue summarization","author":"Tang","year":"2024"},{"key":"10.1016\/j.eswa.2025.130943_bib0050","unstructured":"Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux, M.-A., Lacroix, T., Rozi\u00e9re, B., Goyal, N., Hambro, E., Azhar, F., Rodriguez, A., Joulin, A., Grave, E., & Lample, G. (2023). Llama: Open and efficient foundation language models. 10.48550\/arXiv.2302.13971."},{"key":"10.1016\/j.eswa.2025.130943_bib0051","series-title":"Proceedings of the 58th annual meeting of the association for computational linguistics","first-page":"5008","article-title":"Asking and answering questions to evaluate the factual consistency of summaries","author":"Wang","year":"2020"},{"key":"10.1016\/j.eswa.2025.130943_bib0052","unstructured":"Wang, R., Tang, D., Duan, N., Wei, Z., Huang, X., ji, J., Cao, G., Jiang, D., & Zhou, M. (2020b). K-adapter: Infusing knowledge into pre-trained models with adapters. https:\/\/arxiv.org\/2002.01808."},{"key":"10.1016\/j.eswa.2025.130943_bib0053","series-title":"Proceedings of the 61st annual meeting of the association for computational linguistics (volume 1: Long papers), Toronto, Canada","first-page":"8640","article-title":"Element-aware summarization with large language models: Expert-aligned evaluation and chain-of-thought method","author":"Wang","year":"2023"},{"key":"10.1016\/j.eswa.2025.130943_bib0054","article-title":"Optimal transport with mamba for multimodal inertial signal enhancement","author":"Wang","year":"2025","journal-title":"IEEE Transactions on Industrial Informatics"},{"key":"10.1016\/j.eswa.2025.130943_bib0055","series-title":"Proceedings of the 2018 conference of the north american chapter of the association for computational linguistics: human language technologies, volume 1 (long papers)","first-page":"1112","article-title":"A broad-coverage challenge corpus for sentence understanding through inference","author":"Williams","year":"2018"},{"key":"10.1016\/j.eswa.2025.130943_bib0056","series-title":"Proceedings of the 2024 conference of the north American chapter of the association for computational linguistics: human language technologies (volume 1: Long papers)","first-page":"8665","article-title":"Hallucination diversity-aware active learning for text summarization","author":"Xia","year":"2024"},{"key":"10.1016\/j.eswa.2025.130943_bib0057","unstructured":"Xie, Y., Wang, C., Yan, J., Zhou, J., Deng, F., & Huang, J. (2025). Making small language models better multi-task learners with mixture-of-task-adapters. https:\/\/arxiv.org\/pdf\/2309.11042."},{"key":"10.1016\/j.eswa.2025.130943_bib0058","series-title":"Advances in neural information processing systems","first-page":"27263","article-title":"Bartscore: Evaluating generated text as text generation","volume":"vol. 34","author":"Yuan","year":"2021"},{"key":"10.1016\/j.eswa.2025.130943_bib0059","series-title":"Iclr","article-title":"Pushing mixture of experts to the limit: Extremely parameter efficient moe for instruction tuning","author":"Zadouri","year":"2023"},{"key":"10.1016\/j.eswa.2025.130943_bib0060","series-title":"Proceedings of the 2021 conference on empirical methods in natural language processing","first-page":"4102","article-title":"Gradient-based adversarial factual consistency evaluation for abstractive summarization","author":"Zeng","year":"2021"},{"key":"10.1016\/j.eswa.2025.130943_bib0061","series-title":"Proceedings of the 61st annual meeting of the association for computational linguistics (volume 1: Long papers)","first-page":"11328","article-title":"Alignscore: Evaluating factual consistency with a unified alignment function","author":"Zha","year":"2023"},{"key":"10.1016\/j.eswa.2025.130943_bib0062","series-title":"Proceedings of the 2021 conference on empirical methods in natural language processing","first-page":"107","article-title":"Fine-grained factual consistency assessment for abstractive summarization models","author":"Zhang","year":"2021"},{"key":"10.1016\/j.eswa.2025.130943_bib0063","series-title":"International conference on learning representations","article-title":"Bertscore: Evaluating text generation with bert","author":"Zhang*","year":"2020"},{"key":"10.1016\/j.eswa.2025.130943_bib0064","series-title":"Proceedings of the 2022 conference on empirical methods in natural language processing","first-page":"2023","article-title":"Towards a unified multi-dimensional evaluator for text generation","author":"Zhong","year":"2022"},{"key":"10.1016\/j.eswa.2025.130943_bib0065","series-title":"Proceedings of the 2022 conference of the north American chapter of the association for computational linguistics: human language technologies","first-page":"1610","article-title":"MoEBERT: From BERT to mixture-of-experts via importance-guided adaptation","author":"Zuo","year":"2022"}],"container-title":["Expert Systems with Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0957417425045580?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0957417425045580?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T12:26:21Z","timestamp":1774441581000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0957417425045580"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,4]]},"references-count":65,"alternative-id":["S0957417425045580"],"URL":"https:\/\/doi.org\/10.1016\/j.eswa.2025.130943","relation":{},"ISSN":["0957-4174"],"issn-type":[{"value":"0957-4174","type":"print"}],"subject":[],"published":{"date-parts":[[2026,4]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"AlignExperts: Knowledge alignment experts for summarization faithfulness evaluation","name":"articletitle","label":"Article Title"},{"value":"Expert Systems with Applications","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.eswa.2025.130943","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"130943"}}