{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,21]],"date-time":"2026-03-21T07:14:06Z","timestamp":1774077246111,"version":"3.50.1"},"reference-count":56,"publisher":"Association for Natural Language Processing","issue":"1","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Journal of Natural Language Processing"],"published-print":{"date-parts":[[2026]]},"DOI":"10.5715\/jnlp.33.207","type":"journal-article","created":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T22:12:58Z","timestamp":1773526378000},"page":"207-239","source":"Crossref","is-referenced-by-count":0,"title":["GPT-MM: Improving Multimodal In-context Learning with Task-specific Retrieval and Reasoning"],"prefix":"10.5715","volume":"33","author":[{"given":"Zhen","family":"Wan","sequence":"first","affiliation":[{"name":"Graduate School of Informatics, Kyoto University"}]},{"given":"Fei","family":"Cheng","sequence":"additional","affiliation":[{"name":"Graduate School of Informatics, Kyoto University"}]},{"given":"Sadao","family":"Kurohashi","sequence":"additional","affiliation":[{"name":"Graduate School of Informatics, Kyoto University"},{"name":"National Institute of Informatics"}]}],"member":"3685","reference":[{"key":"1","doi-asserted-by":"crossref","unstructured":"Alayrac, J.-B., Donahue, J., Luc, P., Miech, A., Barr, I., Hasson, Y., Lenc, K., Mensch, A., Millican, K., Reynolds, M., Ring, R., Rutherford, E., Cabi, S., Han, T., Gong, Z., Samangooei, S., Monteiro, M., Menick, J., Borgeaud, S., Brock, A., Nematzadeh, A., Sharifzadeh, S., Binkowski, M., Barreira, R., Vinyals, O., Zisserman, A., and Simonyan, K. (2022). \u201cFlamingo: A Visual Language Model for Few-Shot Learning.\u201d <i>CoRR<\/i>, abs\/2204.14198.","DOI":"10.52202\/068431-1723"},{"key":"2","doi-asserted-by":"crossref","unstructured":"Baldini Soares, L., FitzGerald, N., Ling, J., and Kwiatkowski, T. (2019). \u201cMatching the Blanks: Distributional Similarity for Relation Learning.\u201d In <i>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics<\/i>, pp. 2895\u20132905, Florence, Italy. Association for Computational Linguistics.","DOI":"10.18653\/v1\/P19-1279"},{"key":"3","unstructured":"Banko, M. and Etzioni, O. (2008). \u201cThe Tradeoffs Between Open and Traditional Relation Extraction.\u201d In <i>Proceedings of ACL-08: HLT<\/i>, pp. 28\u201336, Columbus, Ohio. Association for Computational Linguistics."},{"key":"4","doi-asserted-by":"crossref","unstructured":"Beltagy, I., Lo, K., and Cohan, A. (2019). \u201cSciBERT: A Pretrained Language Model for Scientific Text.\u201d In <i>Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)<\/i>, pp. 3615\u20133620, Hong Kong, China. Association for Computational Linguistics.","DOI":"10.18653\/v1\/D19-1371"},{"key":"5","unstructured":"Blevins, T., Gonen, H., and Zettlemoyer, L. (2022). \u201cPrompting Language Models for Linguistic Structure.\u201d <i>CoRR<\/i>, abs\/2211.07830."},{"key":"6","doi-asserted-by":"crossref","unstructured":"Borsos, Z., Marinier, R., Vincent, D., Kharitonov, E., Pietquin, O., Sharifi, M., Roblek, D., Teboul, O., Grangier, D., Tagliasacchi, M., and Zeghidour, N. (2023). \u201cAudioLM: a Language Modeling Approach to Audio Generation.\u201d <i>CoRR<\/i>, abs\/2209.03143.","DOI":"10.1109\/TASLP.2023.3288409"},{"key":"7","doi-asserted-by":"crossref","unstructured":"Bowman, S. R., Angeli, G., Potts, C., and Manning, C. D. (2015). \u201cA Large Annotated Corpus for Learning Natural Language Inference.\u201d In <i>Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing<\/i>, pp. 632\u2013642, Lisbon, Portugal. Association for Computational Linguistics.","DOI":"10.18653\/v1\/D15-1075"},{"key":"8","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., McCandlish, S., Radford, A., Sutskever, I., and Amodei, D. (2020). \u201cLanguage Models are Few-Shot Learners.\u201d In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (Eds.), <i>Advances in Neural Information Processing Systems<\/i>, Vol. 33, pp. 1877\u20131901. Curran Associates, Inc."},{"key":"9","doi-asserted-by":"crossref","unstructured":"Chen, Z., Huang, H., Andrusenko, A., Hrinchuk, O., Puvvada, K. C., Li, J., Ghosh, S., Balam, J., and Ginsburg, B. (2024). \u201cSALM: Speech-Augmented Language Model with in-Context Learning for Speech Recognition and Translation.\u201d In <i>ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)<\/i>, pp. 13521\u201313525.","DOI":"10.1109\/ICASSP48485.2024.10447553"},{"key":"10","unstructured":"Cohen, A. D. N., Rosenman, S., and Goldberg, Y. (2020). \u201cRelation Extraction as Two-way Span-Prediction.\u201d <i>CoRR<\/i>, abs\/2010.04829."},{"key":"11","unstructured":"Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. (2019). \u201cBERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.\u201d In <i>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)<\/i>, pp. 4171\u20134186, Minneapolis, Minnesota. Association for Computational Linguistics."},{"key":"12","doi-asserted-by":"crossref","unstructured":"Gao, T., Yao, X., and Chen, D. (2021). \u201cSimCSE: Simple Contrastive Learning of Sentence Embeddings.\u201d In <i>Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing<\/i>, pp. 6894\u20136910, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2021.emnlp-main.552"},{"key":"13","doi-asserted-by":"crossref","unstructured":"Geng, X., Chen, X., Zhu, K. Q., Shen, L., and Zhao, Y. (2020). \u201cMICK: A Meta-Learning Framework for Few-shot Relation Classification with Small Training Data.\u201d In d\u2019Aquin, M., Dietze, S., Hauff, C., Curry, E., and Cudr\u00e9-Mauroux, P. (Eds.), <i>CIKM \u201920: The 29th ACM International Conference on Information and Knowledge Management, Virtual Event, Ireland, October 19\u201323, 2020<\/i>, pp. 415\u2013424. ACM.","DOI":"10.1145\/3340531.3411858"},{"key":"14","unstructured":"Guti\u00e9rrez, B. J., McNeal, N., Washington, C., Chen, Y., Li, L., Sun, H., and Su, Y. (2022). \u201cThinking about GPT-3 In-Context Learning for Biomedical IE? Think Again.\u201d <i>CoRR<\/i>, abs\/2203.08410."},{"key":"15","unstructured":"Han, P., Pereira, L. K., Cheng, F., She, W. J., and Aramaki, E. (2024). \u201cAMR-RE: Abstract Meaning Representations for Retrieval-Based In-Context Learning in Relation Extraction.\u201d <i>CoRR<\/i>, abs\/2406.10432."},{"key":"16","doi-asserted-by":"crossref","unstructured":"Han, X., Zhu, H., Yu, P., Wang, Z., Yao, Y., Liu, Z., and Sun, M. (2018). \u201cFewRel: A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation.\u201d In <i>Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing<\/i>, pp. 4803\u20134809, Brussels, Belgium. Association for Computational Linguistics.","DOI":"10.18653\/v1\/D18-1514"},{"key":"17","doi-asserted-by":"crossref","unstructured":"Hendrickx, I., Kim, S. N., Kozareva, Z., Nakov, P., \u00d3 S\u00e9aghdha, D., Pad\u00f3, S., Pennacchiotti, M., Romano, L., and Szpakowicz, S. (2010). \u201cSemEval-2010 Task 8: Multi-Way Classification of Semantic Relations between Pairs of Nominals.\u201d In <i>Proceedings of the 5th International Workshop on Semantic Evaluation<\/i>, pp. 33\u201338, Uppsala, Sweden. Association for Computational Linguistics.","DOI":"10.3115\/1621969.1621986"},{"key":"18","unstructured":"Hsu, M.-H., Chang, K.-W., Li, S.-W., and yi Lee, H. (2024). \u201cExploring In-Context Learning of Textless Speech Language Model for Speech Classification Tasks.\u201d <i>CoRR<\/i>, abs\/2310.12477."},{"key":"19","unstructured":"Kojima, T., Gu, S. S., Reid, M., Matsuo, Y., and Iwasawa, Y. (2022). \u201cLarge Language Models are Zero-Shot Reasoners.\u201d <i>CoRR<\/i>, abs\/2205.11916."},{"key":"20","unstructured":"Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., and Soricut, R. (2019). \u201cALBERT: A Lite BERT for Self-supervised Learning of Language Representations.\u201d <i>CoRR<\/i>, abs\/1909.11942."},{"key":"21","doi-asserted-by":"crossref","unstructured":"Liu, F., Lin, H., Han, X., Cao, B., and Sun, L. (2022a). \u201cPre-training to Match for Unified Low-shot Relation Extraction.\u201d In <i>Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)<\/i>, pp. 5785\u20135795, Dublin, Ireland. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2022.acl-long.397"},{"key":"22","doi-asserted-by":"crossref","unstructured":"Liu, J., Shen, D., Zhang, Y., Dolan, B., Carin, L., and Chen, W. (2022b). \u201cWhat Makes Good In-Context Examples for GPT-3?\u201d In <i>Proceedings of Deep Learning Inside Out (DeeLIO 2022): The 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures<\/i>, pp. 100\u2013114, Dublin, Ireland and Online. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2022.deelio-1.10"},{"key":"23","doi-asserted-by":"crossref","unstructured":"Lu, Y., Bartolo, M., Moore, A., Riedel, S., and Stenetorp, P. (2022). \u201cFantastically Ordered Prompts and Where to Find Them: Overcoming Few-Shot Prompt Order Sensitivity.\u201d In <i>Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)<\/i>, pp. 8086\u20138098, Dublin, Ireland. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2022.acl-long.556"},{"key":"24","doi-asserted-by":"crossref","unstructured":"Luan, Y., He, L., Ostendorf, M., and Hajishirzi, H. (2018). \u201cMulti-Task Identification of Entities, Relations, and Coreference for Scientific Knowledge Graph Construction.\u201d In <i>Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing<\/i>, pp. 3219\u20133232, Brussels, Belgium. Association for Computational Linguistics.","DOI":"10.18653\/v1\/D18-1360"},{"key":"25","doi-asserted-by":"crossref","unstructured":"Ma, X., Li, J., and Zhang, M. (2023). \u201cChain of Thought with Explicit Evidence Reasoning for Few-shot Relation Extraction.\u201d In Bouamor, H., Pino, J., and Bali, K. (Eds.), <i>Findings of the Association for Computational Linguistics: EMNLP 2023<\/i>, pp. 2334\u20132352, Singapore. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2023.findings-emnlp.153"},{"key":"26","doi-asserted-by":"crossref","unstructured":"Malkin, N., Wang, Z., and Jojic, N. (2022). \u201cCoherence Boosting: When Your Pretrained Language Model is not Paying Enough Attention.\u201d In <i>Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)<\/i>, pp. 8214\u20138236, Dublin, Ireland. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2022.acl-long.565"},{"key":"27","doi-asserted-by":"crossref","unstructured":"Marino, K., Rastegari, M., Farhadi, A., and Mottaghi, R. (2019). \u201cOK-VQA: A Visual Question Answering Benchmark Requiring External Knowledge.\u201d <i>CoRR<\/i>, abs\/1906.00067.","DOI":"10.1109\/CVPR.2019.00331"},{"key":"28","doi-asserted-by":"crossref","unstructured":"Min, S., Lyu, X., Holtzman, A., Artetxe, M., Lewis, M., Hajishirzi, H., and Zettlemoyer, L. (2022). \u201cRethinking the Role of Demonstrations: What Makes In-Context Learning Work?\u201d <i>CoRR<\/i>, abs\/2202.12837.","DOI":"10.18653\/v1\/2022.emnlp-main.759"},{"key":"29","doi-asserted-by":"crossref","unstructured":"Mo, Y., Liu, J., Yang, J., Wang, Q., Zhang, S., Wang, J., and Li, Z. (2024). \u201cC-ICL: Contrastive In-context Learning for Information Extraction.\u201d In Al-Onaizan, Y., Bansal, M., and Chen, Y.-N. (Eds.), <i>Findings of the Association for Computational Linguistics: EMNLP 2024<\/i>, pp. 10099\u201310114, Miami, Florida, USA. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2024.findings-emnlp.590"},{"key":"30","doi-asserted-by":"crossref","unstructured":"Pan, J., Wu, J., Gaur, Y., Sivasankaran, S., Chen, Z., Liu, S., and Li, J. (2024). \u201cCOSMIC: Data Efficient Instruction-tuning For Speech In-Context Learning.\u201d <i>CoRR<\/i>, abs\/2311.02248.","DOI":"10.21437\/Interspeech.2024-1346"},{"key":"31","doi-asserted-by":"crossref","unstructured":"Pang, C., Cao, Y., Ding, Q., and Luo, P. (2023). \u201cGuideline Learning for In-Context Information Extraction.\u201d In Bouamor, H., Pino, J., and Bali, K. (Eds.), <i>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing<\/i>, pp. 15372\u201315389, Singapore. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2023.emnlp-main.950"},{"key":"32","unstructured":"Perez, E., Kiela, D., and Cho, K. (2021). \u201cTrue Few-Shot Learning with Language Models.\u201d In Ranzato, M., Beygelzimer, A., Dauphin, Y. N., Liang, P., and Vaughan, J. W. (Eds.), <i>Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual<\/i>, pp. 11054\u201311070."},{"key":"33","unstructured":"Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., Krueger, G., and Sutskever, I. (2021). \u201cLearning Transferable Visual Models From Natural Language Supervision.\u201d <i>CoRR<\/i>, abs\/2103.00020."},{"key":"34","unstructured":"Radford, A., Kim, J. W., Xu, T., Brockman, G., McLeavey, C., and Sutskever, I. (2022). \u201cRobust Speech Recognition via Large-Scale Weak Supervision.\u201d."},{"key":"35","unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., and Liu, P. J. (2019). \u201cExploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer.\u201d <i>CoRR<\/i>, abs\/1910.10683."},{"key":"36","doi-asserted-by":"crossref","unstructured":"Reimers, N. and Gurevych, I. (2019). \u201cSentence-BERT: Sentence Embeddings using Siamese BERT-Networks.\u201d In <i>Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)<\/i>, pp. 3982\u20133992, Hong Kong, China. Association for Computational Linguistics.","DOI":"10.18653\/v1\/D19-1410"},{"key":"37","unstructured":"Rio, M. D., Ha, P., McNamara, Q., Miller, C., and Chandra, S. (2022). \u201cEarnings-22: A Practical Benchmark for Accents in the Wild.\u201d <i>CoRR<\/i>, abs\/2203.15591."},{"key":"38","doi-asserted-by":"crossref","unstructured":"Rubin, O., Herzig, J., and Berant, J. (2022). \u201cLearning To Retrieve Prompts for In-Context Learning.\u201d In <i>Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies<\/i>, pp. 2655\u20132671, Seattle, United States. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2022.naacl-main.191"},{"key":"39","doi-asserted-by":"crossref","unstructured":"Shin, R., Lin, C., Thomson, S., Chen, C., Roy, S., Platanios, E. A., Pauls, A., Klein, D., Eisner, J., and Van Durme, B. (2021). \u201cConstrained Language Models Yield Few-Shot Semantic Parsers.\u201d In <i>Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing<\/i>, pp. 7699\u20137715, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2021.emnlp-main.608"},{"key":"40","unstructured":"Sun, X., Dong, L., Li, X., Wan, Z., Wang, S., Zhang, T., Li, J., Cheng, F., Lyu, L., Wu, F., and Wang, G. (2023). \u201cPushing the Limits of ChatGPT on NLP Tasks.\u201d <i>CoRR<\/i>, abs\/2306.09719."},{"key":"41","unstructured":"Tsimpoukelli, M., Menick, J., Cabi, S., Eslami, S. M. A., Vinyals, O., and Hill, F. (2021). \u201cMultimodal Few-Shot Learning with Frozen Language Models.\u201d <i>CoRR<\/i>, abs\/2106.13884."},{"key":"42","doi-asserted-by":"crossref","unstructured":"Wan, Z., Cheng, F., Mao, Z., Liu, Q., Song, H., Li, J., and Kurohashi, S. (2023). \u201cGPT-RE: In-context Learning for Relation Extraction using Large Language Models.\u201d In Bouamor, H., Pino, J., and Bali, K. (Eds.), <i>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing<\/i>, pp. 3534\u20133547, Singapore. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2023.emnlp-main.214"},{"key":"43","doi-asserted-by":"crossref","unstructured":"Wan, Z., Liu, Q., Mao, Z., Cheng, F., Kurohashi, S., and Li, J. (2022). \u201cRescue Implicit and Long-tail Cases: Nearest Neighbor Relation Extraction.\u201d <i>CoRR<\/i>, abs\/2210.11800.","DOI":"10.18653\/v1\/2022.emnlp-main.113"},{"key":"44","doi-asserted-by":"crossref","unstructured":"Wang, C., Liu, X., Chen, Z., Hong, H., Tang, J., and Song, D. (2022). \u201cDeepStruct: Pretraining of Language Models for Structure Prediction.\u201d In <i>Findings of the Association for Computational Linguistics: ACL 2022<\/i>, pp. 803\u2013823, Dublin, Ireland. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2022.findings-acl.67"},{"key":"45","doi-asserted-by":"crossref","unstructured":"Wang, S., Yang, C.-H. H., Wu, J., and Zhang, C. (2024a). \u201cBayesian Example Selection Improves In-Context Learning for Speech, Text, and Visual Modalities.\u201d <i>CoRR<\/i>, abs\/2404.14716.","DOI":"10.18653\/v1\/2024.emnlp-main.1158"},{"key":"46","unstructured":"Wang, S., Yang, C.-H. H., Wu, J., and Zhang, C. (2024b). \u201cCan Whisper Perform Speech-based In-context Learning?\u201d <i>CoRR<\/i>, abs\/2309.07081."},{"key":"47","unstructured":"Wang, X., Wei, J., Schuurmans, D., Le, Q. V., Chi, E. H., and Zhou, D. (2022). \u201cSelf-Consistency Improves Chain of Thought Reasoning in Language Models.\u201d <i>CoRR<\/i>, abs\/2203.11171."},{"key":"48","unstructured":"Wei, J., Wang, X., Schuurmans, D., Bosma, M., Chi, E. H., Le, Q., and Zhou, D. (2022). \u201cChain of Thought Prompting Elicits Reasoning in Large Language Models.\u201d <i>CoRR<\/i>, abs\/2201.11903."},{"key":"49","doi-asserted-by":"crossref","unstructured":"Williams, A., Nangia, N., and Bowman, S. (2018). \u201cA Broad-Coverage Challenge Corpus for Sentence Understanding through Inference.\u201d In <i>Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers)<\/i>, pp. 1112\u20131122, New Orleans, Louisiana. Association for Computational Linguistics.","DOI":"10.18653\/v1\/N18-1101"},{"key":"50","doi-asserted-by":"crossref","unstructured":"Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C., Moi, A., Cistac, P., Rault, T., Louf, R., Funtowicz, M., Davison, J., Shleifer, S., von Platen, P., Ma, C., Jernite, Y., Plu, J., Xu, C., Le Scao, T., Gugger, S., Drame, M., Lhoest, Q., and Rush, A. (2020). \u201cTransformers: State-of-the-Art Natural Language Processing.\u201d In <i>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations<\/i>, pp. 38\u201345, Online. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"51","unstructured":"Xu, J., Guo, Z., He, J., Hu, H., He, T., Bai, S., Chen, K., Wang, J., Fan, Y., Dang, K., Zhang, B., Wang, X., Chu, Y., and Lin, J. (2025). \u201cQwen2.5-Omni Technical Report.\u201d <i>CoRR<\/i>, abs\/2503.20215."},{"key":"52","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Zhong, V., Chen, D., Angeli, G., and Manning, C. D. (2017). \u201cPosition-aware Attention and Supervised Data Improve Slot Filling.\u201d In <i>Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing<\/i>, pp. 35\u201345, Copenhagen, Denmark. Association for Computational Linguistics.","DOI":"10.18653\/v1\/D17-1004"},{"key":"53","doi-asserted-by":"crossref","unstructured":"Zhong, Z. and Chen, D. (2021). \u201cA Frustratingly Easy Approach for Entity and Relation Extraction.\u201d In <i>Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies<\/i>, pp. 50\u201361, Online. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2021.naacl-main.5"},{"key":"54","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Li, X., Wang, Q., and Shen, J. (2024). \u201cVisual In-Context Learning for Large Vision-Language Models.\u201d <i>CoRR<\/i>, abs\/2402.11574.","DOI":"10.18653\/v1\/2024.findings-acl.940"},{"key":"55","unstructured":"Zhu, D., Chen, J., Shen, X., Li, X., and Elhoseiny, M. (2023). \u201cMiniGPT-4: Enhancing Vision-Language Understanding with Advanced Large Language Models.\u201d <i>CoRR<\/i>, abs\/2304.10592."},{"key":"56","unstructured":"Zhuang, L., Wayne, L., Ya, S., and Jun, Z. (2021). \u201cA Robustly Optimized BERT Pre-training Approach with Post-training.\u201d In <i>Proceedings of the 20th Chinese National Conference on Computational Linguistics<\/i>, pp. 1218\u20131227, Huhhot, China. Chinese Information Processing Society of China."}],"container-title":["Journal of Natural Language Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/jnlp\/33\/1\/33_207\/_pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,21]],"date-time":"2026-03-21T03:53:45Z","timestamp":1774065225000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/jnlp\/33\/1\/33_207\/_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"references-count":56,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026]]}},"URL":"https:\/\/doi.org\/10.5715\/jnlp.33.207","relation":{},"ISSN":["1340-7619","2185-8314"],"issn-type":[{"value":"1340-7619","type":"print"},{"value":"2185-8314","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]}}}