{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T11:40:09Z","timestamp":1755862809081,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":16,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,12,8]],"date-time":"2023-12-08T00:00:00Z","timestamp":1701993600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,12,8]]},"DOI":"10.1145\/3659211.3659227","type":"proceedings-article","created":{"date-parts":[[2024,5,29]],"date-time":"2024-05-29T16:22:21Z","timestamp":1716999741000},"page":"90-93","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Advancing Text Analytics: Instruction Fine-Tuning of QianWen-7B for Sentiment Classification"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-4261-9150","authenticated-orcid":false,"given":"Yang","family":"Han","sequence":"first","affiliation":[{"name":"Yan?an University, China"}]}],"member":"320","published-online":{"date-parts":[[2024,5,29]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"J. Bai \"Qwen technical report \" arXiv preprint arXiv:2309.16609 2023."},{"key":"e_1_3_2_1_2_1","volume-title":"Qlora: Efficient finetuning of quantized llms,\" arXiv preprint arXiv:2305.14314","author":"Dettmers T.","year":"2023","unstructured":"T. Dettmers, A. Pagnoni, A. Holtzman, and L. Zettlemoyer, \"Qlora: Efficient finetuning of quantized llms,\" arXiv preprint arXiv:2305.14314, 2023."},{"key":"e_1_3_2_1_3_1","unstructured":"T. Brown \"Language models are few-shot learners \" Advances in neural information processing systems vol. 33 pp. 1877-1901 2020."},{"key":"e_1_3_2_1_4_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding,\" arXiv preprint arXiv:1810.04805","author":"Devlin J.","year":"2018","unstructured":"J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, \"Bert: Pre-training of deep bidirectional transformers for language understanding,\" arXiv preprint arXiv:1810.04805, 2018."},{"key":"e_1_3_2_1_5_1","unstructured":"A. Vaswani \"Attention is all you need \" Advances in neural information processing systems vol. 30 2017."},{"key":"e_1_3_2_1_6_1","volume-title":"Universal language model fine-tuning for text classification,\" arXiv preprint arXiv:1801.06146","author":"Howard J.","year":"2018","unstructured":"J. Howard and S. Ruder, \"Universal language model fine-tuning for text classification,\" arXiv preprint arXiv:1801.06146, 2018."},{"key":"e_1_3_2_1_7_1","unstructured":"J. Wei \"Finetuned language models are zero-shot learners \" arXiv preprint arXiv:2109.01652 2021."},{"key":"e_1_3_2_1_8_1","volume-title":"Adam: A method for stochastic optimization,\" arXiv preprint arXiv:1412.6980","author":"Kingma D. P.","year":"2014","unstructured":"D. P. Kingma and J. Ba, \"Adam: A method for stochastic optimization,\" arXiv preprint arXiv:1412.6980, 2014."},{"key":"e_1_3_2_1_9_1","volume-title":"Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing,\" arXiv preprint arXiv:2111.09543","author":"He P.","year":"2021","unstructured":"P. He, J. Gao, and W. Chen, \"Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing,\" arXiv preprint arXiv:2111.09543, 2021."},{"key":"e_1_3_2_1_10_1","volume-title":"Deberta: Decoding-enhanced bert with disentangled attention,\" arXiv preprint arXiv:2006.03654","author":"He P.","year":"2020","unstructured":"P. He, X. Liu, J. Gao, and W. Chen, \"Deberta: Decoding-enhanced bert with disentangled attention,\" arXiv preprint arXiv:2006.03654, 2020."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1213767109"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1111\/fmii.12002"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.dss.2010.07.012"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/1462198.1462204"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-017-9588-9"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2017.11.196"}],"event":{"name":"BDEIM 2023: 2023 4th International Conference on Big Data Economy and Information Management","acronym":"BDEIM 2023","location":"Zhengzhou China"},"container-title":["Proceedings of the 2023 4th International Conference on Big Data Economy and Information Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3659211.3659227","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3659211.3659227","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T11:01:09Z","timestamp":1755860469000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3659211.3659227"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,8]]},"references-count":16,"alternative-id":["10.1145\/3659211.3659227","10.1145\/3659211"],"URL":"https:\/\/doi.org\/10.1145\/3659211.3659227","relation":{},"subject":[],"published":{"date-parts":[[2023,12,8]]},"assertion":[{"value":"2024-05-29","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}