{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T17:48:16Z","timestamp":1770140896000,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":27,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819722617","type":"print"},{"value":"9789819722594","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-2259-4_21","type":"book-chapter","created":{"date-parts":[[2024,4,24]],"date-time":"2024-04-24T09:02:31Z","timestamp":1713949351000},"page":"277-289","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["DALLMi: Domain Adaption for\u00a0LLM-Based Multi-label Classifier"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-1010-0148","authenticated-orcid":false,"given":"Miruna","family":"Be\u021bianu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4493-7439","authenticated-orcid":false,"given":"Abele","family":"M\u0103lan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8788-0829","authenticated-orcid":false,"given":"Marco","family":"Aldinucci","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1144-3707","authenticated-orcid":false,"given":"Robert","family":"Birke","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4228-6735","authenticated-orcid":false,"given":"Lydia","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,4,25]]},"reference":[{"issue":"4","key":"21_CR1","doi-asserted-by":"publisher","first-page":"719","DOI":"10.1007\/s10994-020-05877-5","volume":"109","author":"J Bekker","year":"2020","unstructured":"Bekker, J., Davis, J.: Learning from positive and unlabeled data: a survey. Mach. Learn. 109(4), 719\u2013760 (2020)","journal-title":"Mach. Learn."},{"key":"21_CR2","doi-asserted-by":"publisher","first-page":"104431","DOI":"10.1016\/j.jbi.2023.104431","volume":"144","author":"TM Buonocore","year":"2023","unstructured":"Buonocore, T.M., Crema, C., Redolfi, A., Bellazzi, R., Parimbelli, E.: Localizing in-domain adaptation of transformer-based biomedical language models. J. Biomed. Informatics 144, 104431 (2023)","journal-title":"J. Biomed. Informatics"},{"key":"21_CR3","doi-asserted-by":"crossref","unstructured":"Cartucho, J., Ventura, R., Veloso, M.: Robust object recognition through symbiotic deep learning in mobile robots. In: IEEE\/RSJ IROS, pp. 2336\u20132341 (2018)","DOI":"10.1109\/IROS.2018.8594067"},{"key":"21_CR4","unstructured":"Chen, H., Liu, F., Wang, Y., Zhao, L., Wu, H.: A variational approach for learning from positive and unlabeled data. In: NeurIPS, vol.\u00a033, pp. 14844\u201314854 (2020)"},{"key":"21_CR5","doi-asserted-by":"crossref","unstructured":"Chronopoulou, A., Peters, M.E., Dodge, J.: Efficient hierarchical domain adaptation for pretrained language models. In: NAACL, pp. 1336\u20131351 (2022)","DOI":"10.18653\/v1\/2022.naacl-main.96"},{"key":"21_CR6","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1186\/s40537-015-0029-9","volume":"2","author":"M Crawford","year":"2015","unstructured":"Crawford, M., Khoshgoftaar, T.M., Prusa, J.D., Richter, A.N., Najada, H.A.: Survey of review spam detection using machine learning techniques. J. Big Data 2, 23 (2015)","journal-title":"J. Big Data"},{"key":"21_CR7","doi-asserted-by":"publisher","first-page":"78","DOI":"10.1016\/j.cag.2021.12.008","volume":"102","author":"A Deng","year":"2022","unstructured":"Deng, A., Wu, Y., Zhang, P., Lu, Z., Li, W., Su, Z.: A weakly supervised framework for real-world point cloud classification. Comput. Graph. 102, 78\u201388 (2022)","journal-title":"Comput. Graph."},{"key":"21_CR8","unstructured":"Eastwood, C., Mason, I., Williams, C.K.I., Sch\u00f6lkopf, B.: Source-free adaptation to measurement shift via bottom-up feature restoration. In: ICLR (2022)"},{"key":"21_CR9","doi-asserted-by":"crossref","unstructured":"Grangier, D., Iter, D.: The trade-offs of domain adaptation for neural language models. In: ACL, pp. 3802\u20133813 (2022)","DOI":"10.18653\/v1\/2022.acl-long.264"},{"key":"21_CR10","doi-asserted-by":"crossref","unstructured":"Guo, Y., Rennard, V., Xypolopoulos, C., Vazirgiannis, M.: Bertweetfr: domain adaptation of pre-trained language models for French tweets. In: W-NUT, pp. 445\u2013450 (2021)","DOI":"10.18653\/v1\/2021.wnut-1.49"},{"issue":"1","key":"21_CR11","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1007\/s10489-011-0314-z","volume":"37","author":"LH Lee","year":"2012","unstructured":"Lee, L.H., Wan, C.H., Rajkumar, R., Isa, D.: An enhanced support vector machine classification framework by using euclidean distance function for text document categorization. Appl. Intell. 37(1), 80\u201399 (2012)","journal-title":"Appl. Intell."},{"key":"21_CR12","unstructured":"Liu, H., Long, M., Wang, J., Wang, Y.: Learning to adapt to evolving domains. In: NeurIPS, vol.\u00a033, pp. 22338\u201322348 (2020)"},{"key":"21_CR13","doi-asserted-by":"crossref","unstructured":"Motiian, S., Piccirilli, M., Adjeroh, D.A., Doretto, G.: Unified deep supervised domain adaptation and generalization. In: IEEE ICCV, pp. 5716\u20135726 (2017)","DOI":"10.1109\/ICCV.2017.609"},{"key":"21_CR14","doi-asserted-by":"crossref","unstructured":"Nasukawa, T., Yi, J.: Sentiment analysis: capturing favorability using natural language processing. In: K-CAP, pp. 70\u201377 (2003)","DOI":"10.1145\/945649.945658"},{"key":"21_CR15","doi-asserted-by":"crossref","unstructured":"Pham, D.D., Koesnadi, S.M., Dovletov, G., Pauli, J.: Unsupervised adversarial domain adaptation for multi-label classification of chest x-ray. In: IEEE ISBI, pp. 1236\u20131240 (2021)","DOI":"10.1109\/ISBI48211.2021.9434003"},{"key":"21_CR16","unstructured":"Quinonero-Candela, J., Sugiyama, M., Schwaighofer, A., Lawrence, N.D.: Dataset Shift in Machine Learning. MIT Press (2022)"},{"key":"21_CR17","unstructured":"Rietzler, A., Stabinger, S., Opitz, P., Engl, S.: Adapt or get left behind: domain adaptation through BERT language model finetuning for aspect-target sentiment classification. In: LREC, pp. 4933\u20134941 (2020)"},{"issue":"11","key":"21_CR18","doi-asserted-by":"publisher","first-page":"3113","DOI":"10.1007\/s10115-022-01736-y","volume":"64","author":"M Ryu","year":"2022","unstructured":"Ryu, M., Lee, G., Lee, K.: Knowledge distillation for BERT unsupervised domain adaptation. Knowl. Inf. Syst. 64(11), 3113\u20133128 (2022)","journal-title":"Knowl. Inf. Syst."},{"key":"21_CR19","doi-asserted-by":"crossref","unstructured":"Sachidananda, V., Kessler, J.S., Lai, Y.: Efficient domain adaptation of language models via adaptive tokenization. In: SustaiNLP@EMNLP, pp. 155\u2013165 (2021)","DOI":"10.18653\/v1\/2021.sustainlp-1.16"},{"key":"21_CR20","doi-asserted-by":"crossref","unstructured":"Singh, I.P., Ghorbel, E., Kacem, A., Rathinam, A., Aouada, D.: Discriminator-free unsupervised domain adaptation for multi-label image classification. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (2023)","DOI":"10.1109\/WACV57701.2024.00389"},{"key":"21_CR21","doi-asserted-by":"publisher","first-page":"6973","DOI":"10.1109\/ACCESS.2023.3237025","volume":"11","author":"P Singhal","year":"2023","unstructured":"Singhal, P., Walambe, R., Ramanna, S., Kotecha, K.: Domain adaptation: challenges, methods, datasets, and applications. IEEE Access 11, 6973\u20137020 (2023)","journal-title":"IEEE Access"},{"key":"21_CR22","doi-asserted-by":"crossref","unstructured":"Sun, X., et al.: Text classification via large language models. In: EMNLP 2023 Findings (2023). https:\/\/aclanthology.org\/2023.findings-emnlp.603\/","DOI":"10.18653\/v1\/2023.findings-emnlp.603"},{"key":"21_CR23","unstructured":"Verma, V., et al.: Manifold Mixup: better representations by interpolating hidden states. In: ICML, vol.\u00a097, pp. 6438\u20136447 (2019)"},{"key":"21_CR24","unstructured":"Wang, D., Shelhamer, E., Liu, S., Olshausen, B.A., Darrell, T.: Tent: fully test-time adaptation by entropy minimization. In: ICLR (2021)"},{"key":"21_CR25","doi-asserted-by":"crossref","unstructured":"Yuan, Z., Zhang, K., Huang, T.: Positive label is all you need for multi-label classification. arXiv preprint arXiv:2306.16016 (2023)","DOI":"10.1109\/ICME57554.2024.10687587"},{"key":"21_CR26","unstructured":"Zhang, H., Ciss\u00e9, M., Dauphin, Y.N., Lopez-Paz, D.: Mixup: beyond empirical risk minimization. In: ICLR (2018)"},{"key":"21_CR27","unstructured":"Zhang, Y., Zhang, H., Deng, B., Li, S., Jia, K., Zhang, L.: Semi-supervised models are strong unsupervised domain adaptation learners. arXiv preprint arXiv:2106.00417 (2021)"}],"container-title":["Lecture Notes in Computer Science","Advances in Knowledge Discovery and Data Mining"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-2259-4_21","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,16]],"date-time":"2024-11-16T23:02:50Z","timestamp":1731798170000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-2259-4_21"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819722617","9789819722594"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-2259-4_21","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"25 April 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PAKDD","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Pacific-Asia Conference on Knowledge Discovery and Data Mining","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Taipei","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Taiwan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7 May 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10 May 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"pakdd2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/pakdd2024.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}