{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T10:10:49Z","timestamp":1766139049175,"version":"3.33.0"},"reference-count":34,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,12,15]],"date-time":"2024-12-15T00:00:00Z","timestamp":1734220800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,12,15]],"date-time":"2024-12-15T00:00:00Z","timestamp":1734220800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,12,15]]},"DOI":"10.1109\/bigdata62323.2024.10825290","type":"proceedings-article","created":{"date-parts":[[2025,1,16]],"date-time":"2025-01-16T18:31:23Z","timestamp":1737052283000},"page":"918-927","source":"Crossref","is-referenced-by-count":1,"title":["Unraveling Intricacies: A Decomposition Approach for Few-Shot Multi-Intent Spoken Language Understanding"],"prefix":"10.1109","author":[{"given":"Wenbin","family":"Hua","sequence":"first","affiliation":[{"name":"Central China Normal University,Hubei Provincial Key Laboratory of Artificial Intelligence and Smart Learning, National Language Resources Monitor Research Center for Network Media,Wuhan,China"}]},{"given":"Yufan","family":"Wang","sequence":"additional","affiliation":[{"name":"Shenzhen Technology University,College of Big Data and Internet,Shenzhen,China"}]},{"given":"Rui","family":"Fan","sequence":"additional","affiliation":[{"name":"Central China Normal University,Hubei Provincial Key Laboratory of Artificial Intelligence and Smart Learning, National Language Resources Monitor Research Center for Network Media,Wuhan,China"}]},{"given":"Xinhui","family":"Tu","sequence":"additional","affiliation":[{"name":"Central China Normal University,Hubei Provincial Key Laboratory of Artificial Intelligence and Smart Learning, National Language Resources Monitor Research Center for Network Media,Wuhan,China"}]},{"given":"Tingting","family":"He","sequence":"additional","affiliation":[{"name":"Central China Normal University,Hubei Provincial Key Laboratory of Artificial Intelligence and Smart Learning, National Language Resources Monitor Research Center for Network Media,Wuhan,China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-1055"},{"article-title":"Stance detection on social media with fine-tuned large language models","year":"2024","author":"G\u00fcl","key":"ref2"},{"article-title":"Navigating the landscape of large language models: A comprehensive review and analysis of paradigms and fine-tuning strategies","year":"2024","author":"Weng","key":"ref3"},{"article-title":"How much reliable is chatgpt\u2019s prediction on information extraction under input perturbations?","year":"2024","author":"Mondal","key":"ref4"},{"article-title":"Deconstructing in-context learning: Understanding prompts via corruption","year":"2024","author":"Shivagunde","key":"ref5"},{"article-title":"Continual few-shot event detection via hierarchical augmentation networks","year":"2024","author":"Zhang","key":"ref6"},{"article-title":"Crosstune: Black-box few-shot classification with label enhancement","year":"2024","author":"Luo","key":"ref7"},{"article-title":"Making pre-trained language models better continual few-shot relation extractors","year":"2024","author":"Ma","key":"ref8"},{"article-title":"Data augmentation is dead, long live data augmentation","year":"2024","author":"Piedboeuf","key":"ref9"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.ijcnlp-main.45"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2023.3237156"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1417"},{"article-title":"A novel bi-directional interrelated model for joint intent detection and slot filling","year":"2019","author":"Niu","key":"ref13"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1097"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2021-95"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1519"},{"article-title":"A prompting-based encoder-decoder approach to intent recognition and slot filling","year":"2021","author":"Tassias","key":"ref17"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP48485.2024.10447804"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-2118"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1214"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.543"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP43922.2022.9747843"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.163"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-acl.853"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.15"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.12"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.263"},{"key":"ref28","first-page":"7203","article-title":"Incorporating instructional prompts into a unified generative framework for joint multiple intent detection and slot filling","volume-title":"Proceedings of the 29th International Conference on Computational Linguistics","author":"Wu"},{"key":"ref29","first-page":"4884","article-title":"A label-aware BERT attention network for zero-shot multi-intent detection in spoken language understanding","volume-title":"Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing","author":"Wu"},{"article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","year":"2018","author":"Devlin","key":"ref30"},{"article-title":"Roberta: A robustly optimized bert pretraining approach","year":"2019","author":"Liu","key":"ref31"},{"issue":"1","key":"ref32","first-page":"5485","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"The Journal of Machine Learning Research"},{"article-title":"Decoupled weight decay regularization","year":"2017","author":"Loshchilov","key":"ref33"},{"article-title":"Review of generative ai methods in cybersecurity","year":"2024","author":"Yigit","key":"ref34"}],"event":{"name":"2024 IEEE International Conference on Big Data (BigData)","start":{"date-parts":[[2024,12,15]]},"location":"Washington, DC, USA","end":{"date-parts":[[2024,12,18]]}},"container-title":["2024 IEEE International Conference on Big Data (BigData)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10824975\/10824942\/10825290.pdf?arnumber=10825290","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,17]],"date-time":"2025-01-17T07:46:08Z","timestamp":1737099968000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10825290\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,15]]},"references-count":34,"URL":"https:\/\/doi.org\/10.1109\/bigdata62323.2024.10825290","relation":{},"subject":[],"published":{"date-parts":[[2024,12,15]]}}}