{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,13]],"date-time":"2025-10-13T00:23:04Z","timestamp":1760314984963,"version":"build-2065373602"},"publisher-location":"Cham","reference-count":31,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032079589","type":"print"},{"value":"9783032079596","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,10,13]],"date-time":"2025-10-13T00:00:00Z","timestamp":1760313600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,13]],"date-time":"2025-10-13T00:00:00Z","timestamp":1760313600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-07959-6_18","type":"book-chapter","created":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T09:22:23Z","timestamp":1760260943000},"page":"242-256","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Sector-Wise Backpropagation for\u00a0Low-Resource Text Classification in\u00a0Deep Models"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9766-4182","authenticated-orcid":false,"given":"Jos\u00e9 Luis","family":"V\u00e1zquez Noguera","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0005-7129-8435","authenticated-orcid":false,"given":"Carlos U.","family":"Valdez","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0910-0310","authenticated-orcid":false,"given":"Marvin M.","family":"Ag\u00fcero","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3698-4043","authenticated-orcid":false,"given":"Julio C.","family":"Mello","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7231-7563","authenticated-orcid":false,"given":"Jos\u00e9 D.","family":"Colbes","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0152-5813","authenticated-orcid":false,"given":"Sebasti\u00e1n A.","family":"Grillo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,13]]},"reference":[{"issue":"4","key":"18_CR1","doi-asserted-by":"publisher","first-page":"1391","DOI":"10.1007\/s12559-023-10165-0","volume":"15","author":"MM Ag\u00fcero-Torales","year":"2023","unstructured":"Ag\u00fcero-Torales, M.M., L\u00f3pez-Herrera, A.G., Vilares, D.: Multidimensional affective analysis for low-resource languages: a use case with Guarani-Spanish code-switching language. Cogn. Comput. 15(4), 1391\u20131406 (2023)","journal-title":"Cogn. Comput."},{"key":"18_CR2","unstructured":"Alain, G., Bengio, Y.: Understanding intermediate layers using linear classifier probes. arXiv preprint arXiv:1610.01644 (2016)"},{"key":"18_CR3","unstructured":"Belilovsky, E., Eickenberg, M., Oyallon, E.: Greedy layerwise learning can scale to ImageNet. In: International Conference on Machine Learning, pp. 583\u2013593. PMLR (2019)"},{"key":"18_CR4","doi-asserted-by":"publisher","unstructured":"Ben\u00a0Zaken, E., Goldberg, Y., Ravfogel, S.: BitFit: simple parameter-efficient fine-tuning for transformer-based masked language-models. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp.\u00a01\u20139. Association for Computational Linguistics, Dublin (2022). https:\/\/doi.org\/10.18653\/v1\/2022.acl-short.1, https:\/\/aclanthology.org\/2022.acl-short.1\/","DOI":"10.18653\/v1\/2022.acl-short.1"},{"issue":"Sep","key":"18_CR5","first-page":"1089","volume":"5","author":"Y Bengio","year":"2004","unstructured":"Bengio, Y., Grandvalet, Y.: No unbiased estimator of the variance of k-fold cross-validation. J. Mach. Learn. Res. 5(Sep), 1089\u20131105 (2004)","journal-title":"J. Mach. Learn. Res."},{"key":"18_CR6","unstructured":"Chollet, F.: Bidirectional LSTM on IMDB (2020). https:\/\/keras.io\/examples\/nlp\/bidirectional_lstm_imdb\/, https:\/\/keras.io\/examples\/nlp\/bidirectional_lstm_imdb\/"},{"key":"18_CR7","doi-asserted-by":"publisher","unstructured":"Clark, J.H., Garrette, D., Turc, I., Wieting, J.: CANINE: pre-training an efficient tokenization-free encoder for language representation. Trans. Assoc. Comput. Linguist. 10, 73\u201391 (2022). https:\/\/doi.org\/10.1162\/tacl_a_00448, https:\/\/aclanthology.org\/2022.tacl-1.5\/","DOI":"10.1162\/tacl_a_00448"},{"issue":"4","key":"18_CR8","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1109\/MCI.2022.3199624","volume":"17","author":"S Duan","year":"2022","unstructured":"Duan, S., Principe, J.C.: Training deep architectures without end-to-end backpropagation: a survey on the provably optimal methods. IEEE Comput. Intell. Mag. 17(4), 39\u201351 (2022)","journal-title":"IEEE Comput. Intell. Mag."},{"key":"18_CR9","doi-asserted-by":"publisher","unstructured":"Gururangan, S., et al.: Don\u2019t stop pretraining: adapt language models to domains and tasks. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 8342\u20138360. Association for Computational Linguistics, Online (2020). https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.740, https:\/\/aclanthology.org\/2020.acl-main.740\/","DOI":"10.18653\/v1\/2020.acl-main.740"},{"key":"18_CR10","unstructured":"Houlsby, N., et al.: Parameter-efficient transfer learning for NLP. In: Proceedings of the 36th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol.\u00a097, pp. 2790\u20132799. PMLR (2019). https:\/\/proceedings.mlr.press\/v97\/houlsby19a.html"},{"key":"18_CR11","doi-asserted-by":"publisher","unstructured":"Howard, J., Ruder, S.: Universal language model fine-tuning for text classification. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 328\u2013339. Association for Computational Linguistics, Melbourne (2018). https:\/\/doi.org\/10.18653\/v1\/P18-1031, https:\/\/aclanthology.org\/P18-1031\/","DOI":"10.18653\/v1\/P18-1031"},{"key":"18_CR12","doi-asserted-by":"publisher","unstructured":"Iandola, F., Shaw, A., Krishna, R., Keutzer, K.: SqueezeBERT: what can computer vision teach NLP about efficient neural networks? In: Proceedings of SustaiNLP: Workshop on Simple and Efficient Natural Language Processing, pp. 124\u2013135. Association for Computational Linguistics, Online (2020). https:\/\/doi.org\/10.18653\/v1\/2020.sustainlp-1.17, https:\/\/aclanthology.org\/2020.sustainlp-1.17\/","DOI":"10.18653\/v1\/2020.sustainlp-1.17"},{"key":"18_CR13","doi-asserted-by":"publisher","unstructured":"Jiao, X., et al.: TinyBERT: distilling BERT for natural language understanding. In: Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 4163\u20134174. Association for Computational Linguistics, Online (2020). https:\/\/doi.org\/10.18653\/v1\/2020.findings-emnlp.372, https:\/\/aclanthology.org\/2020.findings-emnlp.372\/","DOI":"10.18653\/v1\/2020.findings-emnlp.372"},{"key":"18_CR14","unstructured":"Maas, A., Daly, R.E., Pham, P.T., Huang, D., Ng, A.Y., Potts, C.: Learning word vectors for sentiment analysis. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pp. 142\u2013150 (2011)"},{"key":"18_CR15","unstructured":"Marivate, V., et al.: Investigating an approach for low resource language dataset creation, curation and classification: Setswana and Sepedi. In: Proceedings of the first workshop on Resources for African Indigenous Languages, pp. 15\u201320. European Language Resources Association (ELRA), Marseille (2020). https:\/\/aclanthology.org\/2020.rail-1.3\/"},{"key":"18_CR16","unstructured":"Matthew, E., et al: Deep contextualized word representations. In: Proceedings of NAACL, vol.\u00a05 (2018)"},{"key":"18_CR17","unstructured":"Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems, vol. 26 (2013)"},{"issue":"3","key":"18_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3439726","volume":"54","author":"S Minaee","year":"2021","unstructured":"Minaee, S., Kalchbrenner, N., Cambria, E., Nikzad, N., Chenaghlu, M., Gao, J.: Deep learning-based text classification: a comprehensive review. ACM Comput. Surv. (CSUR) 54(3), 1\u201340 (2021)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"18_CR19","unstructured":"Nandan, A.: Text classification with transformer (2020). https:\/\/keras.io\/examples\/nlp\/text_classification_with_transformer\/, https:\/\/keras.io\/examples\/nlp\/text_classification_with_transformer\/"},{"key":"18_CR20","unstructured":"Omernick, M., Chollet, F.: Text classification from scratch (2019). https:\/\/keras.io\/examples\/nlp\/text_classification_from_scratch\/, https:\/\/keras.io\/examples\/nlp\/text_classification_from_scratch\/"},{"key":"18_CR21","doi-asserted-by":"publisher","unstructured":"Rahamim, A., Uziel, G., Goldbraich, E., Anaby\u00a0Tavor, A.: Text augmentation using dataset reconstruction for low-resource classification. In: Findings of the Association for Computational Linguistics: ACL 2023, pp. 7389\u20137402. Association for Computational Linguistics, Toronto (2023). https:\/\/doi.org\/10.18653\/v1\/2023.findings-acl.466, https:\/\/aclanthology.org\/2023.findings-acl.466\/","DOI":"10.18653\/v1\/2023.findings-acl.466"},{"key":"18_CR22","doi-asserted-by":"crossref","unstructured":"Schuster, M., Paliwal, K.K.: Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 45(11), 2673\u20132681 (1997). Introducci\u00f3n formal de redes recursivas bidireccionales (BRNN)","DOI":"10.1109\/78.650093"},{"key":"18_CR23","doi-asserted-by":"crossref","unstructured":"Socher, R., et al.: Recursive deep models for semantic compositionality over a sentiment treebank. In: Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pp. 1631\u20131642 (2013)","DOI":"10.18653\/v1\/D13-1170"},{"key":"18_CR24","doi-asserted-by":"publisher","unstructured":"Sun, Z., Yu, H., Song, X., Liu, R., Yang, Y., Zhou, D.: MobileBERT: a compact task-agnostic BERT for resource-limited devices. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 2158\u20132170. Association for Computational Linguistics, Online (2020). https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.195, https:\/\/aclanthology.org\/2020.acl-main.195\/","DOI":"10.18653\/v1\/2020.acl-main.195"},{"key":"18_CR25","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)"},{"key":"18_CR26","doi-asserted-by":"crossref","unstructured":"Voorhees, E., Harman, D., Wilkinson, R.: The sixth text retrieval conference (TREC-6). In: The Text REtrieval Conference (TREC), vol.\u00a0500, p.\u00a0240. ERIC (1998)","DOI":"10.6028\/NIST.SP.500-240"},{"key":"18_CR27","doi-asserted-by":"crossref","unstructured":"Waibel, A., Hanazawa, M., Hinton, G., Shikano, K., Lang, K.J.: Phoneme recognition using time-delay neural networks. IEEE Trans. Acoust. Speech Signal Process. 37(3), 328\u2013339 (1989). Primer uso de convoluciones temporales 1D con weight sharing y backpropagation","DOI":"10.1109\/29.21701"},{"key":"18_CR28","doi-asserted-by":"crossref","unstructured":"Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., Bowman, S.R.: GLUE: A multi-task benchmark and analysis platform for natural language understanding. In: Proceedings of ICLR (2019)","DOI":"10.18653\/v1\/W18-5446"},{"key":"18_CR29","unstructured":"Zhang, X., Zhao, J., LeCun, Y.: Character-level convolutional networks for text classification (2015)"},{"key":"18_CR30","unstructured":"Zhuang, L., Wayne, L., Ya, S., Jun, Z.: A robustly optimized BERT pre-training approach with post-training. In: Li, S., et al. (eds.) Proceedings of the 20th Chinese National Conference on Computational Linguistics, pp. 1218\u20131227. Chinese Information Processing Society of China, Huhhot (2021). https:\/\/aclanthology.org\/2021.ccl-1.108\/"},{"key":"18_CR31","doi-asserted-by":"crossref","unstructured":"Zulqarnain, M., et al.: Text classification using deep learning models: a comparative review. Cloud Comput. Data Sci. 80\u201396 (2024)","DOI":"10.37256\/ccds.5120243528"}],"container-title":["Lecture Notes in Computer Science","Speech and Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-07959-6_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T09:22:30Z","timestamp":1760260950000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-07959-6_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,13]]},"ISBN":["9783032079589","9783032079596"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-07959-6_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,13]]},"assertion":[{"value":"13 October 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"During the preparation of this work the authors used generative tools in order to fix misspellings and improve writing. After using these tools, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclaimer"}},{"value":"The code for reproducing the experiments presented in this paper is publicly accessible at .","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Code Availability"}},{"value":"SPECOM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Speech and Computer","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Szeged","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Hungary","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"specom2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/specom.inf.u-szeged.hu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}