{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T13:08:10Z","timestamp":1772284090408,"version":"3.50.1"},"publisher-location":"Singapore","reference-count":31,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819543809","type":"print"},{"value":"9789819543816","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,11,16]],"date-time":"2025-11-16T00:00:00Z","timestamp":1763251200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,16]],"date-time":"2025-11-16T00:00:00Z","timestamp":1763251200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-4381-6_9","type":"book-chapter","created":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T06:59:33Z","timestamp":1763189973000},"page":"123-137","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["IR-MHAtt: A Lightweight Skin Lesion Classification Network with\u00a0Improved Inverted Residual and\u00a0Self-attention Mechanism"],"prefix":"10.1007","author":[{"given":"Yang","family":"Lian","sequence":"first","affiliation":[]},{"given":"Ruizhi","family":"Han","sequence":"additional","affiliation":[]},{"given":"Xiaofang","family":"Zhong","sequence":"additional","affiliation":[]},{"given":"Yuehui","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,16]]},"reference":[{"issue":"5","key":"9_CR1","doi-asserted-by":"publisher","first-page":"348","DOI":"10.1016\/j.neurad.2020.12.003","volume":"48","author":"Z Najafpour","year":"2021","unstructured":"Najafpour, Z., Fatemi, A., Goudarzi, Z., Goudarzi, R., Shayanfard, K., Noorizadeh, F.: Cost-effectiveness of neuroimaging technologies in management of psychiatric and insomnia disorders: a meta-analysis and prospective cost analysis. J. Neuroradiol. 48(5), 348\u2013358 (2021)","journal-title":"J. Neuroradiol."},{"key":"9_CR2","doi-asserted-by":"crossref","unstructured":"Bray, F., Laversanne, M., Sung, H., Ferlay, J., Siegel, R.L., Soerjomataram, I., Jemal, A.: Global cancer statistics 2022: Globocan estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA: A Cancer J. Clin. 74(3), 229\u2013263 (2024)","DOI":"10.3322\/caac.21834"},{"key":"9_CR3","doi-asserted-by":"publisher","first-page":"192","DOI":"10.1007\/s10462-024-10762-x","volume":"57","author":"Y Zhang","year":"2024","unstructured":"Zhang, Y., et al.: From single to universal: tiny lesion detection in medical imaging. Artif. Intell. Rev. 57, 192 (2024)","journal-title":"Artif. Intell. Rev."},{"key":"9_CR4","doi-asserted-by":"crossref","unstructured":"Xie, Y., Fei, Z., Deng, D., Meng, L., Sun, J., Niu, F.: A review of advances in deep learning-based small object detection. In: 2024 5th International Conference on Computer Vision, Image and Deep Learning (CVIDL), pp. 1207\u20131215 (2024)","DOI":"10.1109\/CVIDL62147.2024.10603837"},{"issue":"7553","key":"9_CR5","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436\u2013444 (2015)","journal-title":"Nature"},{"key":"9_CR6","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)"},{"key":"9_CR7","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u00e9gou, H.: Training data-efficient image transformers & distillation through attention. In: International Conference on Machine Learning, pp. 10347\u201310357. PMLR (2021)"},{"key":"9_CR8","doi-asserted-by":"crossref","unstructured":"Peng, Z., et al.: Conformer: local features coupling global representations for visual recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 367\u2013376 (2021)","DOI":"10.1109\/ICCV48922.2021.00042"},{"key":"9_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2024.108386","volume":"133","author":"A Li","year":"2024","unstructured":"Li, A., et al.: Residual cosine similar attention and bidirectional convolution in dual-branch network for skin lesion image classification. Eng. Appl. Artif. Intell. 133, 108386 (2024)","journal-title":"Eng. Appl. Artif. Intell."},{"key":"9_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2024.108549","volume":"175","author":"Q Lin","year":"2024","unstructured":"Lin, Q., Guo, X., Feng, B., Guo, J., Ni, S., Dong, H.: A novel multi-task learning network for skin lesion classification based on multi-modal clues and label-level fusion. Comput. Biol. Med. 175, 108549 (2024)","journal-title":"Comput. Biol. Med."},{"key":"9_CR11","unstructured":"Tan, M., Le, Q.: Efficientnet: rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning, pp. 6105\u20136114. PMLR (2019)"},{"key":"9_CR12","doi-asserted-by":"crossref","unstructured":"Howard, A., et\u00a0al.: Searching for mobilenetv3. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1314\u20131324 (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"issue":"3","key":"9_CR13","doi-asserted-by":"publisher","first-page":"619","DOI":"10.1109\/TMI.2022.3215547","volume":"42","author":"Z Liu","year":"2023","unstructured":"Liu, Z., Xiong, R., Jiang, T.: Ci-net: clinical-inspired network for automated skin lesion recognition. IEEE Trans. Med. Imaging 42(3), 619\u2013632 (2023)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"9_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.cmpb.2021.106447","volume":"212","author":"S Ding","year":"2021","unstructured":"Ding, S., et al.: Deep attention branch networks for skin lesion classification. Comput. Methods Programs Biomed. 212, 106447 (2021)","journal-title":"Comput. Methods Programs Biomed."},{"key":"9_CR15","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"9_CR16","doi-asserted-by":"crossref","unstructured":"Liu, Z., Mao, H., Wu, C.Y., Feichtenhofer, C., Darrell, T., Xie, S.: A convnet for the 2020s. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11976\u201311986 (2022)","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"9_CR17","doi-asserted-by":"crossref","unstructured":"Datta, S.K., Shaikh, M.A., Srihari, S.N., Gao, M.: Soft attention improves skin cancer classification performance. In: Interpretability of Machine Intelligence in Medical Image Computing, and Topological Data Analysis and Its Applications for Medical Data, p.\u00a013 (2021)","DOI":"10.1007\/978-3-030-87444-5_2"},{"issue":"1","key":"9_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/sdata.2018.161","volume":"5","author":"P Tschandl","year":"2018","unstructured":"Tschandl, P., Rosendahl, C., Kittler, H.: The ham10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 5(1), 1\u20139 (2018)","journal-title":"Sci. Data"},{"key":"9_CR19","unstructured":"Combalia, M., et\u00a0al.: Bcn20000: Dermoscopic lesions in the wild. arXiv preprint arXiv:1908.02288 (2019)"},{"key":"9_CR20","doi-asserted-by":"crossref","unstructured":"Szegedy, C., et al.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.\u00a01\u20139 (2015)","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"9_CR21","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818\u20132826 (2016)","DOI":"10.1109\/CVPR.2016.308"},{"key":"9_CR22","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Ioffe, S., Vanhoucke, V., Alemi, A.: Inception-v4, inception-resnet and the impact of residual connections on learning. In: Proceedings of the AAAI Conference on Artificial Intelligence (2017)","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"9_CR23","unstructured":"Gulati, A., et\u00a0al.: Conformer: convolution-augmented transformer for speech recognition. ArXiv Preprint ArXiv:2005.08100 (2020)"},{"key":"9_CR24","doi-asserted-by":"publisher","first-page":"76261","DOI":"10.1109\/ACCESS.2022.3181225","volume":"10","author":"Z Lan","year":"2022","unstructured":"Lan, Z., Cai, S., He, X., Wen, X.: Fixcaps: an improved capsules network for diagnosis of skin cancer. IEEE Access 10, 76261\u201376267 (2022)","journal-title":"IEEE Access"},{"key":"9_CR25","doi-asserted-by":"crossref","unstructured":"Qu, X., Zhang, C.: Ef2net: better extracting, fusing and focusing text features for scene text detection. In: 2023 4th International Seminar on Artificial Intelligence, Networking and Information Technology (AINIT), pp. 721\u2013727 (2023)","DOI":"10.1109\/AINIT59027.2023.10212684"},{"key":"9_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2024.124826","volume":"255","author":"X Zheng","year":"2024","unstructured":"Zheng, X., Huang, Y., Liu, W., Cai, C.: Lw-xnet for segmentation and classification of skin lesions from dermoscopy images. Expert Syst. Appl. 255, 124826 (2024)","journal-title":"Expert Syst. Appl."},{"key":"9_CR27","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2023.107385","volume":"127","author":"L Tan","year":"2024","unstructured":"Tan, L., Wu, H., Xia, J., Liang, Y., Zhu, J.: Skin lesion recognition via global-local attention and dual-branch input network. Eng. Appl. Artif. Intell. 127, 107385 (2024)","journal-title":"Eng. Appl. Artif. Intell."},{"key":"9_CR28","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"issue":"9","key":"9_CR29","doi-asserted-by":"publisher","first-page":"2092","DOI":"10.1109\/TMI.2019.2893944","volume":"38","author":"J Zhang","year":"2019","unstructured":"Zhang, J., Xie, Y., Xia, Y., Shen, C.: Attention residual learning for skin lesion classification. IEEE Trans. Med. Imaging 38(9), 2092\u20132103 (2019)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"9_CR30","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. CoRR abs\/1409.1556 (2014)"},{"key":"9_CR31","doi-asserted-by":"crossref","unstructured":"SM, J., P, M., Aravindan, C., Appavu, R.: Classification of skin cancer from dermoscopic images using deep neural network architectures. Multimedia Tools Appl. 82(10), 15763\u201315778 (2023)","DOI":"10.1007\/s11042-022-13847-3"}],"container-title":["Lecture Notes in Computer Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-4381-6_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T12:23:06Z","timestamp":1772281386000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-4381-6_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,16]]},"ISBN":["9789819543809","9789819543816"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-4381-6_9","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,16]]},"assertion":[{"value":"16 November 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors have\u00a0no competing interests to declare that are relevant to the content\u00a0of this article.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Okinawa","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 November 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 November 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"32","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iconip2025.apnns.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}