{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T07:19:54Z","timestamp":1760080794701,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":43,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819916443"},{"type":"electronic","value":"9789819916450"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-981-99-1645-0_50","type":"book-chapter","created":{"date-parts":[[2023,4,13]],"date-time":"2023-04-13T17:03:13Z","timestamp":1681405393000},"page":"598-611","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Double Attention-Based Lightweight Network for\u00a0Plant Pest Recognition"],"prefix":"10.1007","author":[{"given":"Janarthan","family":"Sivasubramaniam","sequence":"first","affiliation":[]},{"given":"Thuseethan","family":"Selvarajah","sequence":"additional","affiliation":[]},{"given":"Sutharshan","family":"Rajasegarar","sequence":"additional","affiliation":[]},{"given":"John","family":"Yearwood","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,14]]},"reference":[{"issue":"2","key":"50_CR1","doi-asserted-by":"publisher","first-page":"711","DOI":"10.1002\/ps.6684","volume":"78","author":"R Wang","year":"2022","unstructured":"Wang, R., et al.: An automatic system for pest recognition and forecasting. Pest Manag. Sci. 78(2), 711\u2013721 (2022)","journal-title":"Pest Manag. Sci."},{"issue":"2","key":"50_CR2","first-page":"189","volume":"4","author":"F Fina","year":"2013","unstructured":"Fina, F., Birch, P., Young, R., Obu, J., Faithpraise, B., Chatwin, C.: Automatic plant pest detection and recognition using k-means clustering algorithm and correspondence filters. Int. J. Adv. Biotechnol. Res. 4(2), 189\u2013199 (2013)","journal-title":"Int. J. Adv. Biotechnol. Res."},{"key":"50_CR3","unstructured":"Alvarez, J.M., Salzmann, M.: Learning the number of neurons in deep networks. In: Advances in Neural Information Processing Systems, vol. 29 (2016)"},{"key":"50_CR4","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, vol. 25 (2012)"},{"key":"50_CR5","doi-asserted-by":"crossref","unstructured":"Gidaris, S., Komodakis, N.: Dynamic few-shot visual learning without forgetting. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4367\u20134375 (2018)","DOI":"10.1109\/CVPR.2018.00459"},{"key":"50_CR6","unstructured":"Lake, B., Salakhutdinov, R., Gross, J., Tenenbaum, J.: One shot learning of simple visual concepts. In: Proceedings of the Annual Meeting of the Cognitive Science Society, vol. 33, no. 33 (2011)"},{"issue":"3","key":"50_CR7","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3386252","volume":"53","author":"Y Wang","year":"2020","unstructured":"Wang, Y., Yao, Q., Kwok, J.T., Ni, L.M.: Generalizing from a few examples: a survey on few-shot learning. ACM Comput. Surv. (CSUR) 53(3), 1\u201334 (2020)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"50_CR8","doi-asserted-by":"crossref","unstructured":"Ye, H.J., Hu, H., Zhan, D.C., Sha, F.: Few-shot learning via embedding adaptation with set-to-set functions. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8808\u20138817 (2020)","DOI":"10.1109\/CVPR42600.2020.00883"},{"key":"50_CR9","unstructured":"Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., Keutzer, K.: SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and $$<$$0.5 MB model size. arXiv preprint arXiv:1602.07360 (2016)"},{"key":"50_CR10","doi-asserted-by":"publisher","first-page":"10858","DOI":"10.1109\/ACCESS.2021.3050628","volume":"9","author":"J Zhang","year":"2021","unstructured":"Zhang, J., Zhu, H., Wang, P., Ling, X.: ATT squeeze U-net: a lightweight network for forest fire detection and recognition. IEEE Access 9, 10858\u201310870 (2021)","journal-title":"IEEE Access"},{"key":"50_CR11","doi-asserted-by":"publisher","first-page":"5573","DOI":"10.1109\/ACCESS.2020.3046715","volume":"9","author":"N Zhou","year":"2020","unstructured":"Zhou, N., Liang, R., Shi, W.: A lightweight convolutional neural network for real-time facial expression detection. IEEE Access 9, 5573\u20135584 (2020)","journal-title":"IEEE Access"},{"issue":"7","key":"50_CR12","doi-asserted-by":"publisher","first-page":"369","DOI":"10.1007\/s12243-019-00731-9","volume":"75","author":"J Zhang","year":"2020","unstructured":"Zhang, J., Wang, W., Lu, C., Wang, J., Sangaiah, A.K.: Lightweight deep network for traffic sign classification. Ann. Telecommun. 75(7), 369\u2013379 (2020)","journal-title":"Ann. Telecommun."},{"key":"50_CR13","series-title":"Communications in Computer and Information Science","doi-asserted-by":"publisher","first-page":"172","DOI":"10.1007\/978-3-030-63823-8_21","volume-title":"Neural Information Processing","author":"AH Rashid","year":"2020","unstructured":"Rashid, A.H., Razzak, I., Tanveer, M., Robles-Kelly, A.: RipNet: a lightweight one-class deep neural network for the identification of RIP currents. In: Yang, H., Pasupa, K., Leung, A.C.-S., Kwok, J.T., Chan, J.H., King, I. (eds.) ICONIP 2020. CCIS, vol. 1333, pp. 172\u2013179. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-63823-8_21"},{"key":"50_CR14","unstructured":"Howard, A.G., et al.: Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)"},{"key":"50_CR15","unstructured":"Tan, M., Le, Q.: Efficientnet: rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning, pp. 6105\u20136114. PMLR (2019)"},{"key":"50_CR16","doi-asserted-by":"crossref","unstructured":"Yang, Z., Yang, X., Li, M., Li, W.: Automated garden-insect recognition using improved lightweight convolution network. Inf. Process. Agric. (2021)","DOI":"10.1016\/j.inpa.2021.12.006"},{"issue":"12","key":"50_CR17","doi-asserted-by":"publisher","first-page":"1587","DOI":"10.3390\/e23121587","volume":"23","author":"M Zha","year":"2021","unstructured":"Zha, M., Qian, W., Yi, W., Hua, J.: A lightweight YOLOv4-Based forestry pest detection method using coordinate attention and feature fusion. Entropy 23(12), 1587 (2021)","journal-title":"Entropy"},{"key":"50_CR18","doi-asserted-by":"crossref","unstructured":"Wang, F., et al.: Residual attention network for image classification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3156\u20133164 (2017)","DOI":"10.1109\/CVPR.2017.683"},{"key":"50_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2020.106566","volume":"212","author":"MA Takalkar","year":"2021","unstructured":"Takalkar, M.A., Thuseethan, S., Rajasegarar, S., Chaczko, Z., Xu, M., Yearwood, J.: LGAttNet: automatic micro-expression detection using dual-stream local and global attentions. Knowl.-Based Syst. 212, 106566 (2021)","journal-title":"Knowl.-Based Syst."},{"key":"50_CR20","doi-asserted-by":"publisher","first-page":"167","DOI":"10.1016\/j.foodcont.2018.07.008","volume":"94","author":"KS Banga","year":"2018","unstructured":"Banga, K.S., Kotwaliwale, N., Mohapatra, D., Giri, S.K.: Techniques for insect detection in stored food grains: an overview. Food Control 94, 167\u2013176 (2018)","journal-title":"Food Control"},{"issue":"4","key":"50_CR21","doi-asserted-by":"publisher","first-page":"635","DOI":"10.1007\/s11119-016-9473-6","volume":"18","author":"H Liu","year":"2017","unstructured":"Liu, H., Lee, S.H., Chahl, J.S.: A review of recent sensing technologies to detect invertebrates on crops. Precision Agric. 18(4), 635\u2013666 (2017)","journal-title":"Precision Agric."},{"issue":"1","key":"50_CR22","first-page":"27","volume":"8","author":"LC Ngugi","year":"2021","unstructured":"Ngugi, L.C., Abelwahab, M., Abo-Zahhad, M.: Recent advances in image processing techniques for automated leaf pest and disease recognition-a review. Inf. Process. Agric. 8(1), 27\u201351 (2021)","journal-title":"Inf. Process. Agric."},{"key":"50_CR23","doi-asserted-by":"crossref","unstructured":"Nagar, H., Sharma, R.S.: A comprehensive survey on pest detection techniques using image processing. In: 2020 4th International Conference on Intelligent Computing and Control Systems (ICICCS), pp. 43\u201348. IEEE (2020)","DOI":"10.1109\/ICICCS48265.2020.9120889"},{"issue":"3","key":"50_CR24","doi-asserted-by":"publisher","first-page":"1636","DOI":"10.3906\/elk-1809-181","volume":"27","author":"M T\u00fcrko\u011flu","year":"2019","unstructured":"T\u00fcrko\u011flu, M., Hanbay, D.: Plant disease and pest detection using deep learning-based features. Turk. J. Electr. Eng. Comput. Sci. 27(3), 1636\u20131651 (2019)","journal-title":"Turk. J. Electr. Eng. Comput. Sci."},{"issue":"3","key":"50_CR25","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1111\/jen.12834","volume":"145","author":"DJA Rustia","year":"2021","unstructured":"Rustia, D.J.A., et al.: Automatic greenhouse insect pest detection and recognition based on a cascaded deep learning classification method. J. Appl. Entomol. 145(3), 206\u2013222 (2021)","journal-title":"J. Appl. Entomol."},{"key":"50_CR26","unstructured":"Faisal, M.S.A.B.: A pest monitoring system for agriculture using deep learning. Res. Progr. Mech. Manuf. Eng. 2(2), 1023\u20131034 (2021)"},{"issue":"11","key":"50_CR27","doi-asserted-by":"publisher","first-page":"2140","DOI":"10.3390\/sym13112140","volume":"13","author":"J Chen","year":"2021","unstructured":"Chen, J., Liu, Q., Gao, L.: Deep convolutional neural networks for tea tree pest recognition and diagnosis. Symmetry 13(11), 2140 (2021)","journal-title":"Symmetry"},{"key":"50_CR28","doi-asserted-by":"crossref","unstructured":"Liu, Y., Zhang, X., Gao, Y., Qu, T., Shi, Y.: Improved CNN method for crop pest identification based on transfer learning. Comput. Intell. Neurosci. (2022)","DOI":"10.1155\/2022\/9709648"},{"issue":"2","key":"50_CR29","doi-asserted-by":"publisher","first-page":"301","DOI":"10.1007\/s11760-021-01909-2","volume":"16","author":"M Turkoglu","year":"2022","unstructured":"Turkoglu, M., Yaniko\u011flu, B., Hanbay, D.: PlantDiseaseNet: convolutional neural network ensemble for plant disease and pest detection. SIViP 16(2), 301\u2013309 (2022)","journal-title":"SIViP"},{"issue":"1","key":"50_CR30","doi-asserted-by":"publisher","first-page":"552","DOI":"10.1080\/21642583.2022.2074168","volume":"10","author":"S Zhang","year":"2022","unstructured":"Zhang, S., Jing, R., Shi, X.: Crop pest recognition based on a modified capsule network. Syst. Sci. Control Eng. 10(1), 552\u2013561 (2022)","journal-title":"Syst. Sci. Control Eng."},{"issue":"10","key":"50_CR31","doi-asserted-by":"publisher","first-page":"1630","DOI":"10.3390\/electronics11101630","volume":"11","author":"C Xu","year":"2022","unstructured":"Xu, C., Yu, C., Zhang, S., Wang, X.: Multi-scale convolution-capsule network for crop insect pest recognition. Electronics 11(10), 1630 (2022)","journal-title":"Electronics"},{"key":"50_CR32","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2020.105240","volume":"169","author":"Y Li","year":"2020","unstructured":"Li, Y., Yang, J.: Few-shot cotton pest recognition and terminal realization. Comput. Electron. Agric. 169, 105240 (2020)","journal-title":"Comput. Electron. Agric."},{"key":"50_CR33","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.biosystemseng.2021.05.006","volume":"208","author":"DJA Rustia","year":"2021","unstructured":"Rustia, D.J.A., et al.: Online semi-supervised learning applied to an automated insect pest monitoring system. Biosys. Eng. 208, 28\u201344 (2021)","journal-title":"Biosys. Eng."},{"key":"50_CR34","doi-asserted-by":"publisher","first-page":"162588","DOI":"10.1109\/ACCESS.2020.3021487","volume":"8","author":"S Janarthan","year":"2020","unstructured":"Janarthan, S., Thuseethan, S., Rajasegarar, S., Lyu, Q., Zheng, Y., Yearwood, J.: Deep metric learning based citrus disease classification with sparse data. IEEE Access 8, 162588\u2013162600 (2020)","journal-title":"IEEE Access"},{"key":"50_CR35","unstructured":"Liang, K., Wang, Y., Sun, L.: Imbalance data set classification of tomato pest based on lightweight CNN model (2021)"},{"key":"50_CR36","doi-asserted-by":"crossref","unstructured":"Zhao, S., Liu, J., Bai, Z., Hu, C., Jin, Y.: Crop pest recognition in real agricultural environment using convolutional neural networks by a parallel attention mechanism. Front. Plant Sci. 13 (2022)","DOI":"10.3389\/fpls.2022.839572"},{"key":"50_CR37","doi-asserted-by":"publisher","first-page":"162448","DOI":"10.1109\/ACCESS.2021.3132486","volume":"9","author":"X Yang","year":"2021","unstructured":"Yang, X., Luo, Y., Li, M., Yang, Z., Sun, C., Li, W.: Recognizing pests in field-based images by combining spatial and channel attention mechanism. IEEE Access 9, 162448\u2013162458 (2021)","journal-title":"IEEE Access"},{"issue":"4","key":"50_CR38","doi-asserted-by":"publisher","first-page":"427","DOI":"10.1007\/s13355-021-00732-y","volume":"56","author":"J Chen","year":"2021","unstructured":"Chen, J., Chen, W., Zeb, A., Zhang, D., Nanehkaran, Y.A.: Crop pest recognition using attention-embedded lightweight network under field conditions. Appl. Entomol. Zool. 56(4), 427\u2013442 (2021). https:\/\/doi.org\/10.1007\/s13355-021-00732-y","journal-title":"Appl. Entomol. Zool."},{"key":"50_CR39","unstructured":"Chen, Y., Kalantidis, Y., Li, J., Yan, S., Feng, J.: A$$^2$$-nets: double attention networks. In: Advances in Neural Information Processing Systems, vol. 31 (2018)"},{"key":"50_CR40","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"50_CR41","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2019.105174","volume":"169","author":"Y Li","year":"2020","unstructured":"Li, Y., Wang, H., Dang, L.M., Sadeghi-Niaraki, A., Moon, H.: Crop pest recognition in natural scenes using convolutional neural networks. Comput. Electron. Agric. 169, 105174 (2020)","journal-title":"Comput. Electron. Agric."},{"key":"50_CR42","doi-asserted-by":"publisher","first-page":"351","DOI":"10.1016\/j.compag.2017.08.005","volume":"141","author":"X Cheng","year":"2017","unstructured":"Cheng, X., Zhang, Y., Chen, Y., Wu, Y., Yue, Y.: Pest identification via deep residual learning in complex background. Comput. Electron. Agric. 141, 351\u2013356 (2017)","journal-title":"Comput. Electron. Agric."},{"key":"50_CR43","doi-asserted-by":"publisher","first-page":"123","DOI":"10.1016\/j.compag.2015.10.015","volume":"119","author":"C Xie","year":"2015","unstructured":"Xie, C., et al.: Automatic classification for field crop insects via multiple-task sparse representation and multiple-kernel learning. Comput. Electron. Agric. 119, 123\u2013132 (2015)","journal-title":"Comput. Electron. Agric."}],"container-title":["Communications in Computer and Information Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-1645-0_50","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,13]],"date-time":"2023-04-13T17:22:21Z","timestamp":1681406541000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-1645-0_50"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9789819916443","9789819916450"],"references-count":43,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-1645-0_50","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"14 April 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Delhi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 November 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 November 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iconip2022.apnns.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easy Chair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"810","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"359","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"44% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.65","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ICONIP 2022 consists of a two-volume set, LNCS & CCIS, which includes 146 and 213 papers","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}