{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T18:08:05Z","timestamp":1774375685233,"version":"3.50.1"},"publisher-location":"Cham","reference-count":22,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031783821","type":"print"},{"value":"9783031783838","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T00:00:00Z","timestamp":1733097600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T00:00:00Z","timestamp":1733097600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-78383-8_20","type":"book-chapter","created":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T21:52:18Z","timestamp":1733089938000},"page":"295-308","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["DG2Net: A MLP-Based Dynamixing Gate and Depthwise Group Norm Network for\u00a0Classification of Glaucoma"],"prefix":"10.1007","author":[{"given":"Yu","family":"Feng","sequence":"first","affiliation":[]},{"given":"Cong","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Yuan","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,2]]},"reference":[{"key":"20_CR1","doi-asserted-by":"crossref","unstructured":"Cao, G., Luo, S., Huang, W., Lan, X., Jiang, D., Wang, Y., Zhang, J.: Strip-mlp: Efficient token interaction for vision mlp. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 1494\u20131504 (2023)","DOI":"10.1109\/ICCV51070.2023.00144"},{"issue":"10","key":"20_CR2","doi-asserted-by":"publisher","first-page":"1516","DOI":"10.1080\/02713683.2021.1900268","volume":"46","author":"H Cho","year":"2021","unstructured":"Cho, H., Hwang, Y.H., Chung, J.K., Lee, K.B., Park, J.S., Kim, H.G., Jeong, J.H.: Deep learning ensemble method for classifying glaucoma stages using fundus photographs and convolutional neural networks. Curr. Eye Res. 46(10), 1516\u20131524 (2021)","journal-title":"Curr. Eye Res."},{"key":"20_CR3","doi-asserted-by":"publisher","first-page":"S50","DOI":"10.1016\/j.optom.2022.08.001","volume":"15","author":"LF Desideri","year":"2022","unstructured":"Desideri, L.F., Rutigliani, C., Corazza, P., Nastasi, A., Roda, M., Nicolo, M., Traverso, C.E., Vagge, A.: The upcoming role of artificial intelligence (ai) for retinal and glaucomatous diseases. Journal of Optometry 15, S50\u2013S57 (2022)","journal-title":"Journal of Optometry"},{"key":"20_CR4","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et\u00a0al.: An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"issue":"1","key":"20_CR5","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1007\/s13534-023-00307-6","volume":"14","author":"G D\u2019Souza","year":"2024","unstructured":"D\u2019Souza, G., Siddalingaswamy, P., Pandya, M.A.: Alternet-k: a small and compact model for the detection of glaucoma. Biomed. Eng. Lett. 14(1), 23\u201333 (2024)","journal-title":"Biomed. Eng. Lett."},{"issue":"1","key":"20_CR6","doi-asserted-by":"publisher","first-page":"483","DOI":"10.1186\/s12886-022-02730-2","volume":"22","author":"KH Hung","year":"2022","unstructured":"Hung, K.H., Kao, Y.C., Tang, Y.H., Chen, Y.T., Wang, C.H., Wang, Y.C., Lee, O.K.S.: Application of a deep learning system in glaucoma screening and further classification with colour fundus photographs: a case control study. BMC Ophthalmol. 22(1), 483 (2022)","journal-title":"BMC Ophthalmol."},{"key":"20_CR7","doi-asserted-by":"crossref","unstructured":"Kiefer, R., Abid, M., Ardali, M.R., Steen, J., Amjadian, E.: Automated fundus image standardization using a dynamic global foreground threshold algorithm. In: 2023 8th International Conference on Image, Vision and Computing (ICIVC). pp. 460\u2013465. IEEE (2023)","DOI":"10.1109\/ICIVC58118.2023.10270429"},{"key":"20_CR8","doi-asserted-by":"crossref","unstructured":"Kiefer, R., Abid, M., Steen, J., Ardali, M.R., Amjadian, E.: A catalog of public glaucoma datasets for machine learning applications: A detailed description and analysis of public glaucoma datasets available to machine learning engineers tackling glaucoma-related problems using retinal fundus images and oct images. In: Proceedings of the 2023 7th International Conference on Information System and Data Mining. pp. 24\u201331 (2023)","DOI":"10.1145\/3603765.3603779"},{"key":"20_CR9","doi-asserted-by":"crossref","unstructured":"Lai, S., Du, X., Guo, J., Zhang, K.: Ramlp: vision mlp via region-aware mixing. In: Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence. pp. 999\u20131007 (2023)","DOI":"10.24963\/ijcai.2023\/111"},{"key":"20_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2023.105607","volume":"88","author":"J Lian","year":"2024","unstructured":"Lian, J., Liu, T.: Lesion identification in fundus images via convolutional neural network-vision transformer. Biomed. Signal Process. Control 88, 105607 (2024)","journal-title":"Biomed. Signal Process. Control"},{"issue":"5","key":"20_CR11","doi-asserted-by":"publisher","first-page":"1405","DOI":"10.1109\/JBHI.2019.2949075","volume":"24","author":"W Liao","year":"2019","unstructured":"Liao, W., Zou, B., Zhao, R., Chen, Y., He, Z., Zhou, M.: Clinical interpretable deep learning model for glaucoma diagnosis. IEEE J. Biomed. Health Inform. 24(5), 1405\u20131412 (2019)","journal-title":"IEEE J. Biomed. Health Inform."},{"key":"20_CR12","first-page":"9204","volume":"34","author":"H Liu","year":"2021","unstructured":"Liu, H., Dai, Z., So, D., Le, Q.V.: Pay attention to mlps. Adv. Neural. Inf. Process. Syst. 34, 9204\u20139215 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"20_CR13","doi-asserted-by":"crossref","unstructured":"Liu, Z., Mao, H., Wu, C.Y., Feichtenhofer, C., Darrell, T., Xie, S.: A convnet for the 2020s. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp. 11976\u201311986 (2022)","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"20_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2021.102559","volume":"67","author":"DR Nayak","year":"2021","unstructured":"Nayak, D.R., Das, D., Majhi, B., Bhandary, S.V., Acharya, U.R.: Ecnet: An evolutionary convolutional network for automated glaucoma detection using fundus images. Biomed. Signal Process. Control 67, 102559 (2021)","journal-title":"Biomed. Signal Process. Control"},{"key":"20_CR15","unstructured":"Tan, M., Le, Q.: Efficientnet: Rethinking model scaling for convolutional neural networks. In: International conference on machine learning. pp. 6105\u20136114. PMLR (2019)"},{"key":"20_CR16","first-page":"24261","volume":"34","author":"IO Tolstikhin","year":"2021","unstructured":"Tolstikhin, I.O., Houlsby, N., Kolesnikov, A., Beyer, L., Zhai, X., Unterthiner, T., Yung, J., Steiner, A., Keysers, D., Uszkoreit, J., et al.: Mlp-mixer: An all-mlp architecture for vision. Adv. Neural. Inf. Process. Syst. 34, 24261\u201324272 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"20_CR17","doi-asserted-by":"crossref","unstructured":"Tu, Z., Talebi, H., Zhang, H., Yang, F., Milanfar, P., Bovik, A., Li, Y.: Maxvit: Multi-axis vision transformer. In: European conference on computer vision. pp. 459\u2013479. Springer (2022)","DOI":"10.1007\/978-3-031-20053-3_27"},{"key":"20_CR18","unstructured":"Wang, Z., Jiang, W., Zhu, Y.M., Yuan, L., Song, Y., Liu, W.: Dynamixer: a vision mlp architecture with dynamic mixing. In: International conference on machine learning. pp. 22691\u201322701. PMLR (2022)"},{"issue":"11","key":"20_CR19","doi-asserted-by":"publisher","first-page":"3217","DOI":"10.1007\/s11517-022-02663-4","volume":"60","author":"C Wu","year":"2022","unstructured":"Wu, C., Li, S., Liu, X., Jiang, F., Shi, B.: Dms-mafm+ efficientnet: a hybrid model for predicting dysthyroid optic neuropathy. Medical & Biological Engineering & Computing 60(11), 3217\u20133230 (2022)","journal-title":"Medical & Biological Engineering & Computing"},{"key":"20_CR20","doi-asserted-by":"crossref","unstructured":"Yu, W., Zhou, P., Yan, S., Wang, X.: Inceptionnext: When inception meets convnext. arXiv preprint arXiv:2303.16900 (2023)","DOI":"10.1109\/CVPR52733.2024.00542"},{"issue":"13","key":"20_CR21","doi-asserted-by":"publisher","first-page":"2180","DOI":"10.3390\/diagnostics13132180","volume":"13","author":"MJ Zedan","year":"2023","unstructured":"Zedan, M.J., Zulkifley, M.A., Ibrahim, A.A., Moubark, A.M., Kamari, N.A.M., Abdani, S.R.: Automated glaucoma screening and diagnosis based on retinal fundus images using deep learning approaches: A comprehensive review. Diagnostics 13(13), 2180 (2023)","journal-title":"Diagnostics"},{"key":"20_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.120389","volume":"228","author":"R Zhang","year":"2023","unstructured":"Zhang, R., Wang, L., Cheng, S., Song, S.: Mlp-based classification of covid-19 and skin diseases. Expert Syst. Appl. 228, 120389 (2023)","journal-title":"Expert Syst. Appl."}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-78383-8_20","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T23:41:15Z","timestamp":1733096475000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-78383-8_20"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,2]]},"ISBN":["9783031783821","9783031783838"],"references-count":22,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-78383-8_20","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,2]]},"assertion":[{"value":"2 December 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICPR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Pattern Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kolkata","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 December 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 December 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icpr2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icpr2024.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}