{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,5]],"date-time":"2025-06-05T04:53:22Z","timestamp":1749099202528,"version":"3.40.3"},"publisher-location":"Cham","reference-count":27,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031159183"},{"type":"electronic","value":"9783031159190"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-15919-0_4","type":"book-chapter","created":{"date-parts":[[2022,9,6]],"date-time":"2022-09-06T04:04:59Z","timestamp":1662437099000},"page":"37-48","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["BiSMSM: A Hybrid MLP-Based Model of\u00a0Global Self-Attention Processes for\u00a0EEG-Based Emotion Recognition"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9235-9429","authenticated-orcid":false,"given":"Wei","family":"Li","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3226-0371","authenticated-orcid":false,"given":"Ye","family":"Tian","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8226-8139","authenticated-orcid":false,"given":"Bowen","family":"Hou","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6484-4363","authenticated-orcid":false,"given":"Jianzhang","family":"Dong","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4689-6140","authenticated-orcid":false,"given":"Shitong","family":"Shao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,7]]},"reference":[{"issue":"4","key":"4_CR1","doi-asserted-by":"publisher","first-page":"1249","DOI":"10.3390\/s21041249","volume":"21","author":"BJ Abbaschian","year":"2021","unstructured":"Abbaschian, B.J., Sierra-Sosa, D., Elmaghraby, A.: Deep learning techniques for speech emotion recognition, from databases to models. Sensors 21(4), 1249 (2021)","journal-title":"Sensors"},{"key":"4_CR2","doi-asserted-by":"publisher","first-page":"143550","DOI":"10.1109\/ACCESS.2019.2944008","volume":"7","author":"F Al-Shargie","year":"2019","unstructured":"Al-Shargie, F., Tariq, U., Alex, M., Mir, H., Al-Nashash, H.: Emotion recognition based on fusion of local cortical activations and dynamic functional networks connectivity: an EEG study. IEEE Access 7, 143550\u2013143562 (2019)","journal-title":"IEEE Access"},{"issue":"3","key":"4_CR3","doi-asserted-by":"publisher","first-page":"374","DOI":"10.1109\/TAFFC.2017.2714671","volume":"10","author":"SM Alarc\u00e3o","year":"2019","unstructured":"Alarc\u00e3o, S.M., Fonseca, M.J.: Emotions recognition using EEG signals: a survey. IEEE Trans. Affect. Comput. 10(3), 374\u2013393 (2019)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"4_CR4","unstructured":"Dosovitskiy, A., et al.: An image is worth 16$$\\times $$16 words: transformers for image recognition at scale. In: 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, 3\u20137 May 2021. OpenReview.net (2021)"},{"key":"4_CR5","doi-asserted-by":"crossref","unstructured":"He, Z., Zhong, Y., Pan, J.: Joint temporal convolutional networks and adversarial discriminative domain adaptation for EEG-based cross-subject emotion recognition. In: ICASSP 2022\u20132022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 3214\u20133218 (2022)","DOI":"10.1109\/ICASSP43922.2022.9746600"},{"issue":"02","key":"4_CR6","doi-asserted-by":"publisher","first-page":"107","DOI":"10.1142\/S0218488598000094","volume":"6","author":"S Hochreiter","year":"1998","unstructured":"Hochreiter, S.: The vanishing gradient problem during learning recurrent neural nets and problem solutions. Internat. J. Uncertain. Fuzziness Knowl.-Based Syst. 6(02), 107\u2013116 (1998)","journal-title":"Internat. J. Uncertain. Fuzziness Knowl.-Based Syst."},{"key":"4_CR7","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1016\/j.patrec.2019.01.008","volume":"120","author":"DK Jain","year":"2019","unstructured":"Jain, D.K., Shamsolmoali, P., Sehdev, P.: Extended deep neural network for facial emotion recognition. Pattern Recogn. Lett. 120, 69\u201374 (2019)","journal-title":"Pattern Recogn. Lett."},{"issue":"3","key":"4_CR8","doi-asserted-by":"publisher","first-page":"327","DOI":"10.1109\/TAFFC.2014.2339834","volume":"5","author":"R Jenke","year":"2014","unstructured":"Jenke, R., Peer, A., Buss, M.: Feature extraction and selection for emotion recognition from EEG. IEEE Trans. Affect. Comput. 5(3), 327\u2013339 (2014)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"4_CR9","doi-asserted-by":"crossref","unstructured":"Jia, Z., Lin, Y., Cai, X., Chen, H., Gou, H., Wang, J.: SST-EmotionNet: spatial-spectral-temporal based attention 3D dense network for EEG emotion recognition. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 2909\u20132917 (2020)","DOI":"10.1145\/3394171.3413724"},{"issue":"1","key":"4_CR10","doi-asserted-by":"publisher","first-page":"98","DOI":"10.1109\/JBHI.2017.2688239","volume":"22","author":"S Katsigiannis","year":"2018","unstructured":"Katsigiannis, S., Ramzan, N.: DREAMER: a database for emotion recognition through EEG and ECG signals from wireless low-cost off-the-shelf devices. IEEE J. Biomed. Health Inform. 22(1), 98\u2013107 (2018)","journal-title":"IEEE J. Biomed. Health Inform."},{"issue":"25","key":"4_CR11","doi-asserted-by":"publisher","first-page":"1359","DOI":"10.1049\/el.2020.2380","volume":"56","author":"S Khare","year":"2020","unstructured":"Khare, S., Nishad, A., Upadhyay, A., Bajaj, V.: Classification of emotions from EEG signals using time-order representation based on the S-transform and convolutional neural network. Electron. Lett. 56(25), 1359\u20131361 (2020)","journal-title":"Electron. Lett."},{"issue":"3","key":"4_CR12","doi-asserted-by":"publisher","first-page":"419","DOI":"10.1007\/BF02344719","volume":"42","author":"KH Kim","year":"2004","unstructured":"Kim, K.H., Bang, S.W., Kim, S.R.: Emotion recognition system using short-term monitoring of physiological signals. Med. Biol. Eng. Comput. 42(3), 419\u2013427 (2004)","journal-title":"Med. Biol. Eng. Comput."},{"issue":"1","key":"4_CR13","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/T-AFFC.2011.15","volume":"3","author":"S Koelstra","year":"2012","unstructured":"Koelstra, S., et al.: DEAP: a database for emotion analysis using physiological signals. IEEE Trans. Affect. Comput. 3(1), 18\u201331 (2012)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"2","key":"4_CR14","doi-asserted-by":"publisher","first-page":"344","DOI":"10.1109\/TCDS.2019.2949306","volume":"12","author":"J Li","year":"2020","unstructured":"Li, J., Qiu, S., Du, C., Wang, Y., He, H.: Domain adaptation for EEG emotion recognition based on latent representation similarity. IEEE Trans. Cogn. Dev. Syst. 12(2), 344\u2013353 (2020)","journal-title":"IEEE Trans. Cogn. Dev. Syst."},{"issue":"10","key":"4_CR15","doi-asserted-by":"publisher","first-page":"2869","DOI":"10.1109\/TBME.2019.2897651","volume":"66","author":"P Li","year":"2019","unstructured":"Li, P., et al.: EEG based emotion recognition by combining functional connectivity network and local activations. IEEE Trans. Biomed. Eng. 66(10), 2869\u20132881 (2019)","journal-title":"IEEE Trans. Biomed. Eng."},{"key":"4_CR16","doi-asserted-by":"publisher","unstructured":"Li, W., Huan, W., Hou, B., Tian, Y., Zhang, Z., Song, A.: Can emotion be transferred?-a review on transfer learning for EEG-based emotion recognition. IEEE Trans. Cogn. Dev. Syst. (2021). https:\/\/doi.org\/10.1109\/TCDS.2021.3098842","DOI":"10.1109\/TCDS.2021.3098842"},{"key":"4_CR17","doi-asserted-by":"publisher","first-page":"109038","DOI":"10.1016\/j.knosys.2022.109038","volume":"250","author":"D Priyasad","year":"2022","unstructured":"Priyasad, D., Fernando, T., Denman, S., Sridharan, S., Fookes, C.: Affect recognition from scalp-EEG using channel-wise encoder networks coupled with geometric deep learning and multi-channel feature fusion. Knowl.-Based Syst. 250, 109038 (2022)","journal-title":"Knowl.-Based Syst."},{"key":"4_CR18","unstructured":"Rached, T.S., Perkusich, A.: Emotion recognition based on brain-computer interface systems. In: Brain-Computer Interface Systems-recent Progress and Future Prospects, pp. 253\u2013270 (2013)"},{"key":"4_CR19","doi-asserted-by":"publisher","unstructured":"Song, T., Zheng, W., Liu, S., Zong, Y., Cui, Z., Li, Y.: Graph-embedded convolutional neural network for image-based EEG emotion recognition. IEEE Trans. Emerg. Top. Comput. 1 (2021). https:\/\/doi.org\/10.1109\/TETC.2021.3087174","DOI":"10.1109\/TETC.2021.3087174"},{"issue":"3","key":"4_CR20","doi-asserted-by":"publisher","first-page":"532","DOI":"10.1109\/TAFFC.2018.2817622","volume":"11","author":"T Song","year":"2020","unstructured":"Song, T., Zheng, W., Song, P., Cui, Z.: EEG emotion recognition using dynamical graph convolutional neural networks. IEEE Trans. Affect. Comput. 11(3), 532\u2013541 (2020)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"4_CR21","doi-asserted-by":"publisher","unstructured":"Tao, W., et al.: EEG-based emotion recognition via channel-wise attention and self attention. IEEE Trans. Affect. Comput. 1\u201312 (2020). https:\/\/doi.org\/10.1109\/TAFFC.2020.3025777","DOI":"10.1109\/TAFFC.2020.3025777"},{"key":"4_CR22","unstructured":"Tolstikhin, H., et al.: MLP-mixer: an all-MLP architecture for vision. Adv. Neural Inf. Process. Syst. 34, 24261\u201324272 (2021)"},{"key":"4_CR23","unstructured":"Vaswani, A., et al.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30, 6000\u20136010 (2017)"},{"key":"4_CR24","doi-asserted-by":"publisher","first-page":"93711","DOI":"10.1109\/ACCESS.2019.2927768","volume":"7","author":"Z Wang","year":"2019","unstructured":"Wang, Z., Tong, Y., Heng, X.: Phase-locking value based graph convolutional neural networks for emotion recognition. IEEE Access 7, 93711\u201393722 (2019)","journal-title":"IEEE Access"},{"key":"4_CR25","doi-asserted-by":"crossref","unstructured":"Xing, X., Li, Z., Xu, T., Shu, L., Hu, B., Xu, X.: SAE+LSTM: a new framework for emotion recognition from multi-channel EEG. Front. Neurorobot. 13, 37(1)\u201337(14) (2019)","DOI":"10.3389\/fnbot.2019.00037"},{"key":"4_CR26","doi-asserted-by":"crossref","unstructured":"Yang, Z., Kay, A., Li, Y., Cross, W., Luo, J.: Pose-based body language recognition for emotion and psychiatric symptom interpretation. In: 2020 25th International Conference on Pattern Recognition (ICPR), pp. 294\u2013301. IEEE, Milan (2021)","DOI":"10.1109\/ICPR48806.2021.9412591"},{"issue":"3","key":"4_CR27","doi-asserted-by":"publisher","first-page":"839","DOI":"10.1109\/TCYB.2017.2788081","volume":"49","author":"T Zhang","year":"2019","unstructured":"Zhang, T., Zheng, W., Cui, Z., Zong, Y., Li, Y.: Spatial-temporal recurrent neural network for emotion recognition. IEEE Trans. Cybern. 49(3), 839\u2013847 (2019)","journal-title":"IEEE Trans. Cybern."}],"container-title":["Lecture Notes in Computer Science","Artificial Neural Networks and Machine Learning \u2013 ICANN 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-15919-0_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,3,29]],"date-time":"2023-03-29T08:14:40Z","timestamp":1680077680000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-15919-0_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031159183","9783031159190"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-15919-0_4","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"7 September 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICANN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Neural Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Bristol","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icann2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/e-nns.org\/icann2022\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"561","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"255","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}