{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T18:56:28Z","timestamp":1763664988149,"version":"3.40.3"},"publisher-location":"Cham","reference-count":31,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031194955"},{"type":"electronic","value":"9783031194962"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19496-2_28","type":"book-chapter","created":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T05:03:19Z","timestamp":1666414999000},"page":"356-368","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["An Intelligent Human Activity Recognizer for\u00a0Visually Impaired People Using VGG-SVM Model"],"prefix":"10.1007","author":[{"given":"Rabeeya","family":"Saleem","sequence":"first","affiliation":[]},{"given":"Tauqir","family":"Ahmad","sequence":"additional","affiliation":[]},{"given":"Muhammad","family":"Aslam","sequence":"additional","affiliation":[]},{"given":"A. M.","family":"Martinez-Enriquez","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,23]]},"reference":[{"key":"28_CR1","unstructured":"Fast facts of common eye disorders, June 2020. https:\/\/www.cdc.gov\/visionhealth\/basics\/ced\/fastfacts.htm"},{"key":"28_CR2","unstructured":"Vision impairment and blindness (2022). https:\/\/www.who.int\/news-room\/fact-sheets\/detail\/blindness-and-visual-impairment"},{"key":"28_CR3","doi-asserted-by":"crossref","unstructured":"Aparna, R., Chitralekha, C., Chaudhari, S.: Comparative study of CNN, VGG16 with LSTM and VGG16 with bidirectional LSTM using kitchen activity dataset. In: 2021 Fifth International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud)(I-SMAC), pp. 836\u2013843. IEEE (2021)","DOI":"10.1109\/I-SMAC52330.2021.9640728"},{"key":"28_CR4","doi-asserted-by":"crossref","unstructured":"Cao, Z., Simon, T., Wei, S.E., Sheikh, Y.: Realtime multi-person 2D pose estimation using part affinity fields. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7291\u20137299 (2017)","DOI":"10.1109\/CVPR.2017.143"},{"key":"28_CR5","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2019.105820","volume":"86","author":"C Dai","year":"2020","unstructured":"Dai, C., Liu, X., Lai, J.: Human action recognition using two-stream attention based LSTM networks. Appl. Soft Comput. 86, 105820 (2020)","journal-title":"Appl. Soft Comput."},{"issue":"1","key":"28_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s41598-022-11173-0","volume":"12","author":"T Debnath","year":"2022","unstructured":"Debnath, T., Reza, M., Rahman, A., Beheshti, A., Band, S.S., Alinejad-Rokny, H., et al.: Four-layer convnet to facial emotion recognition with minimal epochs and the significance of data diversity. Sci. Rep. 12(1), 1\u201318 (2022)","journal-title":"Sci. Rep."},{"key":"28_CR7","doi-asserted-by":"crossref","unstructured":"Deep, S., Zheng, X.: Leveraging CNN and transfer learning for vision-based human activity recognition. In: 2019 29th International Telecommunication Networks and Applications Conference (ITNAC), pp. 1\u20134. IEEE (2019)","DOI":"10.1109\/ITNAC46935.2019.9078016"},{"issue":"5","key":"28_CR8","doi-asserted-by":"publisher","first-page":"1911","DOI":"10.3390\/s22051911","volume":"22","author":"I Dirgov\u00e1 Lupt\u00e1kov\u00e1","year":"2022","unstructured":"Dirgov\u00e1 Lupt\u00e1kov\u00e1, I., Kubov\u010d\u00edk, M., Posp\u00edchal, J.: Wearable sensor-based human activity recognition with transformer model. Sensors 22(5), 1911 (2022)","journal-title":"Sensors"},{"issue":"7","key":"28_CR9","doi-asserted-by":"publisher","first-page":"1461","DOI":"10.1007\/s00607-021-00928-8","volume":"103","author":"N Dua","year":"2021","unstructured":"Dua, N., Singh, S.N., Semwal, V.B.: Multi-input CNN-GRU based human activity recognition using wearable sensors. Computing 103(7), 1461\u20131478 (2021)","journal-title":"Computing"},{"key":"28_CR10","doi-asserted-by":"publisher","first-page":"60736","DOI":"10.1109\/ACCESS.2019.2913393","volume":"7","author":"M Ehatisham-Ul-Haq","year":"2019","unstructured":"Ehatisham-Ul-Haq, M., Javed, A., Azam, M.A., Malik, H.M., Irtaza, A., Lee, I.H., Mahmood, M.T.: Robust human activity recognition using multimodal feature-level fusion. IEEE Access 7, 60736\u201360751 (2019)","journal-title":"IEEE Access"},{"issue":"4","key":"28_CR11","doi-asserted-by":"publisher","first-page":"410","DOI":"10.1080\/17483107.2018.1449018","volume":"14","author":"S Gamache","year":"2019","unstructured":"Gamache, S., Routhier, F., Morales, E., Vandersmissen, M.H., Boucher, N.: Mapping review of accessible pedestrian infrastructures for individuals with physical disabilities. Disabil. Rehabil. Assistive Technol. 14(4), 410\u2013422 (2019)","journal-title":"Disabil. Rehabil. Assistive Technol."},{"issue":"12","key":"28_CR12","doi-asserted-by":"publisher","first-page":"2247","DOI":"10.1109\/TPAMI.2007.70711","volume":"29","author":"L Gorelick","year":"2007","unstructured":"Gorelick, L., Blank, M., Shechtman, E., Irani, M., Basri, R.: Actions as space-time shapes. Trans. Pattern Anal. Mach. Intell. 29(12), 2247\u20132253 (2007)","journal-title":"Trans. Pattern Anal. Mach. Intell."},{"issue":"24","key":"28_CR13","doi-asserted-by":"publisher","first-page":"8404","DOI":"10.3390\/s21248404","volume":"21","author":"Z Hao","year":"2021","unstructured":"Hao, Z., Zhang, D., Dang, X., Liu, G., Bai, Y.: Wi-CAS: a contactless method for continuous indoor human activity sensing using Wi-Fi devices. Sensors 21(24), 8404 (2021)","journal-title":"Sensors"},{"key":"28_CR14","doi-asserted-by":"crossref","unstructured":"Jalal, A., Kamal, S., Kim, D.: A depth video-based human detection and activity recognition using multi-features and embedded hidden markov models for health care monitoring systems (2017)","DOI":"10.9781\/ijimai.2017.447"},{"key":"28_CR15","unstructured":"Kay, W., et al.: The kinetics human action video dataset. arXiv preprint arXiv:1705.06950 (2017)"},{"issue":"28","key":"28_CR16","doi-asserted-by":"publisher","first-page":"35827","DOI":"10.1007\/s11042-020-09408-1","volume":"80","author":"MA Khan","year":"2021","unstructured":"Khan, M.A., Zhang, Y.D., Khan, S.A., Attique, M., Rehman, A., Seo, S.: A resource conscious human action recognition framework using 26-layered deep convolutional neural network. Multimedia Tools Appl. 80(28), 35827\u201335849 (2021)","journal-title":"Multimedia Tools Appl."},{"issue":"6","key":"28_CR17","first-page":"5270","volume":"9","author":"H Kim","year":"2019","unstructured":"Kim, H., Lee, S., Jung, H.: Human activity recognition by using convolutional neural network. Int. J. Electr. Comput. Eng. 9(6), 5270 (2019)","journal-title":"Int. J. Electr. Comput. Eng."},{"key":"28_CR18","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J.F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The extended Cohn-Kanade dataset (CK+): A complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-Workshops, pp. 94\u2013101. IEEE (2010)","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"28_CR19","doi-asserted-by":"crossref","unstructured":"Mmereki, W., Jamisola, R.S., Mpoeleng, D., Petso, T.: YOLOv3-based human activity recognition as viewed from a moving high-altitude aerial camera. In: 2021 7th International Conference on Automation, Robotics and Applications (ICARA), pp. 241\u2013246. IEEE (2021)","DOI":"10.1109\/ICARA51699.2021.9376435"},{"key":"28_CR20","doi-asserted-by":"publisher","first-page":"820","DOI":"10.1016\/j.future.2021.06.045","volume":"125","author":"K Muhammad","year":"2021","unstructured":"Muhammad, K., et al.: Human action recognition using attention based LSTM network with dilated CNN features. Future Gener. Comput. Syst. 125, 820\u2013830 (2021)","journal-title":"Future Gener. Comput. Syst."},{"key":"28_CR21","doi-asserted-by":"crossref","unstructured":"Mutegeki, R., Han, D.S.: A CNN-LSTM approach to human activity recognition. In: 2020 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), pp. 362\u2013366. IEEE (2020)","DOI":"10.1109\/ICAIIC48513.2020.9065078"},{"key":"28_CR22","doi-asserted-by":"crossref","unstructured":"Nadeem, A., Jalal, A., Kim, K.: Human actions tracking and recognition based on body parts detection via artificial neural network. In: 2020 3rd International Conference on Advancements in Computational Sciences (ICACS), pp. 1\u20136. IEEE (2020)","DOI":"10.1109\/ICACS47775.2020.9055951"},{"issue":"14","key":"28_CR23","doi-asserted-by":"publisher","first-page":"21465","DOI":"10.1007\/s11042-021-10687-5","volume":"80","author":"A Nadeem","year":"2021","unstructured":"Nadeem, A., Jalal, A., Kim, K.: Automatic human posture estimation for sport activity recognition with robust body parts detection and entropy Markov model. Multimedia Tools Appl. 80(14), 21465\u201321498 (2021)","journal-title":"Multimedia Tools Appl."},{"issue":"12","key":"28_CR24","doi-asserted-by":"publisher","first-page":"18365","DOI":"10.1007\/s11042-021-10682-w","volume":"80","author":"AJ Naik","year":"2021","unstructured":"Naik, A.J., Gopalakrishna, M.: Deep-violence: individual person violent activity detection in video. Multimedia Tools Appl. 80(12), 18365\u201318380 (2021)","journal-title":"Multimedia Tools Appl."},{"issue":"2","key":"28_CR25","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0247448","volume":"16","author":"S Par\u00e9","year":"2021","unstructured":"Par\u00e9, S., Bleau, M., Djerourou, I., Malotaux, V., Kupers, R., Ptito, M.: Spatial navigation with horizontally spatialized sounds in early and late blind individuals. PloS ONE 16(2), e0247448 (2021)","journal-title":"PloS ONE"},{"key":"28_CR26","doi-asserted-by":"publisher","first-page":"86934","DOI":"10.1109\/ACCESS.2020.2991731","volume":"8","author":"C Pham","year":"2020","unstructured":"Pham, C., et al.: SensCapsNet: deep neural network for non-obtrusive sensing based human activity recognition. IEEE Access 8, 86934\u201386946 (2020)","journal-title":"IEEE Access"},{"issue":"14","key":"28_CR27","doi-asserted-by":"publisher","first-page":"3160","DOI":"10.3390\/s19143160","volume":"19","author":"I Rodr\u00edguez-Moreno","year":"2019","unstructured":"Rodr\u00edguez-Moreno, I., Mart\u00ednez-Otzeta, J.M., Sierra, B., Rodriguez, I., Jauregi, E.: Video activity recognition: state-of-the-art. Sensors 19(14), 3160 (2019)","journal-title":"Sensors"},{"key":"28_CR28","doi-asserted-by":"publisher","first-page":"235","DOI":"10.1016\/j.eswa.2016.04.032","volume":"59","author":"CA Ronao","year":"2016","unstructured":"Ronao, C.A., Cho, S.B.: Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst. Appl. 59, 235\u2013244 (2016)","journal-title":"Expert Syst. Appl."},{"key":"28_CR29","doi-asserted-by":"crossref","unstructured":"Sun, H., et al.: Color correction and repair of haze images under hue-saturation-intensity color space and machine learning (2021)","DOI":"10.21203\/rs.3.rs-170541\/v1"},{"issue":"2","key":"28_CR30","doi-asserted-by":"publisher","first-page":"743","DOI":"10.1007\/s11036-019-01445-x","volume":"25","author":"S Wan","year":"2020","unstructured":"Wan, S., Qi, L., Xu, X., Tong, C., Gu, Z.: Deep learning models for real-time human activity recognition with smartphones. Mob. Netw. Appl. 25(2), 743\u2013755 (2020)","journal-title":"Mob. Netw. Appl."},{"key":"28_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2020.105742","volume":"178","author":"D Wu","year":"2020","unstructured":"Wu, D., Lv, S., Jiang, M., Song, H.: Using channel pruning-based YOLO v4 deep learning algorithm for the real-time and accurate detection of apple flowers in natural environments. Comput. Electron. Agric. 178, 105742 (2020)","journal-title":"Comput. Electron. Agric."}],"container-title":["Lecture Notes in Computer Science","Advances in Computational Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19496-2_28","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,6]],"date-time":"2024-10-06T09:51:36Z","timestamp":1728208296000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19496-2_28"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031194955","9783031194962"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19496-2_28","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"23 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Mexican International Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Monterrey","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Mexico","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"micai2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.micai.org\/2022\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"137","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"63","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"46% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"17 External reviewers","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}