{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,2]],"date-time":"2026-05-02T06:39:51Z","timestamp":1777703991292,"version":"3.51.4"},"reference-count":21,"publisher":"SAGE Publications","issue":"2","license":[{"start":{"date-parts":[[2018,6,13]],"date-time":"2018-06-13T00:00:00Z","timestamp":1528848000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/journals.sagepub.com\/page\/policies\/text-and-data-mining-license"}],"content-domain":{"domain":["journals.sagepub.com"],"crossmark-restriction":true},"short-container-title":["Journal of Intelligent &amp; Fuzzy Systems"],"published-print":{"date-parts":[[2018,8,26]]},"abstract":"<jats:p>Smartphone has been used for recognizing the different motion activities. However, current studies focus on either improving algorithm factor or adjusting neural network structure factor rather than on time cost factor and actual application factor. A novel method to consider these four factors comprehensively enhancing recognition of motion state accuracy is proposed. An architecture of the Bi-LSTM neural network and the TensorFlow machine learning system are used to classify the motion state and evaluate its experimental results. In addition, the Bi-LSTM neural network is compared with other neural network structures. Meanwhile, using the data captured by the accelerometer sensor and gyroscope sensor of the smartphone tests the Bi-LSTM neural network model. Experimental results show that using Bi-LSTM neural network and TensorFlow machine learning system to extract motion state characteristics, this method makes the motion state identification achieve 86.7% accuracy and the Bi-LSTM neural network model is better than other neural network models considering above four factors. The model of Bi-LSTM neural network can be used for other time-series fields such as signal recognition, action analysis, etc. This study provides a new method, which considers the four factors, to enhance the accuracy of the motion state classification.<\/jats:p>","DOI":"10.3233\/jifs-169709","type":"journal-article","created":{"date-parts":[[2018,6,15]],"date-time":"2018-06-15T13:28:33Z","timestamp":1529069313000},"page":"1733-1742","update-policy":"https:\/\/doi.org\/10.1177\/sage-journals-update-policy","source":"Crossref","is-referenced-by-count":2,"title":["Recognition of motion state by smartphone sensors using Bi-LSTM neural network"],"prefix":"10.1177","volume":"35","author":[{"given":"Hong","family":"Zhao","sequence":"first","affiliation":[{"name":"School of Computer and Communication, Lanzhou University of Technology, Lanzhou, China"},{"name":"Information Center, Lanzhou University of Technology, Lanzhou, China"}]},{"given":"Chunning","family":"Hou","sequence":"additional","affiliation":[{"name":"School of Computer and Communication, Lanzhou University of Technology, Lanzhou, China"},{"name":"Information Center, Lanzhou University of Technology, Lanzhou, China"}]}],"member":"179","published-online":{"date-parts":[[2018,6,13]]},"reference":[{"key":"e_1_3_2_2_2","doi-asserted-by":"publisher","DOI":"10.1007\/s12668-013-0088-3"},{"key":"e_1_3_2_3_2","unstructured":"LockhartJ.W. PulickalT. and WeissG.M. Applications of mobile activity recognition International Workshop on Situation Activity and Goal Awareness (2012) pp. 1054\u20131058."},{"key":"e_1_3_2_4_2","first-page":"265","article-title":"TensorFlow: A system for large-scale machine learning","volume":"16","author":"Abadi M.","year":"2016","unstructured":"AbadiM., TensorFlow: A system for large-scale machine learning, OSDI16 (2016), 265\u2013283.","journal-title":"OSDI"},{"key":"e_1_3_2_5_2","doi-asserted-by":"publisher","DOI":"10.1016\/S0747-5632(99)00037-0"},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.3390\/s16060800"},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.medengphy.2014.02.009"},{"key":"e_1_3_2_8_2","doi-asserted-by":"publisher","DOI":"10.1109\/TASE.2013.2256349"},{"key":"e_1_3_2_9_2","doi-asserted-by":"crossref","unstructured":"ZengM. Convolutional Neural Networks for human activity recognition using mobile sensors International Conference on Mobile Computing Applications and Services IEEE (2015) pp. 197\u2013205.","DOI":"10.4108\/icst.mobicase.2014.257786"},{"key":"e_1_3_2_10_2","doi-asserted-by":"crossref","unstructured":"JiangW. and YinZ. Human Activity Recognition Using Wearable Sensors by Deep Convolutional Neural Networks ACM International Conference on Multimedia ACM2015 pp. 1307\u20131310.","DOI":"10.1145\/2733373.2806333"},{"key":"e_1_3_2_11_2","doi-asserted-by":"crossref","unstructured":"ChoK. and ChenX. Classifying and visualizing motion capture sequences using deep neural networks International Conference on Computer Vision Theory and Applications IEEE2015 pp. 122\u2013130.","DOI":"10.5220\/0004718301220130"},{"key":"e_1_3_2_12_2","doi-asserted-by":"crossref","unstructured":"ZhuW. Co-occurrence feature learning for skeleton based action recognition using regularized deep LSTM networks Thirtieth AAAI Conference on Artificial Intelligence AAAI Press2016 pp. 3697\u20133703.","DOI":"10.1609\/aaai.v30i1.10451"},{"issue":"2","key":"e_1_3_2_13_2","first-page":"454","article-title":"Deep, convolutional, and recurrent models for human activity recognition using wearables","volume":"61","author":"Hammerla N.Y.","year":"2016","unstructured":"HammerlaN.Y., HalloranS. and PloetzT., Deep, convolutional, and recurrent models for human activity recognition using wearables, Journal of Scientific Computing61(2) (2016), 454\u2013476.","journal-title":"Journal of Scientific Computing"},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.pmcj.2010.11.004"},{"issue":"5","key":"e_1_3_2_15_2","first-page":"41","article-title":"Human action classification based on LSTM neural network on TensorFlow","volume":"7","author":"Yang Y.","year":"2017","unstructured":"YangY. and weiZ., Human action classification based on LSTM neural network on TensorFlow, Intelligent Computer and Applications7(5) (2017), 41\u201345.","journal-title":"Intelligent Computer and Applications"},{"key":"e_1_3_2_16_2","doi-asserted-by":"crossref","unstructured":"SinghD. MerdivanE. PsychoulaI.. ctivity recognition using recurrent neural networks[C] International Cross-Domain Conference for Machine Learning and Knowledge Extraction Springer Cham 2017 pp. 267\u2013274Humana.","DOI":"10.1007\/978-3-319-66808-6_18"},{"issue":"2","key":"e_1_3_2_17_2","first-page":"4263","article-title":"An end-to-end spatio-temporal attention model for human action recognition from skeleton data[C]","volume":"1","author":"Song S.","year":"2017","unstructured":"SongS., LanC., XingJ.., An end-to-end spatio-temporal attention model for human action recognition from skeleton data[C], AAAI1(2) (2017), 4263\u20134270.","journal-title":"AAAI"},{"key":"e_1_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.3390\/s17020319"},{"key":"e_1_3_2_19_2","first-page":"285","article-title":"Sensor-activity relevance in human activity recognition with wearable motion sensors and mutual information criterion","volume":"264","author":"Dobrucali O.","year":"2013","unstructured":"DobrucaliO. and BarshanB., Sensor-activity relevance in human activity recognition with wearable motion sensors and mutual information criterion, Information Sciences and Systems264 (2013), 285\u2013294.","journal-title":"Information Sciences and Systems"},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"e_1_3_2_21_2","doi-asserted-by":"crossref","unstructured":"SchusterM. and PaliwalK.K. Bidirectional recurrent neural networks IEEE Press1997.","DOI":"10.1109\/78.650093"},{"key":"e_1_3_2_22_2","unstructured":"AnguitaD. A Public Domain Dataset for Human Activity Recognition using Smartphones ESANN (2013) pp. 437\u2013442."}],"container-title":["Journal of Intelligent &amp; Fuzzy Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.3233\/JIFS-169709","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/full-xml\/10.3233\/JIFS-169709","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.3233\/JIFS-169709","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,29]],"date-time":"2026-04-29T09:40:09Z","timestamp":1777455609000},"score":1,"resource":{"primary":{"URL":"https:\/\/journals.sagepub.com\/doi\/10.3233\/JIFS-169709"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,6,13]]},"references-count":21,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2018,8,26]]}},"alternative-id":["10.3233\/JIFS-169709"],"URL":"https:\/\/doi.org\/10.3233\/jifs-169709","relation":{},"ISSN":["1064-1246","1875-8967"],"issn-type":[{"value":"1064-1246","type":"print"},{"value":"1875-8967","type":"electronic"}],"subject":[],"published":{"date-parts":[[2018,6,13]]}}}