Abstract
A hybrid convolutional neural network (CNN) and bidirectional long short-term memory (BLSTM) network for human complex activity recognition with multi-feature fusion is proposed in this paper. Specifically, a new CNN model is designed to extract the spatial features from the sensor data. Considering that in the process of activity recognition, the output at the current moment is not only related to the previous state, but also to the subsequent state. BLSTM network is further used to extract the temporal context of state information to improve the performance of activity recognition. In order to fully mine the features from the sensor data and further improve the performance of activity recognition, a new feature selection method named SFSANW (sequential forward selection and network weights), which is based on sequential forward selection algorithm and network weights is proposed to select features extracted by the traditional methods to obtain dominant features. The dominant features are then fused with the feature vectors extracted by the hybrid CNN and BLSTM network. Experiments are performed on two complex activity datasets, PAMAP2 and UT-Data, and 92.23% and 98.07% F1 scores are obtained, respectively. The experimental results demonstrate that the proposed method can achieve better performance of complex activity recognition, which is superior to the traditional machine learning algorithms and the state-of-the-art deep learning algorithms.
Similar content being viewed by others
Availability of data and material
Data and material are fully available without restriction.
Code availability
Custom code is available without restriction.
References
Blanke U, Schiele B (2009) Daily routine recognition through activity spotting. In Int Symp Loc Context Aware Springer 192–206
Ciabattoni L, Foresi G, Monteriù A, Pagnotta DP, Romeo L, Spalazzi L, De Cesare A (2018) Complex activity recognition system based on cascade classifiers and wearable device data. In 2018 IEEE Int Conf Consum Electron (ICCE) IEEE 1–2
Cvetković B, Mirchevska V, Janko V, Luštrek M (2015) Recognition of high-level activities with a smartphone. In Adjunct Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2015 ACM International Symposium on Wearable Computers, pp 1453–1461
Damirchi H, Khorrambakht R, Taghirad H (2020) Arc-net: Activity recognition through capsules. arXiv preprin
Dernbach S, Das B, Krishnan NC, Thomas BL, Cook DJ (2012) Simple and complex activity recognition through smart phones. In 2012 8th Int Conf Intell Environ IEEE 214–221
Donahue J, Anne Hendricks L, Guadarrama S, Rohrbach M, Venugopalan S, Saenko K, Darrell T (2015) Long-term recurrent convolutional networks for visual recognition and description. In Proc IEEE Conf Comput Vis Pattern Recognit 2625–2634
Edel M, Köppe E (2016) Binarized-blstm-rnn based human activity recognition. In 2016 Int Conf Indoor Position Indoor Navig (IPIN) IEEE 1–7
Friday NH, Al-garadi MA, Mujtaba G, Alo UR, Waqas A (2018) Deep learning fusion conceptual frameworks for complex human activity recognition using mobile and wearable sensors. In 2018 Int Conf Comput Math Eng Technol (iCoMET) IEEE 1–7
Gacav C, Benligiray B, Topal C (2016) Sequential forward feature selection for facial expression recognition. In 2016 24th Signal Process Commun Appl Conf (SIU) IEEE 1481–1484
Gil-Martín M, San-Segundo R, Fernández-Martínez F, de Córdoba R (2020) Human activity recognition adapted to the type of movement. Comput Elect Eng 88:106822
Guan Y, Plötz T (2017) Ensembles of deep lstm learners for activity recognition using wearables. Proc ACM Interact Mob Wearable Ubiquitous Technol 1(2):1–28
Guiry JJ, Van de Ven P, Nelson J (2014) Multi-sensor fusion for enhanced contextual awareness of everyday activities with ubiquitous devices. Sensors 14(3):5687–5701
Ha S, Choi S (2016) Convolutional neural networks for human activity recognition using multiple accelerometer and gyroscope sensors. In 2016 Int Joint Conf Neural Netw (IJCNN) IEEE 381–388
Hammerla NY, Halloran S, Plötz T (2016) Deep, convolutional, and recurrent models for human activity recognition using wearables. In Proc 25th Int Joint Conf Artif Intell AAAI Press IJCAI’16 1533–1540
Haque MN, Tonmoy MTH, Mahmud S, Ali AA, Khan MAH, Shoyaib M (2019) Gru-based attention mechanism for human activity recognition. In 2019 1st Int Conf Adv Sci Eng Robot Technol (ICASERT) IEEE 1–6
Hernández F, Suárez LF, Villamizar J, Altuve M (2019) Human activity recognition on smartphones using a bidirectional lstm network. In 2019 XXII Symp Image Signal Proc Artif Vis (STSIVA) IEEE 1–5
Huynh T, Schiele B (2005) Analyzing features for activity recognition. In Proceedings of the 2005 Joint Conference on Smart Objects and Ambient Intelligence: Innovative Context-Aware Services: Usages and Technologies 159–163
Huynh T, Fritz M, Schiele B (2008) Discovery of activity patterns using topic models. In Proc 10th Int Conf Ubiquitous Comput 10–19
Kim Y, Moon T (2015) Human detection and activity classification based on micro-doppler signatures using deep convolutional neural networks. IEEE Geosci Remote Sens Lett 13(1):8–12
Lee SM, Yoon SM, Cho H (2017) Human activity recognition from accelerometer data using convolutional neural network. In 2017 IEEE Int Conf Big Data Smart Comput (BigComp) IEEE 131–134
Liu L, Peng Y, Liu M, Huang Z (2015) Sensor-based human activity recognition system with a multilayered model using time series shapelets. Knowl Based Syst 90:138–152
Liu L, Peng Y, Wang S, Liu M, Huang Z (2016) Complex activity recognition using time series pattern dictionary learned from ubiquitous sensors. Inf Sci 340:41–57
Liu J, Shahroudy A, Xu D, Wang G (2016b) Spatio-temporal lstm with trust gates for 3d human action recognition. In Eur Conf Comput Vis Springer 816–833
Lv M, Xu W, Chen T (2019) A hybrid deep convolutional and recurrent neural network for complex activity recognition using multimodal sensors. Neurocomputing 362:33–40
Mobark M, Chuprat S, Mantoro T (2017) Improving the accuracy of complex activities recognition using accelerometer-embedded mobile phone classifiers. In 2017 2nd Int Conf Inf Comput (ICIC) IEEE 1–5
Murahari VS, Plötz T (2018) On attention models for human activity recognition. In Proc 2018 ACM Int Symp Wearable Comput 100–103
Münzner S, Schmidt P, Reiss A, Hanselmann M, Stiefelhagen R, Durichen R (2017) CNN-based sensor fusion techniques for multimodal human activity recognition. In Proc 2017 ACM Int Symp Wearable Comput 158–165
Ordóñez FJ, Roggen D (2016) Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 16(1):115
Peng L, Chen L, Wu X, Guo H, Chen G (2016) Hierarchical complex activity representation and recognition using topic model and classifier level fusion. IEEE Trans Biomed Eng 64(6):1369–1379
Peng L, Chen L, Wu M, Chen G (2018) Complex activity recognition using acceleration, vital sign, and location data. IEEE Trans Mob Comput 18(7):1488–1498
Reiss A, Hendeby G, Stricker D (2013) Confidence based multiclass adaboost for physical activity monitoring. In Proc 2013 Int Symp Wearable Comput 13–20
Rueda FM, Grzeszick R, Fink GA, Feldhorst S, Hompel MT (2018) Convolutional neural networks for human activity recognition using body-worn sensors. Informatics 5(2):26
Sabour S, Frosst N, Hinton GE (2017) Dynamic routing between capsules. In Adv Neural Inf Process Syst 3856–3866
Saguna S, Zaslavsky A, Chakraborty D (2013) Complex activity recognition using context-driven activity theory and activity signatures. ACM Trans Comput Hum Interact (TOCHI) 20(6):1–34
Seiter J, Amft O, Rossi M, Tröster G (2014) Discovery of activity composites using topic models: An analysis of unsupervised methods. Pervasive Mob Comput 15:215–227
Shoaib M, Bosch S, Incel OD, Scholten H, Havinga PJ (2016) Complex human activity recognition using smartphone and wrist-worn motion sensors. Sensors 16(4):426
Wan S, Qi L, Xu X, Tong C, Gu Z (2020) Deep learning models for real-time human activity recognition with smartphones. Mob Netw Appl 25(2):743–755
Yang Z, Raymond OI, Zhang C, Wan Y, Long J (2018) Dfternet: Towards 2-bit dynamic fusion networks for accurate human activity recognition. IEEE Access 6:56750–56764
Yu T, Chen J, Yan N, Liu X (2018a) A multi-layer parallel lstm network for human activity recognition with smartphone sensors. In 2018 10th Int Conf Wireless Commun Signal Process (WCSP) IEEE 1–6
Yu S, Qin L, Yin Q (2018b) A c-lstm neural network for human activity recognition using wearables. In 2018 Int Symp Sens Instrum IoT Era (ISSI) IEEE 1–6
Zeng M, Gao H, Yu T, Mengshoel OJ, Langseth H, Lane I, Liu X (2018) Understanding and improving recurrent networks for human activity recognition by continuous attention. In Proc 2018 ACM Int Symp Wearable Comput 56–63
Zheng Z, Shi L, Wang C, Sun L, Pan G (2019) Lstm with uniqueness attention for human activity recognition. In Int Conf Artif Neural Netw Springer 498–509
Zuo Z, Shuai B, Wang G, Liu X, Wang X, Wang B, Chen Y (2016) Learning contextual dependence with convolutional hierarchical recurrent neural networks. IEEE Trans Image Proces 25(7):2983–2996
Acknowledgements
This work was supported by the Zhejiang Provincial Natural Science Foundation of China [grant number LY19F020032], and National Natural Science Foundation of China [grant number 61872322, U1909203, 62036009].
Funding
This study was funded by the Zhejiang Provincial Natural Science Foundation of China (grant number LY19F020032), and National Natural Science Foundation of China (grant number 61872322, U1909203, 62036009).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Huan, R., Zhan, Z., Ge, L. et al. A hybrid CNN and BLSTM network for human complex activity recognition with multi-feature fusion. Multimed Tools Appl 80, 36159–36182 (2021). https://doi.org/10.1007/s11042-021-11363-4
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-021-11363-4