Static Hand Gesture Recognition Using Capacitive Sensing and Machine Learning
<p>The proposed system: (<b>a</b>) Motherboard and connected Sensor Board; (<b>b</b>) Gesture recognition pipeline.</p> "> Figure 2
<p>Sensor Board: (<b>a</b>) The Sensor Board; (<b>b</b>) Illustration of the loading mode capacitance.</p> "> Figure 3
<p>Motherboard. The Power Supply section (red) provides power for the whole board. The MCU section (blue) contains an STM32F7 Nucleo development board, which is used for data processing and communication with the host PC. The Sensor section (green) contains FDC1004 capacitance-to-digital converters and STM32G03 MCUs, which are used to collect capacitance data from the Sensor Board and transfer it to the STM32F7. The female board connectors (yellow) are used to connect the Sensor Board to the Motherboard.</p> "> Figure 4
<p>The LDO regulators powering the Motherboard’s sections. The 5 V LDO (teal) powers the MCU section; 3.3 V LDO 1 and 2 (yellow and white) power the Sensor section’s FDC1004 and STM32G03 Group 1s; 3.3 V LDO 3 and 4 (green and red) power the Sensor section’s FDC1004 and STM32G03 Group 2s.</p> "> Figure 5
<p>MLP classifier’s architecture. The network consists of six layers: an input dense layer, two dense + dropout blocks, and a dense output layer. The input layer has 104 neurons, the first hidden layer has 100 neurons, the second layer has 50 neurons, and the output layer has 5 neurons. A dropout rate of 20% was used for each dropout layer. The network takes a <math display="inline"><semantics> <mrow> <mn>1</mn> <mo>×</mo> <mn>104</mn> </mrow> </semantics></math> vector (x<sub>1</sub>–x<sub>104</sub>) and produces a <math display="inline"><semantics> <mrow> <mn>1</mn> <mo>×</mo> <mn>5</mn> </mrow> </semantics></math> vector (y<sub>1</sub>–y<sub>5</sub>). The network’s architecture was experimentally determined using a grid-search-based approach.</p> "> Figure 6
<p>CNN classifier’s architecture. The network has eight layers: an input convolutional layer, two convolutional + pooling blocks, and an output dense layer. Each layer’s input and output are specified using the “NHWC” format: (batch size, height, width, channels), where appropriate. The network takes in a <math display="inline"><semantics> <mrow> <mn>6</mn> <mo>×</mo> <mn>18</mn> <mo>×</mo> <mn>1</mn> </mrow> </semantics></math> matrix (<b>x</b>) and produces a <math display="inline"><semantics> <mrow> <mn>1</mn> <mo>×</mo> <mn>5</mn> </mrow> </semantics></math> vector (y<sub>1</sub>–y<sub>5</sub>). The network’s architecture was experimentally determined using a grid-search-based approach.</p> "> Figure 7
<p>The five hand gestures used in our work: (<b>a</b>) Palm, (<b>b</b>) Fist, (<b>c</b>) Middle, (<b>d</b>) OK, and (<b>e</b>) Index.</p> "> Figure 7 Cont.
<p>The five hand gestures used in our work: (<b>a</b>) Palm, (<b>b</b>) Fist, (<b>c</b>) Middle, (<b>d</b>) OK, and (<b>e</b>) Index.</p> "> Figure 8
<p>Combined confusion matrix for decision tree model. Rows represent actual classes; columns represent predicted classes. Dark blue shading indicates a large value; light blue shading indicates a small value. The average accuracy, precision, recall, and <math display="inline"><semantics> <mrow> <msub> <mi>F</mi> <mn>1</mn> </msub> </mrow> </semantics></math> score for this model were 91.18%, 77.95%, 77.96%, and 77.85%, respectively.</p> "> Figure 9
<p>Combined confusion matrix for Naive Bayes model. Rows represent actual classes; columns represent predicted classes. Dark blue shading indicates a large value; light blue shading indicates a small value. The average accuracy, precision, recall, and <math display="inline"><semantics> <mrow> <msub> <mi>F</mi> <mn>1</mn> </msub> </mrow> </semantics></math> score for this model were 88.34%, 73.68%, 70.86%, and 71.78%, respectively.</p> "> Figure 10
<p>Combined confusion matrix for MLP model. Rows represent actual classes; columns represent predicted classes. Dark blue shading indicates a large value; light blue shading indicates a small value. The average accuracy, precision, recall, and <math display="inline"><semantics> <mrow> <msub> <mi>F</mi> <mn>1</mn> </msub> </mrow> </semantics></math> score for this model were 96.87%, 92.25%, 92.16%, and 92.16%, respectively.</p> "> Figure 11
<p>Combined confusion matrix for CNN model. Rows represent actual classes; columns represent predicted classes. Dark blue shading indicates a large value; light blue shading indicates a small value. The average accuracy, precision, recall, and <math display="inline"><semantics> <mrow> <msub> <mi>F</mi> <mn>1</mn> </msub> </mrow> </semantics></math> score for this model were 95.94%, 89.75%, 89.84%, and 89.77%, respectively.</p> ">
Abstract
:1. Introduction
- It is one of the first reported works that achieve contactless, static hand gesture recognition without using a camera. We developed a novel, bespoke system using a capacitive sensor array and a neural-network-based classifier.
- The performance, unlike many of the reported works in the literature, was tested with “unseen” subjects and achieved high accuracy.
2. Materials and Methods
2.1. System Overview
2.2. Sensor Board
2.3. Motherboard
2.3.1. Power Supply
2.3.2. MCU Section
2.3.3. Sensor Section
2.3.4. Noise Mitigation
2.3.5. Calibration
2.4. Classifier
2.4.1. Decision Trees
2.4.2. Naïve Bayes
2.4.3. Multi-Layer Perceptron Neural Network
2.4.4. Convolutional Neural Network
2.5. Data Acquisition
2.6. Training
2.7. Evaluation
3. Results
3.1. Decision Tree
3.2. Naïve Bayes
3.3. Multi-Layer Perceptron (MLP)
3.4. Convolution Neural Network (CNN)
3.5. Performance Comparison
4. Conclusion and Future Works
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Yasen, M.; Jusoh, S. A systematic review on hand gesture recognition techniques, challenges and applications. PeerJ Comput. Sci. 2019, 5, e218. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Pramudita, A.A. Contactless hand gesture sensor based on array of CW radar for human to machine interface. IEEE Sens. J. 2021, 21, 15196–15208. [Google Scholar] [CrossRef]
- Moin, A.; Zhou, A.; Rahimi, A.; Menon, A.; Benatti, S.; Alexandrov, G.; Tamakloe, S.; Ting, J.; Yamamoto, N.; Khan, Y.; et al. A wearable biosensing system with in-sensor adaptive machine learning for hand gesture recognition. Nat. Electron. 2021, 4, 54–63. [Google Scholar] [CrossRef]
- Yu, L.; Abuella, H.; Islam, M.; O’Hara, J.; Crick, C.; Ekin, S. Gesture recognition using reflected visible and infrared lightwave signals. IEEE Trans. Hum. Mach. Syst. 2021, 51, 44–55. [Google Scholar] [CrossRef]
- Caeiro-Rodríguez, M.; Otero-González, I.; Mikic-Fonte, F.; Llamas-Nistal, M. A systematic review of commercial smart gloves: Current status and applications. Sensors 2021, 21, 2667. [Google Scholar] [CrossRef]
- Zhang, Y.; Harrison, C. Tomo: Wearable, low-cost electrical impedance tomography for hand gesture recognition. In Proceedings of the 28th Annual ACM Symposium on User Interface Software & Technology, Charlotte, NC, USA, 8–11 November 2015; pp. 167–173. [Google Scholar]
- Oudah, M.; Al-Naji, A.; Chahl, J. Hand gesture recognition based on computer vision: A review of techniques. J. Imaging 2020, 6, 73. [Google Scholar] [CrossRef]
- Alam, F.; Faulkner, N.; Parr, B. Device-free localization: A review of non-RF techniques for unobtrusive indoor positioning. IEEE Internet Things J. 2020, 8, 4228–4249. [Google Scholar] [CrossRef]
- Singh, G.; Nelson, A.; Lu, S.; Robucci, R.; Patel, C.; Banerjee, N. Event-driven low-power gesture recognition using differential capacitance. IEEE Sens. J. 2016, 16, 4955–4967. [Google Scholar] [CrossRef]
- Duan, H.; Huang, M.; Yang, Y.; Hao, J.; Chen, L. Ambient light-based hand gesture recognition enabled by recurrent neural network. IEEE Access 2020, 8, 7303–7312. [Google Scholar] [CrossRef]
- Ma, D.; Lan, G.; Hu, C.; Hassan, M.; Hu, W.; Mushfika, U.; Uddin, A.; Youssef, M. Recognizing Hand Gestures using Solar Cells. In IEEE Transactions on Mobile Computing; IEEE: Piscataway, NJ, USA, 2022. [Google Scholar]
- Sorescu, C.; Meena, Y.; Sahoo, D.R. PViMat: A Self-Powered Portable and Rollable Large Area Gestural Interface Using Indoor Light. In Proceedings of the Adjunct Publication of the 33rd Annual ACM Symposium on User Interface Software and Technology, Online, 20–23 October 2020; pp. 80–83. [Google Scholar]
- Tian, Z.; Wang, J.; Yang, X.; Zhou, M. WiCatch: A Wi-Fi based hand gesture recognition system. IEEE Access 2018, 6, 16911–16923. [Google Scholar] [CrossRef]
- Abdelnasser, H.; Youssef, M.; Harras, K.A. Wigest: A ubiquitous wifi-based gesture recognition system. In Proceedings of the 2015 IEEE Conference on Computer Communications (INFOCOM), Hong Kong, China, 26 April–1 May 2015; pp. 1472–1480. [Google Scholar]
- Kim, Y.; Toomajian, B. Hand gesture recognition using micro-Doppler signatures with convolutional neural network. IEEE Access 2016, 4, 7125–7130. [Google Scholar] [CrossRef]
- Skaria, S.; Al-Hourani, A.; Lech, M.; Evans, R.J. Hand-gesture recognition using two-antenna Doppler radar with deep convolutional neural networks. IEEE Sens. J. 2019, 19, 3041–3048. [Google Scholar] [CrossRef]
- Lien, J.; Gillian, N. Soli: Ubiquitous gesture sensing with millimeter wave radar. ACM Trans. Graph. (TOG) 2016, 35, 1–19. [Google Scholar] [CrossRef] [Green Version]
- Wei, H.; Li, P.; Tang, K.; Wang, W.; Chen, X. Alternating Electric Field-Based Static Gesture-Recognition Technology. Sensors 2019, 19, 2375. [Google Scholar] [CrossRef] [Green Version]
- Pinto, R.F.; Borges, C.; Almeida, A.; Paula, I.C. Static hand gesture recognition based on convolutional neural networks. J. Electr. Comput. Eng. 2019, 2019, 4167890. [Google Scholar] [CrossRef]
- Grosse-Puppendahl, T.; Holz, C.; Cohn, G.; Wimmer, R.; Bechtold, O.; Hodges, S.; Reynolds, M.S.; Smith, J.R. Finding common ground: A survey of capacitive sensing in human-computer interaction. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems, Denver, CO, USA, 6–11 May 2017; pp. 3293–3315. [Google Scholar]
- Faulkner, N.; Parr, B.; Alam, F.; Legg, M.; Demidenko, S. CapLoc: Capacitive sensing floor for device-free localization and fall detection. IEEE Access 2020, 8, 187353–187364. [Google Scholar] [CrossRef]
- Shi, Q.; Zhang, Z.; Yang, Y.; Shan, X.; Salam, B.; Lee, C. Artificial intelligence of things (AIoT) enabled floor monitoring system for smart home applications. ACS Nano 2021, 15, 18312–18326. [Google Scholar] [CrossRef]
- Tang, X.; Mandal, S. Indoor occupancy awareness and localization using passive electric field sensing. IEEE Trans. Instrum. Meas. 2019, 68, 4535–4549. [Google Scholar] [CrossRef]
- Wimmer, R.; Kranz, M.; Boring, S.; Schmidt, A. A Capacitive Sensing Toolkit for Pervasive Activity Detection and Recognition. In Proceedings of the Fifth Annual IEEE International Conference on Pervasive Computing and Communications (PerCom’07), White Plains, NY, USA, 19–23 March 2007. [Google Scholar] [CrossRef] [Green Version]
- Texas Instruments Capacitive Sensing: Ins and Outs of Active Shielding; Texas Instrument: Dallas, TX, USA, 2015.
- Samtec ERF8-060-05.0-L-DV-K-TR; Samtec: New Albany, IN, USA, 2021.
- Charbuty, B.; Abdulazeez, A. Classification Based on Decision Tree Algorithm for Machine Learning. J. Appl. Sci. Technol. Trends 2021, 2, 20–28. [Google Scholar] [CrossRef]
- Rish, I. An empirical study of the naive Bayes classifier. In IJCAI 2001 Workshop on Empirical Methods in Artificial Intelligence; IBM: New, York, NY, USA, 2001; Volume 3, pp. 41–46. [Google Scholar]
- Alnuaim, A.A.; Zakariah, M.; Shukla, P.K.; Alhadlaq, A.; Hatamleh, W.A.; Tarazi, H.; Sureshbabu, R.; Ratna, R. Human-Computer Interaction for Recognizing Speech Emotions Using Multilayer Perceptron Classifier. J. Healthc. Eng. 2022, 2022, e6005446. [Google Scholar] [CrossRef] [PubMed]
- Gadekallu, T.R.; Alazab, M.; Kaluri, R.; Maddikunta, P.K.R.; Bhattacharya, S.; Lakshmanna, K. Hand gesture classification using a novel CNN-crow search algorithm. Complex Intell. Syst. 2021, 7, 1855–1868. [Google Scholar] [CrossRef]
Article | Sensing Method | Participant Details | No. of Gestures (Type) | Accuracy (%) | Sensor–Hand Distance (cm) | Remarks |
---|---|---|---|---|---|---|
Duan et al. [10] (2020) | Ambient light sensing using PD | Trained with 4 subjects, tested on 3 “unseen” | 7 (Dynamic) | 82.57–85.67 | 10 | Accuracy is 96.83% if trained and tested (90–10 split of data) on the same set of 4 subjects. |
Yu et al. [4] (2021) | IR | Trained with 4 and tested on 1 “unseen” | 8 (Dynamic) | 92.13 | 20 | Significantly degraded accuracy with Visible light. |
Ma et al. [11] (2022) | Solar cell | Trained with 2 subjects, tested on 1 “unseen” | 6 (Dynamic) | 94 | 10 | 99% accuracy if trained and tested on the same subject. |
Skaria et al. [16] (2019) | Micro-Doppler Signatures | Trained with 1 subject, tested on 1 “unseen” | 14 (Dynamic) | 45.50–51.20 | 10–30 | When trained and tested on the same subject, accuracy ranged between 90.30% and 95.50%. |
Tian et al. [13] (2018) | Wireless (Wi-Fi) | 10 | 9 (Dynamic) | ~ | 50–250 | Accuracy is 96% when trained and tested (train–test split not given) on the same set of 10 subjects. Can recognize two-handed gestures |
Singh et al. [9] (2016) | Capacitive sensing | 5 | 16 (Dynamic) | ~ | <10 | Accuracy is 93% when trained and tested (train–test split not given) on the same set of 5 subjects. |
Wei et al. [18] (2019) | Capacitive sensing | 4 | 4 (Static) | ~ | 6–9 | Accuracy is 91.6% when trained and tested (50–50 split of data) on the same set of 4 subjects. |
Noble et al. Proposed (2022) | Capacitive sensing | Trained with 4 and tested on 1 “unseen” | 5 (Static) | 96.87 | 5–8 | Accuracy is 100% if trained and tested (90–10 split of data) on the same set of 5 subjects. |
Iteration | Participant Training and Validation Sets | Participant Testing Set |
---|---|---|
1 | 2, 3, 4, 5 | 1 |
2 | 1, 3, 4, 5 | 2 |
3 | 1, 2, 4, 5 | 3 |
4 | 1, 2, 3, 5 | 4 |
5 | 1, 2, 3, 4 | 5 |
Model | Average Accuracy | Average Precision | Average Recall | Average F1 Score |
---|---|---|---|---|
Decision Tree | 91.18% | 77.95% | 77.96% | 77.85% |
Naïve Bayes | 88.34% | 73.68% | 70.86% | 71.78% |
MLP | 96.87% | 92.25% | 92.16% | 92.16% |
CNN | 95.94% | 89.75% | 89.84% | 89.77% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Noble, F.; Xu, M.; Alam, F. Static Hand Gesture Recognition Using Capacitive Sensing and Machine Learning. Sensors 2023, 23, 3419. https://doi.org/10.3390/s23073419
Noble F, Xu M, Alam F. Static Hand Gesture Recognition Using Capacitive Sensing and Machine Learning. Sensors. 2023; 23(7):3419. https://doi.org/10.3390/s23073419
Chicago/Turabian StyleNoble, Frazer, Muqing Xu, and Fakhrul Alam. 2023. "Static Hand Gesture Recognition Using Capacitive Sensing and Machine Learning" Sensors 23, no. 7: 3419. https://doi.org/10.3390/s23073419
APA StyleNoble, F., Xu, M., & Alam, F. (2023). Static Hand Gesture Recognition Using Capacitive Sensing and Machine Learning. Sensors, 23(7), 3419. https://doi.org/10.3390/s23073419