Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Sensing Attribute Weights: A Novel Basic Belief Assignment Method
Next Article in Special Issue
Analysis of Optimal Sensor Positions for Activity Classification and Application on a Different Data Collection Scenario
Previous Article in Journal
Model-Based Estimation of Ankle Joint Stiffness
Previous Article in Special Issue
A Wellness Mobile Application for Smart Health: Pilot Study Design and Results
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Posture Detection Based on Smart Cushion for Wheelchair Users

1
School of Logistics Engineering, Wuhan University of Technology, Wuhan, 430070, China
2
Department of Informatics, Modeling, Electronics and Systems, University of Calabria, Rende, 87036, Italy
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(4), 719; https://doi.org/10.3390/s17040719
Submission received: 1 March 2017 / Revised: 20 March 2017 / Accepted: 25 March 2017 / Published: 29 March 2017
(This article belongs to the Special Issue Advances in Body Sensor Networks: Sensors, Systems, and Applications)

Abstract

:
The postures of wheelchair users can reveal their sitting habit, mood, and even predict health risks such as pressure ulcers or lower back pain. Mining the hidden information of the postures can reveal their wellness and general health conditions. In this paper, a cushion-based posture recognition system is used to process pressure sensor signals for the detection of user’s posture in the wheelchair. The proposed posture detection method is composed of three main steps: data level classification for posture detection, backward selection of sensor configuration, and recognition results compared with previous literature. Five supervised classification techniques—Decision Tree (J48), Support Vector Machines (SVM), Multilayer Perceptron (MLP), Naive Bayes, and k-Nearest Neighbor (k-NN)—are compared in terms of classification accuracy, precision, recall, and F-measure. Results indicate that the J48 classifier provides the highest accuracy compared to other techniques. The backward selection method was used to determine the best sensor deployment configuration of the wheelchair. Several kinds of pressure sensor deployments are compared and our new method of deployment is shown to better detect postures of the wheelchair users. Performance analysis also took into account the Body Mass Index (BMI), useful for evaluating the robustness of the method across individual physical differences. Results show that our proposed sensor deployment is effective, achieving 99.47% posture recognition accuracy. Our proposed method is very competitive for posture recognition and robust in comparison with other former research. Accurate posture detection represents a fundamental basic block to develop several applications, including fatigue estimation and activity level assessment.

1. Introduction

The continuous progress of ICT (Information and Communication Technology) is rapidly changing our daily life. Thanks to the high processing capabilities, reduced dimensions and low cost of modern microelectronics, a plethora of wearable devices are being realized to non-invasively support and monitor several human daily activities. Such devices play a central role in the so-called ubiquitous sensing [1], which aims at gathering data from the sensors and mining knowledge from them. Body Area Networks (BANs) can measure physical activity and biophysical signals [2] and pose the basis for a wide range of new user-centric applications in the domains of health care [3], remote monitoring [4], activity measurement [5], activity level assessment [6], user-centric wheelchairs [7], and health care service security [8].
In order to cope with the problems that elderly and mobility disabled individuals are facing, more effective care services, especially for wheelchair users, need to be provided. Several supportive devices have already been proposed to help such people keep an independent lifestyle. For this purpose, mobility-assistive devices, including smart wheelchairs, have recently begun to emerge. A smart wheelchair augments the functionalities of a traditional wheelchair by means of various technologies such as BANs [9], cloud computing [10,11], and multi sensor fusion [12].
Current research on smart wheelchairs is mainly focused on vital signals and physical activity monitoring [13], human–machine interface to control the wheelchair [14], obstacle detection and navigation [15], and wheelchair movement condition estimation [16].
This work is focused on a wheelchair assist system for mobility-impaired individuals that can recognize postures using a cushion based system. Pressure sensors are used to obtain data of the wheelchair user’s postures, several classification methods have been compared to determine an efficient classifier, and former sensor deployment methods have been compared against our proposed deployment to determine an optimal deployment method. In addition, some practical applications are discussed in the paper. If dangerous situations such as long-term abnormal posture are detected, the system can issue an alarm immediately; in addition, posture detection is a basic block towards activity level assessment.
The main contributions of this work are the following:
  • An optimization method of pressure sensor deployment is proposed to more accurately detect sitting postures.
  • An in-depth comparison of several classification techniques has been carried out to identify the best posture classifier based on pressure sensors’ smart cushion.
  • In contrast with previous studies, the Body Mass Index (BMI) is among the considered parameters to evaluate the generality and robustness of the proposed deployment method across different body shapes.
The reminder of the paper is organized as follows. Section 2 discusses the related work on posture recognition using smart cushion; we describe and analyze the two main production technologies of smart cushions for pressure detection, the first is based on pressure sensor array and the second relies on the use of a few individual sensors deployed on the seat and backrest. Section 3 describes the architecture of the proposed system. Section 4 reports the system evaluation; data collection and experimental settings as well as optimal sensor selection methods are discussed here. Section 5 is about the results and performance analysis; we compared several classifiers and the results of our optimal sensor deployment are analyzed in comparison with the state-of-the-art. Finally, Section 6 concludes the paper and some future directions are anticipated.

2. Related Work

Most previous studies focused on monitoring physical activities of wheelchair users. Postolache et al. [17] proposed a system for monitoring activities performed on wheelchairs by analyzing inertial data. Hiremath et al. [18] developed a similar system but used gyroscopes for capturing wheelchair overturn and a wearable accelerometer to detect physical activity.
Interestingly, although wheelchair users spend most of their time sitting in the wheelchair, and it would be therefore important to analyze sitting postures and behaviors, fewer works have been devoted to these aspects. Wu et al. [19] proposed using ultrasound sensors to measure the distance between the user and wheelchair in order to recognize a number of sitting postures. Nakane et al. [20] focused on fatigue assessment based posturography and developed a pressure sensor sheet. Ding et al. [21] were interested in monitoring the status of the wheelchair including seat tilt, backrest recline, and seat elevation. In a previous work [22], we proposed a cloud-computing infrastructure based on the BodyCloud platform [23,24], and a pressure sensor cushion was used to monitor body postures of wheelchair users.
Due to its unobtrusiveness, several literature works adopted a pressure sensor-based cushion to recognize physical activity, posture and fatigue of individuals sitting on standard chairs or on wheelchairs [25]. Specifically, there are two main production technologies of smart cushions for pressure detection: the first one is based on a pressure sensor array, whereas the second one relies on the use of fewer individual sensors deployed on the seat and backrest.

2.1. Cushion with Pressure Sensor Array

Xu et al. [26] proposed a textile-based sensing system called Smart Cushion used for user’s sitting postures. Binary pressure distribution data was collected and the data model can be interpreted as binary representation of a gray scale image. Dynamic time warping has been applied to pressure distribution in order to recognize postures. There is also a representative sensor matrix called an I-Scan System developed by Tekscan [27]. The versatile I-Scan tactile pressure mapping system consists of over 2000 pressure sensors. It is a powerful tool that accurately measures and analyzes interface pressure between two surfaces. Tan et al. [28] used the commercially available pressure distribution sensor called a Body Pressure Measurement System (BPMS), also manufactured by Tekscan. Using principal components analysis to analyze a pressure distribution map as a gray scale image, sitting posture classification was successfully processed. Mota et al. [29] also used Tekscan; posture features were extracted using a mixture of four Gaussians and served as input to a three-layer feed-forward neural network. Nine different postures were recognized with 87.6% accuracy on average. Meyer et al. [30] presented their own capacitive textile pressure sensor array cushion to measure pressure distribution of the human body. They used a Naive Bayes classifier to identify 16 different sitting postures on the chair, with results similar to Tan. Multu et al. [31] proposed a robust low-cost non-intrusive seat; compared to the commercially-available Tekscan ConforMat system (over 2000 sensors) used in [28,30], they managed to reduce the number of sensors, and an optimal sensor deployment with 19 pressure sensors was adopted to recognize sitting postures. Kamiya et al. [32] used an 8 × 8 pressure sensor matrix inserted inside the foam filling of the chair to recognize sitting postures. Radial basis function and SVM algorithms have been used to recognize various postures. Xu et al. [33] presented the design and implementation of a low-cost smart cushion equipped with sensor array. Binary value of pressure distribution (i.e., when the pressure applied to a sensor is higher than a given threshold, its value is 1; otherwise, it is 0) was analyzed to recognize recognize nine postures in real time. Fard et al. [34] proposed a system for preventing pressure ulcers based on an 8 × 8 pressure sensor matrix for continuous monitoring of surface pressure of the seat.
All of the literature works analyzed, summarized in Table 1, detect and analyze pressure distribution on the cushion/mattress to recognize the postures. On the one hand, pressure sensor array technology allows for accurately measuring multiple pressure points on the seating surface; on the other, however, it leads to costly solutions.

2.2. Smart Cushions Based on Fewer Individual Pressure Sensors

Hu et al. [35] proposed PoSeat, a smart cushion equipped with an accelerometer and pressure sensors for chronic back pain prevention using a hybrid SVM classifier. Benocci et al. [36] proposed a method using five pressure sensors and k-Nearest Neighbour (kNN) was used to classify six different postures. Bao et al. [37] used a pressure cushion to recognize sitting postures by means of a density-based clustering method. Diego et al. [38] proposed a non-invasive system for monitoring pressure changes and tilts during daily use of the wheelchair. Pressure sensors have been deployed on the seat to detect pressure relief tilts and provide reminders to change posture to reduce the long-term risk of causing pressure ulcers. Min et al. [39] proposed a real-time sitting posture monitoring system based on measuring pressure distribution of the human body sitting on the chair. A decision tree is used to recognize five sitting postures; when a user deviates from the correct posture, an alarm function is activated. Zemp et al. [40] developed an instrumented chair with force and acceleration sensors to identify the user’s sitting posture by applying five different machine learning methods (Support Vector Machines, Multinomial Regression, Boosting, Neural Networks and Random Forest). Sixteen force sensor values and the backrest angle were used as the features for the classification, and the recognition results reached average accuracy of 90.9%. Barba et al. [41] used the postural changes of users to detect three kinds of affective states in an e-learning environment. Sixteen pressure sensors were placed on the chair, half on the seat and the other half on the backrest. Fu et al. [42] proposed a robust, low-cost, sensor based system that is capable of recognizing sitting postures and activities. Eight force sensing resistors (FSRs) were placed on chair backrests and seats, and a Hidden Markov Model approach was used to establish the activity model from sitting posture sequences. Kumar et al. [43] designed Care-Chair with just four pressure sensors on the chair backrest. The proposed system can classify 19 kinds of complex user sedentary activities, and it can also detect user functional activities and emotion based activities. Our former research [44] used three pressure sensors, two on the seat and one on the backrest to detect user’s postures in smart wheelchairs. We were able to recognize six different postures; results obtained with experimental evaluation showed high classification accuracy.
There are also two commercial smart cushions called DARMA [45] and SENSIMAT [46] that can monitor posture and sitting habits. Both of them are composed of six sensors and have similar deployment. DARMA is a general-purpose cushion equipped with fiber optic sensors that are much more sensitive and accurate than pressure sensors. It can track user’s posture and sitting time as well as heartbeat and respiration. Based on user’s particular behavior and habits, DARMA provides several functionalities such as actionable coaching, stand-up reminders, and posture advice. SENSIMAT is instead specifically designed for wheelchairs. The proprietary PressureRisk algorithm allows for monitoring sitting pressure and pushing data to the mobile device. It can send reminders to the user when it is time to move and can track movements and progress over time.
Each analyzed literature work, as shown in Table 2, proposes its own sensor deployment, achieving relevant results. However, to the best of our knowledge, a systematic study on the selection of the optimal sensor deployment on the cushion, and which classifier is most suitable for cushion-based posture recognition, is still lacking. In the following, we will introduce an optimal sensor deployment method to detect wheelchair user’s posture; in addition, we evaluate several classification techniques to identify the most effective ones for sitting posture recognition.

3. System Design

In the following, we describe our method for monitoring postures of smart wheelchair users. In particular, we designed a smart cushion to monitor sitting postures. The method serves as a basis for the development of a smart wheelchair system that can e.g., warn the user of long-term wrong/dangerous postures and, in case of emergency situations, alert relatives and caregivers to help the assisted user on time. In the following sections, we will describe the system in detail.

3.1. System Architecture

A layered architecture of the proposed posture detection system is depicted in Figure 1. The posture detection layer represents the core of the proposed system and it has been already implemented; several application services—outside the scope of this work—can be easily developed on top.
It is composed of two main layers:
  • Posture Detection Layer: it includes two subsystems, (i) the Sensing subsystem, which uses pressure sensors deployed on the wheelchair to collect data generated by the weight of the body and (ii) the Recognition subsystem that, based on the Arduino platform (see Section 3.2), runs the posture recognition algorithm on the collected pressure sensor data. Posture detection results can therefore be fed to the application service layer.
  • Application Service Layer: several (mobile and cloud-based) applications can be developed on top of the posture recognition subsystem and exploit the geo-location from dedicated mobile device services (such as GPS, WiFi, or cellular tower signal strength) to locate the user. Applications can display the sensing results, perform activity level assessment, alert the user of dangerous postures and, if necessary, send his/her geo-location and make emergency automatic voice calls to caregivers.

3.2. Hardware Design of the Smart Cushion

The smart cushion is composed of three main modules:
  • data sensing (cushion equipped with pressure sensors);
  • data processing (Arduino-based unit);
  • data transmission (Bluetooth shield for Arduino).
The hardware design of the smart cushion is shown in Figure 2.
The data sensing module is based on Force Sensitive Resistor pressure sensors (Interlink FSR-406) [47]. The resistance value is inversely proportional to the pressure applied on the sensing patch. Such sensors can be easily embedded into the chair textile or foam filling. We chose Arduino DUE [48] to construct the data processing module for its low energy consumption, high sampling rate and processing capabilities. After sensor data collection and processing, results are sent to the smartphone in wireless, using an Arduino Bluetooth HC-06 shield [49] that is attached to the processing module.
As previously mentioned, we deployed 12 pressure sensors on the wheelchair (as shown in Figure 3) to evaluate our system. The list of required components along with its market price has been reported in Table 3. The total price of our prototype is $228. In comparison, the commercially available Body Pressure Measurement System (BPMS) using the sensors array developed by TekScan [27] has a sensor matrix of 42 × 48 , and it is sold for over ten thousand US dollars. PoSeat [35] was reported at a price of $460; another marketed smart wheelchair cushion called SENSIMAT [46] is currently sold at $599.

4. System Evaluation

Instead of using sensor arrays, we chose to design our cushion with fewer pressure sensors individually connected to the processing unit; this has advantages in terms of economic cost and power consumption.
In Section 2.2, we provide an overview of different kinds of pressure sensor deployment methods. Most of the works listed in Table 2 did not systematically analyzed (or at least described) the deployment methods. In this study, we adopt two systematic principles to determine the sensor deployment; one uses empirical analysis to find the most significant sensors and the other tries to take into account all of the analyzed literature sensor deployment methods.
Here, we adopt a sensor deployment configuration as shown in Figure 3. The deployment is suitable for the body mass distribution. Actually, using the pressure sensors on the seat is enough to detect most of the sitting postures; however, we decided to apply pressure sensors on the backrest too, as a supplementary tool to more comprehensively monitor users’ postures.
Assuming a conventional size for the seat and backrest of 40 cm × 40 cm, we split the seat and backrest into 5 × 5 square zones. We performed a preliminary empirical analysis to exclude the least significant pressure points and found that the most effective initial deployment configuration to detect pressure changes due to different postures is the one depicted in Figure 3. It is worth noting that F S R 2 and F S R 4 have been deployed to the left and right sides of the seat to better deal with individual physiognomy differences.
In our previous studies [44], we used two pressure sensors on the seat and one pressure sensor on the backrest. However, we noticed that if the user does not sit perfectly at the center of seat, some posture transitions are not correctly detected. Therefore, with the aim of finding a better deployment to cope with different sitting positions, the smart cushion we designed in this work can record up to 12 pressure points simultaneously (the limitation is due to the number of analog inputs of the Arduino DUE platform). All of the FSR sensors (marked as F S R 0 , , F S R 11 ) were sampled at a frequency of 2 Hz (i.e., sampling rate of two samples per second) because it is enough to capture posture transitions and has the property of low energy consumption [44].
The electrical circuitry highlighting the connections of the FSR sensors to the Arduino platform is shown in Figure 4.

4.1. Posture Definition and Data Collection

We carried out experimental tests in our laboratory involving 12 participants (seven males and five females) with ages in the range of [22, 36] and BMI in the range of [16, 34]. The BMI, whose distribution in our experiment sample is shown in Table 4, is an important parameter to consider in order to realize a user-independent posture recognition system.
To simulate a real-life setting, we asked each participant to perform five different postures that are typical for wheelchair users, as listed in Table 5. When the user sits on the wheelchair, the embedded sensors can measure multiple pressure points due to body weight. The posture data were manually labeled during the experiments; we asked to keep each posture for three minutes; in addition, at the beginning of the experiment and between each posture, we allow 30 s for the participant to adjust the posture in the wheelchair.
The participants were free to choose their favorite seating position and each subject performed the experiment only once. The final dataset, consisting of data collected from all 12 of the test subjects, counts a total of 36,000 different posture recordings (see Table 5).
Pressure sensors signals are marked as { P 0 , P 1 , , P 11 } . At the time t, we obtained the instance vector P t = { P t 0 , P t 1 , , P t 11 } . Using this instance vector, we will recognize the sitting postures defined before.

4.2. Classification

Studies have shown that a lot of different algorithms can be used for the classification of sitting postures with satisfactory accuracy ranging from 90.9% to 99.5% [36,37,40,44]. Since the performance of the classification results is highly dependent on the used data, we compared the five algorithms as referred to before. In order to choose the best classifier for the classification, we analyzed five supervised machine learning algorithms applied for the recognition of sitting postures. Specifically, we compared J48 decision tree, Support Vector Machine, Multilayer Perceptron, Naive Bayes and k-Nearest Neighbors, respectively, briefly described in the following.
J48 [50] is a specific decision tree implementation of the well known C4.5 algorithm using the concept of information entropy. The classifier model is generated by a training procedure that uses a set of pre-classified samples. Each sample is a p-dimensional vector also known as a feature vector. Each node of the tree represents a decision (typically a comparison against a threshold value); at each node, J48 chooses the feature (i.e., an attribute) of the data that most effectively splits its set of samples into distinct subsets according to the normalized information gain (NIG). In particular, the algorithm chooses the feature with the highest NIG to generate the decision node. The main parameters used to tune the classifier generation are the pruning confidence (C) and the minimum number of instances per leaf (M), as summarized in Table 6.
Support Vector Machine (SVM) [51] is a supervised learning model used for both classification and regression analysis with associated learning algorithms that analyze data to recognize patterns. Given a set of training instances, new samples will be classified with one of the defined categories. The standard SVM builds a model that assigns new instances into one of two possible categories, so the SVM is a type of non-probabilistic binary linear classifier. The key parameters of the SVM classifier are reported in Table 6. In our analysis, we used a specific type of SVM model called regularized support vector classification (C-SVC) and we set the C parameter to 1. We used Radial Basis Function (RBF) as the kernel type with its degree set to 3.
Multilayer perceptron (MLP) [52] is a feed-forward artificial neural network model that maps sets of input data onto a set of appropriate outputs. An MLP typically involves three or more layers. MLP adopts a supervised learning technique called back propagation for network training. Except for the input nodes, each node is a neuron (or processing element) with a nonlinear activation function. Nodes are connected with weight values. The network output will be decided by node connection, weight values and activation function. The main parameters of the MLP classifier are reported in Table 6. For our purpose, we need 12 attributes and five classes (corresponding to the five sitting positions), so the model is composed of nine hidden layers. Default Momentum Rate and Learning Rate have been selected for the back-propagation algorithm.
Naive Bayes [53] is a simple technique for constructing classifiers: models that assign class labels to problem instances, represented as vectors of feature values, where the class labels are drawn from some finite set. It consists of a family of algorithms based on the assumption that the value of a particular feature is independent from the value of any other feature, given the class variable. An advantage of Naive Bayes is that it only requires a small amount of training data to estimate the parameters necessary for classification. Using different parameters, we trained two classifiers, one using default parameters and the other is BayesNet.
k-Nearest Neighbors algorithm (k-NN) [54] is a simple supervised learning method method used for classification and regression. In both cases, the input consists of the k closest training examples in the feature space. The parameter k plays the role of a capacity control. If k = 1, it means that the object is simply assigned to the class of that single nearest neighbor.
The five classifiers were set up with various configurations resulting in seven classifier configurations; each classifier was evaluated using the same dataset.

4.3. Sensor Deployment Optimization Using the Backward Selection Method

Our goal is to investigate the contribution of each individual sensor to the recognition accuracy and determine the best sensor deployment. This study was performed using the backward selection procedure [55]. It starts with fitting a model with all the variables of interest (i.e., all the 12 pressure sensors active). Then, sensors are excluded one at a time from the base configuration; we choose the best configuration, and the least significant variable is dropped. We iterate this procedure by successively re-fitting reduced models and applying the same rule until only one sensor is left.

5. Performance Evaluation

In this section, we present and discuss the results obtained from each classifier and choose the best sensor configuration using the backward selection method. To accurately study the performance evaluation, five relevant metrics (Accuracy, Precision, Recall, F-measure, and Model Build Time) have been compared.

5.1. Performance Evaluation of Each Classifier

We analyzed the seven classifiers as in Table 6 using WEKA (Waikato Environment for Knowledge Analysis) data mining toolbox [56] with the dataset described in Section 4.1. Data from all of the 12 pressure sensors are therefore taken into account in this phase. In this work, 10-fold cross validation was used to separate the training from the test set. The classifier models were obviously trained using the training set, while, in the test step, the estimated classes were compared to the true classes in order to compute the classification accuracy.
The Accuracy measure is used to evaluate the classifiers performance; specifically, it measures the proportion of correctly classified instances. In the case of binary classification, the formula of accuracy can be expressed as follows:
A c c u r a c y = T p + T n T p + T n + F p + F n
where T n (true negatives) is the correct classifications of negative examples, T p (true positives) is the correct classifications of positive examples. F n (false negatives) and F p (false positives) are, respectively, the positive examples incorrectly classified into the negative classes and the negative examples incorrectly classified into the positive classes.
For evaluating the classifiers, F-measure accuracy (overall accuracy) of the test data has been computed to evaluate recognition performance. F-measure represents the combination of precision and recall, defined, respectively, as follows:
p r e c i s i o n = T p T p + F p
r e c a l l = T p T p + F n
F m e a s u r e = 2 · r e c a l l · p r e c i s i o n r e c a l l + p r e c i s i o n
Table 7 summaries the results of our comparison. We recorded the same amount of samples for each different posture. The class imbalance problem was not taken into account for these classifiers. However, the main idea of this paper is to determine which classifier is more suitable for our system.
J48 shows better performance than other classifiers. It is a classifier that predicts the posture classes of a new sample based on the threshold feature values. Furthermore, it is lightweight and easy to implement in embedded devices. We will eventually choose a J48 pruned tree for the implementation on the Arduino platform.
The SVM shows the worst performance among the analyzed classifiers; while it is very effective for binary classification, setting its key parameters for multi classification is much harder. In addition, it is computatively demanding and therefore not easily implemented in embedded and mobile devices. The MLP takes a longer time to build the train model than the J48 and its performance results are lower than J48. MLP also requires high processing resources, making it not ideal on embedded and mobile devices. We did not achieve good results with the Naive Bayes either, especially with configuration No. 4 (see Table 6), and it shows the worst performance among the classifiers we took into consideration. kNN achieved slightly worse accuracy than J48. We observed that different k values did not affect the recognition accuracy: our hypothesis is that the clusters identified by our feature set are internally consistent and well separated among each other. kNN requires less time to build the model (essentially the time required to split the dataset to obtain the training set), but it requires high computation load for classification, so it does not fit lightweight requirements imposed by embedded platforms.

5.2. Sensor Selection Using the Backward Selection Method

Given the results described in Section 5.1, we choose J48 as our classifier in the following. It is worth mentioning that the accuracy reported in the table refers to a configuration with all 12 pressure sensors. As described in Section 4.3, sensors were excluded one at a time from the base configuration and accuracy of the recognition was evaluated by a 10-fold cross-validation; then, the configuration with highest accuracy in this step became the base configuration. The procedure then iterates until only one sensor is left. The testing has been performed with a decimation factor d = 1 .
Results of the backward selection of various sensor configurations are shown in Table 8. The column represents the number of active sensors in a given configuration. The row represents the ID excluded sensor in a given configuration (let us mention here that once excluded from a configuration, with the backward selection that sensor cannot be re-added to any following iteration). The initial configuration has 12 active sensors achieving an accuracy of 99.48%, as shown in Table 7. At the first iteration of the backward selection procedure, the least significant sensor is F S R 10 because the best accuracy that can be achieved with 11 sensors is 99.50% and corresponds to the exclusion of F S R 10 . Then, the sensors’ configuration during the second iteration is { F S R 0 , F S R 1 , F S R 2 , F S R 3 , F S R 4 , F S R 5 , F S R 6 , F S R 7 , F S R 8 , F S R 9 , F S R 11 } ; this time, the best configuration can be achieved excluding F S R 6 , with a maximum possible accuracy of 99.51% (see the bold value in column 10). The procedure continues to iterate excluding, sequentially, F S R 8 , F S R 9 , F S R 2 , F S R 5 , and F S R 0 . At this point, the current best configuration has five sensors ( { F S R 1 , F S R 3 , F S R 4 , F S R 7 , F S R 11 } ) with an accuracy of 99.47% as shown in Figure 5. We can conclude that this is the most effective sensor deployment configuration in terms of tradeoff between posture recognition accuracy and number of required sensors (which influences the system complexity and its economic cost). Excluding a further sensor, in fact, would decrease the accuracy below 99%.

5.3. Comparison of Recognition Results with Different BMI Values

Using the best sensor configuration determined in Section 5.2, we analyzed the recognition results grouped by the three weight categories classified by BMI (see Section 4.1). In addition, we analyzed the possible advantage of taking the BMI value as one of the features in the dataset.
The results grouped by BMI category obtained without considering the BMI as a feature of the dataset are shown in Table 9 while Table 10 depicts the result of classification accuracy using the dataset with the added BMI feature. Given that we obtained similar results, we can conclude that our sensor deployment can successfully fit different body physiognomies without the need to include user-dependent information (such as his/her BMI) into the classifier model.

5.4. Comparison of the Proposed Method with Previous Studies

To compare our current work with the state-of-the-art, we applied the J48 classifier to datasets generated using sensor deployment configurations proposed in other literature works, and we summarized the results in Table 11.
Using the same J48 classifier, we compared the quality of different sensor deployments. It can be clearly seen that our novel proposed deployment method allows to achieve the highest accuracy.

6. Conclusions

In this paper, we have proposed an Arduino-based pressure cushion to detect sitting postures of wheelchair users and a sensor selection method to obtain optimal deployment on the wheelchair. We extracted pressure data and performed initial offline analysis using WEKA. Laboratory experiments were run over a diversified sample of subjects (in terms of BMI category) with different sitting habits. Using the obtained dataset, we compared five different machine learning approaches (Decision Tree (J48), Support Vector Machines (SVM), Multilayer Perceptron (MLP), Naive Bayes, and k-Nearest Neighbor (k-NN)) and selected the most effective sensor deployment among the possible configurations given by our smart cushion. We evaluated seven different classifiers to recognize five sitting postures. Results showed that the lightweight J48 algorithm can efficiently and accurately ( 99 . 47 % ) be used for posture recognition of seated users. An advantage of the decision tree-based approach is that trained models are very lightweight and easily implemented on embedded devices. Ongoing efforts are focused on improving current end-user applications based on posture detection. For example, activity level assessment can be achieved based on the analysis of posture transitions. Future research will be devoted to acquiring more data from heterogeneous sensors deployed on the chair, applying multi sensor fusion techniques [57] to better detect the postures, and assessing user activity on a smart wheelchair more accurately.
Finally, we plan to improve our BodyCloud-based Wheelchair System [10,24,58] with the work presented in this paper.

Acknowledgments

The research is financially supported by the China–Italy Science & Technology Cooperation project “Smart Personal Mobility Systems for Human Disabilities in Future Smart Cities” (China-side Project ID: 2015DFG12210, Italy-side Project ID: CN13MO7), and the National Natural Science Foundation of China (Grant Nos: 61571336 and 61502360). This work has been also carried out under the framework of “INTER-IoT” Project financed by the European Union’s Horizon 2020 Research & Innovation Programme under Grant 687283. The authors would like to thank the volunteers who participated in the experiments for their efforts and time.

Author Contributions

Congcong Ma designed the hardware platform and the experiment protocol, implemented the methodologies and drafted the manuscript. Raffaele Gravina contributed to the technical and scientific aspects, to the analysis of the experimental data, and to a substantial revision of the manuscript. Wenfeng Li and Giancarlo Fortino supervised the study and contributed to the overall research planning and assessment. All the authors contributed to the proofreading and approved the final version of the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Perez, A.J.; Labrador, M.A.; Barbeau, S.J. G-sense: A scalable architecture for global sensing and monitoring. IEEE Netw. 2010, 24, 57–64. [Google Scholar] [CrossRef]
  2. Bellifemine, F.; Fortino, G.; Giannantonio, R.; Gravina, R.; Guerrieri, A.; Sgroi, M. SPINE: A domain-specific framework for rapid prototyping of WBSN applications. Softw. Pract. Exp. 2011, 41, 237–265. [Google Scholar] [CrossRef]
  3. Banos, O.; Villalonga, C.; Damas, M.; Gloesekoetter, P.; Pomares, H.; Rojas, I. PhysioDroid: Combining Wearable Health Sensors and Mobile Devices for a Ubiquitous, Continuous, and Personal Monitoring. Sci. World J. 2014, 2014, 490824. [Google Scholar] [CrossRef] [PubMed]
  4. Bourouis, A.; Feham, M.; Bouchachia, A. A new architecture of a ubiquitous health monitoring system: A prototype of cloud mobile health monitoring system. Int. J. Comput. Sci. 2012, 9, 434–438. [Google Scholar]
  5. Alshurafa, N.; Xu, W.; Liu, J.J. Designing a robust activity recognition framework for health and exergaming using wearable sensors. IEEE J. Biomed. Health Inform. 2014, 18, 1636–1646. [Google Scholar] [CrossRef] [PubMed]
  6. Ma, C.; Gravina, R.; Li, W.; Zhang, Y.; Li, Q.; Fortino, G. Activity Level Assessment of Wheelchair Users Using Smart Cushion. In Proceedings of the 11th International Conference on Body Area Networks (BodyNets 2016), Turin, Italy, 15–16 December 2016. [Google Scholar]
  7. Yang, L.; Li, W.; Ge, Y.; Fu, X.; Gravina, R.; Fortino, G. People-Centric Service for mHealth of Wheelchair Users in Smart Cities. In Internet of Things based on Smart Objects: Technology, Middleware and Applications; Springer International Publishing: Cham, Switzerland, 2014; pp. 163–179. [Google Scholar]
  8. Tong, Y.; Sun, J.; Chow, S.M.; Li, P. Cloud-Assisted Mobile-Access of Health Data with Privacy and Auditability. IEEE J. Biomed. Health Inform. 2014, 18, 419–429. [Google Scholar] [CrossRef] [PubMed]
  9. Fortino, G.; Giannantonio, R.; Gravina, R.; Kuryloski, P.; Jafari, R. Enabling Effective Programming and Flexible Management of Efficient Body Sensor Network Applications. IEEE Trans. Hum.-Mach. Syst. 2013, 43, 115–133. [Google Scholar] [CrossRef]
  10. Fortino, G.; Fatta, G.; Pathan, M.; Vasilakos, A. Cloud-assisted Body Area Networks: State-of-the-art and Future Challenges. Wirel. Netw. 2014, 20, 1925–1938. [Google Scholar] [CrossRef]
  11. Gravina, R.; Ma, C.; Pace, P.; Aloi, G.; Russo, W.; Li, W.; Fortino, G. Cloud-based Activity-aaService cyberphysical framework for human activity monitoring in mobility. Futur. Gener. Comput. Syst. 2016. [Google Scholar] [CrossRef]
  12. Fortino, G.; Guerrieri, A.; Bellifemine, F.; Giannantonio, R. Platform-independent development of collaborative wireless body sensor network applications: SPINE2. In Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, 2009 (SMC 2009), San Antonio, TX, USA, 11–14 October 2009; pp. 3144–3150. [Google Scholar]
  13. Chou, H.; Wang, Y.; Chang, H. Design intelligent wheelchair with ECG measurement and wireless transmission function. Technol. Health Care 2015, 24, 345–355. [Google Scholar] [CrossRef] [PubMed]
  14. Srivastava, P.; Chatterjee, S.; Thakur, R. Design and Development of Dual Control System Applied to Smart Wheelchair using Voice and Gesture Control. Int. J. Res. Electr. Electron. Eng. 2014, 2, 1–9. [Google Scholar]
  15. Dryvendra, D.; Ramalingam, M.; Chinnavan, E.; Puviarasi, P. A Better Engineering Design: Low Cost Assistance Kit for Manual Wheelchair Users with Enhanced Obstacle Detection. J. Eng. Technol. Sci. 2015, 47, 389–405. [Google Scholar] [CrossRef]
  16. Sonenblum, S.E.; Sprigle, S.; Caspall, J.; Lopez, R. Validation of an accelerometer-based method to measure the use of manual wheelchairs. Med. Eng. Phys. 2012, 34, 781–786. [Google Scholar] [CrossRef] [PubMed]
  17. Postolache, O.; Viegas, V.; Pereira, J.M.D. Toward developing a smart wheelchair for user physiological stress and physical activity monitoring. In Proceedings of the International Conference on Medical Measurements and Applications (MeMeA), Lisbon, Portugal, 11–12 June 2014; pp. 1–6. [Google Scholar]
  18. Hiremath, S.V.; Intille, S.; Kelleher, A. Detection of physical activities using a physical activity monitor system for wheelchair users. Med. Eng. Phys. 2015, 37, 68–76. [Google Scholar] [CrossRef] [PubMed]
  19. Wu, Y.H.; Wang, C.C.; Chen, T.S. An Intelligent System for Wheelchair Users Using Data Mining and Sensor Networking Technologies. In Proceedings of the 2011 IEEE Asia-Pacific Services Computing Conference (APSCC), Jeju Island, Korea, 12–15 December 2011; pp. 337–344. [Google Scholar]
  20. Nakane, H.; Toyama, J.; Kudo, M. Fatigue detection using a pressure sensor chair. In Proceedings of the International Conference on Granular Computing (GrC2011), Kaohsiung, Taiwan, 8–10 November 2011; pp. 490–495. [Google Scholar]
  21. Ding, D.; Cooper, R.A.; Cooper, R.; Kelleher, A. Monitoring Seat Feature Usage among Wheelchair Users. In Proceedings of the 29th International Conference on Engineering in Medicine and Biology Society (EMBS 2007), Lyon, France, 22–26 August 2007; pp. 4364–4367. [Google Scholar]
  22. Fortino, G.; Gravina, R.; Li, W.; Ma, C. Using Cloud-assisted Body Area Networks to Track People Physical Activity in Mobility. In Proceedings of the 10th International Conference on Body Area Networks (BodyNets 2015), Sydney, Australia, 28–30 September 2015; pp. 85–91. [Google Scholar]
  23. Fortino, G.; Parisi, D.; Pirrone, V.; Fatta, G.D. BodyCloud: A SaaS approach for community Body Sensor Networks. Future Gener. Comput. Syst. 2014, 35, 62–79. [Google Scholar] [CrossRef]
  24. Ma, C.; Li, W.; Cao, J.; Gravina, R.; Fortino, G. Cloud-based Wheelchair Assist System for Mobility Impaired Individuals. In Proceedings of the 9th International Conference on Internet and Distributed Computing Systems (IDCS 2016), Wuhan, China, 28–30 September 2016; pp. 107–118. [Google Scholar]
  25. Ma, C.; Li, W.; Cao, J.; Wang, S.; Wu, L. A Fatigue Detect System Based on Activity Recognition. In Proceedings of the International Conference on Internet and Distributed Computing Systems, Calabria, Italy, 22–24 September 2014; pp. 303–311. [Google Scholar]
  26. Xu, W.; Huang, M.; Amini, N.; He, L.; Sarrafzadeh, M. eCushion: A Textile Pressure Sensor Array Design and Calibration for Sitting Posture Analysis. IEEE Sens. J. 2013, 13, 3926–3934. [Google Scholar] [CrossRef]
  27. Tekscan Website. Available online: http://www.tekscan.com/ (accessed on 20 March 2017).
  28. Tan, H.Z.; Slivovsky, L.A.; Pentland, A. A sensing chair using pressure distribution sensors. IEEE/ASME Trans. Mechatron. 2001, 6, 261–268. [Google Scholar] [CrossRef]
  29. Mota, S.; Picard, R.W. Automated Posture Analysis for Detecting Learner’s Interest Level. In Proceedings of the Conference on Computer Vision and Pattern Recognition Workshop, 2003 (CVPRW ’03), Madison, WI, USA, 16–22 June 2003; Volume 5, p. 49. [Google Scholar]
  30. Meyer, J.; Arnrich, B.; Schumm, J.; Troster, G. Design and modeling of a textile pressure sensor for sitting posture classification. IEEE Sens. J. 2010, 10, 1391–1398. [Google Scholar] [CrossRef]
  31. Mutlu, B.; Krause, A.; Forlizzi, J.; Guestrin, C.; Hodgins, J. Robust, low-cost, non-intrusive sensing and recognition of seated postures. In Proceedings of the 20th Annual ACM Symposium on User Interface Software and Technology, Newport, RI, USA, 7–10 October 2007; pp. 149–158. [Google Scholar]
  32. Kamiya, K.; Kudo, M.; Nonaka, H.; Toyama, J. Sitting posture analysis by pressure sensors. In Proceedings of the 19th International Conference on Pattern Recognition (ICPR2008), Tampa, FL, USA, 8–11 December 2008; pp. 1–4. [Google Scholar]
  33. Xu, L.; Chen, G.; Wang, J.; Shen, R.; Zhao, S. A sensing cushion using simple pressure distribution sensors. In Proceedings of the 2012 IEEE Conference on. Multisensor Fusion and Integration for Intelligent Systems (MFI), Hamburg, Germany, 13–15 September 2012; pp. 451–456. [Google Scholar]
  34. Fard, F.; Moghimi, S.; Lotfi, R. Evaluating Pressure Ulcer Development in Wheelchair-Bound Population Using Sitting Posture Identification. Engineering 2013, 5, 132–136. [Google Scholar] [CrossRef]
  35. Yu, H.; Stoelting, A.; Wang, Y.; Yi, Z.; Sarrafzadeh, M. Providing a cushion for wireless healthcare application development. IEEE Potentials 2010, 29, 19–23. [Google Scholar] [CrossRef]
  36. Benocci, M.; Farella, E.; Benini, L. A context-aware smart seat. In Proceedings of the 2011 4th IEEE International Workshop on Advances in Sensors and Interfaces (IWASI), Brindisi, Italy, 28–29 June 2011; pp. 104–109. [Google Scholar]
  37. Bao, J.; Li, W.; Li, J.; Ge, Y.; Bao, C. Sitting Posture Recognition based on data fusion on pressure cushion. TELKOMNIKA Indones. J. Electr. Eng. 2013, 11, 1769–1775. [Google Scholar] [CrossRef]
  38. Arias, D.E.; Pino, E.J.; Aqueveque, P.; Curtis, D.W. Unobtrusive support system for prevention of dangerous health conditions in wheelchair users. Mobile Inform. Syst. 2016, 2016, 1–14. [Google Scholar] [CrossRef]
  39. Min, S.D. System for Monitoring Sitting Posture in Real-Time Using Pressure Sensors. U.S. Patent 20,160,113,583, 28 April 2016. [Google Scholar]
  40. Zemp, R.; Tanadini, M.; Plüss, S.; Schnüriger, K.; Singh, N.B.; Taylor, W.R.; Lorenzetti, S. Application of Machine Learning Approaches for Classifying Sitting Posture Based on Force and Acceleration Sensors. Biomed Res. Int. 2016, 2016, 1–9. [Google Scholar] [CrossRef] [PubMed]
  41. Barba, R.; Madrid, Á.P.; Boticario, J.G. Development of an inexpensive sensor network for recognition of sitting posture. Int. J. Distrib. Sens. Netw. 2015, 2015, 969237. [Google Scholar] [CrossRef]
  42. Fu, T.; Macleod, A. IntelliChair: An Approach for Activity Detection and Prediction via Posture Analysis. In Proceedings of the International Conference on Intelligent Environments, Washington, DC, USA, 30 June–4 July 2014; pp. 211–213. [Google Scholar]
  43. Kumar, R.; Bayliff, A.; De, D.; Evans, A.; Das, S.K.; Makos, M. Care-Chair: Sedentary Activities and Behavior Assessment with Smart Sensing on Chair Backrest. In Proceedings of the 2016 IEEE International Conference on Smart Computing, St Louis, MO, USA, 18–20 May 2016; pp. 1–8. [Google Scholar]
  44. Ma, C.; Li, W.; Gravina, R.; Fortino, G. Activity Recognition and Monitoring for Smart Wheelchair Users. In Proceedings of the 2016 IEEE Computer Supported Cooperative Work in Design (CSCWD), Nanchang, China, 4–6 May 2016; pp. 664–669. [Google Scholar]
  45. Darma Website. Available online: http://darma.co/ (accessed on 20 March 2017).
  46. Sensimat Website. Available online: http://www.sensimatsystems.com/ (accessed on 20 March 2017).
  47. Fsr406 Website. Available online: http://www.interlinkelectronics.com (accessed on 20 March 2017).
  48. Arduino DUE Website. Available online: https://www.arduino.cc (accessed on 20 March 2017).
  49. HC06 Bluetooth Website. Available online: https://www.olimex.com (accessed on 20 March 2017).
  50. Patil, T.; Sherekar, S. Performance analysis of Naive Bayes and J48 classification algorithm for data classification. Int. J. Comput. Sci. Appl. 2013, 6, 256–261. [Google Scholar]
  51. Chang, C.; Lin, C. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  52. Pirttikangas, S.; Fujinami, K.; Nakajima, T. Feature Selection and Activity Recognition from Wearable Sensors. Ubiquitous Comput. Syst. 2006, 4239, 516–527. [Google Scholar]
  53. Hall, M.A.; Frank, E. Combining Naive Bayes and Decision Tables. FLAIRS Conf. 2008, 2118, 318–319. [Google Scholar]
  54. Zhang, M.; Zhou, Z. ML-KNN: A lazy learning approach to multi-label learning. Pattern Recognit. 2007, 40, 2038–2048. [Google Scholar] [CrossRef]
  55. Sazonov, E.S.; Fulk, G.; Hill, J.; Schütz, Y.; Browning, R. Monitoring of posture allocations and activities by a shoe-based wearable sensor. IEEE Trans. Bio-Med. Eng. 2011, 58, 983–990. [Google Scholar] [CrossRef] [PubMed]
  56. Mark, H.; Eibe, F.; Geoffrey, H.; Bernhard, P.; Peter, R.; H, W.I. The WEKA data mining software: An update. ACM SIGKDD Explor. Newsl. 2009, 11, 10–18. [Google Scholar]
  57. Gravina, R.; Alinia, P.; Ghasemzadeh, H.; Fortino, G. Multi-Sensor Fusion in Body Sensor Networks: State-of-the-art and research challenges. Inf. Fusion 2017, 35, 68–80. [Google Scholar] [CrossRef]
  58. Fortino, G.; Gravina, R.; Guerrieri, A.; Fatta, G.D. Engineering Large-Scale Body Area Networks Applications. In Proceedings of the 8th International Conference on Body Area Networks (BodyNets 2013), Boston, MA, USA, 30 September–2 October 2013; pp. 363–369. [Google Scholar]
Figure 1. Architecture of the proposed system.
Figure 1. Architecture of the proposed system.
Sensors 17 00719 g001
Figure 2. Hardware design of the smart cushion.
Figure 2. Hardware design of the smart cushion.
Sensors 17 00719 g002
Figure 3. Sensor deployment: (a) schematic diagram of the sensor deployment; (b) sensors deployed on the real wheelchair.
Figure 3. Sensor deployment: (a) schematic diagram of the sensor deployment; (b) sensors deployed on the real wheelchair.
Sensors 17 00719 g003
Figure 4. Electrical circuitry of the pressure sensor system.
Figure 4. Electrical circuitry of the pressure sensor system.
Sensors 17 00719 g004
Figure 5. Selected sensors of the best configuration.
Figure 5. Selected sensors of the best configuration.
Sensors 17 00719 g005
Table 1. State-of-the-art on smart cushions based on pressure sensor arrays.
Table 1. State-of-the-art on smart cushions based on pressure sensor arrays.
AuthorSensor Array TypePlacement of the SensorsDetected PosturesClassification Technique/MethodAccuracy
Xu et al. [26]E-textilecushion on the seatSit up, forward, backward, lean left/right, right foot over left, left over rightGray scale image85.9%
Tekscan [27]E-textilecushion on the seat and backrestN/APressure mappingN/A
Tan et al. [28]E-textilecushion on the seat and backrestN/APCA, Grayscale image96%
Mota et al. [29]E-textilecushion on the seat and backrestLean forward/back, lean forward right/left, sit upright/on the edge, etc.Neural Network87.6%
Meyer et al. [30]textile pressure sensorcushion on the seatSeat upright, lean right, left, forward, back, left leg crossed over right etc.Naive Bayes82%
Multu et al. [31]pressure sensor on the seat and backrest19 pressure sensorsLeft leg crossed, right leg crossed lean left, lean back, lean forward etc.Logistic Regression87%
Kamiya et al. [32] 8 × 8 sensor arraycushion on the seatNormal, lean forward, lean backward, lean right, right leg crossed, lean right with right leg crossed etc.SVM98.9%
Xu et al. [33]Seat 6 × 8 , backrest 2 × 8 Cushion on the seat and backrestLean left front, lean front, lean right front, lean left, seat upright, lean right etc.Binary pressure distribution, Naive Bayes82.3%
Fard et al. [34] 8 × 8 sensor arraycushion on the seatSitting straight with bent keens, crossed legs left to right and right to left, stretched legspressure mapping technologyN/A
Table 2. Smart cushions based on fewer individual pressure sensors.
Table 2. Smart cushions based on fewer individual pressure sensors.
AuthorNumber of SensorsPlacement of the SensorsPostures RecognizedClassification TechniquesAccuracy
Hu et al. [35]62 on the seat and 4 on the backrestSit straight, lean left, lean right, lean backSVMN/A
Benocci et al. [36]54 on the seat and 1 on the backrestNormal posture, right side, left side, right/left/both legs extend forwardkNN92.7%
Bao et al. [37]55 on the seatNormal sitting, forward, backward, lean left, lean right, swing, shakeDensity-based cluster94.2%
Diego et al. [38]44 on the seatN/AThreshold-basedN/A
Min et al. [39]64 on the seat and 2 on the backrestCrossing left leg, crossing right leg, forward buttocks, bending down the upper body, correct postureDecision TreeN/A
Zemp et al. [40]1610 on the seat, 2 on the armrests and 4 on the backrestUpright, reclined, forward inclined, laterally right/left, crossed legs, left over right/ right over leftSVM, Multinomial Regression, Boosting, Neural Networks and Random Forest90.9%
Barba et al. [41]168 on the seat and 8 on the backrestStandard, lying, forward, normal position, sitting on the edge, legs crossed, sitting on one/two foot etc.N/AN/A
Fu et al. [42]84 on the seat and 4 on the backrestN/ADecision TreeN/A
Kumar et al. [43]44 on the backrestN/AExtremely Randomized Trees86%
Ma et al. [44]32 on the seat and 1 on the backrestUpright sitting, lean left, right, forward, backwardDecision Tree99.5%
Darma [45]66 on the seatN/AN/AN/A
Sensimat [46]66 on the seatN/AN/AN/A
Table 3. List of the components required by the prototype system.
Table 3. List of the components required by the prototype system.
Part NameDescriptionPrice (USD)
Arduino DUE boardData Processing Unit30
FSR 406 pressure sensor12 pressure sensors applied to the seat and backrest180
Bluetooth shield (HC-06)Bluetooth module to connect the cushion to mobile devices8
Seat CoverA seat cushion10
Total 228
Table 4. Body Mass Index (BMI) distribution of the subjects participating in the experiments.
Table 4. Body Mass Index (BMI) distribution of the subjects participating in the experiments.
DescriptionUnderweightNormalOverweight and Obese
BMI<18.5[18.5, 25)⩾25
Number of subjects444
Table 5. Wheelchair user postures of interest.
Table 5. Wheelchair user postures of interest.
PostureDescriptionSamples of Posture
Proper Sitting (PS)User seated correctly on the wheelchair7200
Lean Left (LL)User seated leaning to the left7200
Lean Right (LR)User seated leaning to the right7200
Lean Forward (LF)User seated leaning forward7200
Lean Backward (LB)User seated leaning backward7200
Table 6. Tested machine learning algorithms.
Table 6. Tested machine learning algorithms.
No.ClassifierParameters
1J48C = 0.25, M = 2
2SVMSVM Type: C-SVC, Kernel Type: RBF, C = 1, Degree = 3
3MLP9 hidden layer neurons
4Naive Bayesdefault
5Naive BayesBayesNet
6kNNk = 1
7kNNk = 5
Table 7. Summary of classifiers’ performance.
Table 7. Summary of classifiers’ performance.
No.ClassifierAccuracyPrecisionRecallF-MeasureModel Build Time (s)
1J4899.48%0.9950.9950.9951.98
2SVM79.08%0.8800.7360.760320.34
3MLP95.5%0.9260.9260.926265.46
4Naive Bayes49.09%0.5850.4910.4270.24
5BayesNet94.06%0.9450.9410.9410.93
6kNN (k = 1)98.53%0.9950.9950.9950.04
7kNN (k = 5)98.52%0.9950.9950.9950.08
Table 8. Backward selection of the best sensor configuration. Bold font highlights the best accuracy for each number of active sensors.
Table 8. Backward selection of the best sensor configuration. Bold font highlights the best accuracy for each number of active sensors.
Number of Active Sensors
1110987654321
Sensor ID099.48%99.49%99.50%99.49%99.50%99.49%99.47%
197.06%97.05%97.11%97.11%97.11%97.13%96.51%92.32%90.54%81.23%63.98%
299.49%99.49%99.49%99.50%99.51%
399.46%99.50%99.50%99.50%99.50%99.42%99.41%98.99%
499.46%99.49%99.50%99.50%99.50%99.46%99.44%98.84%94.36%87.93%48.72%
599.48%99.49%99.49%99.49%99.49%99.50%
699.49%99.51%
799.48%99.49%99.49%99.49%99.49%99.44%99.42%95.80%92.28%89.51%
899.48%99.50%99.51%
999.49%99.50%99.51%99.51%
1099.50%
1199.48%99.30%99.32%99.33%99.27%98.98%98.98%98.90%98.01%
LSS 1 F S R 10 F S R 6 F S R 8 F S R 9 F S R 2 F S R 5 F S R 0 F S R 3 F S R 11 F S R 7 F S R 1
1 Least Significant Sensor.
Table 9. Results with different BMI values without considering the BMI feature.
Table 9. Results with different BMI values without considering the BMI feature.
AccuracyPrecisionRecallF-Measure
Underweight99.92%0.9990.9990.999
Normal98.67%0.9870.9870.987
Overweight and Obese99.82%0.9980.9980.998
All99.47%0.9950.9950.995
Table 10. Results with different BMI values using the BMI feature in the J48 tree.
Table 10. Results with different BMI values using the BMI feature in the J48 tree.
AccuracyPrecisionRecallF-Measure
Underweight99.93%0.9990.9990.999
Normal98.67%0.9870.9870.987
Overweight and Obese99.83%0.9980.9980.998
All99.50%0.9950.9950.995
Table 11. Recognition results of different sensor deployment.
Table 11. Recognition results of different sensor deployment.
AuthorSensor DeploymentAccuracy
Hu et al. [35]6 F S R s (0,1,7,8,10,11)98.70%
Benocci et al. [36]5 F S R s (0,1,5,6,9)97.58%
Bao et al. [37]5 F S R s (0,1,2,3,4)99.16%
Diego et al. [38]4 F S R s (0,1,5,6)97.11%
Min et al. [39]4 F S R s (0,1,5,6,10,11)98.5%
Ma et al. [44]3 F S R s (2,4,9)87.25%
Darma [45], Sensimat [46]6 F S R s (0,1,2,4,5,6)99.14%
Novel proposed method5 F S R s (1,3,4,7,11)99.47%

Share and Cite

MDPI and ACS Style

Ma, C.; Li, W.; Gravina, R.; Fortino, G. Posture Detection Based on Smart Cushion for Wheelchair Users. Sensors 2017, 17, 719. https://doi.org/10.3390/s17040719

AMA Style

Ma C, Li W, Gravina R, Fortino G. Posture Detection Based on Smart Cushion for Wheelchair Users. Sensors. 2017; 17(4):719. https://doi.org/10.3390/s17040719

Chicago/Turabian Style

Ma, Congcong, Wenfeng Li, Raffaele Gravina, and Giancarlo Fortino. 2017. "Posture Detection Based on Smart Cushion for Wheelchair Users" Sensors 17, no. 4: 719. https://doi.org/10.3390/s17040719

APA Style

Ma, C., Li, W., Gravina, R., & Fortino, G. (2017). Posture Detection Based on Smart Cushion for Wheelchair Users. Sensors, 17(4), 719. https://doi.org/10.3390/s17040719

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop