Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Efficient Convolutional Neural Network-Based Keystroke Dynamics for Boosting User Authentication
Previous Article in Journal
TCU-Net: Transformer Embedded in Convolutional U-Shaped Network for Retinal Vessel Segmentation
Previous Article in Special Issue
A Fusion-Assisted Multi-Stream Deep Learning and ESO-Controlled Newton–Raphson-Based Feature Selection Approach for Human Gait Recognition
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Human Activity Recognition in the Presence of Occlusion

by
Ioannis Vernikos
1,
Theodoros Spyropoulos
2,
Evaggelos Spyrou
1,* and
Phivos Mylonas
3
1
Department of Informatics and Telecommunications, University of Thessaly, 35131 Lamia, Greece
2
Department of Digital Systems, University of Piraeus, 18534 Piraeus, Greece
3
Department of Informatics and Computer Engineering, University of West Attica, Egaleo Park, 12243 Athens, Greece
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(10), 4899; https://doi.org/10.3390/s23104899
Submission received: 20 March 2023 / Revised: 28 April 2023 / Accepted: 16 May 2023 / Published: 19 May 2023
Figure 1
<p>The 25 skeletal joints extracted by Microsoft Kinect v2. The joints and edges that correspond to the 5 main body parts, namely head-torso, right arm, left arm, right leg, and left leg are colored in blue, red, green, orange, and purple, respectively.</p> ">
Figure 2
<p>A sequence of an actor performing the activity <span class="html-italic">wearing a hat</span>. Extracted human skeleton 3D joints using the Kinect SDK have been overlaid. Frames have been taken from the NTU dataset [<a href="#B9-sensors-23-04899" class="html-bibr">9</a>] and have been trimmed for illustration purposes.</p> ">
Figure 3
<p>A sequence of an actor performing the activity <span class="html-italic">salute</span>. Extracted human skeleton 3D joints using the Kinect SDK have been overlaid. Frames have been taken from the PKU-MMD dataset [<a href="#B8-sensors-23-04899" class="html-bibr">8</a>] and have been trimmed for illustration purposes.</p> ">
Figure 4
<p>Pseudo-coloured images for the PKU-MMD [<a href="#B8-sensors-23-04899" class="html-bibr">8</a>] and the NTU [<a href="#B9-sensors-23-04899" class="html-bibr">9</a>] datasets. LA, RA, LL, and RL denote the cases of occlusion of left arm, right arm, left leg, and right leg.</p> ">
Figure 5
<p>Example skeleton sequences of the activity <span class="html-italic">salute</span> from the PKU-MMD dataset [<a href="#B8-sensors-23-04899" class="html-bibr">8</a>]. First row: skeletons include all 25 joints; second row: joints corresponding to the left arm have been discarded.</p> ">
Figure 6
<p>Example skeleton sequences of the activity <span class="html-italic">wearing a hat</span> from the NTU dataset [<a href="#B9-sensors-23-04899" class="html-bibr">9</a>].First row: skeletons include all 25 joints; second row: joints corresponding to the left arm have been discarded.</p> ">
Figure 7
<p>The Convolutional Neural Network that has been used in this work.</p> ">
Versions Notes

Abstract

:
The presence of occlusion in human activity recognition (HAR) tasks hinders the performance of recognition algorithms, as it is responsible for the loss of crucial motion data. Although it is intuitive that it may occur in almost any real-life environment, it is often underestimated in most research works, which tend to rely on datasets that have been collected under ideal conditions, i.e., without any occlusion. In this work, we present an approach that aimed to deal with occlusion in an HAR task. We relied on previous work on HAR and artificially created occluded data samples, assuming that occlusion may prevent the recognition of one or two body parts. The HAR approach we used is based on a Convolutional Neural Network (CNN) that has been trained using 2D representations of 3D skeletal motion. We considered cases in which the network was trained with and without occluded samples and evaluated our approach in single-view, cross-view, and cross-subject cases and using two large scale human motion datasets. Our experimental results indicate that the proposed training strategy is able to provide a significant boost of performance in the presence of occlusion.

1. Introduction

Human activity recognition (HAR) from visual data is considered a very challenging computer vision task [1], while it has been continuously attracting increasing attention within the research community [2]. It typically aims to recognize human motion or behavior within a video sequence, using data captured by a camera. In the following, we will define an “activity” as motion performed by a human, taking place within a relatively short (however, not instant) time period and involving multiple body parts [2]. The aforementioned informal definition may be expanded to also include interactions between a human and an object or between two or more humans (group activities). Additionally, note that it does not include gestures, as they require a relatively small (“instant”) amount of time, while typically requiring a single body part.
There exist several HAR applications. The most notable ones include content-based video summarization (e.g., for retrieval of activities), human–computer interaction (e.g., for interaction based on users’ pose and/or motion), education (e.g., for monitoring students’ attendance), healthcare (e.g., for fall detection), video surveillance (e.g., for abnormal activity recognition), and sports (e.g., for identifying players’ actions) [3]. We should note that several HAR applications require the use of either wearable sensors or environmental sensors (i.e., installed in the subject’s environment). Typical wearable sensors include smartwatches, hand/body worn sensors, smartphones, etc. Additionally, typical environmental sensors include video/thermal cameras, microphones, infrared, pressure, magnetic, and RFID sensors [4].
Although wearable sensors may combine adequate performance with low cost, they are not preferred by users and their usability is below average [5,6]. Moreover, it is common sense that, e.g., in the case of a home environment, overloading it with a large number of sensors may be both expensive and unwanted by the users, since it would require many interventions in home furniture or appliances. Thus, a common approach to combining low cost with minimal interventions is to use only cameras, whose role is to capture users’ motion. However, the undesired trade-off is that, although their performance may be excellent in laboratory conditions, in real-life environments they exhibit a performance loss, as they suffer from viewpoint and illumination changes and occlusion.
The problem of viewpoint changes and their effect on HAR has been extensively studied in our previous work [7]. Specifically, we dealt with the problem of viewpoint invariance and demonstrated that the decrease of accuracy due to viewpoint changes may be limited when using more than one camera, allowing for satisfactory results in real-life applications. Moreover, due to recent advances in technology and specifically in camera sensors, recent cameras perform significantly better in low-light conditions compared to previous models. Additionally, some cameras (often called “RGB+depth” cameras), complementary to color video, also provide depth information, resulting in robust extraction of human figures, e.g., as a set of 3D points, and even in poor illumination conditions. Therefore, from the three aforementioned problems, occlusion is the one that introduces the most limitations.
To understand the notion of occlusion, let us consider two objects in a 3D space. Given a certain viewpoint, occlusion occurs when one of the objects partially/fully blocks the view of the second. In the context of HAR, occlusion may occur, e.g., in cases when the subject performs an activity and is occluded by a physical object such as furniture or by another person. Regarding its effect on HAR, let us consider the (extreme) case of a subject performing the activity “handwaving,” while the involved arm and hand are not visible. Clearly, in order to recognize this activity, crucial information regarding the subject’s motion is missing. This case might be excessive, but there are applications such as, e.g., surveillance systems in crowded places such as metro stations where an activity of a subject might not be recognized due to the occlusion of some of their body parts by the crowd surrounding her/him. Occlusion may also be temporally partial, i.e., one or more body parts may be occluded but not during the whole activity. This may happen, e.g., in the case of a moving subject.
In this work, our goal was to assess how partial occlusion of the subject affects the accuracy of HAR and to overcome its effect. To the best of our knowledge, public motion-based datasets that may be used for HAR [8,9] have been created under ideal laboratory conditions, which include the prevention of any type of occlusion. Thus, since the creation of a large scale dataset is a time consuming task, we decided to follow an approach such as that of Gu et al. [10]. More specifically, we simulated occlusion by manually discarding subsets of joints that correspond to one/two body parts of the subject, while she/he is performing a given activity. We assumed that these parts remain occluded during the whole activity, that is, we did not consider the case of temporally partial occlusion. The contribution of this work is twofold: (a) we extended the small-scale evaluation of our previous work [11], where we assessed how partial occlusion of the subject affects the accuracy of HAR focusing on Activities of Daily Living (ADLs) [12]; herein, we evaluated it using two challenging large-scale human motion datasets [8,9]; (b) we tried to overcome the effect of occlusion by incorporating occluded samples in our training procedure.
The rest of this paper is organized as follows: in Section 2, we present research works that aimed to assess or even tackle the effect of occlusion in HAR-related scenarios. Then, in Section 3, we present the proposed methodology for simulating occlusion, for assessing its effect on HAR and for overcoming its effect. Experimental results are presented and discussed in Section 4, while conclusions are drawn in Section 5, wherein plans for future work are also presented.

2. Related Works

As has already been mentioned, the problem of HAR is one of the most challenging research areas in computer vision; therefore, a large amount of research has been conducted on this subject during recent years. The majority of them have been based on 2D representations of skeletal motion which are then used as input to CNNs [13,14,15,16,17,18]. An extensive survey was conducted by Wang et al. [2]. Interestingly, although it is widely accepted, as it is intuitive, that occlusion constitutes one of the most important factors that compromise the performance of HAR approaches [19], rendering them to produce poor and in extreme cases unusable results, there do not exist many studies that focus on studying its effects on the performance of HAR approaches or even attempt to propose methods so as to overcome them.
In previous work [11], we have extensively studied the effect of occlusion in the task of HAR. We relied on the HAR approach of Papadakis et al. [20] and simulated occlusion upon removing one or more body parts from 3D human skeletons and during the whole activity. Throughout our experiments, we used a Convolutional Neural Network that had been trained without using any occluded samples. Specifically, we considered the cases of removing one or both arms, one or both legs, and also the cases of removing the arm and the leg of the same side. Our experimental evaluation proved that the removal of one or both arms led to a drop of performance, which in some cases was considered quite severe. However, we concluded that this was not a surprise, as most of the activities considered therein were expressed mainly by one or both arms’ motion. In the following, we shall attempt to present the approaches that have dealt with occlusion and their main findings.
Angelini et al. [21] proposed an HAR approach that relied on OpenPose [22] for pose extraction. Specifically, they extracted both low- and high-level features from body poses using a 1D CNN and an LSTM for classification. They used real CCTV data where part of the body was occluded; however, they also created synthetic occluded data by removing body parts. They experimented by incorporating occluded samples in training and showed that, in most cases, this strategy is able to provide a boost of performance in cases of both occlusion and missing data. They experimented on multiple datasets, i.e., i3DPost [23], IXMAS [24], KTH [25] etc., while they also created their own dataset called ISLD, comprising classes such as bend, box, hand-clap, one-hand wave, run, walk, etc.
Iosifidis et al. [19] used a multi-camera setup, surrounding the subject from all sides. They trained a model using data from all available cameras and then, in order to simulate occlusion, they used a subset of those cameras. Their main idea was that, due to occlusion, not all cameras are able to simultaneously capture the subject’s motion, yet in all cases more than one camera manages to capture the whole body of the subject. In addition, the recognition of an activity was performed using the combined results of cameras that had not been “affected” by any occlusion, while those that had been omitted “suffered” from occlusion. They conducted experiments using two publicly available datasets, the i3DPost [23], comprising eight daily activities, and the AIIA-MOBISERV, [26] comprising three meal activities. The proposed method has been evaluated for both single-view and multi-view action classification problems, providing results with sufficient accuracy.
Li et al. [27] used manually defined occlusion areas and proposed a new encoding technique to transform a skeleton into a feature matrix and integrated an attention model into a Generative Adversarial Network, focusing on completing the missing data of the feature matrix, rather than the exact skeleton. For classification, they used a Residual Network (ResNet). To evaluate their approach, they used a custom dataset of eight actions of construction workers, such as driving a truck, transporting cement, paving concrete, etc. Additionally, Li et al. [28] proposed an approach based on action graphs so as to capture the dynamics of actions and used a bag-of-3D points so as to describe the salient postures that correspond to the nodes of the action graph. They simulated occlusion by first dividing the depth map into four quadrants and then by ignoring 3D points falling into the specified quadrants that are considered to be affected by occlusion. They did not include any occluded samples in training. They used a dataset comprising 20 actions such as high arm wave, horizontal arm wave, hammer, etc. Their results showed that, unless the critical quadrants were omitted, the performance drop was relatively small.
Similarly, Gu et al. [10] generated occlusion masks which were used both in training and evaluation, so as to simulate occlusion in more than one 2D skeletal joint. Moreover, they attempted to reconstruct the missing skeleton parts using a regression network. They evaluated their work using their own dataset, namely MMHuman, comprising six classes (e.g., Kung Fu, shake hand, etc.). They also demonstrated that their approach is able to outperform previous pose estimation methods on the Human 3.6 M dataset [29]. Finally, Yang et al. [30] proposed to pre-train a single pose embedding network, namely OR-VPE, whose role was to learn occlusion-robust representations for pose sequences. They simulated occlusions using an augmentation mechanism whose goal was to randomly prune the spatio-temporal skeleton structure. They used an encoder in order to create the pose embeddings in a latent space and a contrastive module to render this space occlusion-invariant. They evaluated their approach using the NTU-RGB+D dataset [9] and also the Toyota Smarthome [31], N-UCLA [32], and Penn Action [33] datasets.

3. Proposed Methodology

3.1. Visual Data

In this work, we considered 3D skeleton data that were captured by the Microsoft Kinect v2 camera. Specifically, these data consist of the 3D spatial coordinates of a set of 25 human skeleton joints that were captured during the recording of an action over time, for each subject in the scene and for a maximum of six subjects. The human skeleton comprises these joints and was modelled as a graph; the joints are its nodes, corresponding to parts such as head, shoulders, knees etc. and are connected by edges that follow the body structure. An illustration of the human skeleton joints and edges that were extracted by Kinect v2 are illustrated in Figure 1. Note that, following the example of [10,11], we divided the skeleton into its five main body parts, namely head-torso, right arm, left arm, right leg, and left leg, since one or more will be removed as an effect of occlusion. Therefore, in Figure 1 we illustrate the skeleton, highlighting these parts. In Figure 2 and Figure 3, we illustrate the RGB and the skeletal data of two sequences of actors performing activities.

3.2. Image Representation of Skeletal Data

Since our goal was to use a CNN for the classification of activities, a mandatory step was to represent the 3D skeletal motion with a 2D image, so that the latter would capture and preserve the spatio-temporal properties of this motion. Of course, such representations should be able to discriminate among different activities, while also complying with the graph structure of the skeleton representation that is used. Therefore, and based on our previous work [34], we adopted a pseudo-colored representation of the skeletal motion that aimed to capture the varying inter-joint distances during the performance of an activity.
Specifically, let x ( n ) , y ( n ) and z ( n ) , n = 1 , , N denote the sequences of the x , y , z co-ordinates of the skeletal joint motion at the n-th frame and occurring within a period of time of N frames. Since Kinect v2 extracts 25 joints, this motion “produces” a set of 75 such sequences. Prior to the creation of the image representation, there is the need to address the problem of temporal variability. It should be intuitive that (a) different activities require different amounts of time and (b) the same activity requires different amounts of time both in the cases it is performed by different subjects and even when it is performed by the same subject. As in [20], we used a linear interpolation step, setting the number of frames F as equal for all activity examples. Upon performing several experiments, we set F = 150 . In order to create the pseudo-colored images, spatial co-ordinates x, y, z were assigned to R, G, B color channels of the resulting pseudo-colored image, respectively. The procedure of creating these images was as follows: let R denote the red channel, i the i-th frame, and j the j-th joint. The equation used to create the R-value of pixel ( i , j ) is: R ( i , j ) = R ( i + 1 , j ) R ( i , j ) . Similarly, G and B values were created. In Figure 4, examples of these pseudo-colored images are illustrated.

3.3. Occlusion of Skeletal Data

As we have already discussed, in real-life scenarios, occlusion may be one of the most crucial causes of the poor performance of HAR approaches. Prior to describing our approach to simulating occlusion, we first discuss the main causes. Let us consider an assistive living scenario, where the subject’s behavior is monitored by cameras. In many occasions, actions take place behind furniture or more than one person is present in the same room. Thus, it is expected that occlusion is often occurring, incommoding the task of recognition algorithms, as it is the main factor for the loss of visual information. Even in cases of simple activities that involve one or two body parts, it is obvious that by occluding them, i.e., by losing the information regarding their motion, recognition algorithms are prone to failure.
As we have already mentioned, the goal of this work was to assess whether the performance drop caused by occlusion may be limited upon adding occluded samples to the training process. First, we should emphasize that most publicly available, large scale datasets such as the ones we used for the evaluation of this work, i.e., NTU RGB+D [9] and the PKU-MMD [8], have been created in laboratory (i.e., ideal) conditions. Thus, there exist no data samples that are affected by occlusion, while room illumination is appropriate for video capturing. We assumed that these datasets are “clean,” i.e., they do not contain any samples with missing/noisy data and the only limitations imposed are due to the capabilities of Kinect v2, which was used for data collection.
In order to simulate the effect of occlusion, i.e., to create artificially occluded activity samples, we followed the paradigms of [10,11] and we discarded subsets of the skeleton that correspond to meaningful body parts. Specifically, as we have already discussed in Section 3.1, we considered arms (each comprising shoulder, elbow, wrist, hand, and hand-tip), legs (each comprising hip, knee, ankle, and foot), and the torso (comprising head, neck, spine-shoulder, spine-mid, and spine-base). This way, we formed five body parts, as illustrated in Figure 1. We should herein note that, in our case, we assumed that occlusion affects the whole activity, i.e., the body parts that were occluded remained occluded during the whole activity duration.
We considered the cases of occlusion of: (a) an arm; (b) a leg; (c) both arms; (d) both legs; and (e) an arm and a leg. Intuitively, when two parts are occluded, the most expected case is that they both are from the same side. Moreover, our initial experiments indicated that by removing the torso the accuracy was not significantly affected. Therefore, throughout our experiments, the torso was always present. An example of an activity with and without the occlusion of a body part is illustrated in Figure 5 and Figure 6. In this example, it is evident that a single body part may carry significant information regarding the activity.
Note that several research works among those that have been presented in Section 2 also used occluded samples within the training process [10,21,35]. However, in [21], only random skeletal joints were removed; yet these did not correspond to structured body parts as in our case. Moreover, in [10], occlusion was not continuous as in our case; instead it may go from partial to full during an activity, yet within several temporal periods, no occlusion is present. Additionally, contrary to [19], where due to occlusion only some cameras were affected, in our approach, all available cameras were similarly affected.

3.4. Classification

For classification, we used a Convolutional Neural Network (CNN). Specifically, the architecture of the CNN that has been used throughout our experiments has been experimentally defined and was initially used in previous works [20,34]. It consists of a 2D convolutional layer that filters the 25 × 150 input image with five kernels of 3 × 3 size, a max-pooling layer that performs 2 × 2 subsampling, two consecutive convolutional layers of size 3 × 3 with 10 and 15 kernels, a max-pooling layer performing 2 × 2 subsampling, a flattened layer that transforms the output of the last pooling layer into a vector, which consists of the input to a dense layer upon applying a dropout layer with a dropout rate equal to 0.5 and a second dense layer producing the output of the network. The CNN is illustrated in Figure 7.

4. Experiments and Results

In this section, we present and discuss the results of the experimental evaluation of the proposed approach, using two publicly available and well-known human activity recognition datasets. The goal of this evaluation was to assess whether the inclusion of occluded samples within the training process may improve the performance of classification strategies. We relied on our previous work [34], where we presented an image representation of 3D skeletal motion. Moreover, we used the same CNN architecture for classification as in [36]. Note that the goal of the following experiments was not to provide state-of-the-art results in human activity recognition, but to be used as a baseline so as to verify the effectiveness of our approach.

4.1. Datasets

For the experimental evaluation of our work, we used two widely known and freely available datasets. More specifically, we used:
a.
PKU-MMD [8], which is an action recognition dataset that contains 1076 long video sequences in 51 action categories, performed by 66 subjects and in three camera views. The total amount of the action instances is approximately 20,000. For the recording of the dataset they used the Kinect v2 sensor. The 51 action categories are divided into two parts: 41 daily actions (e.g., brushing hair, brushing teeth, eating a meal, etc.) and 10 interaction actions (e.g., handshaking, hugging another person, slapping another person, etc.). For the data collection, three cameras with different viewpoints ( 45 , 0 , + 45 ) were used.
b.
NTU-RGB+D [9], which is a large-scale RGB+D action recognition dataset containing approx. 57,000 video samples from 60 action classes and from 40 distinct subjects. The dataset contains 40 daily actions (drop, stand up, play with phone, etc.), 11 mutual actions (punch, pushing, hugging, etc.), and nine health-related actions (sneeze, nausea, neck pain, etc.). For data collection, the same camera setup as in PKU-MMD was used.

4.2. Experimental Setup and Implementation Details

The experiments were performed on a workstation with an Intel core™i7 5820K 12-core processor on 3.3 GHz and 16 GB RAM, using NVIDIA™GeForce RTX 2060 SUPER GPU with 8 GB and Ubuntu 20.04 (64 bit). All pipelines were implemented in Python, using Keras 2.6 [37] and Tensorflow 2.3 [38] backend. The batch size used consisted of 32 and 80 epochs. The optimizer used was the RMSprop, while the categorical cross-entropy loss was adopted. The data split for training, validation, and testing followed exactly the one proposed by the datasets’ authors [8,9].

4.3. Evaluation Protocol and Results

For the evaluation of the proposed approach, we studied eight cases of missing body parts of the skeleton, as we have discussed in Section 3.3. Specifically, we performed three types of experiments for comparison purposes:
a.
Training and test sets consisting of full body parts only;
b.
Training set consisting of full body parts only, test set consisting of occluded body parts only;
c.
Training set consisting of full and occluded body parts, test set consisting of occluded body parts.
Following typical evaluation protocols encountered in multi-camera datasets as those used herein [8,9], we further split the aforementioned experiments into three types:
a.
Per camera position (single view), where training and test sets derive from the same camera (viewpoint), e.g., both deriving from the middle camera;
b.
Cross view experiments, where training and test sets derive from different cameras (viewpoints), e.g., training deriving from the middle camera, testing deriving from the left camera;
c.
Cross subject experiments, wherein subjects are split into training and testing groups, i.e., each subject appears in exactly one of these groups.
We performed experiments using the full PKU-MMD [8] and NTU-RGB+D [9] datasets. We also performed experiments using subsets of both these datasets, as follows:
a.
From the PKU-MMD dataset we selected 11 actions related to activities of daily living: eating a meal/snack, falling, handshaking, hugging another person, making a phone call/answering the phone, playing with phone/tablet, reading, sitting down, standing up, typing on a keyboard, and wearing a jacket;
b.
From the NTU-RGB+D dataset we selected 12 medical conditions: sneezing/coughing, staggering, falling down, headache, chest pain, back pain, neck pain, nausea/vomiting, fanning self, yawning, stretching oneself, and blowing the nose.
In all cases, we measured F 1 score and accuracy. Results of the proposed methodology are depicted in Table 1 for the 11 activities of daily living of the PKU-MMD dataset, in Table 2 for the 12 medical conditions of NTU-RGB+D dataset, and in Table 3 and Table 4 for the full PKU-MMD and NTU-RGB+D datasets, respectively.
Starting with the PKU-MMD dataset and based on the results depicted in Table 1 and Table 3, we observe the following:
  • In almost all experiments, both the F 1 score and the accuracy of the proposed approach are improved compared to the “simple” case of occlusion;
  • In general, it is also comparable to the baseline case;
  • The only exception to the above-mentioned are the cases of cross-view experiments when using:
    Only the 11 activities of daily living, and specifically in the case where the left and the middle cameras were used for training, while the right one was used for testing; and
    The full PKU-MMD dataset, and specifically in the case where the left camera was used for training, while the right one was used for testing;
  • As expected, and due to the activities involved, in cases of the occlusion of legs, a drop of performance is not observed, in general.
In the case of the NTU-RGB+D dataset and based on the results depicted in Table 2 and Table 4, we observe the following:
  • In cases where both metrics indicated improved performance, compared to the “simple” case of occlusion:
  • In general, in that case it is also comparable to the baseline case;
  • The only exception to the above-mentioned are the cases of the cross-view experiments when using:
    Only the 12 medical conditions, and specifically in the case where the middle camera was used for training, while the right one was used for testing or vice versa;
    The full NTU-RGB+D dataset, and specifically in the case of single view;
    A few combinations of training and testing in the cross-view case;
  • Additionally, in this case and as expected, in the case of the occlusion of legs, a drop of performance is not observed, in general.
A notable observation of our experimental results indicates that, due to data augmentation that is indirectly caused by the addition of the occluded samples in the training phase, in many cases and in both datasets we may observe that the results of the proposed approach, i.e., when the training set is augmented with occluded samples, typically lead to higher accuracy and F 1 score. However, this result was not anticipated prior to the experimental evaluation. Of course, since it is well-known that data augmentation may typically boost the performance of deep architectures, this result is adequately explained this way.

5. Conclusions and Future Work

In this paper, we dealt with the problem of human activity recognition in the presence of occlusion, which still remains an aspect that has not yet received significant attention, while it compromises performance. Our goal was twofold: (a) we intended to extensively study the effect of occlusion in human activity recognition tasks; and (b) we aimed to assess the effect of the inclusion of occluded samples within the training process. As expected, occlusion is responsible for the loss of parts of motion, which in many cases are crucial for the recognition of an activity. Since to the best of our knowledge there do not exist any publicly available datasets, we used artificially-created occluded data and worked under the hypothesis that one or more body parts remain occluded during the whole duration of the activity. Of course this is often a harder case than those that are encountered in typical, real-life occlusion scenarios, wherein occlusion often does not affect a whole body part, but some of its joints instead. Additionally, in such scenarios, occlusion is not continuous. Some parts may remain occluded for some time intervals. Thus, it is our belief that the proposed approach constitutes a harder problem than typical occlusion cases.
To tackle occlusion, we followed a typical recognition approach. We first created 2D representations of human joint motion. Then we trained a Convolutional Neural Network both with and without artificially occluded samples. We evaluated our approach in typical human activity recognition scenarios, i.e., single-view, cross-view, and cross-subject, using two publicly available, large-scale datasets and upon considering eight occlusion cases involving one or two arms and/or legs. Prior to the experimental evaluation we expected that (a) the inclusion of occluded samples will boost the performance of recognition in the presence of occlusion; and (b) since most activities depend mainly on arms’ motion, recognition will be mostly affected in cases of arm occlusion. Both the aforementioned hypotheses have been experimentally proved. To our surprise, we also observed that the augmentation of the training set with occluded samples led to a boost of performance. We consider this as the most important finding of this work, since it may constitute a simple and effective approach for data augmentation in future work.
Future research efforts could focus on several other important aspects of the problem of occlusion. Firstly, we would like to investigate cases such as temporally partial occlusion. Moreover, we intend to investigate the use of other deep neural network architectures, such as generative adversarial networks (GANs) or approaches such as deep regression, for completing missing information. Finally, we plan to construct a dataset containing “real” occluded samples and also perform real-life experiments.

Author Contributions

Conceptualization E.S.; methodology, I.V.; validation, I.V. and T.S.; data curation, I.V. and T.S.; writing—original draft preparation, I.V. and E.S.; writing—review and editing, E.S. and P.M.; supervision, E.S. and P.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The PKU-MMD dataset is available at https://www.icst.pku.edu.cn/struct/Projects/PKUMMD.html (accessed on 6 July 2022). The NTU-RGB+D dataset is available at https://rose1.ntu.edu.sg/dataset/actionRecognition/ (accessed on 6 July 2022).

Acknowledgments

The implementation of the doctoral thesis was co-financed by Greece and the European Union (European Social Fund-ESF) through the Operational Programme “Human Resources Development, Education and Lifelong Learning” in the context of the Act “Enhancing Human Resources Research Potential by undertaking a Doctoral Research” Sub-action 2: IKY Scholarship Programme for PhD candidates in the Greek Universities.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ADLActivity of Daily Living
CCTVClosed Circuit Television
CNNConvolutional Neural Network
CSCross-Subject
CVCross-View
GPUGraphics Processing Unit
HARHuman Activity Recognition
LALeft Arm
LLLeft Leg
LSTMLong-Short Term Memory
RARight Arm
RLRight Leg
RFIDRadio Frequency Identification
SVSingle-View

References

  1. Vrigkas, M.; Nikou, C.; Kakadiaris, I.A. A review of human activity recognition methods. Front. Robot. AI 2015, 2, 28. [Google Scholar] [CrossRef]
  2. Wang, P.; Li, W.; Ogunbona, P.; Wan, J.; Escalera, S. RGB-D-based human motion recognition with deep learning: A survey. Comput. Vis. Image Underst. 2018, 171, 118–139. [Google Scholar] [CrossRef]
  3. Pareek, P.; Thakkar, A. A survey on video-based human action recognition: Recent updates, datasets, challenges, and applications. Artif. Intell. Rev. 2021, 54, 2259–2322. [Google Scholar] [CrossRef]
  4. Debes, C.; Merentitis, A.; Sukhanov, S.; Niessen, M.; Frangiadakis, N.; Bauer, A. Monitoring activities of daily living in smart homes: Understanding human behavior. IEEE Signal Process. Mag. 2016, 33, 81–94. [Google Scholar] [CrossRef]
  5. Keogh, A.; Dorn, J.F.; Walsh, L.; Calvo, F.; Caulfield, B. Comparing the usability and acceptability of wearable sensors among older irish adults in a real-world context: Observational study. JMIR mHealth uHealth 2020, 8, e15704. [Google Scholar] [CrossRef] [PubMed]
  6. Majumder, S.; Mondal, T.; Deen, M.J. Wearable sensors for remote health monitoring. Sensors 2017, 17, 130. [Google Scholar] [CrossRef] [PubMed]
  7. Papadakis, A.; Mathe, E.; Spyrou, E.; Mylonas, P. A geometric approach for cross-view human action recognition using deep learning. In Proceedings of the 11th International Symposium on Image and Signal Processing and Analysis (ISPA), Dubrovnik, Croatia, 23–25 September 2019; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  8. Liu, C.; Hu, Y.; Li, Y.; Song, S.; Liu, J. Pku-mmd: A large scale benchmark for continuous multi-modal human action understanding. arXiv 2017, arXiv:1703.07475. [Google Scholar]
  9. Shahroudy, A.; Liu, J.; Ng, T.T.; Wang, G. Ntu rgb+ d: A large scale dataset for 3d human activity analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1010–1019. [Google Scholar]
  10. Gu, R.; Wang, G.; Hwang, J.N. Exploring severe occlusion: Multi-person 3d pose estimation with gated convolution. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 8243–8250. [Google Scholar]
  11. Giannakos, I.; Mathe, E.; Spyrou, E.; Mylonas, P. A study on the Effect of Occlusion in Human Activity Recognition. In Proceedings of the 14th PErvasive Technologies Related to Assistive Environments Conference, Corfu, Greece, 29 June–2 July 2021; pp. 473–482. [Google Scholar]
  12. Lawton, M.P.; Brody, E.M. Assessment of older people: Self-maintaining and instrumental activities of daily living. Gerontologist 1969, 9, 179–186. [Google Scholar] [CrossRef] [PubMed]
  13. Du, Y.; Fu, Y.; Wang, L. Skeleton based action recognition with convolutional neural network. In Proceedings of the 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), Kuala Lumpur, Malaysia, 3–6 November 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 579–583. [Google Scholar]
  14. Hou, Y.; Li, Z.; Wang, P.; Li, W. Skeleton optical spectra-based action recognition using convolutional neural networks. IEEE Trans. Circuits Syst. Video Technol. 2016, 28, 807–811. [Google Scholar] [CrossRef]
  15. Ke, Q.; An, S.; Bennamoun, M.; Sohel, F.; Boussaid, F. Skeletonnet: Mining deep part features for 3-d action recognition. IEEE Signal Process. Lett. 2017, 24, 731–735. [Google Scholar] [CrossRef]
  16. Li, C.; Hou, Y.; Wang, P.; Li, W. Joint distance maps based action recognition with convolutional neural networks. IEEE Signal Process. Lett. 2017, 24, 624–628. [Google Scholar] [CrossRef]
  17. Liu, M.; Liu, H.; Chen, C. Enhanced skeleton visualization for view invariant human action recognition. Pattern Recognit. 2017, 68, 346–362. [Google Scholar] [CrossRef]
  18. Wang, P.; Li, W.; Li, C.; Hou, Y. Action recognition based on joint trajectory maps with convolutional neural networks. Knowl.-Based Syst. 2018, 158, 43–53. [Google Scholar] [CrossRef]
  19. Iosifidis, A.; Tefas, A.; Pitas, I. Multi-view human action recognition under occlusion based on fuzzy distances and neural networks. In Proceedings of the 20th European Signal Processing Conference (EUSIPCO), Bucharest, Romania, 27–31 August 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 1129–1133. [Google Scholar]
  20. Papadakis, A.; Mathe, E.; Vernikos, I.; Maniatis, A.; Spyrou, E.; Mylonas, P. Recognizing human actions using 3d skeletal information and CNNs. In Engineering Applications of Neural Networks, Proceedings of the 20th International Conference, EANN 2019, Xersonisos, Crete, Greece, 24–26 May 2019; Springer: Cham, Switzerland, 2019. [Google Scholar]
  21. Angelini, F.; Fu, Z.; Long, Y.; Shao, L.; Naqvi, S.M. 2D pose-based real-time human action recognition with occlusion-handling. IEEE Trans. Multimed. 2019, 22, 1433–1446. [Google Scholar] [CrossRef]
  22. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar]
  23. Gkalelis, N.; Kim, H.; Hilton, A.; Nikolaidis, N.; Pitas, I. The i3dpost multi-view and 3d human action/interaction database. In Proceedings of the 2009 Conference for Visual Media Production, London, UK, 12–13 November 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 159–168. [Google Scholar]
  24. Holte, M.B.; Tran, C.; Trivedi, M.M.; Moeslund, T.B. Human pose estimation and activity recognition from multi-view videos: Comparative explorations of recent developments. IEEE J. Sel. Top. Signal Process. 2012, 6, 538–552. [Google Scholar] [CrossRef]
  25. Schuldt, C.; Laptev, I.; Caputo, B. Recognizing human actions: A local SVM approach. In Proceedings of the 17th International Conference on Pattern Recognition, ICPR 2004, Cambridge, UK, 26 August 2004; IEEE: Piscataway, NJ, USA, 2004; Volume 3, pp. 32–36. [Google Scholar]
  26. Iosifidis, A.; Marami, E.; Tefas, A.; Pitas, I.; Lyroudia, K. The MOBISERV-AIIA eating and drinking multi-view database for vision-based assisted living. J. Inf. Hiding Multimed. Signal Process. 2015, 6, 254–273. [Google Scholar]
  27. Li, Z.; Li, D. Action recognition of construction workers under occlusion. J. Build. Eng. 2022, 45, 103352. [Google Scholar] [CrossRef]
  28. Li, W.; Zhang, Z.; Liu, Z. Action recognition based on a bag of 3d points. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-Workshops, San Francisco, CA, USA, 13–18 June 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 9–14. [Google Scholar]
  29. Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 1325–1339. [Google Scholar] [CrossRef] [PubMed]
  30. Yang, D.; Wang, Y.; Dantcheva, A.; Garattoni, L.; Francesca, G.; Brémond, F. Self-Supervised Video Pose Representation Learning for Occlusion-Robust Action Recognition. In Proceedings of the 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021), Jodhpur, India, 15–18 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–5. [Google Scholar]
  31. Das, S.; Dai, R.; Koperski, M.; Minciullo, L.; Garattoni, L.; Bremond, F.; Francesca, G. Toyota smarthome: Real-world activities of daily living. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 833–842. [Google Scholar]
  32. Wang, J.; Nie, X.; Xia, Y.; Wu, Y.; Zhu, S.C. Cross-view action modeling, learning and recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2649–2656. [Google Scholar]
  33. Zhang, W.; Zhu, M.; Derpanis, K.G. From actemes to action: A strongly-supervised representation for detailed action understanding. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2248–2255. [Google Scholar]
  34. Vernikos, I.; Mathe, E.; Papadakis, A.; Spyrou, E.; Mylonas, P. An image representation of skeletal data for action recognition using convolutional neural networks. In Proceedings of the 12th ACM International Conference on PErvasive Technologies Related to Assistive Environments, Rhodes, Greece, 5–7 June 2009; pp. 325–326. [Google Scholar]
  35. Liu, T.; Sun, J.J.; Zhao, L.; Zhao, J.; Yuan, L.; Wang, Y.; Chen, L.C.; Schroff, F.; Adam, H. View-invariant, occlusion-robust probabilistic embedding for human pose. Int. J. Comput. Vision 2022, 130, 111–135. [Google Scholar] [CrossRef]
  36. Koutrintzes, D.; Spyrou, E.; Mathe, E.; Mylonas, P. A Multimodal Fusion Approach for Human Activity Recognition. Int. J. Neural Syst. 2022, 33, 2350002. [Google Scholar] [CrossRef] [PubMed]
  37. Chollet, F. Keras. Available online: https://github.com/fchollet/keras (accessed on 20 March 2023).
  38. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. TensorFlow: A System for Large-Scale Machine Learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
Figure 1. The 25 skeletal joints extracted by Microsoft Kinect v2. The joints and edges that correspond to the 5 main body parts, namely head-torso, right arm, left arm, right leg, and left leg are colored in blue, red, green, orange, and purple, respectively.
Figure 1. The 25 skeletal joints extracted by Microsoft Kinect v2. The joints and edges that correspond to the 5 main body parts, namely head-torso, right arm, left arm, right leg, and left leg are colored in blue, red, green, orange, and purple, respectively.
Sensors 23 04899 g001
Figure 2. A sequence of an actor performing the activity wearing a hat. Extracted human skeleton 3D joints using the Kinect SDK have been overlaid. Frames have been taken from the NTU dataset [9] and have been trimmed for illustration purposes.
Figure 2. A sequence of an actor performing the activity wearing a hat. Extracted human skeleton 3D joints using the Kinect SDK have been overlaid. Frames have been taken from the NTU dataset [9] and have been trimmed for illustration purposes.
Sensors 23 04899 g002
Figure 3. A sequence of an actor performing the activity salute. Extracted human skeleton 3D joints using the Kinect SDK have been overlaid. Frames have been taken from the PKU-MMD dataset [8] and have been trimmed for illustration purposes.
Figure 3. A sequence of an actor performing the activity salute. Extracted human skeleton 3D joints using the Kinect SDK have been overlaid. Frames have been taken from the PKU-MMD dataset [8] and have been trimmed for illustration purposes.
Sensors 23 04899 g003
Figure 4. Pseudo-coloured images for the PKU-MMD [8] and the NTU [9] datasets. LA, RA, LL, and RL denote the cases of occlusion of left arm, right arm, left leg, and right leg.
Figure 4. Pseudo-coloured images for the PKU-MMD [8] and the NTU [9] datasets. LA, RA, LL, and RL denote the cases of occlusion of left arm, right arm, left leg, and right leg.
Sensors 23 04899 g004
Figure 5. Example skeleton sequences of the activity salute from the PKU-MMD dataset [8]. First row: skeletons include all 25 joints; second row: joints corresponding to the left arm have been discarded.
Figure 5. Example skeleton sequences of the activity salute from the PKU-MMD dataset [8]. First row: skeletons include all 25 joints; second row: joints corresponding to the left arm have been discarded.
Sensors 23 04899 g005
Figure 6. Example skeleton sequences of the activity wearing a hat from the NTU dataset [9].First row: skeletons include all 25 joints; second row: joints corresponding to the left arm have been discarded.
Figure 6. Example skeleton sequences of the activity wearing a hat from the NTU dataset [9].First row: skeletons include all 25 joints; second row: joints corresponding to the left arm have been discarded.
Sensors 23 04899 g006
Figure 7. The Convolutional Neural Network that has been used in this work.
Figure 7. The Convolutional Neural Network that has been used in this work.
Sensors 23 04899 g007
Table 1. Results of the evaluation using the 11 activities of daily living of the PKU-MMD dataset. Setup cases are Cross-Subject (CS), Single-View (SV), and Cross-View (CV). Camera views are Left (L), Right (R), and Middle (M). In the baseline case (B), all body parts are included in both training and testing, while in the “simple” case of occlusion (S), the training set consists of full body parts and the test set consists of occluded body parts, and in the “augmented” case of occlusion (A), the training set consists of full and occluded body parts, while the test set consists of occluded body parts. Moreover, LA, RA, LL, and RL denote the occlusion of left arm, right arm, left leg, and right leg, respectively. F 1 and Acc. denote F 1 score and accuracy, respectively. In all cases, bold numbers indicate the best result between S and A cases.
Table 1. Results of the evaluation using the 11 activities of daily living of the PKU-MMD dataset. Setup cases are Cross-Subject (CS), Single-View (SV), and Cross-View (CV). Camera views are Left (L), Right (R), and Middle (M). In the baseline case (B), all body parts are included in both training and testing, while in the “simple” case of occlusion (S), the training set consists of full body parts and the test set consists of occluded body parts, and in the “augmented” case of occlusion (A), the training set consists of full and occluded body parts, while the test set consists of occluded body parts. Moreover, LA, RA, LL, and RL denote the occlusion of left arm, right arm, left leg, and right leg, respectively. F 1 and Acc. denote F 1 score and accuracy, respectively. In all cases, bold numbers indicate the best result between S and A cases.
SetupTrainTestMetricBLALA + LLLLLA + RALL + RLRARA + RLRL
SASASASASASASASA
CSLRMLRMF 1 0.680.560.710.530.710.620.740.180.590.470.690.530.700.490.690.660.73
Acc.0.700.590.740.550.750.640.770.280.640.520.720.550.730.500.710.670.75
SVLLF 1 0.680.510.760.490.770.700.770.180.680.580.760.400.700.370.720.660.75
Acc.0.690.540.760.540.770.710.780.290.690.600.760.420.700.400.710.660.76
MMF 1 0.770.660.810.620.790.740.810.270.720.630.760.630.800.570.750.740.79
Acc.0.780.670.810.640.790.750.810.370.720.660.760.650.790.590.750.740.79
RRF 1 0.750.540.750.490.710.710.710.160.630.640.700.590.710.590.730.720.73
Acc.0.760.570.760.530.720.720.720.290.660.650.710.620.720.600.740.720.74
CVLMF 1 0.710.590.780.580.770.690.790.220.700.600.760.610.750.600.740.700.79
Acc0.0.720.620.780.600.780.700.790.320.700.630.770.640.750.620.740.710.79
LRF 1 0.700.500.640.460.630.660.670.150.650.540.670.500.680.470.690.640.69
Acc.0.700.520.650.480.640.660.680.240.660.580.670.530.690.510.690.650.70
MLF 1 0.710.580.770.560.770.680.760.190.670.610.710.480.710.440.690.700.74
Acc.0.720.620.780.600.770.690.760.320.690.630.720.500.710.470.680.700.74
MRF 1 0.550.350.660.340.630.520.690.150.510.400.640.420.700.380.700.500.70
Acc.0.560.410.670.390.640.530.700.260.560.430.660.440.710.400.720.510.71
RLF 1 0.600.530.620.480.600.550.590.180.480.440.560.490.570.480.550.580.57
Acc.0.590.540.640.490.620.540.590.290.530.470.560.500.570.490.550.580.57
RMF 1 0.730.600.700.520.690.650.740.250.610.570.720.600.710.570.700.720.74
Acc.0.730.610.710.540.700.660.740.340.630.600.720.600.720.570.710.720.74
LRMF 1 0.870.700.880.670.870.840.900.230.800.680.880.600.880.560.880.830.89
Acc.0.870.710.880.680.860.850.900.360.790.710.880.640.880.590.880.830.89
LMRF 1 0.790.670.650.650.620.760.680.210.520.640.640.560.660.530.670.770.70
Acc.0.790.680.660.660.630.760.690.340.540.660.650.590.660.550.670.770.70
RMLF 1 0.740.680.790.690.790.740.780.220.680.620.710.550.730.540.710.720.76
Acc.0.750.690.800.700.800.740.770.340.690.630.710.570.730.560.700.720.76
Table 2. Results of the evaluation using the 12 medical conditions of the NTU-RGB+d dataset. Setup cases are Cross-Subject (CS), Single-View (SV), and Cross-View (CV). Camera views are Left (L), Right (R), and Middle (M). In the baseline case (B), all body parts are included in both training and testing, while in the “simple” case of occlusion (S) the training set consists of full body parts and the test set consists of occluded body parts, and in the “augmented” case of occlusion (A), the training set consists of full and occluded body parts, while the test set consists of occluded body parts. LA, RA, LL, and RL denote the occlusion of left arm, right arm, left leg, and right leg, respectively. F 1 and Acc. denote F 1 score and accuracy, respectively. In all cases, bold numbers indicate the best result between S and A cases.
Table 2. Results of the evaluation using the 12 medical conditions of the NTU-RGB+d dataset. Setup cases are Cross-Subject (CS), Single-View (SV), and Cross-View (CV). Camera views are Left (L), Right (R), and Middle (M). In the baseline case (B), all body parts are included in both training and testing, while in the “simple” case of occlusion (S) the training set consists of full body parts and the test set consists of occluded body parts, and in the “augmented” case of occlusion (A), the training set consists of full and occluded body parts, while the test set consists of occluded body parts. LA, RA, LL, and RL denote the occlusion of left arm, right arm, left leg, and right leg, respectively. F 1 and Acc. denote F 1 score and accuracy, respectively. In all cases, bold numbers indicate the best result between S and A cases.
SetupTrainTestMetricBLALA + LLLLLA + RALL + RLRARA + RLRL
SASASASASASASASA
CSLRMLRMF 1 0.610.300.550.320.540.590.640.180.440.510.640.430.610.430.620.580.66
Acc.0.610.320.540.330.540.620.640.220.420.580.650.440.620.440.620.580.66
SVLLF 1 0.600.350.540.370.540.580.640.170.410.460.630.450.610.430.620.570.66
Acc.0.610.380.550.390.540.590.640.240.440.480.640.470.620.450.620.570.66
MMF 1 0.490.280.500.300.490.490.590.120.420.360.610.380.590.370.600.460.61
Acc.0.500.300.500.320.490.490.600.190.430.380.610.410.590.380.600.460.62
RRF 1 0.530.290.480.300.480.510.590.110.370.440.620.390.540.370.550.520.61
Acc.0.540.310.480.320.480.520.590.190.370.460.620.420.540.400.550.520.61
CVLMF 1 0.500.250.500.260.510.430.630.140.380.360.630.390.620.360.640.470.64
Acc.0.500.280.500.270.510.430.630.190.400.390.630.390.620.370.640.460.64
LRF 1 0.500.330.500.340.510.490.580.180.370.420.580.380.530.370.540.480.58
Acc.0.500.350.500.350.510.490.580.220.370.430.580.390.530.380.550.470.59
MLF 1 0.590.310.510.330.510.550.640.130.360.470.630.450.580.440.580.550.64
Acc.0.590.330.510.340.510.550.640.190.390.490.630.470.590.440.580.550.65
MRF 1 0.610.370.420.360.450.570.530.120.300.490.540.420.490.400.500.560.53
Acc.0.610.380.430.380.450.570.540.180.310.510.540.440.490.410.500.560.54
RLF 1 0.590.280.510.310.510.570.610.110.390.490.610.430.560.420.580.550.62
Acc.0.590.310.520.320.520.570.620.180.410.500.610.440.560.430.580.550.63
RMF 1 0.550.340.390.340.390.520.470.170.290.460.470.430.460.410.490.490.49
Acc.0.550.350.400.340.390.520.470.210.310.490.470.420.460.420.480.510.49
LRMF 1 0.570.320.510.310.520.530.610.170.440.440.640.440.620.440.650.550.64
Acc.0.580.330.520.310.520.530.620.220.450.460.640.460.630.460.650.560.65
LMRF 1 0.600.280.510.290.520.540.610.140.370.470.610.440.550.430.560.560.60
Acc.0.600.310.520.300.530.540.610.180.380.500.610.450.550.430.560.560.61
RMLF 1 0.650.380.580.390.590.620.720.140.440.520.720.480.680.460.690.620.72
Acc.0.650.390.580.390.590.620.730.210.450.550.720.500.690.480.690.620.73
Table 3. Results of the evaluation using the PKU-MMD dataset. Setup cases are Cross-Subject (CS), Single-View (SV), and Cross-View (CV). Camera views are Left (L), Right (R), and Middle (M). In the baseline case (B), all body parts are included in both training and testing, while in the “simple” case of occlusion (S) the training set consists of full body parts and the test set consists of occluded body parts, and in the “augmented” case of occlusion (A), the training set consists of full and occluded body parts, while the test set consists of occluded body parts. LA, RA, LL, and RL denote the occlusion of left arm, right arm, left leg, and right leg, respectively. F 1 and Acc. denote F 1 score and accuracy, respectively. In all cases, bold numbers indicate the best result between S and A cases.
Table 3. Results of the evaluation using the PKU-MMD dataset. Setup cases are Cross-Subject (CS), Single-View (SV), and Cross-View (CV). Camera views are Left (L), Right (R), and Middle (M). In the baseline case (B), all body parts are included in both training and testing, while in the “simple” case of occlusion (S) the training set consists of full body parts and the test set consists of occluded body parts, and in the “augmented” case of occlusion (A), the training set consists of full and occluded body parts, while the test set consists of occluded body parts. LA, RA, LL, and RL denote the occlusion of left arm, right arm, left leg, and right leg, respectively. F 1 and Acc. denote F 1 score and accuracy, respectively. In all cases, bold numbers indicate the best result between S and A cases.
SetupTrainTestMetricBLALA + LLLLLA + RALL + RLRARA + RLRL
SASASASASASASASA
CSLRMLRMF 1 0.690.290.510.250.480.650.640.100.340.600.620.390.590.360.580.650.65
Acc.0.700.330.530.290.510.660.660.170.370.620.650.450.610.420.600.660.67
SVLLF 1 0.640.290.520.27490.580.650.090.370.500.630.300.600.280.600.610.65
Acc.0.650.330.530.300.500.590.660.160.390.520.640.360.600.350.600.610.66
MMF 1 0.630.260.540.210.510.590.670.09380.500.670.360.650.330.650.620.68
Acc.0.640.290.550.260.520.600.680.150.410.530.680.410.650.390.650.620.69
RRF 1 0.630.290.530.260.530.600.650.090.370.560.620.380.640.330.630.610.65
Acc.0.630.320.540.300.540.600.650.150.390.570.630.420.640.380.630.620.65
CVLMF 1 0.610.230.500.220.470.550.640.100.340.490.630.340.600.320.600.590.65
Acc.0.620.270.510.260.480.560.650.150.370.530.640.390.610.370.610.590.66
LRF 1 0.580.230.340.210.330.530.440.080.270.450.440.360.420.320.420.560.46
Acc.0.590.280.370.250.350.540.460.140.300.490.460.400.440.360.450.570.48
MLF 1 0.610.280.530.260.520.540.630.080.340.490.620.340.580.330.580.580.64
Acc.0.620.310.540.280.530.560.640.140.370.510.630.390.590.380.580.600.65
MRF 1 0.500.200.470.180.460.450.620.090.340.390.620.290.610.240.610.470.63
Acc.0.510.240.490.220.470.470.630.140.360.410.620.320.610.270.610.470.63
RLF 1 0.470.180.380.160.370.410.470.080.250.370.440.260.440.260.430.440.48
Acc.0.480.220.400.190.390.430.490.160.280.400.470.320.460.330.450.460.50
RMF 1 0.600.250.480.220.470.540.620.080.330.460.610.380.580.330.570.570.64
Acc.0.610.300.500.240.480.540.630.160.350.480.620.420.580.370.580.570.64
LRMF 1 0.700.310.580.260.560.650.740.090.400.600.730.420.690.380.690.680.75
Acc.0.700.340.580.300.560.670.740.150.410.630.730.460.690.430.680.680.75
LMRF 1 0.650.260.500.250.490.610.630.110.360.560.620.360.610.350.620.600.64
Acc.0.660.300.520.290.500.620.640.170.390.58640.420.620.410.620.610.65
RMLF 1 0.650.270.500.260.490.600.620.090.340.550.600.360.560.350.550.610.62
Acc.0.650.300.510.300.490.610.630.150.360.560.600.400.570.400.560.620.63
Table 4. Results of the evaluation using the NTU-RGB+D dataset. Setup cases are Cross-Subject (CS), Single-View (SV), and Cross-View (CV). Camera views are Left (L), Right (R), and Middle (M). In the baseline case (B), all body parts are included in both training and testing, while in the “simple” case of occlusion (S) the training set consists of full body parts and the test set consists of occluded body parts, and in the “augmented” case of occlusion (A), the training set consists of full and occluded body parts, while the test set consists of occluded body parts. LA, RA, LL, and RL denote the occlusion of left arm, right arm, left leg, and right leg, respectively. F 1 and Acc. denote F 1 score and accuracy, respectively. In all cases, bold numbers indicate the best result between S and A cases.
Table 4. Results of the evaluation using the NTU-RGB+D dataset. Setup cases are Cross-Subject (CS), Single-View (SV), and Cross-View (CV). Camera views are Left (L), Right (R), and Middle (M). In the baseline case (B), all body parts are included in both training and testing, while in the “simple” case of occlusion (S) the training set consists of full body parts and the test set consists of occluded body parts, and in the “augmented” case of occlusion (A), the training set consists of full and occluded body parts, while the test set consists of occluded body parts. LA, RA, LL, and RL denote the occlusion of left arm, right arm, left leg, and right leg, respectively. F 1 and Acc. denote F 1 score and accuracy, respectively. In all cases, bold numbers indicate the best result between S and A cases.
SetupTrainTestMetricBLALA + LLLLLA + RALL + RLRARA + RLRL
SASASASASASASASA
CSLRMLRMF 1 0.490.250.370.220.360.460.450.090.250.390.450.310.440.290.440.470.47
Acc.0.490.260.380.230.370.470.470.120.270.410.470.320.450.310.450.470.48
SVLLF 1 0.600.250.390.200.380.500.490.080.280.370.500.340.460.290.480.530.51
Acc.0.600.260.400.210.390.510.500.110.290.410.500.360.460.320.480.550.51
MMF 1 0.570.270.360.240.350.500.460.090.240.420.460.360.440.310.430.510.48
Acc.0.580.280.370.250.360.510.470.110.260.450.470.380.440.320.440.520.49
RRF 1 0.580.270.350.220.350.510.440.080.250.400.460.320.380.270.400.520.47
Acc.0.590.290.360.240.350.520.450.100.270.430.460.340.380.300.400.530.47
CVLMF 1 0.380.140.350.110.340.350.450.040.240.280.460.230.430.200.440.350.47
Acc.0.370.160.360.130.350.350.460.060.260.300.460.230.430.200.440.340.47
LRF 1 0.410.240.350.200.350.370.430.090.230.310.420.260.390.250.390.390.44
Acc.0.410.240.350.200.350.370.440.120.230.330.430.290.390.280.390.400.44
MLF 1 0.460.230.390.200.380.440.490.080.250.350.490.310.450.290.450.450.50
Acc.0.460.240.390.210.380.440.500.100.260.370.490.320.450.320.450.450.51
MRF 1 0.460.260.330.250.320.430.400.090.220.370.390.300.370.280.370.450.41
Acc.0.460.280.330.260.320.440.400.110.230.390.400.320.370.300.370.460.41
RLF 1 0.440.230.330.220.330.410.440.080.210.350.430.270.390.250.390.420.45
Acc.0.430.240.340.230.330.410.440.100.220.370.440.280.380.270.390.420.46
RMF 1 0.500.230.310.200.300.470.380.090.230.420.390.320.380.280.390.470.40
Acc.0.500.240.320.210.310.480.390.100.250.430.390.330.390.300.390.470.41
LRMF 1 0.420.220.370.210.360.400.460.070.260.350.460.270.440.260.440.410.48
Acc.0.430.230.380.220.370.420.470.100.280.370.470.300.440.280.450.430.49
LMRF 1 0.460.240.340.220.330.430.420.090.220.380.410.270.370.260.370.440.42
Acc.0.450.250.350.230.340.430.430.120.240.400.420.280.380.270.380.440.43
RMLF 1 0.520.290.370.240.360.480.480.070.250.400.480.320.450.280.460.480.51
Acc.0.520.290.380.250.360.490.490.100.260.430.480.340.450.310.460.500.51
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Vernikos, I.; Spyropoulos, T.; Spyrou, E.; Mylonas, P. Human Activity Recognition in the Presence of Occlusion. Sensors 2023, 23, 4899. https://doi.org/10.3390/s23104899

AMA Style

Vernikos I, Spyropoulos T, Spyrou E, Mylonas P. Human Activity Recognition in the Presence of Occlusion. Sensors. 2023; 23(10):4899. https://doi.org/10.3390/s23104899

Chicago/Turabian Style

Vernikos, Ioannis, Theodoros Spyropoulos, Evaggelos Spyrou, and Phivos Mylonas. 2023. "Human Activity Recognition in the Presence of Occlusion" Sensors 23, no. 10: 4899. https://doi.org/10.3390/s23104899

APA Style

Vernikos, I., Spyropoulos, T., Spyrou, E., & Mylonas, P. (2023). Human Activity Recognition in the Presence of Occlusion. Sensors, 23(10), 4899. https://doi.org/10.3390/s23104899

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop