PSUDO-CODE-Paper 107-A Novel Hybrid Deep Neural Network Classifier
PSUDO-CODE-Paper 107-A Novel Hybrid Deep Neural Network Classifier
PSUDO-CODE-Paper 107-A Novel Hybrid Deep Neural Network Classifier
net/publication/381836178
A Novel Hybrid Deep Neural Network Classifier for EEG Emotional Brain Signals
Article in International Journal of Advanced Computer Science and Applications · June 2024
DOI: 10.14569/IJACSA.2024.01506107
CITATIONS READS
0 3
3 authors, including:
SEE PROFILE
All content following this page was uploaded by Abdelrahman Tarek Elgohr on 30 June 2024.
Abstract—The field of brain computer interface (BCI) is one of BCI technology can be used to control assistive devices such as
the most exciting areas in the field of scientific research, as it can wheelchairs, prostheses and communication systems, as well as
overlap with all fields that need intelligent control, especially the to monitor brain activity and diagnose neurological diseases.
field of the medical industry. In order to deal with the brain and Moreover, BCI technology can be used to provide a more
its different signals, there are many ways to collect a dataset of natural form of human-computer interaction, allowing users to
brain signals, the most important of which is the collection of control computers with thoughts [1]. BCI technology can be
signals using the non-invasive EEG method. This group of data divided into two main categories as shown in Fig.1: invasive
that has been collected must be classified, and the features and noninvasive. Invasive BCI requires the insertion of
affecting changes in it must be selected to become useful for use in
electrodes into the brain in order to capture brain signals, which
different control capabilities. Due to the need for some fields used
in BCI to have high accuracy and speed in order to comply with
is a risky and complicated process. On the other hand,
the environment's motion sequences, this paper explores the noninvasive BCI relies on measuring signals from the scalp or
classification of brain signals for their usage as control signals in other parts of the body to detect brain activities. Noninvasive
Brain Computer Interface research, with the aim of integrating BCI is more commonly used and includes
them into different control systems. The objective of the study is electroencephalography (EEG), magnetoencephalography
to investigate the EEG brain signal classification using different (MEG), and functional near-infrared spectroscopy (fNIRS).
techniques such as Long Short-Term Memory (LSTM), EEG is the most widely used BCI technique and is based on
Convolutional Neural Networks (CNN), as well as the machine electrical signals generated by the brain [2].
learning approach represented by the Support Vector Machine
(SVM). We also present a novel hybrid classification technique
EEG signals are a type of electrical activity that can be
called CNN-LSTM which combines CNNs with LSTM networks. measured from the brain. They are used in a variety of
This proposed model processes the input data through one or more engineering fields, including medical, robotics, and computer
of the CNN’s convolutional layers to identify spatial patterns and engineering. In medical engineering, EEG signals are used to
the output is fed into the LSTM layers to capture temporal diagnose and monitor neurological conditions. EEGs can be
dependencies and sequential patterns. This proposed combination used to detect seizures, diagnose sleep disorders, and monitor
uses CNNs’ spatial feature extraction and LSTMs’ temporal brain activity during surgery. EEGs can also be used to measure
modelling to achieve high efficacy across domains. A test was done brain activity during cognitive tasks, such as memory tests. This
to determine the most effective approach for classifying emotional can help doctors better understand how the brain works and
brain signals that indicate the user's emotional state. The dataset how to treat neurological conditions [3]. In robotics
used in this research was generated from a widely available MUSE engineering, EEG signals are used to control robotic devices.
EEG headgear with four dry extra-cranial electrodes. The By measuring the electrical activity of the brain, robots can be
comparison came in favor of the proposed hybrid model (CNN- programmed to respond to certain commands. This can be used
LSTM) in first place with an accuracy of 98.5% and a step speed to create robots that can interact with humans in a more natural
of 244 milliseconds/step; the CNN model came in the second place way. For example, robots can be programmed to respond to
with an accuracy of 98.03% and a step speed of 58 facial expressions or voice commands [4]. In computer
milliseconds/step; and in the third place, the LSTM model engineering, EEG signals are used to create brain-computer
recorded an accuracy of 97.35% and a step speed of 2 sec/step;
interfaces. These allow users to control computers with their
finally, in last place, SVM came with 87.5% accuracy and 39
milliseconds/step running speed.
thoughts. This technology is still in its early stages, but it has
the potential to revolutionize the way we interact with
Keywords—BCI; EEG; Brain Signals Classification; SVM; computers [5].
LSTM, CNN; CNN-LSTM The most common EEG signal classification methods as
I. INTRODUCTION shown in Fig. 2 are supervised learning algorithms, such as
Support Vector Machines (SVMs), Artificial Neural Networks
Brain Computer Interface (BCI) is a technology that enables (ANNs), and decision trees. These algorithms are used to
direct communication between the brain and an external device identify patterns in EEG signals that can be used to diagnose
using signals generated from the brain. It has been proposed as and monitor neurological conditions [6]. For example, SVMs
a potential therapeutic treatment for various neurological can be used to classify EEG signals into different categories,
disorders and a tool for efficient human-computer interaction. such as normal or abnormal, or to detect changes in EEG signals
1045 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 6, 2024
over time. ANNs, which include Convolutional Neural basis function kernel that outperforms prior methods on the
Network (CNN), can be used to identify patterns in EEG signals same dataset achieves an accuracy of 95.70%. This article
that can be used to diagnose and monitor neurological presents a nonparametric method for decomposing EEG signals
conditions. Decision trees can be used to identify patterns in to improve efficiency. This approach can enhance the progress
EEG signals that can be used to diagnose and monitor of BCI system development by utilizing machine learning
neurological conditions [7]. techniques [11].
Chowdary MK, et al., (2022) aim to classify emotions from
electroencephalogram signals by utilizing different recurrent
neural network structures. Three architectures employed in this
study for emotion recognition using EEG signals are RNN
(recurrent neural network), LSTM (long short-term memory
network), and GRU (gated recurrent unit). Experimental data
confirmed the efficiency of these networks in terms of
performance measures. The study utilized the EEG Brain Wave
Dataset: Feeling Emotions and obtained an average accuracy of
95% for RNN, 97% for LSTM, and 96% for GRU in detecting
emotions [12].
Fig. 1. Brain computer interface technology [1]. EEG capture and emotion categorization in a simulated
driving environment is suggested by Chen J. et al. (2024) to
study panic emotion and accident-avoidance skills. The
program models obstacle avoidance at different risk levels
using vehicle speed. The system models the brain's
physiological structure for data processing using graph neural
networks (GNN) with functional connection and attention
mechanisms. Various research compared entropy and power
properties. The top single-label F1 score was 76.7%, and the
three-class classification was 75.26 % accurate. Binary
classification had 91.5% accuracy and the highest F1 score for
Fig. 2. Dataset classification techniques [8]. a single label was 91.86%. Deep learning algorithms can
accurately mimic hazardous events, record the driver's EEG
A. Related Work data, and quickly track emotional states, according to
Z. -T. Liu et al. (2019) tested a proposed approach on DEAP experiments [13].
dataset, classifying Valance and Arousal emotional states using This research investigates classifying brain signals for use
K-nearest neighbor and support vector machine. The as control signals in Brain-Computer Interface (BCI) systems
experiments compare temporal windows of different lengths designed for various robotic applications. The aim is to
and three EEG signal rhythms. The results show that the EEG compare four methods for multi-class classification: Long
signal with one temporal window has the highest recognition Short-Term Memory (LSTM) and Convolutional Neural
accuracy of 86.46%. A multimodal emotional communication- Networks (CNN) from deep learning, a proposed hybrid CNN-
based humans-robots interaction system would use the LSTM approach, and Support Vector Machine (SVM) from
suggested approach for real-time emotion identification [9]. machine learning. Ultimately, this research seeks to determine
T. Song et al. (2020) to recognize emotions in multichannel the most effective method for classifying emotional brain
EEG data used a dynamical graph convolutional neural network signals that reflect the user's emotional state.
(DGCNN). Our EEG emotion recognition method uses a graph The rest of this paper is organized as follows: Section II
to describe multichannel EEG data and classify emotions using demonstrates the main concepts for signals classification
this model. EEG emotion recognition is improved by learning overview; Section III presents the classification models;
new features from the adjacency matrix. Emotion EEG datasets Section IV describes the dataset; Sections V and VI elaborate
SEED and DREAMER were extensively studied. The proposed the classification results and a discussion of the results
recognition method is more accurate than current methods. On generated from the tests; Section VII mentions the applications
SEED, it averaged 90.4% in subject-dependent experiments that can benefit from this research topic; Section VIII concludes
and 79.95% in subject-independent cross-validation [10]. the paper and presents the future work.
In 2020, S. K. Khare and colleagues introduced an adaptive II. CLASSIFICATION OVERVIEW
tunable Q wavelet transform for selecting tuning parameters
automatically. Grey wolf optimization identifies the best tuning Dataset classification is a process of organizing data into
parameters. GWO tuning parameters divide EEG signals into categories based on certain characteristics. It is a way of
sub bands. Time-domain properties of SB are inputted into a organizing data into meaningful groups so that it can be more
multiclass least-squares support vector machine. Evaluating the easily analyzed and understood. Dataset classification is used
classification accuracy of four main emotions - happiness, fear, in a variety of fields, including data mining, machine learning,
sadness, and relaxation - compared to current methods. A radial and artificial intelligence.
1046 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 6, 2024
1047 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 6, 2024
B. Long Short Term Memory (LSTM) the optimizer, and the metrics to be used for evaluating the
The Long Short-Term Memory (LSTM) classifier is a model. Finally, the model can be trained. This involves
powerful deep learning algorithm that can be used to classify specifying the number of epochs, the batch size, and the
data. It is a type of recurrent neural network (RNN) that is validation split. It is important to monitor the training process
capable of learning long-term dependencies in data. The LSTM to ensure that the model is not overfitting or underfitting the
classifier is a powerful tool for predicting and classifying data, data [22].
and it has been used in a variety of applications, such as natural C. Convolutional Neural Network (CNN)
language processing, speech recognition, time series
forecasting, and classifying sequences of data, such as text, A convolutional neural network (CNN) is a type of artificial
audio, and video. As shown in Fig. 4, the LSTM classifier is neural network used in deep learning that is specifically
composed of a series of memory cells, each of which contains designed to process data that has a grid-like structure, such as
a set of weights and biases. The weights and biases are adjusted tabular and images datasets. CNNs are composed of multiple
during the training process to learn the patterns in the data [21]. layers of neurons that each perform a specific task as shown in
The memory cells are connected in a chain, and each cell is Fig. 5. The first layer of neurons is responsible for detecting
connected to the next cell in the chain. This allows the network edges and other basic features in the input image. The second
to remember information from previous cells and use it to make layer of neurons is responsible for detecting more complex
predictions [22]. The LSTM classifier is trained using a features, such as shapes and patterns. The third layer of neurons
supervised learning algorithm. During the training process, the is responsible for recognizing objects in the image. The fourth
network is presented with a set of input data and the desired layer of neurons is responsible for recognizing more complex
output. The network then adjusts the weights and biases of the objects, such as faces or animals [24].
memory cells to learn the patterns in the data. Once the training CNNs are particularly useful in robotics because they are
is complete, the network can be used to make predictions on able to process large amounts of data quickly and accurately.
new data [23]. For example, a CNN can be used to identify objects in an image
or video feed. It can also be used to analyze cognitive data
represented in a database that enables robots to understand the
surrounding environment and also understand the commands
stored within it and classify them according to the event the
robot is exposed to. This is useful for robots that need to identify
objects in their environment in order to navigate or interact with
them. CNNs can also be used to classify objects in a scene,
which is useful for robots that need to recognize and interact
with objects in their environment [25].
1048 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 6, 2024
determines the complexity of the model and how well it spatial and temporal data. This powerful combination leverages
will perform. CNNs' ability to extract spatial features and LSTMs' strength in
temporal modeling, leading to high effectiveness across various
Training: Once the model architecture is decided, the domains.
next step is to train the model. This involves feeding the
data into the model and adjusting the weights and biases This hybrid model finds applications in tasks involving
of the neurons in order to minimize the error. This step complex sequential data. It utilizes CNNs for spatial analysis
is important as it ensures that the model is able to and LSTMs for understanding temporal sequences. The CNN-
accurately predict the output given the input. LSTM model processes input data through one or more
convolutional layers to identify spatial patterns. The output
Evaluation: After the model is trained, the next step is to from these layers then feeds into LSTM layers to capture
evaluate the model. This involves testing the model on temporal dependencies and sequential patterns. Finally, dense
unseen data and measuring its performance. This step is layers are often used for classification or regression tasks.
important as it allows us to determine how well the Algorithm 3 lists the whole process of proposed model.
model is performing and if it needs to be improved.
The model's strength lies in the specialized functions of its
Deployment: The final step is to deploy the model. This layers. CNNs excel at extracting features from spatial data,
involves making the model available to users so that while LSTMs represent complex temporal connections. This
they can use it to make predictions. This step is combination allows the model to learn both spatial and
important as it allows the model to be used in real-world temporal characteristics simultaneously, enabling a
applications. comprehensive interpretation of the data. However, achieving
These are the basic steps for building a CNN model. optimal performance requires careful hyperparameter tuning
However, there are many other steps that can be taken to for both CNN and LSTM components, and ensuring
improve the model, such as hyperparameter tuning, compatibility between the input data shape and both layer types.
regularization, and data augmentation. With the right guidance A small code example using the Keras library shows how to
and experience, building a CNN model can be a relatively sequentially add CNN and LSTM layers for spatiotemporal
straightforward process. modelling [27].
1049 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 6, 2024
different groups of people or different types of data, store identify patterns and relationships in data. Classification results
information about people, places, events, and other topics, and are used to make predictions about future data points, and can
create visualizations like charts, graphs, and maps. They can be used to make decisions about how to best utilize resources.
even be used to generate reports and presentations. Classification results are typically presented in the form of a
confusion matrix, which is a table that shows the number of true
Datasets can be used to analyze trends, identify patterns, positives, false positives, true negatives, and false negatives
and make predictions. They can also be used to compare [30]. The confusion matrix is used to evaluate the accuracy of
different groups of people or to compare different types of data. the classification model, and can be used to identify areas where
They can also be used to store information about people, places, the model is performing well or poorly. Classification results
events, and other topics. Datasets can be used to create can also be used to identify important features in the data that
visualizations, such as charts, graphs, and maps. They can also are driving the model’s predictions. This can be done by
be used to create reports and presentations. They can also be looking at the feature importance scores, which are calculated
used to store information about people, places, events, and other by the model and indicate how important each feature is in
topics [28]. making the prediction. This can be used to identify which
The dataset used in this research is a mental emotional features are most important for making accurate predictions,
sentiment dataset that was collected by other researchers using and can be used to inform decisions about which features to
a commercial MUSE EEG headband which was used with a focus on when building a model [31], [32].
resolution of four (TP9, AF7, AF8, TP10) electrodes. To collect
the data, researchers used a widely available MUSE EEG
headgear with four dry extra-cranial electrodes. As can be seen
in Fig. 6, micro voltage readings are taken from electrodes TP9,
AF7, AF8, and TP10. Two individuals (1 male, 1 female, aged
20-22) each provided 60 seconds of data for each of the 6 film
segments, for a total of 12 minutes (720 seconds) of brain
activity data (6 minutes for each emotional state). A total of 36
minutes of EEG data was obtained from each individual,
including six minutes of "neutral brainwave" data. The brain's
waves were captured at a variable frequency and then
resampled to 150Hz, yielding a collection of 324,000 data Fig. 7. Classes appearance analysis.
points. The positive and negative valence descriptors were
evaluated instead of the emotions themselves to determine Finally, classification results can be used to compare
which activities were most likely to elicit. For a third category, different models and determine which one is the best for a given
representing the subject's baseline emotional state, neutral data task. This can be done by looking at the accuracy scores of each
were also obtained before any data on emotions were gathered model, as well as other metrics such as precision, recall, and F1
(to prevent contamination from the latter). We only gathered score. Comparing the results of different models can help
data from each participant for three minutes every day to identify which model is best suited for a given task, and can
minimize the influence of a baseline emotional state [29]. help inform decisions about which model to use [32].
To classify the dataset, it must first understand its details,
the resultant classes from each row of input, and the number of
instances of each class over the whole dataset. As the
information from the dataset were analyzed, it was discovered
that there are three separate classes as a consequence of all the
input rows, which are positive, negative, and neutral, as they
represent an indicator of the subject's emotional state. After
each class was counted, it was discovered that the positive case
occurred 708 times, the negative case appeared 708 times, and
the neutral case appeared 716 times as shown in Fig. 7. These
statistics reveal the dataset's balance, from which the difference
in findings may be precisely calculated. As the last stage in
studying the dataset, a sample may be obtained for each class
using a variety of inputs, as illustrated in Fig. 8.
A. SVM Results
When implementing the SVM algorithm, the tensorflow
library was used for deep learning in Python, on the Kaggle
Fig. 6. Position of used EEG electrodes on human skull [29]. coding website. And by classifying the studied dataset, and
specifying each of the training data percentage as 50%, the
V. CLASSIFICATION RESULTS testing data percentage as 25%, the validation data percentage
Classification results are the outcomes of a classification as 25%, and 100 epochs to test the algorithm and conclude the
process, which is a type of data mining technique used to best classification result.
1050 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 6, 2024
1051 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 6, 2024
conducting 50 epochs for model evaluation, and choosing the VI. RESULT AND DISCUSSION
optimal classification outcome. This model was constructed When comparing the results of different models to classify
with an input layer, three CNN layers utilizing the ReLU the studied dataset (SVM, LSTM, CNN, and CNN-LSTM),
activation function and progressively increasing filter sizes more than one aspect can be relied on for comparison, the first
starting from 64 bits. It also includes LSTM layers and a final and most important of which is the accuracy of the
output layer. Since the dataset consists of three distinct classes, classification when implementing the model, the second is the
the Softmax Activation Function is employed. confusion matrix, which is related in one way or another to the
first factor in the comparison, and the third factor that was taken
into account when comparing is speed of implementation of the
model, as measured by the speed of the test steps and also the
speed of early stopping when testing the model. Table I
compiles these features for all models utilized in the paper.
Comparison factors
Model
Accuracy Test speed Early stop
SVM 87.5 % 39 ms/step 15 epochs
LSTM 97.35 % 2 sec/step 15 epochs
CNN 98.03 % 58 ms/step 38 epochs
CNN-LSTM 98.50 % 244 ms/step No
1052 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 6, 2024
previous results are dependent on the input factors of each VII. APPLICATIONS
model, which were fixed in all cases, so that comparison can be Brain Computer Interfacing (BCI) is a technology, which
made based on the equality of classification characteristics. enables communication between humans and machines through
The CNN-LSTM was the best of them, as it has the highest the direct interpretation of brain activities. This technology has
accuracy, which is the most important factor in the comparison. a wide range of potential applications as shown in Fig. 13 and
In addition, the execution speed (step speed) was not bad has been explored by researchers in fields such as medical
(between the other models), but despite that, it was the most in diagnostics, prosthetics, human-machine interaction, and
the number of epochs that the model needed to infer the best communication aids [36].
classification result (training the model) but this did not
significantly affect the outcome of the total execution time of
the model.
And if we exclude the factor of execution speed, then CNN-
LSTM, CNN and LSTM are very close in the result of
classification accuracy, then these models can be equally
reliable on the classification of the data set. In contrast, if the
accuracy factor is excluded, the SVM model is the fastest in
step speed and the least in the number of epochs required to
train the model equally with the LSTM model, so SVM can be
said that it is the best in the speed factor.
Considering that the CNN-LSTM result is the best model
studied in this paper in terms of accuracy, which is the most
reliable factor in the comparison. In the end, this model can also
be compared with the similar models previously published on a
similar database and with the different models. Through the Fig. 13. BCI applications.
previous literature study during previous years, published
research showed limited accuracy of the techniques used, as K- After classifying the data set, the classification model may
nearest neighbor recorded an accuracy of 86.46% in Z-T. Liu. be used to connect the model's inputs, which are brain signals,
et al.'s 2019 research. T. Song et al. also recorded in their and any application that can be controlled by the dataset's
research published in 2020, an accuracy of 90.4% using distinct classifications. When looking at the field of medical
DGCNN. Also, it was followed by the accuracy of the research industries, it is possible to coordinate between commands to
of S.K. Khare et al. 2020, which reached 95.7% using the control prosthetic limbs through brain signals directly, through
LSSVM, and then the RNN, LSTM, and GRU models obtained three different commands linked to the three groups of the data
an accuracy of 95%, 97%, and 96% respectively in the research set, for patients with paraplegia or total paralysis who are
of Chowdary MK et al. 2022. Finally, Chen J. et al. 2024 are unable to move their natural organs, or move the muscles to
achieved 75.26% multi-classification accuracy by using GNN control the Industrial limb [37]–[39]. Among the applications
model in their published study. On the other hand, expectations is Brain control in industrial robots in smart industries. It is also
were raised for an impressive result using the studied model possible to link the results of classification (one of the classes
(CNN-LSTM), where a classification accuracy was obtained for the data set) and a set of successive commands that include
that achieved a gorgeous mark in prediction. A summary of the a path for a complete industrial process, so that the controller
results of the reviewed researches appears in Table II, where has the ability to control the brain in three separate industrial
each row indicates the research person, the model used in it, and processes [40]. In addition, brain control technology can be
the model’s accuracy result during testing. used to control laboratory robots, and this can be used in
research and scientific projects that allow the formation of
TABLE II. LITERATURE RESULTS SUMMARY COMPARED WITH PROPOSED innovative systems of intelligent control [41].
MODEL
One of the most exciting applications of BCI in gaming and
Comparison entertainment is the ability to control game characters and
Author
Model Accuracy objects with your thoughts. This could allow players to control
Z-T. Liu et al. (2019) [9] K-NN 86.46 % their characters in a more natural and intuitive way, as well as
allowing for more complex interactions with the game world.
T. Song et al. (2020) [10] DGCNN 90.40 %
For example, a player could use their thoughts to control a
S.K. Khare et al. (2020) [11] LSSVM 95.70 % character’s movements, or to manipulate objects in the game
RNN 95.00 % world. This could open up a completely new level of immersion
and interaction with games [42].
Chowdary MK et al. (2022) [12] LSTM 97.00 %
GRU 96.00 % Brain-Computer Interface (BCI) technology might change
how individuals see themselves and their surroundings. BCI
Chen J. et al. (2024) [13] GNN 75.26 % technology raises ethical and legal issues. BCI technology may
Proposed model CNN-LSTM 98.50 % enhance quality of life and give therapeutic advantages, but also
1053 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 6, 2024
1054 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 6, 2024
[17] T. N. T. Thu and V. D. Xuan, “Using support vector machine in FoRex Machine Learning Section A-Research paper Eur,” Chem. Bull, vol. 2023,
predicting,” in 2018 IEEE International Conference on Innovative no. 8, pp. 7648–7656, doi: 10.48047/ecb/2023.12.8.619.
Research and Development (ICIRD), 2018, pp. 1–5. doi: [31] T. Kaur and T. K. Gandhi, “Deep convolutional neural networks with
10.1109/ICIRD.2018.8376303. transfer learning for automated brain image classification,” Mach Vis
[18] K. Harimoorthy and M. T, “Multi-disease prediction model using Appl, vol. 31, no. 3, p. 20, 2020, doi: 10.1007/s00138-020-01069-2.
improved SVM-radial bias technique in healthcare monitoring system,” J [32] F. Hemmatian and M. K. Sohrabi, “A survey on classification techniques
Ambient Intell Humaniz Comput, vol. 12, Feb. 2021, doi: for opinion mining and sentiment analysis,” Artif Intell Rev, vol. 52, no.
10.1007/s12652-019-01652-0. 3, pp. 1495–1545, 2019, doi: 10.1007/s10462-017-9599-6.
[19] M. A. Hearst, S. T. Dumais, E. Osuna, J. Platt, and B. Scholkopf, “Support [33] Y.-W. Chang, C.-J. Hsieh, K.-W. Chang, M. Ringgaard, and C.-J. Lin,
vector machines,” IEEE Intelligent Systems and their Applications, vol. “Training and Testing Low-degree Polynomial Data Mappings via Linear
13, no. 4, pp. 18–28, 1998, doi: 10.1109/5254.708428. SVM,” 2010.
[20] D. Tomar and S. Agarwal, “A comparison on multi-class classification [34] B. Gao and L. Pavel, “On the Properties of the Softmax Function with
methods based on least squares twin support vector machine,” Knowl Application in Game Theory and Reinforcement Learning,” Apr. 2017,
Based Syst, vol. 81, pp. 131–147, 2015, doi: [Online]. Available: http://arxiv.org/abs/1704.00805
https://doi.org/10.1016/j.knosys.2015.02.009.
[35] V. Nair and G. E. Hinton, “Rectified Linear Units Improve Restricted
[21] J. Hernandez, D. López, and N. Vera, “Primary user characterization for Boltzmann Machines,” in Proceedings of the 27th International
cognitive radio wireless networks using long short-term memory,” Int J Conference on International Conference on Machine Learning, in
Distrib Sens Netw, vol. 14, p. 155014771881182, Feb. 2018, doi: ICML’10. Madison, WI, USA: Omnipress, 2010, pp. 807–814.
10.1177/1550147718811828.
[36] S. N. Abdulkader, A. Atia, and M.-S. M. Mostafa, “Brain computer
[22] S. Tortora, S. Ghidoni, C. Chisari, S. Micera, and F. Artoni, “Deep interfacing: Applications and challenges,” Egyptian Informatics Journal,
learning-based BCI for gait decoding from EEG with LSTM recurrent vol. 16, no. 2, pp. 213–230, 2015, doi:
neural network,” J Neural Eng, vol. 17, no. 4, Aug. 2020, doi: https://doi.org/10.1016/j.eij.2015.06.002.
10.1088/1741-2552/ab9842.
[37] J. H. Jeong, K. H. Shim, D. J. Kim, and S. W. Lee, “Brain-Controlled
[23] H. Almutairi, G. M. Hassan, and A. Datta, “Classification of Obstructive Robotic Arm System Based on Multi-Directional CNN-BiLSTM
Sleep Apnoea from single-lead ECG signals using convolutional neural Network Using EEG Signals,” IEEE Transactions on Neural Systems and
and Long Short Term Memory networks,” Biomed Signal Process Rehabilitation Engineering, vol. 28, no. 5, pp. 1226–1238, May 2020, doi:
Control, vol. 69, p. 102906, 2021, doi: 10.1109/TNSRE.2020.2981659.
https://doi.org/10.1016/j.bspc.2021.102906.
[38] M. A. A. Mousa, A. Elgohr, and H. Khater, “Path Planning for a 6 DoF
[24] E. Pranav, S. Kamal, C. Satheesh Chandran, and M. H. Supriya, “Facial Robotic Arm Based on Whale Optimization Algorithm and Genetic
Emotion Recognition Using Deep Convolutional Neural Network,” in Algorithm,” Journal of Engineering Research, vol. 7, no. 5, pp. 160–168,
2020 6th International Conference on Advanced Computing and Nov. 2023, doi: 10.21608/erjeng.2023.237586.1256.
Communication Systems (ICACCS), 2020, pp. 317–320. doi:
10.1109/ICACCS48705.2020.9074302. [39] M. A. A. Mousa, A. T. Elgohr, and H. A. Khater, “Trajectory
Optimization for a 6 DOF Robotic Arm Based on Reachability Time,”
[25] K. Noda, H. Arie, Y. Suga, and T. Ogata, “Multimodal integration Annals of Emerging Technologies in Computing, vol. 8, no. 1, pp. 22–35,
learning of robot behavior using deep neural networks,” Rob Auton Syst, Jan. 2024, doi: 10.33166/AETiC.2024.01.003.
vol. 62, no. 6, pp. 721–736, 2014, doi: 10.1016/j.robot.2014.03.003.
[40] M. A. Elazab, hamouda Abueldahab, A. Elgohr, and M. S. Elhadidy, “A
[26] B. Cao, H. Niu, J. Hao, and G. Wang, “Building EEG-based CAD object Comprehensive Review on Hybridization in Sustainable Desalination
selection intention discrimination model using convolutional neural Systems,” Journal of Engineering Research, vol. 7, no. 5, pp. 89–99, Nov.
network (CNN),” Advanced Engineering Informatics, vol. 52, p. 101548, 2023, doi: 10.21608/erjeng.2023.235480.1238.
2022, doi: https://doi.org/10.1016/j.aei.2022.101548.
[41] B. Zhang, J. Wang, and T. Fuhlbrigge, “A review of the commercial brain-
[27] G. A. Marcoulides, “Discovering Knowledge in Data: an Introduction to computer interface technology from perspective of industrial robotics,” in
Data Mining,” J Am Stat Assoc, vol. 100, no. 472, pp. 1465–1465, Dec. 2010 IEEE International Conference on Automation and Logistics, 2010,
2005, doi: 10.1198/jasa.2005.s61. pp. 379–384. doi: 10.1109/ICAL.2010.5585311.
[28] Daniel T. Larose, Discovering Knowledge in Data : An Introduction to [42] S. Dutta, T. Banerjee, N. D. Roy, B. Chowdhury, and A. Biswas,
Data Mining, Second edition. Wiley-Interscience, Hoboken, N.J., ©2005, “Development of a BCI-based gaming application to enhance cognitive
2017. control in psychiatric disorders,” Innov Syst Softw Eng, vol. 17, no. 2, pp.
[29] J. Bird, A. Ekart, C. Buckingham, and D. Faria, “Mental Emotional 99–107, 2021, doi: 10.1007/s11334-020-00370-7.
Sentiment Classification with an EEG-based Brain-machine Interface,” [43] S. Burwell, M. Sample, and E. Racine, “Ethical aspects of brain computer
Feb. 2019. interfaces: a scoping review,” BMC Med Ethics, vol. 18, no. 1, p. 60,
[30] R. Ahmed Hegazii, E. Abdelhalim, and H. El-Din Mostafa, “A Proposed 2017, doi: 10.1186/s12910-017-0220-y.
Technique for Breast Cancer Prediction and Classification Based on
1055 | P a g e
www.ijacsa.thesai.org
View publication stats