Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (599)

Search Parameters:
Keywords = resampling methods

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 6169 KiB  
Article
Uncertain Particle Filtering: A New Real-Time State Estimation Method for Failure Prognostics
by Jingyu Liang, Yinghua Shao, Waichon Lio, Jie Liu and Rui Kang
Mathematics 2025, 13(5), 48; https://doi.org/10.3390/math13050702 - 21 Feb 2025
Viewed by 164
Abstract
Particle filtering (PF) has become a state-of-the-art method in predicting the future degradation trend of the target equipment based on its current state, with its advantage in sequentially processing the observed data for continual state estimation. The convergence speed is important in PF [...] Read more.
Particle filtering (PF) has become a state-of-the-art method in predicting the future degradation trend of the target equipment based on its current state, with its advantage in sequentially processing the observed data for continual state estimation. The convergence speed is important in PF for real-time state estimation. However, the Bayesian theorem can only converge when sufficient observations are available, which does not always fulfill the requirement in time-varying scenarios with abrupt changes in health state. In this work, based on the newly proposed Uncertainty Theory, Uncertain Particle Filtering (UPF) is derived for the first time. The initialization, prediction, update, and resampling processes are explained in detail in the scope of Uncertainty Theory. The UPF method significantly improves the performance of traditional particle filters by enhancing the speed of convergence in dynamic parameter estimation. Resampling techniques are introduced to mitigate particle phagocytosis, thereby improving computational accuracy and efficiency. Two case studies, addressing the degradation of the capacitor in an enhanced electromagnetic railgun and the degradation of the battery, are conducted to verify the effectiveness of the proposed UPF method. The results show that the UPF method achieves a faster convergence speed compared to traditional approaches. Full article
Show Figures

Figure 1

Figure 1
<p>The uncertainty distribution function of the <math display="inline"><semantics> <mi>δ</mi> </semantics></math> function.</p>
Full article ">Figure 2
<p>An example of the uncertainty distribution function of <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>θ</mi> <mo>^</mo> </mover> <mi>t</mi> </msub> </mrow> </semantics></math>. The <a href="#mathematics-13-00702-f002" class="html-fig">Figure 2</a> demonstrates the process of discretizing a continuous curve using uncertain particles. The red line indicates the resulting discretized distribution, while the other colored lines represent the individual uncertain particles that collectively form the discrete approximation of the original curve.</p>
Full article ">Figure 3
<p>The implementation process of Uncertain Particle Filtering.</p>
Full article ">Figure 4
<p>Schematic diagram of the electromagnetic railgun [<a href="#B31-mathematics-13-00702" class="html-bibr">31</a>].</p>
Full article ">Figure 5
<p>The simulation model of the enhanced electromagnetic railgun.</p>
Full article ">Figure 6
<p>State estimation results by UPF with resampling when <math display="inline"><semantics> <mrow> <msub> <mi>N</mi> <mrow> <mi>e</mi> <mi>f</mi> <mi>f</mi> </mrow> </msub> <mo>&lt;</mo> <mn>0.1</mn> <mi>N</mi> <mo>=</mo> <mn>100</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 7
<p>State estimation results by UPF with resampling under the 95% confidence interval.</p>
Full article ">Figure 8
<p>The degradation trajectory of the capacitor by using comparative methods.</p>
Full article ">Figure 9
<p>The dynamic parameter <math display="inline"><semantics> <mrow> <msub> <mi>p</mi> <mn>1</mn> </msub> </mrow> </semantics></math> estimation by using the UPF method.</p>
Full article ">Figure 10
<p>The dynamic parameter <math display="inline"><semantics> <mrow> <msub> <mi>p</mi> <mn>1</mn> </msub> </mrow> </semantics></math> estimation by using comparative methods.</p>
Full article ">
15 pages, 3085 KiB  
Article
Early Detection of Skin Diseases Across Diverse Skin Tones Using Hybrid Machine Learning and Deep Learning Models
by Akasha Aquil, Faisal Saeed, Souad Baowidan, Abdullah Marish Ali and Nouh Sabri Elmitwally
Information 2025, 16(2), 48; https://doi.org/10.3390/info16020152 - 19 Feb 2025
Viewed by 225
Abstract
Skin diseases in melanin-rich skin often present diagnostic challenges due to the unique characteristics of darker skin tones, which can lead to misdiagnosis or delayed treatment. This disparity impacts millions within diverse communities, highlighting the need for accurate, AI-based diagnostic tools. In this [...] Read more.
Skin diseases in melanin-rich skin often present diagnostic challenges due to the unique characteristics of darker skin tones, which can lead to misdiagnosis or delayed treatment. This disparity impacts millions within diverse communities, highlighting the need for accurate, AI-based diagnostic tools. In this paper, we investigated the performance of three machine learning methods -Support Vector Machines (SVMs), Random Forest (RF), and Decision Trees (DTs)-combined with state-of-the-art (SOTA) deep learning models, EfficientNet, MobileNetV2, and DenseNet121, for predicting skin conditions using dermoscopic images from the HAM10000 dataset. The features were extracted using the deep learning models, with the labels encoded numerically. To address the data imbalance, SMOTE and resampling techniques were applied. Additionally, Principal Component Analysis (PCA) was used for feature reduction, and fine-tuning was performed to optimize the models. The results demonstrated that RF with DenseNet121 achieved a superior accuracy of 98.32%, followed by SVM with MobileNetV2 at 98.08%, and Decision Tree with MobileNetV2 at 85.39%. The proposed methods overcome the SVM with the SOTA EfficientNet model, validating the robustness of the proposed approaches. Evaluation metrics such as accuracy, precision, recall, and F1-score were used to benchmark performance, showcasing the potential of these methods in advancing skin disease diagnostics for diverse populations. Full article
(This article belongs to the Special Issue AI-Based Image Processing and Computer Vision)
Show Figures

Figure 1

Figure 1
<p>Phases of CRISP-DM.</p>
Full article ">Figure 2
<p>Frequency of skin lesion types.</p>
Full article ">Figure 3
<p>Distribution of age.</p>
Full article ">Figure 4
<p>Sex distribution.</p>
Full article ">Figure 5
<p>Distribution of lesion localization.</p>
Full article ">Figure 6
<p>Age distribution across lesion types.</p>
Full article ">Figure 7
<p>Distribution of lesion types by sex.</p>
Full article ">Figure 8
<p>Model accuracy graph for SVM-MobileNetV2.</p>
Full article ">Figure 9
<p>Random Forest DenseNet121 Accuracy Model.</p>
Full article ">Figure 10
<p>Validation Accuracy of Decision Tree.</p>
Full article ">
34 pages, 2988 KiB  
Article
Improving Surgical Site Infection Prediction Using Machine Learning: Addressing Challenges of Highly Imbalanced Data
by Salha Al-Ahmari and Farrukh Nadeem
Diagnostics 2025, 15(4), 48; https://doi.org/10.3390/diagnostics15040501 - 19 Feb 2025
Viewed by 264
Abstract
Background: Surgical site infections (SSIs) lead to higher hospital readmission rates and healthcare costs, representing a significant global healthcare burden. Machine learning (ML) has demonstrated potential in predicting SSIs; however, the challenge of addressing imbalanced class ratios remains. Objectives: The aim [...] Read more.
Background: Surgical site infections (SSIs) lead to higher hospital readmission rates and healthcare costs, representing a significant global healthcare burden. Machine learning (ML) has demonstrated potential in predicting SSIs; however, the challenge of addressing imbalanced class ratios remains. Objectives: The aim of this study is to evaluate and enhance the predictive capabilities of machine learning models for SSIs by assessing the effects of feature selection, resampling techniques, and hyperparameter optimization. Methods: Using routine SSI surveillance data from multiple hospitals in Saudi Arabia, we analyzed a dataset of 64,793 surgical patients, of whom 1632 developed SSI. Seven machine learning algorithms were created and tested: Decision Tree (DT), Gaussian Naive Bayes (GNB), Support Vector Machine (SVM), Logistic Regression (LR), Random Forest (RF), Stochastic Gradient Boosting (SGB), and K-Nearest Neighbors (KNN). We also improved several resampling strategies, such as undersampling and oversampling. Grid search five-fold cross-validation was employed for comprehensive hyperparameter optimization, in conjunction with balanced sampling techniques. Features were selected using a filter method based on their relationships with the target variable. Results: Our findings revealed that RF achieves the highest performance, with an MCC of 0.72. The synthetic minority oversampling technique (SMOTE) is the best-performing resampling technique, consistently enhancing the performance of most machine learning models, except for LR and GNB. LR struggles with class imbalance due to its linear assumptions and bias toward the majority class, while GNB’s reliance on feature independence and Gaussian distribution make it unreliable for under-represented minority classes. For computational efficiency, the Instance Hardness Threshold (IHT) offers a viable alternative undersampling technique, though it may compromise performance to some extent. Conclusions: This study underscores the potential of ML models as effective tools for assessing SSI risk, warranting further clinical exploration to improve patient outcomes. By employing advanced ML techniques and robust validation methods, these models demonstrate promising accuracy and reliability in predicting SSI events, even in the face of significant class imbalances. In addition, using MCC in this study ensures a more reliable and robust evaluation of the model’s predictive performance, particularly in the presence of an imbalanced dataset, where other metrics may fail to provide an accurate evaluation. Full article
(This article belongs to the Special Issue Artificial Intelligence for Clinical Diagnostic Decision Making)
Show Figures

Figure 1

Figure 1
<p>Data distribution in the two classes.</p>
Full article ">Figure 2
<p>Flow diagram of the research methodology.</p>
Full article ">Figure 3
<p>Numerical features selected based on Spearman correlation.</p>
Full article ">Figure 4
<p>Categorial features selected based on chi-square test.</p>
Full article ">Figure 5
<p>Confusion matrix.</p>
Full article ">Figure 6
<p>Confusion matrices for the six ML models with ADASYN oversampling.</p>
Full article ">Figure 7
<p>Confusion matrices for the six ML models with SMOTE oversampling.</p>
Full article ">Figure 8
<p>Confusion matrices for the six ML models with random oversampling.</p>
Full article ">Figure 9
<p>AUC and PR_AUC plots for all six ML models. (<b>a</b>) Random oversampling; (<b>b</b>) SMOTE oversampling; (<b>c</b>) ADASYN oversampling.</p>
Full article ">Figure 10
<p>Confusion matrices for the seven ML models with random undersampling.</p>
Full article ">Figure 11
<p>Confusion matrices for the seven ML models with IHT undersampling.</p>
Full article ">Figure 12
<p>Confusion matrices for the seven ML models with CC undersampling.</p>
Full article ">Figure 13
<p>AUC and PR_AUC plots for all six ML models. (<b>a</b>) Random undersampling; (<b>b</b>) IHT undersampling; (<b>c</b>) CC undersampling.</p>
Full article ">Figure 14
<p>MCC results from all ML models.</p>
Full article ">
29 pages, 6722 KiB  
Article
Framework for Addressing Imbalanced Data in Aviation with Federated Learning
by Igor Kabashkin
Information 2025, 16(2), 48; https://doi.org/10.3390/info16020147 - 16 Feb 2025
Viewed by 295
Abstract
The aviation industry generates vast amounts of data across multiple stakeholders, but critical faults and anomalies occur rarely, creating inherently imbalanced datasets that complicate machine learning applications. Traditional centralized approaches are further constrained by privacy concerns and regulatory requirements that limit data sharing [...] Read more.
The aviation industry generates vast amounts of data across multiple stakeholders, but critical faults and anomalies occur rarely, creating inherently imbalanced datasets that complicate machine learning applications. Traditional centralized approaches are further constrained by privacy concerns and regulatory requirements that limit data sharing among stakeholders. This paper presents a novel framework for addressing imbalanced data challenges in aviation through federated learning, focusing on fault detection, predictive maintenance, and safety management. The proposed framework combines specialized techniques for handling imbalanced data with privacy-preserving federated learning to enable effective collaboration while maintaining data security. The framework incorporates local resampling methods, cost-sensitive learning, and weighted aggregation mechanisms to improve minority class detection performance. The framework is validated through extensive experiments involving multiple aviation stakeholders, demonstrating a 23% improvement in fault detection accuracy and a 17% reduction in remaining useful life prediction error compared to conventional models. Results show the enhanced detection of rare but critical faults, improved maintenance scheduling accuracy, and effective risk assessment across distributed aviation datasets. The proposed framework provides a scalable and practical solution for using distributed aviation data while addressing both class imbalance and privacy concerns, contributing to improved safety and operational efficiency in the aviation industry. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Aviation ecosystem with unbalanced data sources.</p>
Full article ">Figure 2
<p>Imbalanced data generated by AHMS.</p>
Full article ">Figure 3
<p>Imbalanced data generated by stakeholders.</p>
Full article ">Figure 4
<p>Data integration in ATSaaS platform.</p>
Full article ">Figure 5
<p>Framework for addressing imbalanced data in aviation using FL.</p>
Full article ">Figure 6
<p>Federated learning workflow.</p>
Full article ">Figure 7
<p>Weighted loss values for each stakeholder.</p>
Full article ">Figure 8
<p>Normalized loss for normal and fault classes.</p>
Full article ">Figure 9
<p>Federated learning benchmark analysis.</p>
Full article ">Figure 10
<p>Average RUL error.</p>
Full article ">Figure 11
<p>Comparison of predictive performance across different models.</p>
Full article ">Figure 12
<p>Risk scores: (<b>a</b>) Predicted fault probability; (<b>b</b>) Predicted RUL; (<b>c</b>) Risk scores.</p>
Full article ">Figure 13
<p>Feature contributions analysis.</p>
Full article ">
21 pages, 8936 KiB  
Article
A Minority Sample Enhanced Sampler for Crop Classification in Unmanned Aerial Vehicle Remote Sensing Images with Class Imbalance
by Jiapei Cheng, Liang Huang, Bohui Tang, Qiang Wu, Meiqi Wang and Zixuan Zhang
Agriculture 2025, 15(4), 48; https://doi.org/10.3390/agriculture15040388 - 12 Feb 2025
Viewed by 348
Abstract
Deep learning techniques have become the mainstream approach for fine-grained crop classification in unmanned aerial vehicle (UAV) remote sensing imagery. However, a significant challenge lies in the long-tailed distribution of crop samples. This imbalance causes neural networks to focus disproportionately on majority class [...] Read more.
Deep learning techniques have become the mainstream approach for fine-grained crop classification in unmanned aerial vehicle (UAV) remote sensing imagery. However, a significant challenge lies in the long-tailed distribution of crop samples. This imbalance causes neural networks to focus disproportionately on majority class features during training, leading to biased decision boundaries and weakening model performance. We designed a minority sample enhanced sampling (MES) method with the goal of addressing the performance limitations that are caused by class imbalance in many crop classification models. The main principle of MES is to relate the re-sampling probability of each class to the sample pixel frequency, thereby achieving intensive re-sampling of minority classes and balancing the training sample distribution. Meanwhile, during re-sampling, data augmentation is performed on the sampled images to improve the generalization. MES is simple to implement, is highly adaptable, and can serve as a general-purpose sampler for semantic segmentation tasks, functioning as a plug-and-play component within network models. To validate the applicability of MES, experiments were conducted on four classic semantic segmentation networks. The results showed that MES achieved mIoU improvements of +1.54%, +4.14%, +2.44%, and +7.08% on the Dali dataset and +2.36%, +0.86%, +4.26%, and +2.75% on the Barley Remote Sensing Dataset compared with the respective benchmark models. Additionally, our hyperparameter sensitivity analysis confirmed the stability and reliability of the method. MES mitigates the impact of class imbalance on network performance, which facilitates the practical application of deep learning in fine-grained crop classification. Full article
(This article belongs to the Special Issue Applications of Remote Sensing in Agricultural Soil and Crop Mapping)
Show Figures

Figure 1

Figure 1
<p>The class distribution of a long-tailed dataset. The head class feature space that is learned on these samples is often larger than the tail classes, while the decision boundary is usually biased towards dominant classes.</p>
Full article ">Figure 2
<p>Dali dataset: (<b>a</b>) geographical location; (<b>b</b>) diagram illustrating UAV orthophoto image.</p>
Full article ">Figure 3
<p>Dali dataset: (<b>a</b>) manual annotation labels; (<b>b</b>) pixel statistics of training sample classes.</p>
Full article ">Figure 4
<p>Barley Remote Sensing Dataset: (<b>a</b>) manual annotation labels; (<b>b</b>) pixel statistics of training sample classes.</p>
Full article ">Figure 5
<p>Baseline network architecture: (<b>a</b>) Deeplabv3+ [<a href="#B26-agriculture-15-00388" class="html-bibr">26</a>]; (<b>b</b>) SegNeXt [<a href="#B27-agriculture-15-00388" class="html-bibr">27</a>]; (<b>c</b>) Segformer [<a href="#B28-agriculture-15-00388" class="html-bibr">28</a>]; (<b>d</b>) Swin Transformer [<a href="#B29-agriculture-15-00388" class="html-bibr">29</a>].</p>
Full article ">Figure 6
<p>Visualization of experimental results when using MES with benchmark networks on the Dali dataset.</p>
Full article ">Figure 7
<p>Training loss curves of each method with iteration periods.</p>
Full article ">Figure 8
<p>Variation curves of training IoU for Brassica chinensis (minority class) with iteration period.</p>
Full article ">Figure 9
<p>Examples of experimental results of different methods based on Swin Transformer baseline network on Dali dataset.</p>
Full article ">Figure 10
<p>Visualization of the experimental results of using MES on benchmark networks in the Barley Remote Sensing Dataset.</p>
Full article ">Figure 11
<p>Examples of experimental results for different methods based on the Swin Transformer baseline network and the Barley Remote Sensing Dataset, the details of how each method was addressed are highlighted in the red circles.</p>
Full article ">Figure 12
<p>Re-sampling frequencies for different <span class="html-italic">t</span> values in the Dali dataset.</p>
Full article ">Figure 13
<p>Number of sampled pixels for each crop class in Dali dataset: (<b>a</b>) number of sampled pixels corresponding to different <span class="html-italic">T</span> values (<span class="html-italic">α</span> = 1); (<b>b</b>) number of sampled pixels corresponding to different <span class="html-italic">α</span> values (<span class="html-italic">T</span> = 0.05).</p>
Full article ">
19 pages, 779 KiB  
Article
An Analytical Prior Selection Procedure for Empirical Bayesian Analysis Using Resampling Techniques: A Simulation-Based Approach Using the Pancreatic Adenocarcinoma Data from the SEER Database
by Aditya Chakraborty and Mohan D. Pant
Computation 2025, 13(2), 48; https://doi.org/10.3390/computation13020051 - 12 Feb 2025
Viewed by 446
Abstract
Introduction: In the field of medical research, empirical Bayesian analysis has emerged as an increasingly applicable approach. This statistical framework offers greater flexibility, enabling researchers to incorporate prior information and rigorously estimate parameters of interest. However, the selection of suitable prior distributions can [...] Read more.
Introduction: In the field of medical research, empirical Bayesian analysis has emerged as an increasingly applicable approach. This statistical framework offers greater flexibility, enabling researchers to incorporate prior information and rigorously estimate parameters of interest. However, the selection of suitable prior distributions can be a challenging endeavor, with profound implications for the resulting inferences. To address this challenge, this study proposes a new analytical procedure that leverages resampling techniques to guide the choice of priors in Bayesian analysis. Subject and Methods: The study group consisted of patients who had been diagnosed and had died of pancreatic adenocarcinoma (cause-specific death) who had undergone both chemotherapy and radiation at stage IV of cancer. The data were collected from the Surveillance, Epidemiology, and End Results (SEER) database. Initially, the most suitable probabilistic behavior of the survival times of patients was identified parametrically via goodness-of-fit (GOF) tests, and afterward, empirical Bayesian analysis (EBA) was performed using resampling techniques (bootstrapping and the jackknife method). The Hamiltonian Monte Carlo (HMC) method was used to obtain the posterior distribution. Results: The most appropriate data distribution was found to be a two-parameter log-normal via GOF tests. A sensitivity analysis, followed by a simulation study, was performed to validate the analytical method. The performance of bootstrapped and jackknifed empirical Bayesian estimates was compared with maximum likelihood (ML) methods at each simulation stage. The empirical Bayesian estimates were found to be consistent with the ML estimates. Finally, a comparison was made among the parametric, Kaplan–Meier and empirical Bayesian survival estimates at different time points to illustrate the validity of the method. Conclusions: Determining the appropriate prior distribution is one of the crucial components in Bayesian analysis, as it can significantly influence the resulting inferences. The cautious selection of the prior information is essential, as it encapsulates the researcher’s beliefs or external prior knowledge about the parameters of interest. In the Bayesian framework, empirical resampling methods, such as bootstrapping and jackknifing, can offer valuable insights into the significance of prior selection, thus improving the consistency of statistical inferences. However, the analytical procedure is based on the time-to-event data, and the prior selection procedure can be extended to any real data, where Bayesian analysis is needed for decision-making and uncertainty quantification. Full article
Show Figures

Figure 1

Figure 1
<p>Probability density of the study data.</p>
Full article ">Figure 2
<p>Log-normal distribution for different scale parameters <math display="inline"><semantics> <mi>σ</mi> </semantics></math>.</p>
Full article ">Figure 3
<p>Analytical process to identify the prior via bootstrapping: the method.</p>
Full article ">Figure 4
<p>Distribution of bootstrapped empirical prior (BEP) (<b>left</b>) and jackknifed empirical prior (JEP) (<b>Right</b>).</p>
Full article ">Figure 5
<p>Trace plot showing the convergence of four Markov chains.</p>
Full article ">Figure 6
<p>Autocorrelation plot for intercept and variance.</p>
Full article ">Figure 7
<p>Comparing the SDs of ML, bootstrap, and jackknife estimates.</p>
Full article ">Figure 8
<p>Steps performed at each stage of the analysis.</p>
Full article ">Figure 9
<p>NP, P, and B are indicated as non-parametric KM, parametric, and Bayesian survival curve, respectively.</p>
Full article ">
14 pages, 590 KiB  
Article
Optimizing Multiclass Classification Using Convolutional Neural Networks with Class Weights and Early Stopping for Imbalanced Datasets
by Muhammad Nazim Razali, Nureize Arbaiy, Pei-Chun Lin and Syafikrudin Ismail
Electronics 2025, 14(4), 48; https://doi.org/10.3390/electronics14040705 - 12 Feb 2025
Viewed by 441
Abstract
Multiclass classification in machine learning often faces significant challenges due to unbalanced datasets. This situation leads to biased predictions and reduced model performance. This research addresses this issue by proposing a novel approach that combines convolutional neural networks (CNNs) with class weights and [...] Read more.
Multiclass classification in machine learning often faces significant challenges due to unbalanced datasets. This situation leads to biased predictions and reduced model performance. This research addresses this issue by proposing a novel approach that combines convolutional neural networks (CNNs) with class weights and early-stopping techniques. The motivation behind this study stems from the need to improve model performance, especially for minority classes, which are often neglected in existing methodologies. Although various strategies such as resampling, ensemble methods, and data augmentation have been explored, they frequently have limitations based on the characteristics of the data and the specific model type. Our approach focuses on optimizing the loss function via class weights to give greater importance to minority classes. Therefore, it reduces bias and improves overall accuracy. Furthermore, we implement early stopping to avoid overfitting and improve generalization by continuously monitoring the validation performance during training. This study contributes to the body of knowledge by demonstrating the effectiveness of this combined technique in improving multiclass classification in unbalanced scenarios. The proposed model is tested for oil palm leaves analysis to identify deficiencies in nitrogen (N), boron (B), magnesium (Mg), and potassium (K). The CNN model with three layers and a SoftMax activation function was trained for 200 epochs each. The analysis compared three scenarios: training with the imbalanced dataset, training with class weights, and training with class weights and early stopping. The results showed that applying class weights significantly improved the classification accuracy, with a trade-off in other class predictions. This indicates that, while class weight has a positive overall impact, further strategies are necessary to improve model performance across all categories in this study. Full article
Show Figures

Figure 1

Figure 1
<p>CNN with class weighting and early stopping.</p>
Full article ">
25 pages, 34424 KiB  
Article
Resampling Point Clouds Using Series of Local Triangulations
by Vijai Kumar Suriyababu, Cornelis Vuik and Matthias Möller
J. Imaging 2025, 11(2), 48; https://doi.org/10.3390/jimaging11020049 - 8 Feb 2025
Viewed by 568
Abstract
The increasing reliance on 3D scanning and meshless methods highlights the need for algorithms optimized for point-cloud geometry representations in CAE simulations. While voxel-based binning methods are simple, they often compromise geometry and topology, particularly with coarse voxelizations. We propose an algorithm based [...] Read more.
The increasing reliance on 3D scanning and meshless methods highlights the need for algorithms optimized for point-cloud geometry representations in CAE simulations. While voxel-based binning methods are simple, they often compromise geometry and topology, particularly with coarse voxelizations. We propose an algorithm based on a Series of Local Triangulations (SOLT) as an intermediate representation for point clouds, enabling efficient upsampling and downsampling. This robust and straightforward approach preserves the integrity of point clouds, ensuring resampling without feature loss or topological distortions. The proposed techniques integrate seamlessly into existing engineering workflows, avoiding complex optimization or machine learning methods while delivering reliable, high-quality results for a large number of examples. Resampled point clouds produced by our method can be directly used for solving PDEs or as input for surface reconstruction algorithms. We demonstrate the effectiveness of this approach with examples from mechanically sampled point clouds and real-world 3D scans. Full article
(This article belongs to the Special Issue Exploring Challenges and Innovations in 3D Point Cloud Processing)
Show Figures

Figure 1

Figure 1
<p>Workflow of the overall methodology. Optional modules are highlighted in light orange.</p>
Full article ">Figure 2
<p>Point cloud (blue) converted to a Series of Local Triangulations (SOLT) representation. The SOLT is shown in yellow, both with and without edges.</p>
Full article ">Figure 3
<p>Different views of an eagle point cloud (from Open3D’s datasets [<a href="#B18-jimaging-11-00049" class="html-bibr">18</a>]). The point cloud (796,825 points) contains intricate features, making it an excellent candidate for evaluating reconstruction algorithms.</p>
Full article ">Figure 4
<p>SOLT reconstruction of the eagle point cloud (Time taken: 35.8 s). The SOLT algorithm effectively captures the intricate features of the point cloud while being computationally efficient.</p>
Full article ">Figure 5
<p>BPA reconstruction of the eagle point cloud (Time taken: 26.91 min). This method is 62 times slower than the SOLT algorithm, achieving a similar reconstruction quality.</p>
Full article ">Figure 6
<p>Poisson reconstruction of the eagle point cloud (Time taken: 87.9 s). This method is 2.46 times slower than the SOLT algorithm, achieving comparable quality.</p>
Full article ">Figure 7
<p>Feature distance fields for selected geometries (purple indicates a distance field value of zero).</p>
Full article ">Figure 8
<p>Point cloud (blue) meshed using SOLT (yellow), downsampled in two stages (pink and green), and reconstructed using the SOLT representation (purple).</p>
Full article ">Figure 9
<p>Point clouds synthesized from the SimJEB dataset. Point cloud (blue) meshed using SOLT (yellow), downsampled in two stages (pink and green), and reconstructed using the SOLT representation (purple).</p>
Full article ">Figure 10
<p>A screw geometry resampled using our algorithm (geometry from the Thingi10k dataset).</p>
Full article ">Figure 11
<p>A mixture of smooth and sharp geometries with twist-like features (geometries from the Thingi10k dataset).</p>
Full article ">Figure 12
<p>Mechanical components from the Thingi10k dataset. Sharp creases were recovered perfectly.</p>
Full article ">Figure 13
<p>Selected geometries from the Thingi10k dataset, resampled using our algorithm and reconstructed using a simple Ball-Pivoting Algorithm (BPA) [<a href="#B28-jimaging-11-00049" class="html-bibr">28</a>]. The results demonstrate the uniformity and quality of the reconstructed meshes.</p>
Full article ">Figure 14
<p>Histograms showing the triangle area distribution for the reconstructed geometries presented in <a href="#jimaging-11-00049-f013" class="html-fig">Figure 13</a>.</p>
Full article ">Figure 14 Cont.
<p>Histograms showing the triangle area distribution for the reconstructed geometries presented in <a href="#jimaging-11-00049-f013" class="html-fig">Figure 13</a>.</p>
Full article ">Figure 15
<p>Input bunny point cloud along with a 5000-point resample produced by [<a href="#B12-jimaging-11-00049" class="html-bibr">12</a>]. These results were provided by the authors.</p>
Full article ">Figure 16
<p>Bunny resampled at various sizes using SOLT, along with corresponding sampling times. The results demonstrate that SOLT maintains consistent efficiency and quality as sample size increases, comparable to the algorithms proposed in [<a href="#B12-jimaging-11-00049" class="html-bibr">12</a>].</p>
Full article ">Figure 17
<p>Chair reconstruction from input point cloud using the RepKPU workflow (results shared by the authors). The reconstruction contains multiple holes and is of poor quality. For comparison, the SOLT reconstruction of the same chair geometry is shown, demonstrating significantly higher quality and robustness.</p>
Full article ">Figure 18
<p>Chair resampled at various sizes using SOLT, along with corresponding sampling times. The results demonstrate that SOLT maintains consistent efficiency and high-quality output as the sample size increases.</p>
Full article ">
20 pages, 1553 KiB  
Review
A Review of Deep Learning Applications in Intrusion Detection Systems: Overcoming Challenges in Spatiotemporal Feature Extraction and Data Imbalance
by Ya Zhang, Ravie Chandren Muniyandi and Faizan Qamar
Appl. Sci. 2025, 15(3), 48; https://doi.org/10.3390/app15031552 - 3 Feb 2025
Viewed by 1071
Abstract
In the rapid development of the Internet of Things (IoT) and large-scale distributed networks, Intrusion Detection Systems (IDS) face significant challenges in handling complex spatiotemporal features and addressing data imbalance issues. This article systematically reviews recent advancements in applying deep learning techniques in [...] Read more.
In the rapid development of the Internet of Things (IoT) and large-scale distributed networks, Intrusion Detection Systems (IDS) face significant challenges in handling complex spatiotemporal features and addressing data imbalance issues. This article systematically reviews recent advancements in applying deep learning techniques in IDS, focusing on the core challenges of spatiotemporal feature extraction and data imbalance. First, this article analyzes the spatiotemporal dependencies of Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN) in network traffic feature extraction and examines the main methods these models use to solve this problem. Next, the impact of data imbalance on IDS performance is explored, and the effectiveness of various data augmentation and handling techniques, including Generative Adversarial Networks (GANs) and resampling methods, in improving the detection of minority class attacks is assessed. Finally, the paper highlights the current research gaps and proposes future research directions to optimize deep learning models further to enhance the detection capabilities and robustness of IDS in complex network environments. This review provides researchers with a comprehensive perspective, helping them identify the challenges in the current field and laying a foundation for future research efforts. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Flow In IDS.</p>
Full article ">Figure 2
<p>Number of the schemes.</p>
Full article ">Figure 3
<p>Challenges in Intrusion Detection Systems.</p>
Full article ">
26 pages, 5439 KiB  
Article
Particle Filter Tracking System Based on Digital Zoom and Regional Image Measure
by Qisen Zhao, Liquan Dong, Xuhong Chu, Ming Liu, Lingqin Kong and Yuejin Zhao
Sensors 2025, 25(3), 48; https://doi.org/10.3390/s25030880 - 31 Jan 2025
Viewed by 527
Abstract
To address the challenges of low accuracy and the difficulty in balancing a large field of view and long distance when tracking high-speed moving targets with a single sensor, an ROI adaptive digital zoom tracking method is proposed. In this paper, we discuss [...] Read more.
To address the challenges of low accuracy and the difficulty in balancing a large field of view and long distance when tracking high-speed moving targets with a single sensor, an ROI adaptive digital zoom tracking method is proposed. In this paper, we discuss the impact of ROI on image processing and describe the design of the ROI adaptive digital zoom tracking system. Additionally, we construct an adaptive ROI update model based on normalized target information. To capture target changes effectively, we introduce the multi-scale regional measure and propose an improved particle filter algorithm, referred to as the improved multi-scale regional measure resampling particle filter (IMR-PF). This method enables high temporal resolution processing efficiency within a high-resolution large field of view, which is particularly beneficial for high-resolution videos. The IMR-PF can maintain high temporal resolution within a wide field of view with high resolution. Simulation results demonstrate that the improved target tracking method effectively improves tracking robustness to target motion changes and reduces the tracking center error by 20%, as compared to other state-of-the-art methods. The IMR-PF still maintains good performance even when confronted with various interference factors and in real-world scenario applications. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>The overall structure of the proposed method.</p>
Full article ">Figure 2
<p>MENS filter diagram.</p>
Full article ">Figure 3
<p>Tracking results of the top six methods on the U-skier dataset (different video clips from top to bottom).</p>
Full article ">Figure 4
<p>Evaluation result on the U-skin dataset. (<b>a</b>), Success rate of different methods on the U-skier dataset; (<b>b</b>), Precision of different methods on the U-skier dataset.</p>
Full article ">Figure 5
<p>System schematic: large FOV camera is used to capture object images, PC processor is used to target detection and tracking on the captured images, and the tracking results are used to drive the Pan-tilt Platform to rotate. Video camera is a zoom camera used to capture close-up images of the target and evaluate the results. <math display="inline"><semantics> <mi>L</mi> </semantics></math> is the depth range of the entire system.</p>
Full article ">Figure 6
<p>Tracking results of the top five tracking methods on the U-skier dataset in an indoor simulation environment (The following line shows the tracking results on a high-resolution image with a large field of view. The top line shows the close-up camera captured images corresponding to each tracking moment. The trajectory curve in the figure was formed by manually registering the front and rear frames and connecting the predicted center positions of different methods. The images in column (<b>a</b>) depict a scenario where the athlete’s target leaps in the distance, with interference from similar targets in the background. The images in column (<b>b</b>) show the athlete’s target descending and encountering a sudden change in trajectory due to the ski slope. The images in column (<b>c</b>) illustrate the athlete’s target leaping into the air from a close distance. The images in column (<b>d</b>) capture the athlete’s target descending at a close distance, experiencing a sudden change in trajectory upon encountering the ski slope).</p>
Full article ">Figure 7
<p>Tracking results on the U-skier dataset in an indoor simulation environment. (<b>a</b>) Ground truth of the target trajectory; (<b>b</b>) Changes in the target trajectory and target area measure; (<b>c</b>,<b>d</b>) Tracking error curves of different methods as the video frame changes.</p>
Full article ">Figure 8
<p>Results of the top six different methods on the VOT2021 dataset for tracking different data segments on the screen. (<b>a</b>) represents the success rate of these methods on the graduate_set of the VOT2021; (<b>b</b>) represents the success rate of these methods on the matrix_set of the VOT2021; (<b>c</b>) represents the success rate of these methods on the pedestrian_set of the VOT2021; (<b>d</b>) represents the success rate of these methods on the road_set of the VOT2021; and (<b>e</b>) represents the success rate of these methods on the shaking_set of the VOT2021.</p>
Full article ">Figure 9
<p>Success rate of different methods on the different sub-datasets of the VOT2021 dataset. (<b>a</b>) represents the Success rate of these methods at graduate_set in the VOT2021 dataset; (<b>b</b>) represents the Success rate of these methods at matrix_set in the VOT2021 dataset; (<b>c</b>) represents the Success rate of these methods at soccer_set in the VOT2021 dataset; (<b>d</b>) represents the Success rate of these methods at nature_set in the VOT2021 dataset; (<b>e</b>) represents the Success rate of these methods at road_set in the VOT2021 dataset; (<b>f</b>) represents the Success rate of these methods at racing_set in the VOT2021 dataset; (<b>g</b>) represents the Success rate of these methods at pedestrian_set in the VOT2021 dataset; and (<b>h</b>) represents the Success rate of these methods at shaking_set in the VOT2021 dataset.</p>
Full article ">Figure 10
<p>Precision of different methods on data with different properties in the VOT2021 dataset. (<b>a</b>) represents the precision of these methods at size variations properties; (<b>b</b>) represents the precision of these methods at occlusion properties; (<b>c</b>) represents the precision of these methods at background clutter properties; (<b>d</b>) represents the precision of these methods at camera movement properties; (<b>e</b>) represents the precision of these methods at target fast motion properties; and (<b>f</b>) represents the precision of these methods at illumination variable properties.</p>
Full article ">Figure 11
<p>Tracking error and success rate curves of different module improvement methods on the U-skier dataset. (<b>a</b>) Representing the central error of methods with different structures at Uskier sequence; (<b>b</b>) Repeat the success rate of methods with different structures at Uskier sequence.</p>
Full article ">
11 pages, 848 KiB  
Article
Evaluation of Sampling Precision for Native and Nonnative Fish in the Gila River Basin, New Mexico
by Zach Klein and Matt Zeigler
Hydrobiology 2025, 4(1), 48; https://doi.org/10.3390/hydrobiology4010003 - 31 Jan 2025
Viewed by 389
Abstract
Biologists commonly use standard sampling protocols to ensure that data are comparable spatiotemporally. Data also need to be precise to allow for statistically meaningful comparisons. However, the effort needed to precisely sample desert fishes is unclear. We used a resampling approach to evaluate [...] Read more.
Biologists commonly use standard sampling protocols to ensure that data are comparable spatiotemporally. Data also need to be precise to allow for statistically meaningful comparisons. However, the effort needed to precisely sample desert fishes is unclear. We used a resampling approach to evaluate the effort requirements needed to precisely sample fishes among six wadeable rivers in the Gila River basin, New Mexico. We evaluated the number of samples that are necessary to obtain relative density estimates that had 25% relative standard error 80% of the time. We also estimated the effort needed to precisely characterize species richness. Our results indicate that precisely sampling fish in the Gila River basin is difficult. Sonora Sucker Catostomus insignis, Desert Sucker C. clarkii, Longfin Dace Agosia chrysogaster, and Speckled Dace Rhinicthyes osculus were generally the only species that could be precisely sampled. Characterizing the native species assemblage in the Gila River basin required between two and seven reaches, whereas the entire species assemblage could only be characterized in 50% of the study systems. The challenge of precisely sampling fish in the Gila River basin suggests that alternative sampling methods may be required to characterize changes in density or species distribution in desert Southwest systems. Full article
Show Figures

Figure 1

Figure 1
<p>The location of permanent sampling sites (squares) and the Heart Bar sampling site (star) in the Gila River basin, southwestern New Mexico.</p>
Full article ">Figure 2
<p>The proportion of total species richness by the number of sampling reaches for all species in the Gila River basin.</p>
Full article ">Figure 3
<p>The proportion of total species richness by the number of sampling reaches for all native species in the Gila River basin.</p>
Full article ">
14 pages, 2079 KiB  
Article
The Problem with Time: Application of Partial Least Squares Analysis on Time-Frequency Plots to Account for Varying Time Intervals with Applied EEG Data
by Jessie M. H. Szostakiwskyj, Filomeno Cortese, Raneen Abdul-Rhaman, Sarah J. Anderson, Amy L. Warren, Rebecca Archer, Emma Read and Kent G. Hecker
Brain Sci. 2025, 15(2), 48; https://doi.org/10.3390/brainsci15020135 - 30 Jan 2025
Viewed by 512
Abstract
Background/Objectives: When attempting to study neurocognitive mechanisms with electroencephalography (EEG) in applied ecologically valid settings, responses to stimuli may differ in time, which presents challenges to traditional EEG averaging methods. In this proof-of-concept paper, we present a method to normalize time over unequal [...] Read more.
Background/Objectives: When attempting to study neurocognitive mechanisms with electroencephalography (EEG) in applied ecologically valid settings, responses to stimuli may differ in time, which presents challenges to traditional EEG averaging methods. In this proof-of-concept paper, we present a method to normalize time over unequal trial lengths while preserving frequency content. Methods: Epochs are converted to time-frequency space where they are resampled to contain an equal number of timepoints representing the proportion of trial complete rather than true time. To validate this method, we used EEG data recorded from 8 novices and 4 experts in veterinary medicine while completing decision-making tasks using two question types: multiple-choice and script concordance questions used in veterinary school exams. Results: The resulting resampled time-frequency data were analyzed with partial least squares (PLS), a multivariate technique that extracts patterns of data that support a contrast between conditions and groups while controlling for Type I error. We found a significant latent variable representing a difference between question types for experts only. Conclusions: Despite within and between subject differences in timing, we found consistent differences between question types in experts in gamma and beta bands that are consistent with changes resulting from increased information load and decision-making. This novel analysis method may be a viable path forward to preserve ecological validity in EEG studies. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Multiple choice question sample; (<b>b</b>) Script concordance question sample. Questions are designed to be parallel in content.</p>
Full article ">Figure 2
<p>Box plot depicting the variability in length of trials based on response times (RT; in seconds) for each question in (<b>a</b>) the MCQ test condition and (<b>b</b>) the SCT condition as performed by experts (red) and novices (blue). Thick horizontal lines are the median RT, whiskers indicate the range of RTs (minimum to maximum), and the black dots indicate outlier RTs.</p>
Full article ">Figure 3
<p>Averaged brain scores for PLS analysis comparing novice and expert time-frequency data while answering MCQ and SCT questions.</p>
Full article ">Figure 4
<p>Time-frequency spectral power plots at electrode locations (<b>a</b>) Fz, (<b>b</b>) Cz, (<b>c</b>) Pz, and (<b>d</b>) Oz, respectively. Positive bootstrap ratios (&gt;2) indicate regions that demonstrate a stable relationship between conditions (i.e., greater spectral power for SCT in experts). Negative bootstrap ratios (&lt;−2) indicate regions that demonstrate a stable relationship between conditions (i.e., greater spectral power for MCQ in experts). The black outlined boxes highlight clusters of significant brain activity.</p>
Full article ">
31 pages, 499 KiB  
Article
A Survey of Methods for Addressing Imbalance Data Problems in Agriculture Applications
by Tajul Miftahushudur, Halil Mertkan Sahin, Bruce Grieve and Hujun Yin
Remote Sens. 2025, 17(3), 48; https://doi.org/10.3390/rs17030454 - 29 Jan 2025
Viewed by 534
Abstract
This survey explores recent advances in addressing class imbalance issues for developing machine learning models in precision agriculture, with a focus on techniques used for plant disease detection, soil management, and crop classification. We examine the impact of class imbalance on agricultural data [...] Read more.
This survey explores recent advances in addressing class imbalance issues for developing machine learning models in precision agriculture, with a focus on techniques used for plant disease detection, soil management, and crop classification. We examine the impact of class imbalance on agricultural data and evaluate various resampling methods, such as oversampling and undersampling, as well as algorithm-level approaches, to mitigate this challenge. The paper also highlights the importance of evaluation metrics, including F1-score, G-mean, and MCC, in assessing the performance of machine learning models under imbalanced conditions. Additionally, the review provides an in-depth analysis of emerging trends in the use of generative models, like GANs and VAEs, for data augmentation in agricultural applications. Despite the significant progress, challenges such as noisy data, incomplete datasets, and lack of publicly available datasets remain. This survey concludes with recommendations for future research directions, including the need for robust methods that can handle high-dimensional agricultural data effectively. Full article
(This article belongs to the Section Remote Sensing in Agriculture and Vegetation)
Show Figures

Figure 1

Figure 1
<p>Difference between undersampling and oversampling [<a href="#B102-remotesensing-17-00454" class="html-bibr">102</a>]. (Blue and orange colors represent majority and minority classes, respectively).</p>
Full article ">Figure 2
<p>GAN model, adapted from [<a href="#B48-remotesensing-17-00454" class="html-bibr">48</a>].</p>
Full article ">Figure 3
<p>cGAN model, adapted from [<a href="#B128-remotesensing-17-00454" class="html-bibr">128</a>].</p>
Full article ">Figure 4
<p>Variational autoencoder model, a type of generative model that learns to encode input data into a probabilistic latent space (adapted from [<a href="#B135-remotesensing-17-00454" class="html-bibr">135</a>]).</p>
Full article ">Figure 5
<p>Confusion matrix.</p>
Full article ">
21 pages, 5110 KiB  
Article
Leveraging Quantum Machine Learning to Address Class Imbalance: A Novel Approach for Enhanced Predictive Accuracy
by Seongjun Kwon, Jihye Huh, Sang Ji Kwon, Sang-ho Choi and Ohbyung Kwon
Symmetry 2025, 17(2), 48; https://doi.org/10.3390/sym17020186 - 25 Jan 2025
Viewed by 601
Abstract
The class imbalance problem presents a critical challenge in real-world applications, particularly in high-stakes domains such as healthcare, finance, disaster management, and fault diagnosis, where accurate anomaly detection is paramount. Class imbalance often disrupts the inherent symmetry of data distributions, resulting in suboptimal [...] Read more.
The class imbalance problem presents a critical challenge in real-world applications, particularly in high-stakes domains such as healthcare, finance, disaster management, and fault diagnosis, where accurate anomaly detection is paramount. Class imbalance often disrupts the inherent symmetry of data distributions, resulting in suboptimal performance of traditional machine learning models. Conventional approaches such as undersampling and oversampling are commonly employed to address this issue; however, these methods can introduce additional asymmetries, including information loss and overfitting, which ultimately compromise model efficacy. This study introduces an innovative approach leveraging quantum machine learning (QML), specifically the Variational Quantum Classifier (VQC), to restore and capitalize on the symmetrical properties of data distributions without relying on resampling techniques. By employing quantum circuits optimized to mitigate the asymmetries inherent in imbalanced datasets, the proposed method demonstrates consistently superior performance across diverse datasets, with notable improvements in Recall for minority classes. These findings underscore the potential of quantum machine learning as a robust alternative to classical methods, offering a symmetry-aware solution to class imbalance and advancing QML-driven technologies in fields where equitable representation and symmetry are of critical importance. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

Figure 1
<p>Comparison between conventional machine learning and QML.</p>
Full article ">Figure 2
<p>Example of data transformation through quantum gate.</p>
Full article ">Figure 3
<p>AI GPU server (Nvidia Tesla A100).</p>
Full article ">Figure 4
<p>Illustration of VQC algorithm.</p>
Full article ">Figure 5
<p>Illustration of Variational Quantum Circuit.</p>
Full article ">Figure 6
<p>Example of measurement using qubits.</p>
Full article ">Figure 7
<p>Example of loss function.</p>
Full article ">Figure 8
<p>Optimization condition exploration of the VQC algorithm.</p>
Full article ">Figure 9
<p>Illustration of feature space: 4 × 4 matrix.</p>
Full article ">Figure 10
<p>Illustrations of Bloch sphere visualizations before and after rotation through encoding.</p>
Full article ">
24 pages, 2674 KiB  
Article
Achieving Excellence in Cyber Fraud Detection: A Hybrid ML+DL Ensemble Approach for Credit Cards
by Eyad Btoush, Xujuan Zhou, Raj Gururajan, Ka Ching Chan and Omar Alsodi
Appl. Sci. 2025, 15(3), 48; https://doi.org/10.3390/app15031081 - 22 Jan 2025
Viewed by 992
Abstract
The rapid advancement of technology has increased the complexity of cyber fraud, presenting a growing challenge for the banking sector to efficiently detect fraudulent credit card transactions. Conventional detection approaches face challenges in adapting to the continuously evolving tactics of fraudsters. This study [...] Read more.
The rapid advancement of technology has increased the complexity of cyber fraud, presenting a growing challenge for the banking sector to efficiently detect fraudulent credit card transactions. Conventional detection approaches face challenges in adapting to the continuously evolving tactics of fraudsters. This study addresses these limitations by proposing an innovative hybrid model that integrates Machine Learning (ML) and Deep Learning (DL) techniques through a stacking ensemble and resampling strategies. The hybrid model leverages ML techniques including Decision Tree (DT), Random Forest (RF), Support Vector Machine (SVM), eXtreme Gradient Boosting (XGBoost), Categorical Boosting (CatBoost), and Logistic Regression (LR) alongside DL techniques such as Convolutional Neural Network (CNN) and Bidirectional Long Short-Term Memory Network (BiLSTM) with attention mechanisms. By utilising the stacking ensemble method, the model consolidates predictions from multiple base models, resulting in improved predictive accuracy compared to individual models. The methodology incorporates robust data pre-processing techniques. Experimental evaluations demonstrate the superior performance of the hybrid ML+DL model, particularly in handling class imbalances and achieving a high F1 score, achieving an F1 score of 94.63%. This result underscores the effectiveness of the proposed model in delivering reliable cyber fraud detection, highlighting its potential to enhance financial transaction security. Full article
Show Figures

Figure 1

Figure 1
<p>Change in Australians affected by personal fraud.</p>
Full article ">Figure 2
<p>Hybrid ML+DL model.</p>
Full article ">Figure 3
<p>The percentage of fraudulent vs non-fraudulent transactions.</p>
Full article ">Figure 4
<p>The plot of the credit card dataset.</p>
Full article ">Figure 5
<p>The data pre-processing steps.</p>
Full article ">Figure 6
<p>Performance of ML algorithms.</p>
Full article ">Figure 7
<p>Confusion matrix of CNN model.</p>
Full article ">Figure 8
<p>Confusion matrix of BiLSTM model.</p>
Full article ">Figure 9
<p>Confusion matrix hybrid ML+DL model.</p>
Full article ">Figure 10
<p>F1 score of sampling with the novel hybrid ML+DL model.</p>
Full article ">Figure 11
<p>Performance of sampling with the hybrid ML+DL model.</p>
Full article ">
Back to TopTop