Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (7,583)

Search Parameters:
Keywords = feature vectors

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 4753 KiB  
Article
Power Quality Disturbance Classification Method Based on Unscented Kalman Filter and Kernel Extreme Learning Machine
by Yanjun Jiao, Haoyu Cao, Linke Wang, Jiahui Wei, Yansong Zhu and Hucheng He
Appl. Sci. 2025, 15(5), 2721; https://doi.org/10.3390/app15052721 (registering DOI) - 4 Mar 2025
Abstract
The power quality index is an important index in the power industry. Power quality disturbances (PQDs) have a great impact on the power grid. It is important to identify the type of PQDs accurately. However, it is difficult to analyze a large number [...] Read more.
The power quality index is an important index in the power industry. Power quality disturbances (PQDs) have a great impact on the power grid. It is important to identify the type of PQDs accurately. However, it is difficult to analyze a large number of PQDs, especially in more complex power systems. Considering the limitations of the traditional time–frequency domain method and the complexity of the optimization algorithm in extracting features, a novel algorithm is proposed to classify the PQDs in this paper, which is based on the unscented Kalman filter (UKF) and the kernel extreme learning machine (KELM). The UKF is used to detect and process the original disturbances, and anti-noise detection performance is analyzed by tracking the amplitude change in voltage swell under different signal–noise ratios (SNRs). The amplitudes of the fundamental wave, third harmonic, fifth harmonic, seventh harmonic, and oscillatory wave are tracked in real time, and their minimum and peak indexes are taken as the optimal feature vector set. The feature vector set is classified by the KELM. The performance of the proposed method has also been evaluated with simulated and experimental results. Full article
Show Figures

Figure 1

Figure 1
<p>The flowchart of UKF algorithm.</p>
Full article ">Figure 2
<p>Tracking voltage surge disturbance by UKF. (<b>a</b>) In total, 40 db amplitude variation; (<b>b</b>) 20 db amplitude variation.</p>
Full article ">Figure 3
<p>The index distribution of different feature quantities with single disturbance. (<b>a</b>) Minimum feature distribution; (<b>b</b>) peak index characteristic distribution.</p>
Full article ">Figure 4
<p>KELM topological structure.</p>
Full article ">Figure 5
<p>The structure diagram of KELM to identify PQDs.</p>
Full article ">Figure 6
<p>The classification results of fourteen kinds of PQDs by KELM.</p>
Full article ">Figure 7
<p>The success rate of voltage swell detection and recognition in different noise environments.</p>
Full article ">Figure 8
<p>The overall scheme of PQDs classification system.</p>
Full article ">Figure 9
<p>The experimental platform of PQDs classification system.</p>
Full article ">Figure 10
<p>The test results of the hardware platform. (<b>a</b>) Harmonic disturbance; (<b>b</b>) voltage sag; (<b>c</b>) harmonics and voltage swell; (<b>d</b>) voltage swell; (<b>e</b>) voltage pulse; (<b>f</b>) harmonics and voltage sag; (<b>g</b>) harmonics and voltage pulse.</p>
Full article ">
3062 KiB  
Proceeding Paper
The Use of Support Vector Machine to Classify Potential Customers for the Wealth Management of a Bank
by Chien-Hung Lai, Yi Lin, Ju-Wen Hsieh and Yuh-Shyan Hwang
Eng. Proc. 2025, 89(1), 32; https://doi.org/10.3390/engproc2025089032 - 3 Mar 2025
Abstract
We developed a method for the evaluation and selection of customer business analysis in two stages. First, using the bank’s existing expert model, artificial rules of thumb were used to evaluate the value of each field of the data and establish screening rules. [...] Read more.
We developed a method for the evaluation and selection of customer business analysis in two stages. First, using the bank’s existing expert model, artificial rules of thumb were used to evaluate the value of each field of the data and establish screening rules. Secondly, the machine learning feature screening method was applied based on the customer’s transaction data to find out whether the customer’s contribution to the bank had a significant impact as a feature of the model. Based on the results, the best classification model was selected through data verification. The effectiveness of the proposed model was validated through actual case analysis, taking wealth management in banks as an example. The classification method, using support vector machines (SVMs), effectively assists banks in identifying potential customers efficiently and in planning to manage customers. This method helps to avoid the traditional blind spots, which emerge based on subjective judgment, and allows bank wealth managers to promote customer relationship management (CRM). Full article
Show Figures

Figure 1

Figure 1
<p>ROC of rules of thumb, dashed line represents the threshold.</p>
Full article ">Figure 2
<p>ROC of correlation, dashed line represents the threshold.</p>
Full article ">Figure 3
<p>ROC of feature importance, dashed line represents the threshold.</p>
Full article ">Figure 4
<p>ROC of L1 regularization, dashed line represents the threshold.</p>
Full article ">Figure 5
<p>Confusion matrix of rules of thumb.</p>
Full article ">Figure 6
<p>Confusion matrix of correlation.</p>
Full article ">Figure 7
<p>Confusion matrix of feature importance.</p>
Full article ">Figure 8
<p>Confusion matrix of L1 regularization.</p>
Full article ">
24 pages, 3438 KiB  
Article
AE-XGBoost: A Novel Approach for Machine Tool Machining Size Prediction Combining XGBoost, AE and SHAP
by Mu Gu, Shuimiao Kang, Zishuo Xu, Lin Lin and Zhihui Zhang
Mathematics 2025, 13(5), 835; https://doi.org/10.3390/math13050835 - 2 Mar 2025
Viewed by 89
Abstract
To achieve intelligent manufacturing and improve the machining quality of machine tools, this paper proposes an interpretable machining size prediction model combining eXtreme Gradient Boosting (XGBoost), autoencoder (AE), and Shapley additive explanation (SHAP) analysis. In this study, XGBoost was used to establish an [...] Read more.
To achieve intelligent manufacturing and improve the machining quality of machine tools, this paper proposes an interpretable machining size prediction model combining eXtreme Gradient Boosting (XGBoost), autoencoder (AE), and Shapley additive explanation (SHAP) analysis. In this study, XGBoost was used to establish an evaluation system for the actual machining size of computer numerical control (CNC) machine tools. The XGBoost model was combined with SHAP approximation to effectively capture local and global features in the data using autoencoders and transform the preprocessed data into more representative feature vectors. Grey correlation analysis (GRA) and principal component analysis (PCA) were used to reduce the dimensions of the original data features, and the synthetic minority overstimulation technique of the Gaussian noise regression (SMOGN) method was used to deal with the problem of data imbalance. Taking the actual size of the machine tool as the response parameter, based on the size parameters in the milling process of the CNC machine tool, the effectiveness of the model is verified. The experimental results show that the proposed AE-XGBoost model is superior to the traditional XGBoost method, and the prediction accuracy of the model is 7.11% higher than that of the traditional method. The subsequent SHAP analysis reveals the importance and interrelationship of features and provides a reliable decision support system for machine tool processing personnel, helping to improve processing quality and achieve intelligent manufacturing. Full article
(This article belongs to the Special Issue Applied Mathematics to Mechanisms and Machines II)
Show Figures

Figure 1

Figure 1
<p>Overall model flowchart.</p>
Full article ">Figure 2
<p>SMOGN sampling.</p>
Full article ">Figure 3
<p>Convolutional neural network diagram.</p>
Full article ">Figure 4
<p>SHAP visual interpretation of machine learning models.</p>
Full article ">Figure 5
<p>GRA heat map.</p>
Full article ">Figure 6
<p>Comparison of quality_real distribution in raw data and SMOGN processed data.</p>
Full article ">Figure 7
<p>Comparison between real values and predicted values of two methods.</p>
Full article ">Figure 8
<p>Mean (|SHAP value|) and SHAP summary plot.</p>
Full article ">Figure 9
<p>SHAP partial dependence plots of two parameters.</p>
Full article ">Figure 10
<p>SHAP dependence plots of two parameters.</p>
Full article ">
20 pages, 6040 KiB  
Article
Two-Tower Structure Recommendation Method Fusing Multi-Source Data
by Yaning Su, Yuxiang Li and Zhiyong Zhang
Electronics 2025, 14(5), 1003; https://doi.org/10.3390/electronics14051003 - 2 Mar 2025
Viewed by 39
Abstract
In view of the problem that the recommendation system is not good at integrating multi-source information and user sentiment, this paper proposes a BERT-LSTM Dual-Tower Recommendation Method for Sequential Feature Extraction (BLDRM-SFE). This method uses BERT to extract semantic features from user reviews [...] Read more.
In view of the problem that the recommendation system is not good at integrating multi-source information and user sentiment, this paper proposes a BERT-LSTM Dual-Tower Recommendation Method for Sequential Feature Extraction (BLDRM-SFE). This method uses BERT to extract semantic features from user reviews and item details and obtains vector representations of item IDs and their groups through embedding. The user tower combines user review features with item group features to generate a user vector, while the item tower integrates item detail features with item ID vectors to generate an item vector. The outputs of the two towers are processed by LSTM to handle item ID sequence information, uncover potential sequence relationships, and generate rating predictions, thereby constructing a personalized recommendation list. The experimental results show that this method significantly outperforms four baseline models—BERT4Rec, PRM, BST, and ComiRec—on the Amazon Review Data and Yelp datasets. On the Amazon dataset, BLDRM-SFE improves by 10.39%, 8.08%, 10.78%, 10.59%, and 5.49% across five metrics compared to the baseline models; on the Yelp dataset, the improvements are 10.95%, 10.06%, 13.04%, 12.59%, and 10.8%, respectively. In addition, ablation experiments confirmed the positive impact of item ID sequence information on the method’s performance. The results show that the incorporation of sequence information significantly enhanced the recommendation performance. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

Figure 1
<p>Overall architecture diagram of the method.</p>
Full article ">Figure 2
<p>Comparison of the six models on NDCG@5 and NDCG@10 evaluation metrics on the Amazon Review Data-Music dataset.</p>
Full article ">Figure 3
<p>Comparison of the six models on the Amazon Review Data-Music dataset based on HR@5 and HR@10 evaluation metrics.</p>
Full article ">Figure 4
<p>Comparison of the six models on the Amazon Review Data-Music dataset based on the MRR evaluation metric.</p>
Full article ">Figure 5
<p>The comparison chart of six models on the Yelp dataset based on the evaluation metrics NDCG@5 and NDCG@10.</p>
Full article ">Figure 6
<p>The comparison chart of six models on the Yelp dataset based on the evaluation metrics HR@5 and HR@10.</p>
Full article ">Figure 7
<p>Comparison of the six models based on the MRR evaluation metric on the Yelp dataset.</p>
Full article ">
14 pages, 3470 KiB  
Article
Predicting the Pathway Involvement of Compounds Annotated in the Reactome Knowledgebase
by Erik D. Huckvale and Hunter N. B. Moseley
Metabolites 2025, 15(3), 161; https://doi.org/10.3390/metabo15030161 - 1 Mar 2025
Viewed by 151
Abstract
Background/Objectives: Pathway annotations of non-macromolecular (relatively small) biomolecules facilitate biological and biomedical interpretation of metabolomics datasets. However, low pathway annotation levels of detected biomolecules hinder this type of interpretation. Thus, predicting the pathway involvement of detected but unannotated biomolecules has a high potential [...] Read more.
Background/Objectives: Pathway annotations of non-macromolecular (relatively small) biomolecules facilitate biological and biomedical interpretation of metabolomics datasets. However, low pathway annotation levels of detected biomolecules hinder this type of interpretation. Thus, predicting the pathway involvement of detected but unannotated biomolecules has a high potential to improve metabolomics data analysis and omics integration. Past publications have only made use of the Kyoto Encyclopedia of Genes and Genomes-derived datasets to develop machine learning models to predict pathway involvement. However, to our knowledge, the Reactome knowledgebase has not been utilized to develop these types of predictive models. Methods: We created a dataset ready for machine learning using chemical representations of all pathway-annotated compounds available from the Reactome knowledgebase. Next, we trained and evaluated a multilayer perceptron binary classifier using combined metabolite-pathway paired feature vectors engineered from this new dataset. Results: While models trained on a prior corresponding KEGG dataset with 502 pathways scored a mean Matthew’s correlation coefficient (MCC) of 0.847 and a 0.0098 standard deviation, the models trained on the Reactome dataset with 3985 pathways demonstrated improved performance with a mean MCC of 0.916, but with a higher standard deviation of 0.0149. Conclusions: These results indicate that the pathways in Reactome can also be effectively predicted, greatly increasing the number of human-defined pathways available for prediction. Full article
(This article belongs to the Section Bioinformatics and Data Analysis)
Show Figures

Figure 1

Figure 1
<p>Number of pathways per hierarchy level. Reactome organizes its pathways into a hierarchical structure beginning at the L1 pathways, which have L2 pathways under them, which have L3 under them, etc. L6+ refers to hierarchy levels L6, L7, L8, and L9 combined.</p>
Full article ">Figure 2
<p>Distribution of MCC of the full dataset. The full dataset contained all pathway hierarchy levels. The distribution of MCC for 200 CV iterations is displayed.</p>
Full article ">Figure 3
<p>oMCC by dataset and hierarchy level. The L1+ dataset is the full dataset, while L2+ excludes the L1 pathways, and L3+ excludes the L1 and L2 pathways. MCC is calculated from the sum of TP, TN. FP, and FN across all pathways and CV iterations in each hierarchy level.</p>
Full article ">Figure 4
<p>oMCC by hierarchy level and dataset. The L1+ dataset is the full dataset, while L2+ excludes the L1 pathways, and L3+ excludes the L1 and L2 pathways. oMCC is calculated from the sum of TP, TN. FP, and FN across all pathways and CV iterations in each hierarchy level.</p>
Full article ">Figure 5
<p>Distributions of compound size and pathway size.</p>
Full article ">Figure 6
<p>Distributions of individual compound oMCC and individual pathway oMCC.</p>
Full article ">Figure 7
<p>Violin plot displaying the distribution of the sizes of pathways in each hierarchy level.</p>
Full article ">Figure 8
<p>Individual compound and pathway compared to oMCC.</p>
Full article ">Figure 9
<p>The distribution and probability density of pathway size among KEGG and Reactome pathways.</p>
Full article ">Figure 10
<p>Distribution of MCCs for models trained on the KEGG and Reactome datasets. Notice the <span class="html-italic">y</span>-axis range between 0.75 and 0.975.</p>
Full article ">
25 pages, 17052 KiB  
Article
Spatial Orientation Relation Recognition for Water Surface Targets
by Peiyong Gong, Kai Zheng, Yi Jiang, Huixuan Zhao, Xiao Liang, Zhiwen Feng and Wenbin Huang
J. Mar. Sci. Eng. 2025, 13(3), 482; https://doi.org/10.3390/jmse13030482 - 28 Feb 2025
Viewed by 189
Abstract
Recently, extensive research efforts have concentrated on comprehending the semantic features of images in the field of computer vision. In order to address the spatial orientation relations among water surface targets (WSTs) in an image, which is a fundamental semantic feature, this paper [...] Read more.
Recently, extensive research efforts have concentrated on comprehending the semantic features of images in the field of computer vision. In order to address the spatial orientation relations among water surface targets (WSTs) in an image, which is a fundamental semantic feature, this paper focused on the recognition of spatial orientation relations. We first developed the water surface target spatial orientation vector field (WST-SOVF) algorithm, a novel end-to-end methodology, to recognize these spatial orientation relations among WSTs in an image. The WST-SOVF algorithm encodes the spatial orientation relation into the learning framework of a new deep convolutional neural network model, which comprises two distinct branches: the T-branch and the S-branch, both designed for the spatial feature extraction. The T-branch employs keypoint estimation to identify central points and classify the WST categories, while the S-branch constructs a spatial orientation vector field between WSTs, where each pixel in the field encodes the spatial orientation angle between two separated WSTs and collectively determines the category of spatial orientation. A fusion module was also designed to integrate the spatial feature obtained from both branches, thereby generating a comprehensive triple list that provides not only all the WSTs and their spatial orientation relations, but also their associated confidence levels. We performed a comparative evaluation of our WST-SOVF algorithm based on Huawei’s “Typical Surface/Underwater Target Recognition” dataset and the results demonstrated the outstanding performance of WST-SOVF algorithm. Full article
Show Figures

Figure 1

Figure 1
<p>Example images of misclassification.</p>
Full article ">Figure 2
<p>Example images of the water surface scene.</p>
Full article ">Figure 3
<p>Four categories of fundamental WST spatial orientation relations.</p>
Full article ">Figure 4
<p>WST spatial patterns in the images.</p>
Full article ">Figure 5
<p>The relation between the pixel coordinate system and image coordinate system.</p>
Full article ">Figure 6
<p>Conceptualization of spatial orientation relation recognition.</p>
Full article ">Figure 7
<p>The output of the T-branch.</p>
Full article ">Figure 8
<p>Schematic diagram of SOVF encoding.</p>
Full article ">Figure 9
<p>The output of the S-branch.</p>
Full article ">Figure 10
<p>Overview network architecture of the spatial orientation relations recognition model.</p>
Full article ">Figure 11
<p>Details of the network architecture of our spatial orientation relation recognition model.</p>
Full article ">Figure 12
<p>Demonstration of the test set based on SUTR construction.</p>
Full article ">Figure 13
<p>Visualization of the ground-truth constructed by the SUTR dataset.</p>
Full article ">Figure 14
<p>Classification results of the WST-SOVF algorithm for spatial orientation relations.</p>
Full article ">Figure 15
<p>Recognition results for the spatial orientation relation in the test set.</p>
Full article ">Figure 16
<p>Confusion matrix for the DCNN model of the WST-SOVF algorithm evaluation.</p>
Full article ">Figure 17
<p>Recognition results in complex marine environments.</p>
Full article ">Figure 18
<p>Recognition results in occlusion scenarios.</p>
Full article ">Figure 19
<p>Recognition results in other broader datasets.</p>
Full article ">
53 pages, 35092 KiB  
Article
Alteration Lithogeochemistry of an Archean Porphyry-Type Au(-Cu) Setting: The World-Class Côté Gold Deposit, Canada
by Laura R. Katz, Daniel J. Kontak and Benoit Dubé
Minerals 2025, 15(3), 256; https://doi.org/10.3390/min15030256 - 28 Feb 2025
Viewed by 236
Abstract
Charactering alteration and its geochemical signature provides critical information relevant to ore-deposit genesis and its related footprint; for porphyry-type deposits, zoned potassic-phyllic-propylitic alteration and metal enrichment are critical features. Here we integrate earlier lithological and mineralogical studies of the (10+ Moz Au) Archean [...] Read more.
Charactering alteration and its geochemical signature provides critical information relevant to ore-deposit genesis and its related footprint; for porphyry-type deposits, zoned potassic-phyllic-propylitic alteration and metal enrichment are critical features. Here we integrate earlier lithological and mineralogical studies of the (10+ Moz Au) Archean Côté Gold porphyry-type Au(-Cu) deposit (Ontario, Canada) with identified alteration types to provide exploration vectors. The ca. 2740 tonalite-quartz diorite-diorite intrusive complex and co-temporal Au(-Cu) mineralization as disseminations, breccias and veins are co-spatial with ore-related alteration types (amphibole, biotite, muscovite). An early, locally developed amphibole event coring the deposit is followed by emplacement of a Au(-Cu) mineralized biotite-rich magmatic-hydrothermal breccia body and broad halo of disseminated biotite and quartz veining. These rocks record gains via mass balance calculations of K, Fe, Mg, LILE, and LREE with Au, Cu, Mo, Ag, Se and Bi. Later muscovite alteration is enriched in K, Rb, Cs, Ba, CO2, and LOI with varied Au, Cu, Mo, Te, As, and Bi values. A strong albite overprint records extreme Na gains with the loss of most other elements, including ore metals (i.e., Au, Cu). Together these data define an Au-Cu-Mo-Ag-Te-Bi-Se core co-spatial with biotite breccia versus a peripheral stockwork and sheeted vein zone with a Te-Se-Zn-Pb-As association. These features further support the posited porphyry-type model for the Côté Gold Au(-Cu) deposit. Full article
26 pages, 22509 KiB  
Article
Interpretable Machine Learning for Explaining and Predicting Collapse Hazards in the Changbai Mountain Region
by Xiangyang He, Qiuling Lang, Jiquan Zhang, Yichen Zhang, Qingze Jin and Jinyuan Xu
Sensors 2025, 25(5), 1512; https://doi.org/10.3390/s25051512 - 28 Feb 2025
Viewed by 144
Abstract
This study analyzes collapse hazards for complex interactions between geology, meteorology, and human activities in the Changbai Mountain region, focusing on how to cope with these features through machine learning. Using a dataset of 651 collapse events, this study evaluates four machine learning [...] Read more.
This study analyzes collapse hazards for complex interactions between geology, meteorology, and human activities in the Changbai Mountain region, focusing on how to cope with these features through machine learning. Using a dataset of 651 collapse events, this study evaluates four machine learning methods, Support Vector Machine (SVM), Random Forest (RF), Extreme Gradient Boosting (XGBoost), and Light Gradient Boosting Machine (LightGBM), to deal with complex nonlinear data structures. To overcome the limitations of a single-feature selection method, a variance inflation factor is introduced to optimize the selection of collapse risk factors. The transparency and interpretability of the modeling results are enhanced by combining the Shapley Additive Explanations (SHAP) with interpretable artificial intelligence. Model performance is evaluated on a test set by several statistical metrics, which shows that the optimized random forest model performs best and outperforms SVM, XGBoost, and LightGBM. The SHAP analysis results indicate that distance from the road is a key factor for collapse hazard. This study emphasizes the need for collapse management strategies that provide interpretable solutions for collapse hazard assessment. Full article
(This article belongs to the Section Environmental Sensing)
Show Figures

Figure 1

Figure 1
<p>Spatial distribution of collapse sites within the study area.</p>
Full article ">Figure 2
<p>Field photos of collapse geological disaster sites in the study area.</p>
Full article ">Figure 3
<p>Flowchart for prediction of collapse.</p>
Full article ">Figure 4
<p>(<b>a</b>) Elevation; (<b>b</b>) distance from faults; (<b>c</b>) distance from river; (<b>d</b>) aspect; (<b>e</b>) distance from road; (<b>f</b>) NDVI; (<b>g</b>) annual average rainfall; (<b>h</b>) slope; (<b>i</b>) lithology; (<b>j</b>) curvature; (<b>k</b>) TWI; (<b>l</b>) exceedance probability.</p>
Full article ">Figure 4 Cont.
<p>(<b>a</b>) Elevation; (<b>b</b>) distance from faults; (<b>c</b>) distance from river; (<b>d</b>) aspect; (<b>e</b>) distance from road; (<b>f</b>) NDVI; (<b>g</b>) annual average rainfall; (<b>h</b>) slope; (<b>i</b>) lithology; (<b>j</b>) curvature; (<b>k</b>) TWI; (<b>l</b>) exceedance probability.</p>
Full article ">Figure 5
<p>Pearson’s correlation coefficient.</p>
Full article ">Figure 6
<p>Collapse susceptibility mapping: (<b>a</b>) Random Forest; (<b>b</b>) SVM; (<b>c</b>) XGBoost; (<b>d</b>) LightGBM.</p>
Full article ">Figure 6 Cont.
<p>Collapse susceptibility mapping: (<b>a</b>) Random Forest; (<b>b</b>) SVM; (<b>c</b>) XGBoost; (<b>d</b>) LightGBM.</p>
Full article ">Figure 7
<p>ROC curves for the four models and the AUC values.</p>
Full article ">Figure 8
<p>Hazard zoning.</p>
Full article ">Figure 9
<p>Hazard subdivision and raster occupancy map.</p>
Full article ">Figure 10
<p>SHAP summary map.</p>
Full article ">Figure 11
<p>Shap dependency diagram.</p>
Full article ">Figure 12
<p>Shap explains the extent to which the collapse contributes to the distance from the road and the NDVI.</p>
Full article ">
19 pages, 2262 KiB  
Article
Development of Advanced Machine Learning Models for Predicting CO2 Solubility in Brine
by Xuejia Du and Ganesh C. Thakur
Energies 2025, 18(5), 1202; https://doi.org/10.3390/en18051202 - 28 Feb 2025
Viewed by 204
Abstract
This study explores the application of advanced machine learning (ML) models to predict CO2 solubility in NaCl brine, a critical parameter for effective carbon capture, utilization, and storage (CCUS). Using a comprehensive database of 1404 experimental data points spanning temperature (−10 to [...] Read more.
This study explores the application of advanced machine learning (ML) models to predict CO2 solubility in NaCl brine, a critical parameter for effective carbon capture, utilization, and storage (CCUS). Using a comprehensive database of 1404 experimental data points spanning temperature (−10 to 450 °C), pressure (0.098 to 140 MPa), and salinity (0.017 to 6.5 mol/kg), the research evaluates the predictive capabilities of five ML algorithms: Decision Tree, Random Forest, XGBoost, Multilayer Perceptron, and Support Vector Regression with a radial basis function kernel. Among these, XGBoost demonstrated the highest overall accuracy, achieving an R2 value of 0.9926, with low root mean square error (RMSE) and mean absolute error (MAE) of 0.0655 and 0.0191, respectively. A feature importance analysis revealed that pressure has the most impactful effect and positively correlates with CO2 solubility, while temperature generally exhibits a negative effect. A higher accuracy was found when the developed model was compared with one well-established empirical model and one ML-based model from the literature. The results underscore the potential of ML models to significantly enhance prediction accuracy over a wide data range, reduce computational costs, and improve the efficiency of CCUS operations. This work demonstrates the robustness and adaptability of ML approaches for modeling complex subsurface conditions, paving the way for optimized carbon sequestration strategies. Full article
(This article belongs to the Special Issue Sustainable Energy, Environment and Low-Carbon Development)
Show Figures

Figure 1

Figure 1
<p>Schematic diagram among DT, RF, XGBoost (different colors refer to different group).</p>
Full article ">Figure 2
<p>Boxplot distribution of key parameters in CO<sub>2</sub> solubility dataset.</p>
Full article ">Figure 3
<p>Distributions of key parameters in the database.</p>
Full article ">Figure 4
<p>Heatmap of correlation coefficients.</p>
Full article ">Figure 5
<p>Flowchart of proposed ML models.</p>
Full article ">Figure 6
<p>Architecture of the MLP solubility model.</p>
Full article ">Figure 7
<p>Actual vs. predicted CO<sub>2</sub> solubility for (<b>a</b>) decision tree, (<b>b</b>) random forest, (<b>c</b>) XGBoost, (<b>d</b>) MLP, (<b>e</b>) SVR-RBF.</p>
Full article ">Figure 8
<p>Feature importance for XGBoost model.</p>
Full article ">Figure 9
<p>SHAP values analysis for the XGBoost model.</p>
Full article ">Figure 10
<p>Solubility of CO<sub>2</sub> for different models. (<b>a</b>) Salinity = 1 mol/kg, pressure = 10 MPa. (<b>b</b>) Temperature = −5 C, salinity = 4.82 mol/kg [<a href="#B9-energies-18-01202" class="html-bibr">9</a>,<a href="#B30-energies-18-01202" class="html-bibr">30</a>,<a href="#B36-energies-18-01202" class="html-bibr">36</a>,<a href="#B37-energies-18-01202" class="html-bibr">37</a>].</p>
Full article ">
18 pages, 575 KiB  
Article
A Deep Learning Method of Credit Card Fraud Detection Based on Continuous-Coupled Neural Networks
by Yanxi Wu, Liping Wang, Hongyu Li and Jizhao Liu
Mathematics 2025, 13(5), 819; https://doi.org/10.3390/math13050819 - 28 Feb 2025
Viewed by 168
Abstract
With the widespread use of credit cards in online and offline transactions, credit card fraud has become a significant challenge in the financial sector. The rapid advancement of payment technologies has led to increasingly sophisticated fraud techniques, necessitating more effective detection methods. While [...] Read more.
With the widespread use of credit cards in online and offline transactions, credit card fraud has become a significant challenge in the financial sector. The rapid advancement of payment technologies has led to increasingly sophisticated fraud techniques, necessitating more effective detection methods. While machine learning has been extensively applied in fraud detection, the application of deep learning methods remains relatively limited. Inspired by brain-like computing, this work employs the Continuous-Coupled Neural Network (CCNN) for credit card fraud detection. Unlike traditional neural networks, the CCNN enhances the representation of complex temporal and spatial patterns through continuous neuron activation and dynamic coupling mechanisms. Using the Kaggle Credit Card Fraud Detection (CCFD) dataset, we mitigate data imbalance via the Synthetic Minority Oversampling Technique (SMOTE) and transform sample feature vectors into matrices for training. Experimental results show that our method achieves an accuracy of 0.9998, precision of 0.9996, recall of 1.0000, and an F1-score of 0.9998, surpassing traditional machine learning models, which highlight CCNN’s potential to enhance the security and efficiency of fraud detection in the financial industry. Full article
Show Figures

Figure 1

Figure 1
<p>Comparison before and after dataset balancing. (<b>a</b>) original; (<b>b</b>) after SOMTE.</p>
Full article ">Figure 2
<p>CCNN neuron schematic.</p>
Full article ">Figure 3
<p>CCNN model for credit card fraudulent detection.</p>
Full article ">Figure 4
<p>CCNN model training results.</p>
Full article ">Figure 5
<p>Confusion matrix.</p>
Full article ">
9 pages, 1106 KiB  
Article
Automatic Movement Recognition for Evaluating the Gross Motor Development of Infants
by Yin-Zhang Yang, Jia-An Tsai, Ya-Lan Yu, Mary Hsin-Ju Ko, Hung-Yi Chiou, Tun-Wen Pai and Hui-Ju Chen
Children 2025, 12(3), 310; https://doi.org/10.3390/children12030310 - 28 Feb 2025
Viewed by 112
Abstract
Objective: The objective of this study was to early-detect gross motor abnormalities through video detection in Taiwanese infants aged 2–6 months. Background: The current diagnosis of infant developmental delays primarily relies on clinical examinations. However, during clinical visits, infants may show atypical behaviors [...] Read more.
Objective: The objective of this study was to early-detect gross motor abnormalities through video detection in Taiwanese infants aged 2–6 months. Background: The current diagnosis of infant developmental delays primarily relies on clinical examinations. However, during clinical visits, infants may show atypical behaviors due to unfamiliar environments, which might not truly reflect their true developmental status. Methods: This study utilized videos of infants recorded in their home environments. Two pediatric neurologists manually annotated these clips to identify whether an infant possessed the characteristics of gross motor delays through an assessment of his/her gross motor movements. Using transfer learning techniques, four pose recognition models, including ViTPose, HRNet, DARK, and UDP, were applied to the infant gross motor dataset. Four machine learning classification models, including random forest, support vector machine, logistic regression, and XGBoost, were used to predict the developmental status of infants. Results: The experimental results of pose estimation and tracking indicate that the ViTPose model provided the best performance for pose recognition. A total of 227 features related to kinematics, motions, and postures were extracted and calculated. A one-way ANOVA analysis revealed 106 significant features that were retained for constructing prediction models. The results show that a random forest model achieved the best performance with an average F1-score of 0.94, a weighted average AUC of 0.98, and an average accuracy of 94%. Full article
(This article belongs to the Section Pediatric Neurology & Neurodevelopmental Disorders)
Show Figures

Figure 1

Figure 1
<p>System architecture.</p>
Full article ">Figure 2
<p>Example of pose estimation annotation. Red bounding box: encompassing the infant’s body; Color lines: movement lines of 17 body joints.</p>
Full article ">Figure 3
<p>Eight joint angle features of the shoulders (a) and (b), elbows (c) and (d), hips (e) and (f), and knees (g) and (h).</p>
Full article ">
24 pages, 945 KiB  
Article
MD-Former: Multiscale Dual Branch Transformer for Multivariate Time Series Classification
by Yanling Du, Shuhao Chu, Jintao Wang, Manli Shi, Dongmei Huang and Wei Song
Sensors 2025, 25(5), 1487; https://doi.org/10.3390/s25051487 - 28 Feb 2025
Viewed by 110
Abstract
Multivariate Time Series Classification (MTSC) is a challenging task in real-world applications. Current approaches emphasize modeling multiscale relationships over time. However, the Multivariate Time Series (MTS) also exhibits multiscale cross-channel relationships. Furthermore, the long-term temporal relationships in time series are difficult to capture. [...] Read more.
Multivariate Time Series Classification (MTSC) is a challenging task in real-world applications. Current approaches emphasize modeling multiscale relationships over time. However, the Multivariate Time Series (MTS) also exhibits multiscale cross-channel relationships. Furthermore, the long-term temporal relationships in time series are difficult to capture. In this paper, we introduce MD-Former, a Multiscale Dual-Branch Attention network leveraging the Transformer architecture to capture multiscale relationships across time and channels for MTSC. In MD-Former, MTS is embedded into 2D vectors using Channel-Patching (CP) to retain channel information. Following this, we develop two branches: the Interlaced Attention Branch (IAB) and the Channel-Independent Attention Branch (CIAB). The IAB facilitates the fusion of information across channels and time, while the CIAB prevents the loss of information resulting from excessive fusion. Both the IAB and CIAB consist of multiple layers, each representing a distinct time scale. Finally, we utilize features from each layer of both IAB and CIAB as inputs to the Multiscale Classification Head (MCH) for feature fusion and classification. Experimental results show that MD-Former achieves performance levels that are comparable to SOTA methods in MTSC. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Figure 1
<p>The overall framework of MD-Former.</p>
Full article ">Figure 2
<p>Illustration for Channel-Patching: (<b>a</b>) Embedding method of previous Transformer-based models: all channels at the same time step are embedded into a vector. (<b>b</b>) Channel-Patching: in different channels, neighboring time steps are embedded into a vector, either overlapping or non overlapping.</p>
Full article ">Figure 3
<p>Relationship modeling of different data embedding methods.</p>
Full article ">Figure 4
<p>Interlaced Attention Branch.</p>
Full article ">Figure 5
<p>Interlaced Attention Branch.</p>
Full article ">Figure 6
<p>Multiscale Classification Head.</p>
Full article ">Figure 7
<p>The CD (Critical Difference) plot for 12 multivariate time series classification methods across 30 datasets, with a confidence level of 0.95. Thick horizontal lines indicate groups of methods that do not have a significant difference.</p>
Full article ">Figure 8
<p>A “standing” instance from the StandWalkJump dataset. The blue lines represent the actual data, with each line corresponding to a different channel (4 channels in total). The background color represents the attention scores of the last layer in the IAB, with darker shades indicating higher scores.</p>
Full article ">Figure 9
<p>A “walking” instance from the StandWalkJump dataset. (<b>a</b>) Original data, which consist of four channels. (<b>b</b>) Attention scores from different layers of IAB. Time scale increases from top to bottom.</p>
Full article ">
26 pages, 4362 KiB  
Article
EQLC-EC: An Efficient Voting Classifier for 1D Mass Spectrometry Data Classification
by Lin Guo, Yinchu Wang, Zilong Liu, Fengyi Zhang, Wei Zhang and Xingchuang Xiong
Electronics 2025, 14(5), 968; https://doi.org/10.3390/electronics14050968 - 28 Feb 2025
Viewed by 107
Abstract
Mass spectrometry (MS) data present challenges for machine learning (ML) classification due to their high dimensionality, complex feature distributions, batch effects, and intensity discrepancies, often hindering model generalization and efficiency. To address these issues, this study introduces the Efficient Quick 1D Lite Convolutional [...] Read more.
Mass spectrometry (MS) data present challenges for machine learning (ML) classification due to their high dimensionality, complex feature distributions, batch effects, and intensity discrepancies, often hindering model generalization and efficiency. To address these issues, this study introduces the Efficient Quick 1D Lite Convolutional Neural Network (CNN) Ensemble Classifier (EQLC-EC), integrating 1D convolutional networks with reshape layers and dual voting mechanisms for enhanced feature representation and classification performance. Validation was performed on five publicly available MS datasets, each featured in high-impact publications. EQLC-EC underwent comprehensive evaluation against classical machine learning (ML) models (e.g., support vector machine (SVM), random forest) and the leading deep learning methods reported in these studies. EQLC-EC demonstrated dataset-specific improvements, including enhanced classification accuracy (1–5% increase) and reduced standard deviation (1–10% reduction). Performance differences between soft and hard voting mechanisms were negligible (<1% variation in accuracy and standard deviation). EQLC-EC presents a powerful and efficient tool for MS data analysis with potential applications across metabolomics and proteomics. Full article
Show Figures

Figure 1

Figure 1
<p>One-Dimensional Mass Spectrometry Data: Sources, Characteristics, and Challenges: (<b>a</b>) This figure shows the sources of mass spectrometry data from various fields, including environmental science, metabolomics, single-cell analysis, and medicine. These sources generate a large amount of diverse one-dimensional mass spectrometry data, encompassing a wide range of samples from aerosol particles to blood and urine. (<b>b</b>) One-dimensional mass spectrum and its three characteristics: Extreme range, highly scattered, and high-dimensional.</p>
Full article ">Figure 2
<p>Sample distribution of the development dataset CHD. (<b>a</b>) Distribution of CHD patient samples and control group samples across the four batches. The sectors corresponding to CHD patients are shown in their original colors, while the sectors representing the control group samples in the corresponding batch are shown with 40% transparency. Different batches represent variations in experimental conditions or sample sources. (<b>b</b>) The CHD dataset’s training set consists of the sum of data from two randomly selected batches, and the test set consists of one of the other two batches. There are twelve possible data partitioning methods, and this figure shows two of them. Split 1’s training set is the sum of all data from batch 1 and batch 2, and its test set is the data from batch 3. Split 2’s training set is the sum of data from batch 3 and batch 4, and its test set is the data from batch 2. These two partitioning methods are examples of random sampling. The sector colors in (<b>a</b>,<b>b</b>) are consistent to facilitate understanding of the data partitioning method.</p>
Full article ">Figure 3
<p>Sample distributions of the validation datasets: (<b>a</b>) MI, this dataset shares a similar structure with the CHD dataset in <a href="#electronics-14-00968-f002" class="html-fig">Figure 2</a>, featuring three batches. Data partitioning also follows the same approach: The sum of two batches forms the training set, while the remaining batch constitutes the test set, resulting in three possible partitioning methods. This dataset is used to evaluate the model’s applicability to datasets exhibiting batch effects. (<b>b</b>) ICC, the test set accounts for 20% of the total data. (<b>c</b>) HIP_CER, the test set accounts for 20% of the total data. (<b>d</b>) TOMATO, the test set accounts for 29% of the total data. Since each sample in the TOMATO dataset underwent 10 data acquisitions, a 30% split was used to create the test set, yielding 470 data points, in order to prevent data leakage. This approach ensures that the test set’s sample distribution closely aligns with that of the entire dataset.</p>
Full article ">Figure 4
<p>Network architecture of the EQLC base model. The input is One-Dimensional Mass Spectrometry Input (1DMS INPUT). Key components include convolutional layers Convolutional Layer 1 (env1/conv1) and Convolutional Layer 2 (env2/conv2), followed by max-pooling layers Max Pooling Layer 1 (maxpool-1) and Max Pooling Layer 2 (maxpool-2), and Rectified Linear Unit (ReLU) activation functions. Dropout 0.2 indicates a Dropout layer with a 0.2 dropout rate for regularization. Feature processing concludes with a flatten layer and Fully Connected layers Fully Connected Layer 1 (FC1) and Fully Connected Layer 2 (FC2). The output Classification employs Hard Voting (HardVoting) and Soft Voting (SoftVoting) for ensemble decisions, providing class label and class probability (proba), respectively.</p>
Full article ">Figure 5
<p>Structure of the EQLC-EC ensemble model.</p>
Full article ">Figure 6
<p>Test results corresponding to 12 dataset partitioning methods in the CHD dataset. Taking 12_3 as an example, it represents that the training set is composed of the sum of batch 1 and batch 2, and the test set is batch 3. Box plots show the F1-scores of two voting methods, EQLC and EQLC-EC. Each plot contains 100 F1-scores, derived from 100 test results of the test set reconstructed by the bootstrap method. The narrower the box body of the box plot, the more stable the model; the higher the horizontal line of the median, the better the classification performance of the model. Bar charts show the standard deviation of 100 F1-scores. The shorter the bar, the better the stability of the model.</p>
Full article ">Figure 7
<p>MI dataset results. Test result comparison between EQLC-EC and methods reported in high-impact studies. The three bar graphs represent algorithm performance variability across test sets, with shaded areas included for visual clarity only. Dashed lines denote the mean values reported in the literature, while solid circles and hollow circles represent EQLC-EC’s mean performance.</p>
Full article ">Figure 8
<p>Box plots depict the F1-score performance comparison of EQLC-EC with other methods on the ICC dataset. The blue XGBoost method represents the current state-of-the-art (SOTA) method. Red boxes denote the two voting modes of EQLC-EC. Bar charts illustrate the standard deviation of F1-scores for EQLC-EC and other methods on the ICC dataset.</p>
Full article ">Figure 9
<p>Box plots depict the F1-score performance comparison of EQLC-EC with other methods on the HIP_CER dataset. The blue XGBoost method represents the current state-of-the-art (SOTA) method. Red boxes denote the two voting modes of EQLC-EC. Bar charts illustrate the standard deviation of F1-scores for EQLC-EC and other methods on the HIP_CER dataset.</p>
Full article ">Figure 10
<p>For the TOMATO dataset, the state-of-the-art (SOTA) method in the literature is Random Forest, and only its best classification accuracy is available, which is represented as a dashed line in the figure. Therefore, the comparison of methods on the TOMATO dataset can only be conducted through accuracy. Box plots represent accuracy, and bar charts represent the standard deviation of accuracy.</p>
Full article ">
17 pages, 13551 KiB  
Article
Lithology Identification of Buried Hill Reservoir Based on XGBoost with Optimized Interpretation
by Bin Zhao and Wenlong Liao
Processes 2025, 13(3), 682; https://doi.org/10.3390/pr13030682 - 27 Feb 2025
Viewed by 89
Abstract
Buried hill reservoirs are characterized by complex formation conditions and highly heterogeneous rock structures, which result in the poor performance of traditional crossplot methods in stratigraphic lithology classification. Logging curves, as comprehensive reflections of various petrophysical properties, are influenced by complex geological factors, [...] Read more.
Buried hill reservoirs are characterized by complex formation conditions and highly heterogeneous rock structures, which result in the poor performance of traditional crossplot methods in stratigraphic lithology classification. Logging curves, as comprehensive reflections of various petrophysical properties, are influenced by complex geological factors, leading to overlapping response values even among different lithologies with similar physical properties. This overlap negatively impacts the accuracy of intelligent lithology identification methods. To address this challenge, this study leverages logging response data, experimental data, and mud logging data to propose an optimized inversion method for mineral content, introducing mineral curves to resolve the curve overlap issue. By analyzing six wells in the study area, models were constructed using the calculated mineral content curves and conventional logging features to mitigate the feature overlap. The XGBoost algorithm was employed to identify lithologies by addressing the nonlinear relationships inherent in complex reservoirs. The experimental results indicate that the optimized mineral curves significantly enhance the model’s discriminative capability, effectively addressing the decline in identification accuracy due to feature overlap. Compared to models such as Random Forest (RF) and Support Vector Machine (SVM), the XGBoost model demonstrated superior accuracy and stability, providing a reliable basis for precise reservoir identification in the study area. Full article
(This article belongs to the Section Energy Systems)
Show Figures

Figure 1

Figure 1
<p>XGBoost model flowchart.</p>
Full article ">Figure 2
<p>Relationship between lithology and mineral content in the study area.</p>
Full article ">Figure 3
<p>Relationship between quartz, potassium feldspar, and plagioclase content in the buried hill.</p>
Full article ">Figure 4
<p>Relationship between quartz, feldspar, and clay content in granite buried hill.</p>
Full article ">Figure 5
<p>Mineral Content calculation results for Well H1.</p>
Full article ">Figure 6
<p>Overview of the study area.</p>
Full article ">Figure 7
<p>Typical logging response curve characteristics of different lithologies in the study area.</p>
Full article ">Figure 8
<p>Gamma and density crossplot for different lithologies.</p>
Full article ">Figure 9
<p>Gamma and neutron crossplot for different lithologies.</p>
Full article ">Figure 10
<p>Density–sonic crossplot of different lithologies in the study area.</p>
Full article ">Figure 11
<p>Optimal parameter combination for the model obtained through optimization and cross-validation.</p>
Full article ">Figure 12
<p>t-SNE visualization of the distribution of lithology samples: (<b>a</b>) dataset distribution without mineral curve data, (<b>b</b>) dataset distribution with mineral curve data.</p>
Full article ">Figure 13
<p>Comparison of confusion matrices for model prediction results with and without mineral content curves: (<b>a</b>) model prediction results without using mineral content curves for training, (<b>b</b>) model prediction results using mineral content curves for training.</p>
Full article ">Figure 14
<p>Lithology identification results for Well H2.</p>
Full article ">
26 pages, 7963 KiB  
Article
Pig Face Open Set Recognition and Registration Using a Decoupled Detection System and Dual-Loss Vision Transformer
by Ruihan Ma, Hassan Ali, Malik Muhammad Waqar, Sang Cheol Kim and Hyongsuk Kim
Animals 2025, 15(5), 691; https://doi.org/10.3390/ani15050691 - 27 Feb 2025
Viewed by 153
Abstract
Effective pig farming relies on precise and adaptable animal identification methods, particularly in dynamic environments where new pigs are regularly added to the herd. However, pig face recognition is challenging due to high individual similarity, lighting variations, and occlusions. These factors hinder accurate [...] Read more.
Effective pig farming relies on precise and adaptable animal identification methods, particularly in dynamic environments where new pigs are regularly added to the herd. However, pig face recognition is challenging due to high individual similarity, lighting variations, and occlusions. These factors hinder accurate identification and monitoring. To address these issues under Open-Set conditions, we propose a three-phase Pig Face Open-Set Recognition (PFOSR) system. In the Training Phase, we adopt a decoupled design, first training a YOLOv8-based pig face detection model on a small labeled dataset to automatically locate pig faces in raw images. We then refine a Vision Transformer (ViT) recognition model via a dual-loss strategy—combining Sub-center ArcFace and Center Loss—to enhance both inter-class separation and intra-class compactness. Next, in the Known Pig Registration Phase, we utilize the trained detection and recognition modules to extract representative embeddings from 56 identified pigs, storing these feature vectors in a Pig Face Feature Gallery. Finally, in the Unknown and Known Pig Recognition and Registration Phase, newly acquired pig images are processed through the same detection–recognition pipeline, and the resulting embeddings are compared against the gallery via cosine similarity. If the system classifies a pig as unknown, it dynamically assigns a new ID and updates the gallery without disrupting existing entries. Our system demonstrates strong Open-Set recognition, achieving an AUROC of 0.922, OSCR of 0.90, and F1-Open of 0.94. In the closed set, it attains a precision@1 of 0.97, NMI of 0.92, and mean average precision@R of 0.96. These results validate our approach as a scalable, efficient solution for managing dynamic farm environments with high recognition accuracy, even under challenging conditions. Full article
Show Figures

Figure 1

Figure 1
<p>Overview of the proposed three-phase PFOSR pipeline. In the Training Phase, labeled pig images (image + label: pig + ID) are used to develop a robust detection model and a recognition model featuring a dual-loss design (SubCenterArcFace + Center Loss). In the Known Pig Registration Phase, images accompanied by known pig IDs pass through the pig face detection and recognition modules, and the resulting feature embeddings are registered in a Face Gallery. In the Unknown and Known Pig Recognition and Registration Phase, unlabeled images are again processed by the same detection and recognition models; if, for a new embedding, the similarity score with all existing gallery entries falls below a specified threshold, a new pig ID is assigned, and the Face Gallery is updated accordingly. This iterative process enables Open-Set pig face recognition by dynamically integrating newly encountered pigs.</p>
Full article ">Figure 2
<p>This is the application of our PFOSR system in inference time. During the real-time detection phase of PFOSR, all unknown and known pigs have already been registered. Incoming test images are first processed by the pig face detection and recognition model to extract face features, which are then matched in a 1:N manner against all registered pigs in the feature gallery. The system retrieves the top five matches based on similarity scores and assigns the label of the highest-scoring match to the test image, completing the real-time identification process.</p>
Full article ">Figure 3
<p>The visualization of the sample images of the Small-Scale Pig Face Detection Dataset.</p>
Full article ">Figure 4
<p>The visualization of the sample images of the Unknown Pig Face Test Dataset.</p>
Full article ">Figure 5
<p>Pig Face Recognition Module in our proposed PFOSR System. This figure illustrates the pig face recognition pipeline in our PFOSR system. The input image is resized to 224 × 224 and processed through a ViT-based backbone for feature extraction. The extracted features are refined through an embedding layer consisting of linear layers, batch normalization (BN), ReLU activation, and dropout, producing the final embedding vector. BS (Batch Size) represents the number of images processed at once, while ES (Embedding Size) refers to the dimensionality of the output feature vector.</p>
Full article ">Figure 6
<p>YOLOv8 Performance on Labeled and Unlabeled Datasets. This figure illustrates YOLOv8’s pig face detection across two datasets. In the first row, images from the Small-Scale Pig Face Detection Dataset are shown, with red boxes indicating predictions and blue boxes representing ground truth labels. The second row displays images from the Known Pig Face Recognition Dataset, where red boxes denote YOLOv8’s predictions without ground truth labels. This comparison highlights the model’s effectiveness in both labeled and unlabeled scenarios.</p>
Full article ">Figure 7
<p>F1-Open, CCR and FAR curves of ViT-DL-IN21K model at different thresholds in PFOSR.</p>
Full article ">Figure 8
<p>Visualization of ViT-DL-IN21K Model Performance on the 65 known Pig Face Testing Dataset. Each panel displays a test image with its true ID, followed by the top five most similar gallery images with their cosine similarity scores and true IDs. High similarity scores indicate correct matches for known classes, including newly registered ones.</p>
Full article ">Figure 9
<p>UMAP Visualization of Test Features for Model Performance on Dataset1 and Dataset2 in PFOSR System. This figure presents the UMAP projections of test feature embeddings for three models (ResNet18-DL-IN21K, ResNet50-DL-IN21K, and ViT-DL-IN21K) on two datasets. (<b>A</b>–<b>C</b>) correspond to Dataset1 (Known Pig Face Recognition Dataset test set, 56 known classes), while (<b>D</b>–<b>F</b>) correspond to Dataset2 (65-Known Pig Face Testing Dataset, including 56 known classes and 9 newly registered unknown classes). Each color represents a different pig identity. A more compact and well-separated clustering indicates better feature representation. ViT-DL-IN21K (<b>C</b>,<b>F</b>) shows improved feature clustering, demonstrating its superior ability to distinguish different pig identities, especially in Dataset2, where new classes have been introduced.</p>
Full article ">Figure 10
<p>Confusion Matrices for Models Performance on testing Dataset1 and testing Dataset2 in PFCSR System. This figure presents the confusion matrices for three models (ResNet18-DL-IN21K, ResNet50-DL-IN21K, and ViT-DL-IN21K) evaluated on two test datasets. (<b>A</b>–<b>C</b>) correspond to Dataset1 (Known Pig Face Recognition Dataset test set, 56 known classes), while (<b>D</b>–<b>F</b>) correspond to Dataset2 (65-Known Pig Face Testing Dataset, including 56 known classes and 9 newly registered unknown classes). The diagonal elements represent correct classifications, while off-diagonal elements indicate misclassification. Comparing these matrices, ViT-DL-IN21K (<b>C</b>,<b>F</b>) achieves the lowest misclassification rates, as seen by the darker diagonal and lighter off-diagonal cells. This demonstrates its superior performance in both standard Closed-Set recognition (Dataset1) and recognizing newly registered pigs in Dataset2, proving its robustness and adaptability in pig face recognition.</p>
Full article ">
Back to TopTop