Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,132)

Search Parameters:
Keywords = deep metric learning

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 5484 KiB  
Article
LMPSeizNet: A Lightweight Multiscale Pyramid Convolutional Neural Network for Epileptic Seizure Detection on EEG Brain Signals
by Arwa Alsaadan, Mai Alzamel and Muhammad Hussain
Mathematics 2024, 12(23), 3648; https://doi.org/10.3390/math12233648 - 21 Nov 2024
Abstract
Epilepsy is a chronic disease and one of the most common neurological disorders worldwide. Electroencephalogram (EEG) signals are widely used to detect epileptic seizures, which provide specialists with essential information about the brain’s functioning. However, manual screening of EEG signals is laborious, time-consuming, [...] Read more.
Epilepsy is a chronic disease and one of the most common neurological disorders worldwide. Electroencephalogram (EEG) signals are widely used to detect epileptic seizures, which provide specialists with essential information about the brain’s functioning. However, manual screening of EEG signals is laborious, time-consuming, and subjective. The rapid detection of epilepsy seizures is important to reduce the risk of seizure-related implications. The existing automatic machine learning techniques based on deep learning techniques are characterized by automatic extraction and selection of the features, leading to better performance and increasing the robustness of the systems. These methods do not consider the multiscale nature of EEG signals, eventually resulting in poor sensitivity. In addition, the complexity of deep models is relatively high, leading to overfitting issues. To overcome these problems, we proposed an efficient and lightweight multiscale convolutional neural network model (LMPSeizNet), which performs multiscale temporal and spatial analysis of an EEG trial to learn discriminative features relevant to epileptic seizure detection. To evaluate the proposed method, we employed 10-fold cross-validation and three evaluation metrics: accuracy, sensitivity, and specificity. The method achieved an accuracy of 97.42%, a sensitivity of 99.33%, and a specificity of 96.51% for inter-ictal and ictal classes outperforming the state-of-the-art methods. The analysis of the features and the decision-making of the method shows that it learns the features that clearly discriminate the two classes. It will serve as a useful tool for helping neurologists and epilepsy patients. Full article
(This article belongs to the Special Issue Signal Processing and Machine Learning in Real-Life Processes)
Show Figures

Figure 1

Figure 1
<p>Seizure detection problem specification, where 1 and 2 are labels that stand for inter-ictal and ictal classes. Here, <math display="inline"><semantics> <mi mathvariant="bold-italic">x</mi> </semantics></math> is an EEG signal epoch to be passed to a seizure detection method (<math display="inline"><semantics> <mi mathvariant="script">H</mi> </semantics></math>) for its classification.</p>
Full article ">Figure 2
<p>An example EEG epoch showing that each of its channels is the superposition of several frequencies along the time dimension, and it is necessary to analyze it using multiscale convolution. Each EGG channel represent in different color.</p>
Full article ">Figure 3
<p>The architecture of LMPSeizNet; it takes an EEG trial as input and processes it using three main modules, defining the mappings of <math display="inline"><semantics> <mrow> <msup> <mi>ϕ</mi> <mn>1</mn> </msup> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <msup> <mi>ϕ</mi> <mn>2</mn> </msup> </mrow> </semantics></math>, and <math display="inline"><semantics> <mrow> <msup> <mi>ϕ</mi> <mn>3</mn> </msup> </mrow> </semantics></math>, and yields the predicted label <math display="inline"><semantics> <mover accent="true"> <mi>y</mi> <mo>^</mo> </mover> </semantics></math>. Where (<b>a</b>) is the first temporal convolution block, (<b>b</b>) is temporal separable convolutional block, (<b>c</b>) is spatial convolutional block, (<b>d</b>) is the classifier.</p>
Full article ">Figure 4
<p>Results of Experiment 1 showing the effect of different choices of filter sizes <math display="inline"><semantics> <mrow> <mo stretchy="false" mathvariant="bold">[</mo> <msub> <mi mathvariant="bold-italic">T</mi> <mn mathvariant="bold">0</mn> </msub> </mrow> </semantics></math>,<math display="inline"><semantics> <mrow> <msub> <mi mathvariant="bold-italic">T</mi> <mn mathvariant="bold">1</mn> </msub> <mo>,</mo> <msub> <mi mathvariant="bold-italic">T</mi> <mn mathvariant="bold">2</mn> </msub> <mo>,</mo> <msub> <mi mathvariant="bold-italic">T</mi> <mn mathvariant="bold">3</mn> </msub> <mo stretchy="false">]</mo> </mrow> </semantics></math>. The experiments were performed with the channels {FP1-F7, F7-T7, T7-P7} and a fixed number of filters.</p>
Full article ">Figure 5
<p>Results of Experiment 2 depicting the effects of varying numbers of filters [<math display="inline"><semantics> <mrow> <msub> <mi>K</mi> <mn>0</mn> </msub> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <msub> <mi>K</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>K</mi> <mn>2</mn> </msub> <mo>,</mo> <mo> </mo> <msub> <mi>K</mi> <mn>3</mn> </msub> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <msub> <mi>K</mi> <mn>4</mn> </msub> <mo>,</mo> <msub> <mi>K</mi> <mn>5</mn> </msub> <mo>,</mo> <mo> </mo> <msub> <mi>K</mi> <mn>6</mn> </msub> </mrow> </semantics></math>] of each block [TConv, TSepConv-1, TSepConv-2, TSepConv-3, SConv-1, SConv-2, SConv-3]. The experiments were performed with the channels {FP1-F7, F7-T7, T7-P7} and fixed filter sizes.</p>
Full article ">Figure 6
<p>The results of Experiment 3 showing the effects of different combinations of channels. The experiments were performed with fixed numbers of filters and sizes.</p>
Full article ">Figure 7
<p>Results of Experiment 4 showing the effects of different stride values for sampling the EEG trials for each class. The experiments were performed with the channels {F7-T7, F3-C3, FT9-FT10} and fixed filter sizes and number of filters.</p>
Full article ">Figure 8
<p>Visualization of the extracted features of the two classes (inter-ictal and ictal) using a t−SNE plot. The clustering of features belonging to each indicates that the two classes have a large inter-class difference, and the learned features are discriminative.</p>
Full article ">Figure 9
<p>Confusion matrix of the best configuration of LMPSeizNet with channels {F7-T7, F3-C3, FT9-FT10}. It shows the potential of the model to classify each class.</p>
Full article ">
32 pages, 7730 KiB  
Article
High-Fidelity Infrared Remote Sensing Image Generation Method Coupled with the Global Radiation Scattering Mechanism and Pix2PixGAN
by Yue Li, Xiaorui Wang, Chao Zhang, Zhonggen Zhang and Fafa Ren
Remote Sens. 2024, 16(23), 4350; https://doi.org/10.3390/rs16234350 - 21 Nov 2024
Abstract
To overcome the problems in existing infrared remote sensing image generation methods, which make it difficult to combine high fidelity and high efficiency, we propose a High-Fidelity Infrared Remote Sensing Image Generation Method Coupled with the Global Radiation Scattering Mechanism and Pix2PixGAN (HFIRSIGM_GRSMP) [...] Read more.
To overcome the problems in existing infrared remote sensing image generation methods, which make it difficult to combine high fidelity and high efficiency, we propose a High-Fidelity Infrared Remote Sensing Image Generation Method Coupled with the Global Radiation Scattering Mechanism and Pix2PixGAN (HFIRSIGM_GRSMP) in this paper. Firstly, based on the global radiation scattering mechanism, the HFIRSIGM_GRSMP model is constructed to address the problem of accurately characterizing factors that affect fidelity—such as the random distribution of the radiation field, multipath scattering, and nonlinear changes—through the innovative fusion of physical models and deep learning. This model accurately characterizes the complex radiation field distribution and the image detail-feature mapping relationship from visible-to-infrared remote sensing. Then, 8000 pairs of image datasets were constructed based on Landsat 8 and Sentinel-2 satellite data. Finally, the experiment demonstrates that the average SSIM of images generated using HFIRSIGM_GRSMP reaches 89.16%, and all evaluation metrics show significant improvement compared to the contrast models. More importantly, this method demonstrates high accuracy and strong adaptability in generating short-wave, mid-wave, and long-wave infrared remote sensing images. This method provides a more comprehensive solution for generating high-fidelity infrared remote sensing images. Full article
(This article belongs to the Special Issue Deep Learning Innovations in Remote Sensing)
21 pages, 6945 KiB  
Article
Automatic Segmentation and Statistical Analysis of the Foveal Avascular Zone
by Geanina Totolici, Mihaela Miron and Anisia-Luiza Culea-Florescu
Technologies 2024, 12(12), 235; https://doi.org/10.3390/technologies12120235 - 21 Nov 2024
Abstract
This study facilitates the extraction of foveal avascular zone (FAZ) metrics from optical coherence tomography angiography (OCTA) images, offering valuable clinical insights and enabling detailed statistical analysis of FAZ size and shape across three patient groups: healthy, type II diabetes mellitus and both [...] Read more.
This study facilitates the extraction of foveal avascular zone (FAZ) metrics from optical coherence tomography angiography (OCTA) images, offering valuable clinical insights and enabling detailed statistical analysis of FAZ size and shape across three patient groups: healthy, type II diabetes mellitus and both type II diabetes mellitus (DM) and high blood pressure (HBP). Additionally, it evaluates the performance of four deep learning (DL) models—U-Net, U-Net with DenseNet121, U-Net with MobileNetV2 and U-Net with VGG16—in automating the segmentation of the FAZ. Manual segmentation of the images by ophthalmological clinicians was performed initially, and data augmentation was used to enhance the dataset for robust model training and evaluation. Consequently, the original set of 103 full retina OCTA images was extended to 672 cases, including 42 images from normal patients, 357 images from DM patients, and 273 images from patients with both DM and HBP. Among the models, U-Net with DenseNet outperformed the others, achieving the highest accuracy, Intersection over Union (IoU), and Dice coefficient across all patient groups. This research is distinct in its focus on full retina OCTA images and its inclusion of patients with both hypertension and diabetes, an area that is less studied in the existing literature. Full article
(This article belongs to the Special Issue Medical Imaging & Image Processing III)
Show Figures

Figure 1

Figure 1
<p>Block diagram of the proposed framework.</p>
Full article ">Figure 2
<p>The extraction of the ground-truth mask of FAZ.</p>
Full article ">Figure 3
<p>U-Net model with encoder highlighted by dashed box.</p>
Full article ">Figure 4
<p>U-Net variant with DenseNet121 model, backbone highlighted by dashed box.</p>
Full article ">Figure 5
<p>U-Net variant with MobileNetV2, backbone highlighted by dashed box.</p>
Full article ">Figure 6
<p>U-Net variant with VGG16, backbone highlighted by dashed box.</p>
Full article ">Figure 7
<p>Boxplot analysis of FAZ metrics across patient groups with type II diabetes and type II diabetes + high blood pressure: (<b>a</b>) perimeter (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mn>1</mn> </mrow> </msub> </mrow> </semantics></math>); (<b>b</b>) area (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>A</mi> </mrow> <mrow> <mn>1</mn> </mrow> </msub> </mrow> </semantics></math>).</p>
Full article ">Figure 8
<p>Boxplot analysis of FAZ metrics across patient groups with type II diabetes and type II diabetes + high blood pressure: (<b>a</b>) equivalent circle perimeter (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mn>2</mn> </mrow> </msub> </mrow> </semantics></math>); (<b>b</b>) acircularity index (<span class="html-italic">A_index</span>).</p>
Full article ">Figure 9
<p>Boxplot analysis of FAZ metrics across patient groups with type II diabetes and type II diabetes + high blood pressure: (<b>a</b>) angle (θ); (<b>b</b>) axis ratio (<span class="html-italic">A_ratio</span>).</p>
Full article ">Figure 10
<p>Relationship between central vascular density (DVF C) and (<b>a</b>) perimeter (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mn>1</mn> </mrow> </msub> </mrow> </semantics></math>); (<b>b</b>) area (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>A</mi> </mrow> <mrow> <mn>1</mn> </mrow> </msub> </mrow> </semantics></math>).</p>
Full article ">Figure 11
<p>Relationship between central vascular density (DVF C) and (<b>a</b>) equivalent circle perimeter (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>P</mi> </mrow> <mrow> <mn>2</mn> </mrow> </msub> </mrow> </semantics></math>); (<b>b</b>) acircularity index (<math display="inline"><semantics> <mrow> <mi>A</mi> <mo>_</mo> <mi>i</mi> <mi>n</mi> <mi>d</mi> <mi>e</mi> <mi>x</mi> </mrow> </semantics></math>).</p>
Full article ">Figure 12
<p>Relationship between central vascular density (DVF C) and (<b>a</b>) angle (θ); (<b>b</b>) axis ratio (A_ratio).</p>
Full article ">Figure 13
<p>Training and validation accuracy for normal group: (<b>a</b>) U-Net; (<b>b</b>) U-Net variant with DenseNet121; (<b>c</b>) U-Net variant with MobileNetV2; (<b>d</b>) U-Net variant with VGG16.</p>
Full article ">Figure 14
<p>Training and validation accuracy for type II DM group: (<b>a</b>) U-Net; (<b>b</b>) U-Net variant with DenseNet121; (<b>c</b>) U-Net variant with MobileNetV2; (<b>d</b>) U-Net variant with VGG16.</p>
Full article ">Figure 14 Cont.
<p>Training and validation accuracy for type II DM group: (<b>a</b>) U-Net; (<b>b</b>) U-Net variant with DenseNet121; (<b>c</b>) U-Net variant with MobileNetV2; (<b>d</b>) U-Net variant with VGG16.</p>
Full article ">Figure 15
<p>Training and validation accuracy for type II DM + HBP group: (<b>a</b>) U-Net; (<b>b</b>) U-Net variant with DenseNet121; (<b>c</b>) U-Net variant with MobileNetV2; (<b>d</b>) U-Net variant with VGG16.</p>
Full article ">Figure 15 Cont.
<p>Training and validation accuracy for type II DM + HBP group: (<b>a</b>) U-Net; (<b>b</b>) U-Net variant with DenseNet121; (<b>c</b>) U-Net variant with MobileNetV2; (<b>d</b>) U-Net variant with VGG16.</p>
Full article ">
27 pages, 28012 KiB  
Article
A Model Development Approach Based on Point Cloud Reconstruction and Mapping Texture Enhancement
by Boyang You and Barmak Honarvar Shakibaei Asli
Big Data Cogn. Comput. 2024, 8(11), 164; https://doi.org/10.3390/bdcc8110164 - 20 Nov 2024
Viewed by 231
Abstract
To address the challenge of rapid geometric model development in the digital twin industry, this paper presents a comprehensive pipeline for constructing 3D models from images using monocular vision imaging principles. Firstly, a structure-from-motion (SFM) algorithm generates a 3D point cloud from photographs. [...] Read more.
To address the challenge of rapid geometric model development in the digital twin industry, this paper presents a comprehensive pipeline for constructing 3D models from images using monocular vision imaging principles. Firstly, a structure-from-motion (SFM) algorithm generates a 3D point cloud from photographs. The feature detection methods scale-invariant feature transform (SIFT), speeded-up robust features (SURF), and KAZE are compared across six datasets, with SIFT proving the most effective (matching rate higher than 0.12). Using K-nearest-neighbor matching and random sample consensus (RANSAC), refined feature point matching and 3D spatial representation are achieved via antipodal geometry. Then, the Poisson surface reconstruction algorithm converts the point cloud into a mesh model. Additionally, texture images are enhanced by leveraging a visual geometry group (VGG) network-based deep learning approach. Content images from a dataset provide geometric contours via higher-level VGG layers, while textures from style images are extracted using the lower-level layers. These are fused to create texture-transferred images, where the image quality assessment (IQA) metrics SSIM and PSNR are used to evaluate texture-enhanced images. Finally, texture mapping integrates the enhanced textures with the mesh model, improving the scene representation with enhanced texture. The method presented in this paper surpassed a LiDAR-based reconstruction approach by 20% in terms of point cloud density and number of model facets, while the hardware cost was only 1% of that associated with LiDAR. Full article
Show Figures

Figure 1

Figure 1
<p>Samples from Dataset 1 (Source: <a href="https://github.com/Abhishek-Aditya-bs/MultiView-3D-Reconstruction/tree/main/Datasets" target="_blank">https://github.com/Abhishek-Aditya-bs/MultiView-3D-Reconstruction/tree/main/Datasets</a> accessed on 18 November 2024) and samples from Dataset 2.</p>
Full article ">Figure 2
<p>Demonstration of Dataset 3.</p>
Full article ">Figure 3
<p>Diagram of SFM algorithm.</p>
Full article ">Figure 4
<p>Camera imaging model.</p>
Full article ">Figure 5
<p>Coplanarity condition of photogrammetry.</p>
Full article ">Figure 6
<p>Process of surface reconstruction.</p>
Full article ">Figure 7
<p>Demonstration of isosurface.</p>
Full article ">Figure 8
<p>Demonstration of VGG network.</p>
Full article ">Figure 9
<p>Demonstration of Gram matrix.</p>
Full article ">Figure 10
<p>Style transformation architecture.</p>
Full article ">Figure 11
<p>Texture mapping process.</p>
Full article ">Figure 12
<p>Demonstration of the three kinds of feature descriptors used on Dataset 1 and Dataset 2.</p>
Full article ">Figure 13
<p>Matching rate fitting of three kinds of image descriptors.</p>
Full article ">Figure 14
<p>SIFT point matching for <span class="html-italic">CNC1</span> object under different thresholds.</p>
Full article ">Figure 15
<p>SIFT point matching for <span class="html-italic">Fountain</span> object under different thresholds.</p>
Full article ">Figure 16
<p>Matching result of Dataset 2 using RANSAC method.</p>
Full article ">Figure 17
<p>Triangulation presentation of feature points obtained from objects in Dataset 1.</p>
Full article ">Figure 18
<p>Triangulation presentation of feature points obtained from objects in Dataset 2.</p>
Full article ">Figure 19
<p>Point cloud data of objects in Dataset 1.</p>
Full article ">Figure 20
<p>Point cloud data of objects in Dataset 2.</p>
Full article ">Figure 21
<p>Normal vector presentation of the points set obtained from objects in Dataset 1.</p>
Full article ">Figure 22
<p>Normal vector of the points set obtained from objects in Dataset 2.</p>
Full article ">Figure 23
<p>Poisson surface reconstruction results of objects in Dataset 1.</p>
Full article ">Figure 24
<p>Poisson surface reconstruction results of objects in Dataset 2.</p>
Full article ">Figure 25
<p>Style transfer result of <span class="html-italic">Statue</span> object.</p>
Full article ">Figure 26
<p>Style transfer result of <span class="html-italic">Fountain</span> object.</p>
Full article ">Figure 27
<p>Style transfer result of <span class="html-italic">Castle</span> object.</p>
Full article ">Figure 28
<p>Style transfer result of <span class="html-italic">CNC1</span> object.</p>
Full article ">Figure 29
<p>Style transfer result of <span class="html-italic">CNC2</span> object.</p>
Full article ">Figure 30
<p>Style transfer result of <span class="html-italic">Robot</span> object.</p>
Full article ">Figure 31
<p>Training loss in style transfer for <b>CNC1</b> object.</p>
Full article ">Figure 32
<p>IQA assessment for <b>CNC1</b> images after style transfer.</p>
Full article ">Figure 33
<p>Results of texture mapping for Dataset 1.</p>
Full article ">Figure 34
<p>Results of texture mapping for Dataset 2.</p>
Full article ">Figure A1
<p>Results of Camera calibration.</p>
Full article ">
21 pages, 23509 KiB  
Article
Learning Airfoil Flow Field Representation via Geometric Attention Neural Field
by Li Xiao, Mingjie Zhang and Xinghua Chang
Appl. Sci. 2024, 14(22), 10685; https://doi.org/10.3390/app142210685 - 19 Nov 2024
Viewed by 254
Abstract
Numerical simulation in fluid dynamics can be computationally expensive and difficult to achieve. To enhance efficiency, developing high-performance and accurate surrogate models is crucial, where deep learning shows potential. This paper introduces geometric attention (GeoAttention), a method that leverages attention mechanisms to encode [...] Read more.
Numerical simulation in fluid dynamics can be computationally expensive and difficult to achieve. To enhance efficiency, developing high-performance and accurate surrogate models is crucial, where deep learning shows potential. This paper introduces geometric attention (GeoAttention), a method that leverages attention mechanisms to encode geometry represented by point cloud, thereby enhancing the neural network’s generalizability across different geometries. Furthermore, by integrating GeoAttention with neural field, we propose the geometric attention neural field (GeoANF), specifically for learning representations of airfoil flow fields. The GeoANF embeds observational data independently of the specific discretization process into a latent space, constructing a mapping that relates geometric shape to the corresponding flow fields under given initial conditions. We use the public dataset AirfRANS to evaluate our approach, GeoANF significantly surpasses the baseline models on four key performance metrics, particularly in volume flow field and surface pressure measurements, achieving mean squared errors of 0.0038 and 0.0089, respectively. Full article
Show Figures

Figure 1

Figure 1
<p>CFD grid data with pixel representation.</p>
Full article ">Figure 2
<p>Implicit neural representation.</p>
Full article ">Figure 3
<p>Airfoil flow field visualization.</p>
Full article ">Figure 4
<p>Airfoil optimization and GeoANF.</p>
Full article ">Figure 5
<p>GeoAttention neural field overview. (<b>a</b>) provides an overview of GeoANF, including both the inference and training phases. Airfoil boundary points (assumed to consist of <span class="html-italic">m</span> points) and <span class="html-italic">n</span> feature space query points, are each represented in a <math display="inline"><semantics> <msub> <mi>d</mi> <mi>p</mi> </msub> </semantics></math>-dimensional space. These inputs are processed through distinct NN block to the neural representation space <math display="inline"><semantics> <mi mathvariant="script">H</mi> </semantics></math>. GeoAttention processes the query points, in conjunction with the airfoil boundary points, obtain a geometric representation <span class="html-italic">g</span>. Subsequently, we concatenate <span class="html-italic">z</span> and <span class="html-italic">g</span>, which passing the neural field to decode into the flow field. (<b>b</b>) shows that the basic module used by the neural network is three-layer ReLU MLP, which has <math display="inline"><semantics> <msub> <mi>c</mi> <mi>i</mi> </msub> </semantics></math> input and <math display="inline"><semantics> <msub> <mi>c</mi> <mi>o</mi> </msub> </semantics></math> output channels. (<b>c</b>) Shows the process of GeoAttention. A point representation <math display="inline"><semantics> <msub> <mi>z</mi> <mi>i</mi> </msub> </semantics></math>, calculates attention weights with geometric representations of the airfoil boundary points, obtaining <math display="inline"><semantics> <msub> <mi>g</mi> <mi>i</mi> </msub> </semantics></math>. (<b>d</b>) shows the structure of the neural field. The concatenated query point and geometric information representations enter an NN block, are processed by a ReLU MLP with batch normalization, and are then decoded into the flow field output by another NN block.</p>
Full article ">Figure 6
<p>Multi-head attention.</p>
Full article ">Figure 7
<p>GeoAttention 2 situations.</p>
Full article ">Figure 8
<p>GeoANF training process.</p>
Full article ">Figure 9
<p>Flow field comparison between the prediction and CFD.</p>
Full article ">Figure 10
<p>Streamtrace visualization comparison.</p>
Full article ">Figure 11
<p>Spearman correlation visualization (full data regime).</p>
Full article ">Figure 12
<p>Error comparison. Visualization of the error results with the real flow field, from top to bottom, the total error, the error of velocity-x, the error of velocity-y, the error of pressure and the error of <math display="inline"><semantics> <mi>μ</mi> </semantics></math>t.</p>
Full article ">Figure 13
<p>The pressure coefficient distribution on the airfoil surface for two CFD cases.</p>
Full article ">Figure 14
<p>Normalized GeoAttention weight visualization.</p>
Full article ">
18 pages, 1508 KiB  
Article
Adversarial Validation in Image Classification Datasets by Means of Cumulative Spectral Gradient
by Diego Renza, Ernesto Moya-Albor and Adrian Chavarro
Algorithms 2024, 17(11), 531; https://doi.org/10.3390/a17110531 - 19 Nov 2024
Viewed by 257
Abstract
The main objective of a machine learning (ML) system is to obtain a trained model from input data in such a way that it allows predictions to be made on new i.i.d. (Independently and Identically Distributed) data with the lowest possible error. However, [...] Read more.
The main objective of a machine learning (ML) system is to obtain a trained model from input data in such a way that it allows predictions to be made on new i.i.d. (Independently and Identically Distributed) data with the lowest possible error. However, how can we assess whether the training and test data have a similar distribution? To answer this question, this paper presents a proposal to determine the degree of distribution shift of two datasets. To this end, a metric for evaluating complexity in datasets is used, which can be applied in multi-class problems, comparing each pair of classes of the two sets. The proposed methodology has been applied to three well-known datasets: MNIST, CIFAR-10 and CIFAR-100, together with corrupted versions of these. Through this methodology, it is possible to evaluate which types of modification have a greater impact on the generalization of the models without the need to train multiple models multiple times, also allowing us to determine which classes are more affected by corruption. Full article
(This article belongs to the Special Issue Machine Learning Algorithms for Image Understanding and Analysis)
Show Figures

Figure 1

Figure 1
<p>Outline of the proposed adversarial validation methodology.</p>
Full article ">Figure 2
<p>Examples of the 15 corruptions included in the MNIST-C dataset.</p>
Full article ">Figure 3
<p>Examples of the 19 corruptions included in the CIFAR-10-C and CIFAR-100-C datasets.</p>
Full article ">Figure 4
<p>Flowchart diagram of the adversarial validation methodology. The blue lines represent the original datasets, and the red line represents the corrupted datasets.</p>
Full article ">Figure 5
<p>Class-level adversarial validation on the MNIST dataset and its corrupted version (MNIST-C). (<b>a</b>) CSG metric ordered by the average value of all classes (dataset-level adversarial validation). Lower values relate to data that deviate from the original distribution. (<b>b</b>) Variability of the CSG metric between classes in each of the corruption types (corruption-level adversarial validation).</p>
Full article ">Figure 6
<p>Class-level adversarial validation on the CIFAR-10 dataset and its corrupted version (CIFAR-10-C). (<b>a</b>) CSG metric ordered by the average value of all classes. Lower values relate to data that deviate from the original distribution (dataset-level adversarial validation). (<b>b</b>) Variability of the CSG metric between classes in each of the corruption types (corruption-level adversarial validation).</p>
Full article ">Figure 7
<p>Class-level adversarial validation on CIFAR-100 dataset and its corrupted version (CIFAR-100-C). CSG metric ordered by the average value of all classes (dataset-level adversarial validation). Lower values relate to data that deviate from the original distribution.</p>
Full article ">Figure 8
<p>Class-level adversarial validation on CIFAR-100 dataset and its corrupted version (CIFAR-100-C). Variability of the CSG metric between classes in each of the corruption types (corruption-level adversarial validation).</p>
Full article ">Figure 9
<p>Example of a CIFAR image with the four types of modifications that have the greatest impact on the distribution of data (<b>b</b>–<b>d</b>) and the four types of modifications that have the least impact (<b>e</b>–<b>h</b>).</p>
Full article ">
15 pages, 1737 KiB  
Article
The Lightweight Fracture Segmentation Algorithm for Logging Images Based on Fully 3D Attention Mechanism and Deformable Convolution
by Qishun Yang, Liyan Zhang, Zihan Xi, Yu Qian and Ang Li
Appl. Sci. 2024, 14(22), 10662; https://doi.org/10.3390/app142210662 - 18 Nov 2024
Viewed by 359
Abstract
The challenge of fracture segmentation remains a significant obstacle in imaging logging interpretation within the current oil and gas exploration and development field. However, existing image segmentation algorithms still encounter issues related to accuracy, speed, and robustness, as well as a tendency to [...] Read more.
The challenge of fracture segmentation remains a significant obstacle in imaging logging interpretation within the current oil and gas exploration and development field. However, existing image segmentation algorithms still encounter issues related to accuracy, speed, and robustness, as well as a tendency to misdetect or overlook small fractures when applied to logging image fracture segmentation tasks. To address these challenges comprehensively, this paper proposes an end-to-end fracture segmentation algorithm named SWSDS-Net. This algorithm is built upon the UNet architecture and incorporates the SimAM with slicing (SWS) attention mechanism along with the deformable strip convolution (DSCN) module. The SWS introduces a fully 3D attention mechanism that effectively learns the weights of each neuron in the feature map, enabling better capture of fracture features while ensuring fair attention and enhancement for both large and small objects. Additionally, the deformable properties of DSCN allow for adaptive sampling based on fracture shapes, effectively tackling challenges posed by varying fracture shapes and enhancing segmentation robustness. Experimental results demonstrate that SWSDS-Net achieves optimal performance across all evaluation metrics in this task, delivering superior visual results in fracture segmentation while successfully overcoming limitations present in existing algorithms such as complex shapes, noise interference, and low-quality images. Moreover, serving as a lightweight network solution enables SWSDS-Net’s deployment on mobile devices at remote sites—an advancement that lays a solid foundation for interpreting logging data and promotes deep learning technology application within traditional industrial scenarios. Full article
Show Figures

Figure 1

Figure 1
<p>SWSDS-Net network architecture.</p>
Full article ">Figure 2
<p>Comparison of bilinear interpolation and linear interpolation.</p>
Full article ">Figure 3
<p>Comparison of different attentions: (<b>a</b>) channel-wise attention, (<b>b</b>) spatial-wise attention, and (<b>c</b>) full 3-D weights for attention.</p>
Full article ">Figure 4
<p>The segmentation result of logging fracture images: (<b>a</b>) original picture, (<b>b</b>) the result of UNet, (<b>c</b>) the result of UNetPlus, (<b>d</b>) the result of UNet++, and (<b>e</b>) the result of SWSDS-Net.</p>
Full article ">
21 pages, 1716 KiB  
Article
AI-Driven Neuro-Monitoring: Advancing Schizophrenia Detection and Management Through Deep Learning and EEG Analysis
by Elena-Anca Paraschiv, Lidia Băjenaru, Cristian Petrache, Ovidiu Bica and Dragoș-Nicolae Nicolau
Future Internet 2024, 16(11), 424; https://doi.org/10.3390/fi16110424 - 16 Nov 2024
Viewed by 414
Abstract
Schizophrenia is a complex neuropsychiatric disorder characterized by disruptions in brain connectivity and cognitive functioning. Continuous monitoring of neural activity is essential, as it allows for the detection of subtle changes in brain connectivity patterns, which could provide early warnings of cognitive decline [...] Read more.
Schizophrenia is a complex neuropsychiatric disorder characterized by disruptions in brain connectivity and cognitive functioning. Continuous monitoring of neural activity is essential, as it allows for the detection of subtle changes in brain connectivity patterns, which could provide early warnings of cognitive decline or symptom exacerbation, ultimately facilitating timely therapeutic interventions. This paper proposes a novel approach for detecting schizophrenia-related abnormalities using deep learning (DL) techniques applied to electroencephalogram (EEG) data. Using an openly available EEG dataset on schizophrenia, the focus is on preprocessed event-related potentials (ERPs) from key electrode sites and applied transfer entropy (TE) analysis to quantify the directional flow of information between brain regions. TE matrices were generated to capture neural connectivity patterns, which were then used as input for a hybrid DL model, combining convolutional neural networks (CNNs) and Bidirectional Long Short-Term Memory (BiLSTM) networks. The model achieved a performant accuracy of 99.94% in classifying schizophrenia-related abnormalities, demonstrating its potential for real-time mental health monitoring. The generated TE matrices revealed significant differences in connectivity between the two groups, particularly in frontal and central brain regions, which are critical for cognitive processing. These findings were further validated by correlating the results with EEG data obtained from the Muse 2 headband, emphasizing the potential for portable, non-invasive monitoring of schizophrenia in real-world settings. The final model, integrated into the NeuroPredict platform, offers a scalable solution for continuous mental health monitoring. By incorporating EEG data, heart rate, sleep patterns, and environmental metrics, NeuroPredict facilitates early detection and personalized interventions for schizophrenia patients. Full article
(This article belongs to the Special Issue eHealth and mHealth)
Show Figures

Figure 1

Figure 1
<p>Heatmap representation of directional connectivity between brain regions based on TE values (unitless)<span class="html-italic">:</span> (<b>a</b>) schizophrenia and (<b>b</b>) HCs.</p>
Full article ">Figure 2
<p>Training and validation accuracy (<b>a</b>) and loss (<b>b</b>) plots for the proposed model.</p>
Full article ">Figure 3
<p>The classification report.</p>
Full article ">Figure 4
<p>The proposed integration of the DL model into the NeuroPredict platform.</p>
Full article ">
12 pages, 621 KiB  
Systematic Review
Systematic Review of AI-Assisted MRI in Prostate Cancer Diagnosis: Enhancing Accuracy Through Second Opinion Tools
by Saeed Alqahtani
Diagnostics 2024, 14(22), 2576; https://doi.org/10.3390/diagnostics14222576 - 15 Nov 2024
Viewed by 615
Abstract
Background: Prostate cancer is a leading cause of cancer-related deaths in men worldwide, making accurate diagnosis critical for effective treatment. Recent advancements in artificial intelligence (AI) and machine learning (ML) have shown promise in improving the diagnostic accuracy of prostate cancer. Objectives: This [...] Read more.
Background: Prostate cancer is a leading cause of cancer-related deaths in men worldwide, making accurate diagnosis critical for effective treatment. Recent advancements in artificial intelligence (AI) and machine learning (ML) have shown promise in improving the diagnostic accuracy of prostate cancer. Objectives: This systematic review aims to evaluate the effectiveness of AI-based tools in diagnosing prostate cancer using MRI, with a focus on accuracy, specificity, sensitivity, and clinical utility compared to conventional diagnostic methods. Methods: A comprehensive search was conducted across PubMed, Embase, Ovid MEDLINE, Web of Science, Cochrane Library, and Institute of Electrical and Electronics Engineers (IEEE) Xplore for studies published between 2019 and 2024. Inclusion criteria focused on full-text, English-language studies involving AI for Magnetic Resonance Imaging (MRI) -based prostate cancer diagnosis. Diagnostic performance metrics such as area under curve (AUC), sensitivity, and specificity were analyzed, with risk of bias assessed using the Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) tool. Results: Seven studies met the inclusion criteria, employing various AI techniques, including deep learning and machine learning. These studies reported improved diagnostic accuracy (with AUC scores of up to 97%) and moderate sensitivity, with performance varying based on training data quality and lesion characteristics like Prostate Imaging Reporting and Data System (PI-RADS) scores. Conclusions: AI has significant potential to enhance prostate cancer diagnosis, particularly when used for second opinions in MRI interpretations. While these results are promising, further validation in diverse populations and clinical settings is necessary to fully integrate AI into standard practice. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

Figure 1
<p>Flowchart for search results.</p>
Full article ">
32 pages, 3323 KiB  
Systematic Review
Artificial Intelligence Applied to Support Agronomic Decisions for the Automatic Aerial Analysis Images Captured by UAV: A Systematic Review
by Josef Augusto Oberdan Souza Silva, Vilson Soares de Siqueira, Marcio Mesquita, Luís Sérgio Rodrigues Vale, Jhon Lennon Bezerra da Silva, Marcos Vinícius da Silva, João Paulo Barcelos Lemos, Lorena Nunes Lacerda, Rhuanito Soranz Ferrarezi and Henrique Fonseca Elias de Oliveira
Agronomy 2024, 14(11), 2697; https://doi.org/10.3390/agronomy14112697 - 15 Nov 2024
Viewed by 468
Abstract
Integrating advanced technologies such as artificial intelligence (AI) with traditional agricultural practices has changed how activities are developed in agriculture, with the aim of automating manual processes and improving the efficiency and quality of farming decisions. With the advent of deep learning models [...] Read more.
Integrating advanced technologies such as artificial intelligence (AI) with traditional agricultural practices has changed how activities are developed in agriculture, with the aim of automating manual processes and improving the efficiency and quality of farming decisions. With the advent of deep learning models such as convolutional neural network (CNN) and You Only Look Once (YOLO), many studies have emerged given the need to develop solutions to problems and take advantage of all the potential that this technology has to offer. This systematic literature review aims to present an in-depth investigation of the application of AI in supporting the management of weeds, plant nutrition, water, pests, and diseases. This systematic review was conducted using the PRISMA methodology and guidelines. Data from different papers indicated that the main research interests comprise five groups: (a) type of agronomic problems; (b) type of sensor; (c) dataset treatment; (d) evaluation metrics and quantification; and (e) AI technique. The inclusion (I) and exclusion (E) criteria adopted in this study included: (I1) articles that obtained AI techniques for agricultural analysis; (I2) complete articles written in English; (I3) articles from specialized scientific journals; (E1) articles that did not describe the type of agrarian analysis used; (E2) articles that did not specify the AI technique used and that were incomplete or abstract; (E3) articles that did not present substantial experimental results. The articles were searched on the official pages of the main scientific bases: ACM, IEEE, ScienceDirect, MDPI, and Web of Science. The papers were categorized and grouped to show the main contributions of the literature to support agricultural decisions using AI. This study found that AI methods perform better in supporting weed detection, classification of plant diseases, and estimation of agricultural yield in crops when using images captured by Unmanned Aerial Vehicles (UAVs). Furthermore, CNN and YOLO, as well as their variations, present the best results for all groups presented. This review also points out the limitations and potential challenges when working with deep machine learning models, aiming to contribute to knowledge systematization and to benefit researchers and professionals regarding AI applications in mitigating agronomic problems. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>Flowchart of the systematic review selection steps according to the PRISMA methodology, according to the PRISMA 2020 statement from Page et al. [<a href="#B25-agronomy-14-02697" class="html-bibr">25</a>].</p>
Full article ">Figure 2
<p>Flowchart of the systematic literature review data extraction and sequence highlights, adapted from Siqueira et al. [<a href="#B23-agronomy-14-02697" class="html-bibr">23</a>]. Data extraction steps: (a) list of articles divided by the type of agronomic problem that each proposed to solve; (b) list of articles, divided by type of agronomic problem, that used sensors to acquire the dataset; (c) list of articles, divided by type of agronomic problem, that used image improvement techniques in the dataset; (d) number of articles that used evaluation metrics; (e) list of the main machine learning models used by each article in this study.</p>
Full article ">Figure 3
<p>Example of data output after training the YOLOv7 model for weed segmentation in commercial crops.</p>
Full article ">Figure 4
<p>Number of articles and timeline of publications per type of agronomic problems.</p>
Full article ">Figure 5
<p>Number of articles published and scientific platforms per type of agronomic problems.</p>
Full article ">Figure 6
<p>Number of articles per country included in this SLR.</p>
Full article ">
18 pages, 746 KiB  
Article
Evaluating Anomaly Explanations Using Ground Truth
by Liat Antwarg Friedman, Chen Galed, Lior Rokach and Bracha Shapira
AI 2024, 5(4), 2375-2392; https://doi.org/10.3390/ai5040117 - 15 Nov 2024
Viewed by 389
Abstract
The widespread use of machine and deep learning algorithms for anomaly detection has created a critical need for robust explanations that can identify the features contributing to anomalies. However, effective evaluation methodologies for anomaly explanations are currently lacking, especially those that compare the [...] Read more.
The widespread use of machine and deep learning algorithms for anomaly detection has created a critical need for robust explanations that can identify the features contributing to anomalies. However, effective evaluation methodologies for anomaly explanations are currently lacking, especially those that compare the explanations against the true underlying causes, or ground truth. This paper aims to address this gap by introducing a rigorous, ground-truth-based framework for evaluating anomaly explanation methods, which enables the assessment of explanation correctness and robustness—key factors for actionable insights in anomaly detection. To achieve this, we present an innovative benchmark dataset of digital circuit truth tables with model-based anomalies, accompanied by local ground truth explanations. These explanations were generated using a novel algorithm designed to accurately identify influential features within each anomaly. Additionally, we propose an evaluation methodology based on correctness and robustness metrics, specifically tailored to quantify the reliability of anomaly explanations. This dataset and evaluation framework are publicly available to facilitate further research and standardize evaluation practices. Our experiments demonstrate the utility of this dataset and methodology by evaluating common model-agnostic explanation methods in an anomaly detection context. The results highlight the importance of ground-truth-based evaluation for reliable and interpretable anomaly explanations, advancing both theory and practical applications in explainable AI. This work establishes a foundation for rigorous, evidence-based assessments of anomaly explanations, fostering greater transparency and trust in AI-driven anomaly detection systems. Full article
(This article belongs to the Special Issue Interpretable and Explainable AI Applications)
Show Figures

Figure 1

Figure 1
<p>A flow chart describing two processes: (1) data generation (<b>above</b>), in which the data set with anomalies and the ground truth explanations are created, and (2) explanation evaluation (<b>below</b>), in which anomalies are detected and explained and the local explanations are evaluated against the ground truth.</p>
Full article ">Figure 2
<p><b>Left</b>—C17 original diagram, along with sample instances. <b>Right</b>—C17 anomalous diagram after changing gate z4 from NAND to AND, along with sample instances, and by doing that, we created an anomaly. The instances containing outputs marked in red are anomalous.</p>
Full article ">
21 pages, 7459 KiB  
Article
Deep Learning for Urban Tree Canopy Coverage Analysis: A Comparison and Case Study
by Grayson R. Morgan, Danny Zlotnick, Luke North, Cade Smith and Lane Stevenson
Geomatics 2024, 4(4), 412-432; https://doi.org/10.3390/geomatics4040022 - 14 Nov 2024
Viewed by 354
Abstract
Urban tree canopy (UTC) coverage, or area, is an important metric for monitoring changes in UTC over large areas within a municipality. Several methods have been used to obtain these data, but remote sensing image classification is one of the fastest and most [...] Read more.
Urban tree canopy (UTC) coverage, or area, is an important metric for monitoring changes in UTC over large areas within a municipality. Several methods have been used to obtain these data, but remote sensing image classification is one of the fastest and most reliable over large areas. However, most studies have tested only one or two classification methods to accomplish this while using costly satellite imagery or LiDAR data. This study seeks to compare three urban tree canopy cover classifiers by testing a deep learning U-Net convolutional neural network (CNN), support vector machine learning classifier (SVM) and a random forests machine learning classifier (RF) on cost-free 2012 aerial imagery over a small southern USA city and midsize, growing southern USA city. The results of the experiment are then used to decide the best classifier and apply it to more recent aerial imagery to determine canopy changes over a 10-year period. The changes are subsequently compared visually and statistically with recent urban heat maps derived from thermal Landsat 9 satellite data to compare the means of temperatures within areas of UTC loss and no change. The U-Net CNN classifier proved to provide the best overall accuracy for both cities (89.8% and 91.4%), while also requiring the most training and classification time. When compared spatially with city heat maps, city periphery regions were most impacted by substantial changes in UTC area as cities grow and the outer regions get warmer. Furthermore, areas of UTC loss had higher temperatures than those areas with no canopy change. The broader impacts of this study reach the urban forestry managers at the local, state/province, and national levels as they seek to provide data-driven decisions for policy makers. Full article
(This article belongs to the Topic Geocomputation and Artificial Intelligence for Mapping)
Show Figures

Figure 1

Figure 1
<p>Study areas (<b>A</b>) Georgetown TX and (<b>B</b>) Laurel MS within the United States of America.</p>
Full article ">Figure 2
<p>Land surface temperature (LST) maps for Georgetown (<b>A</b>) and Laurel (<b>B</b>) and the quality assessment overlay in pink. The pink colors indicate regions where the uncertainty is higher.</p>
Full article ">Figure 3
<p>A general workflow of the experiment and case study.</p>
Full article ">Figure 4
<p>Training and validation samples used for classification and accuracy assessment of the 2012 NAIP images (water included for map purpose only).</p>
Full article ">Figure 5
<p>Comparison of the original NAIP image (<b>A</b>), the RF classifier results (<b>B</b>), SVM classifier results (<b>C</b>), and the U-Net classifier results (<b>D</b>) for Laurel, MS. Light green represents grass, dark green is urban tree canopy, and yellow is urban and other classes combined.</p>
Full article ">Figure 6
<p>Comparison of the original NAIP image (<b>A</b>), the RF classifier results (<b>B</b>), SVM classifier results (<b>C</b>), and the U-Net classifier results (<b>D</b>) for Georgetown, TX. Light green represents grass, dark green is urban tree canopy, and yellow is urban and other classes combined.</p>
Full article ">Figure 7
<p>Canopy changes (in blue) overlaid on the heat maps for Georgetown (<b>A</b>) and Laurel (<b>B</b>).</p>
Full article ">Figure 8
<p>Laurel Mississippi example of tree canopy loss or change using the 2012 NAIP image (<b>A</b>), 2023 NAIP image (<b>B</b>), heat map (<b>C</b>), and heat map showing detected canopy loss (<b>D</b>).</p>
Full article ">Figure 9
<p>Georgetown, TX first example of tree canopy loss or change using the 2012 NAIP image (<b>A</b>), 2023 NAIP image (<b>B</b>), heat map (<b>C</b>), and heat map showing detected canopy loss (<b>D</b>).</p>
Full article ">Figure 10
<p>Georgetown, TX second example of tree canopy loss or change using the 2012 NAIP image (<b>A</b>), 2023 NAIP image (<b>B</b>), heat map (<b>C</b>), and heat map showing detected canopy loss (<b>D</b>).</p>
Full article ">
9 pages, 1543 KiB  
Article
DQN-Based Shaped Reward Function Mold for UAV Emergency Communication
by Chenhao Ye, Wei Zhu, Shiluo Guo and Jinyin Bai
Appl. Sci. 2024, 14(22), 10496; https://doi.org/10.3390/app142210496 - 14 Nov 2024
Viewed by 358
Abstract
Unmanned aerial vehicles (UAVs) have emerged as pivotal tools in emergency communication scenarios. In the aftermath of disasters, UAVs can be communication nodes to provide communication services for users in the area. In this paper, we establish a meticulously crafted virtual simulation environment [...] Read more.
Unmanned aerial vehicles (UAVs) have emerged as pivotal tools in emergency communication scenarios. In the aftermath of disasters, UAVs can be communication nodes to provide communication services for users in the area. In this paper, we establish a meticulously crafted virtual simulation environment and leverage advanced deep reinforcement learning algorithms to train UAVs agents. Notwithstanding, the development of reinforcement learning algorithms is beset with challenges such as sparse rewards and protracted training durations. To mitigate these issues, we devise an enhanced reward function aimed at bolstering training efficiency. Initially, we delineate a specific mountainous emergency communication scenario and integrate it with the particularized application of UAVs to undertake virtual simulations, constructing a realistic virtual environment. Furthermore, we introduce a supplementary shaped reward function tailored to alleviate the problem of sparse rewards. By refining the DQN algorithm and devising a reward structure grounded on potential functions, we observe marked improvements in the final evaluation metrics, substantiating the efficacy of our approach. The experimental outcomes underscore the prowess of our methodology in effectively curtailing training time while augmenting convergence rates. In summary, our work underscores the potential of leveraging sophisticated virtual environments and refined reinforcement learning techniques to optimize UAVs deployment in emergency communication contexts. Full article
(This article belongs to the Section Aerospace Science and Engineering)
Show Figures

Figure 1

Figure 1
<p>DQN.</p>
Full article ">Figure 2
<p>Typical communication scenario.</p>
Full article ">Figure 3
<p>Main work chart.</p>
Full article ">Figure 4
<p>Reward shaping.</p>
Full article ">Figure 5
<p>Training reward.</p>
Full article ">Figure 6
<p>Training reward over episodes.</p>
Full article ">
19 pages, 602 KiB  
Article
WKNN-Based Wi-Fi Fingerprinting with Deep Distance Metric Learning via Siamese Triplet Network for Indoor Positioning
by Jae-Hyeon Park, Dongdeok Kim and Young-Joo Suh
Electronics 2024, 13(22), 4448; https://doi.org/10.3390/electronics13224448 - 13 Nov 2024
Viewed by 331
Abstract
Weighted k-nearest neighbor (WKNN)-based Wi-Fi fingerprinting is popular in indoor location-based services due to its ease of implementation and low computational cost. KNN-based methods rely on distance metrics to select the nearest neighbors. However, traditional metrics often fail to capture the complexity of [...] Read more.
Weighted k-nearest neighbor (WKNN)-based Wi-Fi fingerprinting is popular in indoor location-based services due to its ease of implementation and low computational cost. KNN-based methods rely on distance metrics to select the nearest neighbors. However, traditional metrics often fail to capture the complexity of indoor environments and have limitations in identifying non-linear relationships. To address these issues, we propose a novel WKNN-based Wi-Fi fingerprinting method that incorporates distance metric learning. In the offline phase, our method utilizes a Siamese network with a triplet loss function to learn a meaningful distance metric from training fingerprints (FPs). This process employs a unique triplet mining strategy to handle the inherent noise in FPs. Subsequently, in the online phase, the learned metric is used to calculate the embedding distance, followed by a signal-space distance filtering step to optimally select neighbors and estimate the user’s location. The filtering step mitigates issues from an overfitted distance metric influenced by hard triplets, which could lead to incorrect neighbor selection. We evaluate the proposed method on three benchmark datasets, UJIIndoorLoc, Tampere, and UTSIndoorLoc, and compare it with four WKNN models. The results show a mean positioning error reduction of 3.55% on UJIIndoorLoc, 16.21% on Tampere, and 16.49% on UTSIndoorLoc, demonstrating enhanced positioning accuracy. Full article
(This article belongs to the Special Issue Next-Generation Indoor Wireless Communication)
Show Figures

Figure 1

Figure 1
<p>Overview of our proposed method.</p>
Full article ">Figure 2
<p>The Siamese network for distance metric learning. The network learns an embedding space where similar fingerprints are positioned closer together.</p>
Full article ">Figure 3
<p>Triplets in the position space and embedding space (A: anchor; N: negative; EP: easy positive; HP: hard positive; EN: easy negative; SHN: semi-hard negative; HN: hard negative).</p>
Full article ">Figure 4
<p>Cumulative distribution function of positioning errors with UJIIndoorLoc.</p>
Full article ">Figure 5
<p>Cumulative distribution function of positioning errors with Tampere.</p>
Full article ">Figure 6
<p>Cumulative distribution function of positioning errors with UTSIndoorLoc.</p>
Full article ">
16 pages, 6553 KiB  
Article
Cucumber Leaf Segmentation Based on Bilayer Convolutional Network
by Tingting Qian, Yangxin Liu, Shenglian Lu, Linyi Li, Xiuguo Zheng, Qingqing Ju, Yiyang Li, Chun Xie and Guo Li
Agronomy 2024, 14(11), 2664; https://doi.org/10.3390/agronomy14112664 - 12 Nov 2024
Viewed by 445
Abstract
When monitoring crop growth using top-down images of the plant canopies, leaves in agricultural fields appear very dense and significantly overlap each other. Moreover, the image can be affected by external conditions such as background environment and light intensity, impacting the effectiveness of [...] Read more.
When monitoring crop growth using top-down images of the plant canopies, leaves in agricultural fields appear very dense and significantly overlap each other. Moreover, the image can be affected by external conditions such as background environment and light intensity, impacting the effectiveness of image segmentation. To address the challenge of segmenting dense and overlapping plant leaves under natural lighting conditions, this study employed a Bilayer Convolutional Network (BCNet) method for accurate leaf segmentation across various lighting environments. The major contributions of this study are as follows: (1) Utilized Fully Convolutional Object Detection (FCOS) for plant leaf detection, incorporating ResNet-50 with the Convolutional Block Attention Module (CBAM) and Feature Pyramid Network (FPN) to enhance Region of Interest (RoI) feature extraction from canopy top-view images. (2) Extracted the sub-region of the RoI based on the position of the detection box, using this region as input for the BCNet, ensuring precise segmentation. (3) Utilized instance segmentation of canopy top-view images using BCNet, improving segmentation accuracy. (4) Applied the Varifocal Loss Function to improve the classification loss function in FCOS, leading to better performance metrics. The experimental results on cucumber canopy top-view images captured in glass greenhouse and plastic greenhouse environments show that our method is highly effective. For cucumber leaves at different growth stages and under various lighting conditions, the Precision, Recall and Average Precision (AP) metrics for object recognition are 97%, 94% and 96.57%, respectively. For instance segmentation, the Precision, Recall and Average Precision (AP) metrics are 87%, 83% and 84.71%, respectively. Our algorithm outperforms commonly used deep learning algorithms such as Faster R-CNN, Mask R-CNN, YOLOv4 and PANet, showcasing its superior capability in complex agricultural settings. The results of this study demonstrate the potential of our method for accurate recognition and segmentation of highly overlapping leaves in diverse agricultural environments, significantly contributing to the application of deep learning algorithms in smart agriculture. Full article
(This article belongs to the Special Issue AI, Sensors and Robotics for Smart Agriculture—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Bilayer decomposition diagram.</p>
Full article ">Figure 2
<p>Flow chart of image segmentation based on improved BCNet.</p>
Full article ">Figure 3
<p>Schematic diagram of image annotation method. <span class="html-fig-inline" id="agronomy-14-02664-i001"><img alt="Agronomy 14 02664 i001" src="/agronomy/agronomy-14-02664/article_deploy/html/images/agronomy-14-02664-i001.png"/></span> means labeled, <span class="html-fig-inline" id="agronomy-14-02664-i002"><img alt="Agronomy 14 02664 i002" src="/agronomy/agronomy-14-02664/article_deploy/html/images/agronomy-14-02664-i002.png"/></span> means unlabeled.</p>
Full article ">Figure 4
<p>Schematic diagram of image expansion scheme.</p>
Full article ">Figure 5
<p>Segmentation effect of cucumber plant images in glass greenhouses. (<b>a</b>) Early Growth Stage, Sunny. (<b>b</b>) Early Growth Stage, Cloudy. (<b>c</b>) Metaphase Growth Stage, Sunny. (<b>d</b>) Metaphase Growth Stage. Cloudy. (<b>e</b>) Terminal Growth Stage, Sunny. (<b>f</b>) Terminal Growth Stage, Cloudy.</p>
Full article ">Figure 6
<p>Effect of cucumber plant image segmentation in plastic greenhouses. (<b>a</b>) Early Growth Stage, Sunny. (<b>b</b>) Early Growth Stage, Cloudy. (<b>c</b>) Terminal Growth Stage, Sunny. (<b>d</b>) Terminal Growth Stage, Cloudy.</p>
Full article ">Figure 7
<p>Target recognition and instance segmentation P-R curve of six models: (<b>a</b>) Object detection P-R curve; (<b>b</b>) Example split P-R curve.</p>
Full article ">Figure 8
<p>Effect of cucumber plant image segmentation in glass greenhouse: (<b>a</b>) improved BCNet; (<b>b</b>) BCNet; (<b>c</b>) PANet; (<b>d</b>) Mask R-CNN; (<b>e</b>) YOLOv4; (<b>f</b>) Faster R-CNN.</p>
Full article ">
Back to TopTop