Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (46)

Search Parameters:
Keywords = Gabor-based textures

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 13590 KiB  
Article
Fast and Nondestructive Proximate Analysis of Coal from Hyperspectral Images with Machine Learning and Combined Spectra-Texture Features
by Jihua Mao, Hengqian Zhao, Yu Xie, Mengmeng Wang, Pan Wang, Yaning Shi and Yusen Zhao
Appl. Sci. 2024, 14(17), 7920; https://doi.org/10.3390/app14177920 - 5 Sep 2024
Abstract
Proximate analysis, including ash, volatile matter, moisture, fixed carbon, and calorific value, is a fundamental aspect of fuel testing and serves as the primary method for evaluating coal quality, which is critical for the processing and utilization of coal. The traditional analytical methods [...] Read more.
Proximate analysis, including ash, volatile matter, moisture, fixed carbon, and calorific value, is a fundamental aspect of fuel testing and serves as the primary method for evaluating coal quality, which is critical for the processing and utilization of coal. The traditional analytical methods involve time-consuming and costly combustion processes, particularly when applied to large volumes of coal that need to be sampled in massive batches. Hyperspectral imaging is promising for the rapid and nondestructive determination of coal quality indices. In this study, a fast and nondestructive coal proximate analysis method with combined spectral-spatial features was developed using a hyperspectral imaging system in the 450–2500 nm range. The processed spectra were evaluated using PLSR, with the most effective MSC spectra selected. To reduce the spectral redundancy and improve the accuracy, the SPA, Boruta, iVISSA, and CARS algorithms were adopted to extract the characteristic wavelengths, and 16 prediction models were constructed and optimized based on the PLSR, RF, BPNN, and LSSVR algorithms within the Optuna framework for each quality indicator. For spatial information, the histogram statistics, gray-level covariance matrix, and Gabor filters were employed to extract the texture features within the characteristic wavelengths. The texture feature-based and combined spectral-texture feature-based prediction models were constructed by applying the spectral modeling strategy, respectively. Compared with the models based on spectral or texture features only, the LSSVR models with combined spectral-texture features achieved the highest prediction accuracy in all quality metrics, with Rp2 values of 0.993, 0.989, 0.979, 0.948, and 0.994 for Ash, VM, MC, FC, and CV, respectively. This study provides a technical reference for hyperspectral imaging technology as a new method for the rapid, nondestructive proximate analysis and quality assessment of coal. Full article
(This article belongs to the Section Optics and Lasers)
Show Figures

Figure 1

Figure 1
<p>Research flow chart of the study.</p>
Full article ">Figure 2
<p>Pseudo-color images (817 nm, 661 nm, and 549 nm) of four coal samples and the measured values of quality indices. The samples are arranged from left to right by CV, from highest to lowest.</p>
Full article ">Figure 3
<p>Reflectance spectra and characteristic wavelengths obtained by averaging pixels in the region of interest from hyperspectral images of 61 coal samples. Each different colored curve represents each coal sample.</p>
Full article ">Figure 4
<p>Evolution of the noise level with wavelength evaluated based on spectra of all coal samples. Wavelengths with prominent noise spikes (&gt;0.5%) have been excluded from further analysis.</p>
Full article ">Figure 5
<p>The spectral curves of (<b>a</b>) raw and preprocessed reflectance of all coal samples using (<b>b</b>) SG, (<b>c</b>) FD, and (<b>d</b>) MSC methods. The red curves indicate the mean spectra of all coal samples, and the gray shadows represent the spectra reflectance range.</p>
Full article ">Figure 6
<p>Results of characteristic wavelength extraction (marked by red square). The characteristic wavelengths of Ash, VM, MC, and CV were extracted by CARS. The characteristic wavelengths of FC were extracted by Boruta.</p>
Full article ">Figure 7
<p>Scatter plots of actual and predicted coal quality indices values obtained using the optimal LSSVR model based on the combined spectra-texture features. (<b>a</b>) Ash; (<b>b</b>) VM; (<b>c</b>) MC; (<b>d</b>) FC; and (<b>e</b>) CV.</p>
Full article ">Figure 8
<p>Contributions of the top ten significant variables by SHAP values in the coal quality indices optimal prediction models.</p>
Full article ">Figure 9
<p>Relative contribution of coal quality indices based on mean absolute SHAP values.</p>
Full article ">Figure 10
<p>Predictive distribution of coal quality indices by combined spectra-textual feature based on hyperspectral images. The four samples in each index correspond to the minimum, 25%, 75%, and maximum values in the dataset from left to right, respectively.</p>
Full article ">
18 pages, 16408 KiB  
Article
Enhanced Scratch Detection for Textured Materials Based on Optimized Photometric Stereo Vision and Fast Fourier Transform–Gabor Filtering
by Yaoshun Yue, Wenpeng Sang, Kaiwei Zhai and Maohai Lin
Appl. Sci. 2024, 14(17), 7812; https://doi.org/10.3390/app14177812 - 3 Sep 2024
Viewed by 291
Abstract
In the process of scratch defect detection in textured materials, there are often problems of low efficiency in traditional manual detection, large errors in machine vision, and difficulty in distinguishing defective scratches from the background texture. In order to solve these problems, we [...] Read more.
In the process of scratch defect detection in textured materials, there are often problems of low efficiency in traditional manual detection, large errors in machine vision, and difficulty in distinguishing defective scratches from the background texture. In order to solve these problems, we developed an enhanced scratch defect detection system for textured materials based on optimized photometric stereo vision and FFT-Gabor filtering. We designed and optimized a novel hemispherical image acquisition device that allows for selective lighting angles. This device integrates images captured under multiple light sources to obtain richer surface gradient information for textured materials, overcoming issues caused by high reflections or dark shadows under a single light source angle. At the same time, for the textured material, scratches and a textured background are difficult to distinguish; therefore, we introduced a Gabor filter-based convolution kernel, leveraging the fast Fourier transform (FFT), to perform convolution operations and spatial domain phase subtraction. This process effectively enhances the defect information while suppressing the textured background. The effectiveness and superiority of the proposed method were validated through material applicability experiments and comparative method evaluations using a variety of textured material samples. The results demonstrated a stable scratch capture success rate of 100% and a recognition detection success rate of 98.43% ± 1.0%. Full article
(This article belongs to the Section Applied Industrial Technologies)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Due to the micro-geometry, some micro-planes are occluded and do not receive light (shadowing). (<b>b</b>) Light reflected from micro-planes that cannot be seen from the observation direction is also not visible (masking).</p>
Full article ">Figure 2
<p>Enhanced scratch detection for textured materials based on optimized photometric stereo vision and FFT-Gabor filtering.</p>
Full article ">Figure 3
<p>Light-source-selective image acquisition device based on photometric stereo vision: (<b>A</b>) presents the principle of photometric stereo vision, and (<b>B</b>) presents the principle of application based on photometric stereo vision.</p>
Full article ">Figure 4
<p>The impact of different numbers of light sources on image quality evaluated.</p>
Full article ">Figure 5
<p>Photometric stereo vision based on input images from 8 light sources.</p>
Full article ">Figure 6
<p>Photometric stereo vision image acquisition for different textured materials. (<b>A</b>) represents coarse textured leather; (<b>B</b>) represents fine textured leather; (<b>C</b>) represents textile fabric; and (<b>D</b>) represents textured kraft paper.</p>
Full article ">Figure 7
<p>Framework for scratch defect detection on textured material surface based on image enhancement algorithm.</p>
Full article ">Figure 8
<p>The effect of image contrast enhancement after fast Fourier transform based on Gabor filter. Sample A represents fine textured leather, sample B represents coarse textured leather, and sample C represents light textured leather.</p>
Full article ">Figure 9
<p>Validation of textured material applicability of detection methods. Sample A represents coarse-textured kraft paper; Sample B represents fine-textured leather; Sample C represents coarse linen fabric; Sample D represents fine-textured kraft paper; and Sample E represents a textured wood panel.</p>
Full article ">Figure 10
<p>Validation of methodological superiority. (<b>A</b>) represents a fine-textured leather material with deep and dense self-texture; (<b>B</b>) represents a coarse-textured leather material with deep and irregular self-texture; (<b>C</b>) represents a fine-textured leather material with shallow and relatively regular self-texture.</p>
Full article ">
33 pages, 30114 KiB  
Article
Exploring the Influence of Object, Subject, and Context on Aesthetic Evaluation through Computational Aesthetics and Neuroaesthetics
by Fangfu Lin, Wanni Xu, Yan Li and Wu Song
Appl. Sci. 2024, 14(16), 7384; https://doi.org/10.3390/app14167384 - 21 Aug 2024
Viewed by 504
Abstract
Background: In recent years, computational aesthetics and neuroaesthetics have provided novel insights into understanding beauty. Building upon the findings of traditional aesthetics, this study aims to combine these two research methods to explore an interdisciplinary approach to studying aesthetics. Method: Abstract artworks were [...] Read more.
Background: In recent years, computational aesthetics and neuroaesthetics have provided novel insights into understanding beauty. Building upon the findings of traditional aesthetics, this study aims to combine these two research methods to explore an interdisciplinary approach to studying aesthetics. Method: Abstract artworks were used as experimental materials. Based on traditional aesthetics and in combination, features of composition, tone, and texture were selected. Computational aesthetic methods were then employed to correspond these features to physical quantities: blank space, gray histogram, Gray Level Co-occurrence Matrix (GLCM), Local Binary Pattern (LBP), and Gabor filters. An electroencephalogram (EEG) experiment was carried out, in which participants conducted aesthetic evaluations of the experimental materials in different contexts (genuine, fake), and their EEG data were recorded to analyze the impact of various feature classes in the aesthetic evaluation process. Finally, a Support Vector Machines (SVMs) was utilized to model the feature data, Event-Related Potentials (ERPs), context data, and subjective aesthetic evaluation data. Result: Behavioral data revealed higher aesthetic ratings in the genuine context. ERP data indicated that genuine contexts elicited more negative deflections in the prefrontal lobes between 200 and 1000 ms. Class II compositions demonstrated more positive deflections in the parietal lobes at 50–120 ms, while Class I tones evoked more positive amplitudes in the occipital lobes at 200–300 ms. Gabor features showed significant variations in the parieto-occipital area at an early stage. Class II LBP elicited a prefrontal negative wave with a larger amplitude. The results of the SVM models indicated that the model incorporating aesthetic subject and context data (ACC = 0.76866) outperforms the model using only parameters of the aesthetic object (ACC = 0.68657). Conclusion: A positive context tends to provide participants with a more positive aesthetic experience, but abstract artworks may not respond to this positivity. During aesthetic evaluation, the ERP data activated by different features show a trend from global to local. The SVM model based on multimodal data fusion effectively predicts aesthetics, further demonstrating the feasibility of the combined research approach of computational aesthetics and neuroaesthetics. Full article
Show Figures

Figure 1

Figure 1
<p>The calculation of blank space in Suprematist Composition: Airplane Flying (images processed by authors as fair use from wikiart.org) <a href="https://www.wikiart.org/en/kazimir-malevich/aeroplane-flying-1915" target="_blank">https://www.wikiart.org/en/kazimir-malevich/aeroplane-flying-1915</a> (accessed on 4 March 2024).</p>
Full article ">Figure 2
<p>Kernels of different wavelengths <math display="inline"><semantics> <mrow> <mi>λ</mi> </mrow> </semantics></math> and angles <math display="inline"><semantics> <mrow> <mi>θ</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 3
<p>Example images of features.</p>
Full article ">Figure 4
<p>Illustration of the stimulus paradigm applied.</p>
Full article ">Figure 5
<p>Grand–average event–related brain potentials and isopotential contour plot (200–1000 ms) for genuine and fake context. <span class="html-italic">N</span> = 12.</p>
Full article ">Figure 6
<p>Grand–average event–related brain potentials and isopotential contour plot (50–120 ms) for context (genuine, fake) × composition (Class I, Class II). <span class="html-italic">N</span> = 12.</p>
Full article ">Figure 7
<p>Grand–average event–related brain potentials and isopotential contour plot (200–300 ms) for context (genuine, fake) × tone (Class I, Class II). <span class="html-italic">N</span> = 12.</p>
Full article ">Figure 8
<p>Grand–average event–related brain potentials and isopotential contour plot (70–130 ms) for context (genuine, fake) × Gabor–Mean (Class I, Class II). <span class="html-italic">N</span> = 12.</p>
Full article ">Figure 9
<p>Grand–average event–related brain potentials and isopotential contour plot (70–130 ms) for context (genuine, fake) × Gabor–Variance (Class I, Class II). <span class="html-italic">N</span> = 12.</p>
Full article ">Figure 10
<p>Grand–average event–related brain potentials and isopotential contour plot (70–130 ms and 200–300 ms) for context (genuine, fake) × Gabor–Energy (Class I, Class II). <span class="html-italic">N</span> = 12.</p>
Full article ">Figure 11
<p>Grand–average event–related brain potentials and isopotential contour plot (500–1000 ms) for context (genuine, fake) × horizontal GLCM (Class I, Class II). <span class="html-italic">N</span> = 12.</p>
Full article ">Figure 12
<p>Grand–average event–related brain potentials and isopotential contour plot (70–140 ms and 500–1000 ms) for context (genuine, fake) × diagonal GLCM (Class I, Class II). <span class="html-italic">N</span> = 12.</p>
Full article ">Figure 13
<p>Grand–average event–related brain potentials and isopotential contour plot (300–1000 ms) for context (genuine, fake) × LBP (Class I, Class II). <span class="html-italic">N</span> = 12.</p>
Full article ">Figure 14
<p>Performance of SVM models with varying C and γ values: (<b>a</b>) the ACC with different C and γ combinations; (<b>b</b>) the AUC with different C and γ combinations. (The closer the color is to red, the higher the value; the closer it is to blue, the lower the value).</p>
Full article ">
16 pages, 3627 KiB  
Article
New Approach for Brain Tumor Segmentation Based on Gabor Convolution and Attention Mechanism
by Yuan Cao and Yinglei Song
Appl. Sci. 2024, 14(11), 4919; https://doi.org/10.3390/app14114919 - 6 Jun 2024
Viewed by 710
Abstract
In the treatment process of brain tumors, it is of great importance to develop a set of MRI image segmentation methods with high accuracy and low cost. In order to extract the feature information for each region of the brain tumor more effectively, [...] Read more.
In the treatment process of brain tumors, it is of great importance to develop a set of MRI image segmentation methods with high accuracy and low cost. In order to extract the feature information for each region of the brain tumor more effectively, this paper proposes a new model Ga-U-Net based on Gabor convolution and an attention mechanism. Based on 3D U-Net, Gabor convolution is added at the shallow layer of the encoder, which is able to learn the local structure and texture information of the tumor better. After that, the CBAM attention mechanism is added after the output of each layer of the encoder, which not only enhances the network’s ability to perceive the brain tumor boundary information but also reduces some redundant information by allocating the attention to the two dimensions of space and channel. Experimental results show that the model performs well for multiple tumor regions (WT, TC, ET) on the brain tumor dataset BraTS 2021, with Dice coefficients of 0.910, 0.897, and 0.856, respectively, which are improved by 0.3%, 2%, and 1.7% compared to the base network, the U-Net network, with an average Dice of 0.887 and an average Hausdorff distance of 9.12, all of which are better than a few other state-of-the-art deep models for biomedical image segmentation. Full article
Show Figures

Figure 1

Figure 1
<p>The overall structure of the proposed Ga-U-Net network, where the meanings of modules and arrows in different colors are shown in the bottom of the figure.</p>
Full article ">Figure 2
<p>The structure of GA module, where the component for Gabor convolution is shown in green color and the other two ordinary convolutional components are shown in blue color.</p>
Full article ">Figure 3
<p>The general structure of a CBAM module, where the input and output feature maps are shown in blue color, the channel attention module and the spatial attention module are shown in green and purple colors respectively.</p>
Full article ">Figure 4
<p>The structure of the attention channel module. The MaxPool layer is shown in dark blue color and the output of the shared MLP network on its output is shown in light blue color; The AvgPool layer is shown in dark orange color and the output of the shared MLP network on its output is shown in light orange color; input feature maps <math display="inline"><semantics> <mrow> <mi>F</mi> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>M</mi> </mrow> <mrow> <mi>c</mi> </mrow> </msub> </mrow> </semantics></math> are shown in light and dark green colors respectively.</p>
Full article ">Figure 5
<p>The structure of the spatial channel module, where the layer for MaxPool is shown in blue color and the layer for AvgPool is shown in orange color. In addition, the input feature and output feature maps are shown in light and dark green colors respectively.</p>
Full article ">Figure 6
<p>Examples of the four modalities of a case: (<b>a</b>–<b>d</b>) are the FLAIR, T1, T1CE, and T2 modalities for the case.</p>
Full article ">Figure 7
<p>The Hausdorff distance.</p>
Full article ">Figure 8
<p>Examples of segmentation results using the Ga-U-Net model.</p>
Full article ">Figure 9
<p>Loss curves obtained for the training set.</p>
Full article ">Figure 10
<p>Loss curves obtained for the validation set.</p>
Full article ">
22 pages, 15475 KiB  
Article
Background Subtraction for Dynamic Scenes Using Gabor Filter Bank and Statistical Moments
by Julio-Alejandro Romero-González, Diana-Margarita Córdova-Esparza, Juan Terven, Ana-Marcela Herrera-Navarro and Hugo Jiménez-Hernández
Algorithms 2024, 17(4), 133; https://doi.org/10.3390/a17040133 - 25 Mar 2024
Viewed by 1228
Abstract
This paper introduces a novel background subtraction method that utilizes texture-level analysis based on the Gabor filter bank and statistical moments. The method addresses the challenge of accurately detecting moving objects that exhibit similar color intensity variability or texture to the surrounding environment, [...] Read more.
This paper introduces a novel background subtraction method that utilizes texture-level analysis based on the Gabor filter bank and statistical moments. The method addresses the challenge of accurately detecting moving objects that exhibit similar color intensity variability or texture to the surrounding environment, which conventional methods struggle to handle effectively. The proposed method accurately distinguishes between foreground and background objects by capturing different frequency components using the Gabor filter bank and quantifying the texture level through statistical moments. Extensive experimental evaluations use datasets featuring varying lighting conditions, uniform and non-uniform textures, shadows, and dynamic backgrounds. The performance of the proposed method is compared against other existing methods using metrics such as sensitivity, specificity, and false positive rate. The experimental results demonstrate that the proposed method outperforms other methods in accuracy and robustness. It effectively handles scenarios with complex backgrounds, lighting changes, and objects that exhibit similar texture or color intensity as the background. Our method retains object structure while minimizing false detections and noise. This paper provides valuable insights into computer vision and object detection, offering a promising solution for accurate foreground detection in various applications such as video surveillance and motion tracking. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Gabor filter <math display="inline"><semantics> <msub> <mi>G</mi> <mrow> <mi>λ</mi> <mo>,</mo> <mi>θ</mi> <mo>,</mo> <mi>ϕ</mi> </mrow> </msub> </semantics></math>.</p>
Full article ">Figure 2
<p>The procedure is as follows: (1) capture images from a dataset or a camera, (2) build the Gabor kernel, (3) obtain intensities as the texture level, (4) texture-level quantization, and (5) foreground detection.</p>
Full article ">Figure 3
<p>Gabor filter size–frequency ratio. This figure shows the comparison between the size of the envelope function (the Gaussian distribution in red) and the response of <math display="inline"><semantics> <msub> <mi>G</mi> <mrow> <mi>λ</mi> <mo>,</mo> <mi>θ</mi> <mo>,</mo> <mi>ϕ</mi> </mrow> </msub> </semantics></math> (the blue distribution) relative to the size of the image.</p>
Full article ">Figure 4
<p>Scene intensity levels. The image shows a scene with intensity values similar to the moving object.</p>
Full article ">Figure 5
<p>Gabor kernel 3–D view.</p>
Full article ">Figure 6
<p>Periodic and non–periodic texture of the scene [<a href="#B66-algorithms-17-00133" class="html-bibr">66</a>]: (<b>a</b>) non–periodic texture; (<b>b</b>) periodic texture.</p>
Full article ">Figure 7
<p>Scene’s texture. The texture level is expressed as an edge in this image, obtained by characterizing the image using the Gabor function.</p>
Full article ">Figure 8
<p>Distributions of the statistical moments in the scene. Objects <math display="inline"><semantics> <msub> <mi>O</mi> <mi>k</mi> </msub> </semantics></math> show a greater dispersion, while <math display="inline"><semantics> <msub> <mi>B</mi> <mi>k</mi> </msub> </semantics></math> remains more homogeneous.</p>
Full article ">Figure 9
<p>The typical deviation of the quantified texture. The standard deviation <math display="inline"><semantics> <mfenced open="(" close=")"> <mi>σ</mi> </mfenced> </semantics></math> is taken as the segmentation threshold.</p>
Full article ">Figure 10
<p>Homogeneous region segmentation by texture analysis.</p>
Full article ">Figure 11
<p>Adjustment of the <math display="inline"><semantics> <mi>λ</mi> </semantics></math> value to characterize the light changes of objects on the scene. (<b>a</b>) Original image, in (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mn>0.95</mn> </mrow> </semantics></math>, (<b>c</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mn>0.6</mn> </mrow> </semantics></math>, (<b>d</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mn>0.367</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 12
<p>Adjustment of the <math display="inline"><semantics> <mi>λ</mi> </semantics></math> value to focus on the structure, texture and edge of the object. In (<b>a</b>) original image, in (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mn>3</mn> </mrow> </semantics></math>, for (<b>c</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mn>1.2</mn> </mrow> </semantics></math> y and (<b>d</b>) <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mn>0.5</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 13
<p>Percentage of wrong classifications comparison. (<b>a</b>) Scene <math display="inline"><semantics> <msub> <mi>S</mi> <mn>3</mn> </msub> </semantics></math>; (<b>b</b>) Scene <math display="inline"><semantics> <msub> <mi>S</mi> <mn>4</mn> </msub> </semantics></math>; (<b>c</b>) Scene <math display="inline"><semantics> <msub> <mi>S</mi> <mn>5</mn> </msub> </semantics></math>; (<b>d</b>) Scene <math display="inline"><semantics> <msub> <mi>S</mi> <mn>6</mn> </msub> </semantics></math>.</p>
Full article ">Figure 14
<p>Comparison of the methods precision. (<b>a</b>) Scene <math display="inline"><semantics> <msub> <mi>S</mi> <mn>3</mn> </msub> </semantics></math>; (<b>b</b>) Scene <math display="inline"><semantics> <msub> <mi>S</mi> <mn>4</mn> </msub> </semantics></math>; (<b>c</b>) Scene <math display="inline"><semantics> <msub> <mi>S</mi> <mn>5</mn> </msub> </semantics></math>; (<b>d</b>) Scene <math display="inline"><semantics> <msub> <mi>S</mi> <mn>6</mn> </msub> </semantics></math>.</p>
Full article ">
13 pages, 4654 KiB  
Article
Estimating Winter Wheat Plant Nitrogen Content by Combining Spectral and Texture Features Based on a Low-Cost UAV RGB System throughout the Growing Season
by Liyuan Zhang, Xiaoying Song, Yaxiao Niu, Huihui Zhang, Aichen Wang, Yaohui Zhu, Xingye Zhu, Liping Chen and Qingzhen Zhu
Agriculture 2024, 14(3), 456; https://doi.org/10.3390/agriculture14030456 - 11 Mar 2024
Cited by 1 | Viewed by 1088
Abstract
As prior information for precise nitrogen fertilization management, plant nitrogen content (PNC), which is obtained timely and accurately through a low-cost method, is of great significance for national grain security and sustainable social development. In this study, the potential of the low-cost unmanned [...] Read more.
As prior information for precise nitrogen fertilization management, plant nitrogen content (PNC), which is obtained timely and accurately through a low-cost method, is of great significance for national grain security and sustainable social development. In this study, the potential of the low-cost unmanned aerial vehicle (UAV) RGB system was investigated for the rapid and accurate estimation of winter wheat PNC across the growing season. Specifically, texture features were utilized as complements to the commonly used spectral information. Five machine learning regression algorithms, including support vector machines (SVMs), classification and regression trees, artificial neural networks, K-nearest neighbors, and random forests, were employed to establish the bridge between UAV RGB image-derived features and ground-truth PNC, with multivariate linear regression serving as the reference. The results show that both spectral and texture features had significant correlations with ground-truth PNC, indicating the potential of low-cost UAV RGB images to estimate winter wheat PNC. The H channel, S4O6, and R_SE and R_EN had the highest correlation among the spectral indices, Gabor texture features, and grey level co-occurrence matrix texture features, with absolute Pearson’s correlation coefficient values of 0.63, 0.54, and 0.69, respectively. When the texture features were used together with spectral indices, the PNC estimation accuracy was enhanced, with the root mean square error (RMSE) decreasing from 2.56 to 2.24 g/kg, for instance, when using the SVM regression algorithm. The SVM regression algorithm with validation achieved the highest estimation accuracy, with a coefficient of determination (R2) of 0.62 and an RMSE of 2.15 g/kg based on the optimal feature combination of B_CON, B_M, G_DIS, H, NGBDI, R_EN, R_M, R_SE, S3O7, and VEG. Overall, this study demonstrated that the low-cost UAV RGB system could be successfully used to map the PNC of winter wheat across the growing season. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>Overview of the winter wheat experimental design. R1–R5 presents replication areas of 1 to 5, N0–N4 represents nitrogen levels of 0 to 4, and S1–S5 represents winter wheat species of 1 to 5. GCP is the abbreviation used for ground control point.</p>
Full article ">Figure 2
<p>Plant nitrogen content (PNC) bar plots, which were drawn for the five key growth stages of winter wheat, with nitrogen application levels and winter wheat species as two factors.</p>
Full article ">Figure 3
<p>Pearson’s correlation coefficient (r) between PNC and each individual feature derived from the UAV RGB images. (<b>a</b>) Spectral indices, (<b>b</b>) Gabor texture features, and (<b>c</b>) GLCM texture features.</p>
Full article ">Figure 4
<p>PNC estimation accuracy derived from the six regression algorithms based on the optimal feature combination for each of the seven feature categories.</p>
Full article ">Figure 5
<p>Estimation performance of the PNC models, which were established based on corresponding optimal features. The black dashed line is 1:1 line.</p>
Full article ">Figure 6
<p>PNC maps of winter wheat obtained from the UAV RGB images and the SVM regression algorithm.</p>
Full article ">
16 pages, 1633 KiB  
Article
Feature Extraction Algorithm of Massive Rainstorm Debris Flow Based on Ecological Environment Telemetry
by Jun Li, Yuandi Zhao, Na He and Filip Gurkalo
Water 2023, 15(21), 3807; https://doi.org/10.3390/w15213807 - 31 Oct 2023
Viewed by 1022
Abstract
In order to accurately extract the characteristics of debris flow caused by group rainstorms, effectively identify the on-site information of debris flow, and provide a scientific basis for debris flow monitoring, early warning and disaster control, this paper proposes a method for extracting [...] Read more.
In order to accurately extract the characteristics of debris flow caused by group rainstorms, effectively identify the on-site information of debris flow, and provide a scientific basis for debris flow monitoring, early warning and disaster control, this paper proposes a method for extracting the characteristics of heavy rainstorm debris flow using multiregional ecological environment remote sensing. In the ecological environment where debris flows occur frequently, remote sensing data of heavy rainstorm debris flows are preprocessed using remote sensing technology, providing an important basis for the feature extraction of debris flows. The kernel principal component analysis method and Gabor filters are innovatively used to extract the spectral and texture features of rainstorm and debris flow remote sensing images, and the convolutional neural network structure is improved based on the open source deep learning framework, integrating multilevel features to generate debris flow feature maps. The improved convolution neural network is then used to extract the secondary features of the fusion feature map, and the feature extraction of heavy rainstorm debris flow is realized. The experiment shows that this method can accurately extract the characteristics of heavy rainstorm debris flow. Fused remote sensing images of debris flow effectively ameliorate the problem of insufficient informational content in a single image and improve image clarity. When the Gabor kernel function has eight different directions, the feature extraction effect of the debris flow image in each direction of the heavy rainstorm is the best. Full article
Show Figures

Figure 1

Figure 1
<p>Multilevel feature extraction network framework of mass rainstorm debris flow.</p>
Full article ">Figure 2
<p>Hyperspectral remote sensing image of the original debris flow.</p>
Full article ">Figure 3
<p>Image of debris flow after feature extraction using the proposed method.</p>
Full article ">Figure 4
<p>Information entropy of fused images under different definition conditions.</p>
Full article ">Figure 5
<p>Feature extraction effect when different values are taken.</p>
Full article ">
20 pages, 25160 KiB  
Article
Edge Consistency Feature Extraction Method for Multi-Source Image Registration
by Yang Zhou, Zhen Han, Zeng Dou, Chengbin Huang, Li Cong, Ning Lv and Chen Chen
Remote Sens. 2023, 15(20), 5051; https://doi.org/10.3390/rs15205051 - 21 Oct 2023
Cited by 1 | Viewed by 1331
Abstract
Multi-source image registration has often suffered from great radiation and geometric differences. Specifically, grayscale and texture from similar landforms in different source images often show significantly different visual features, and these differences disturb the corresponding point extraction in the following image registration process. [...] Read more.
Multi-source image registration has often suffered from great radiation and geometric differences. Specifically, grayscale and texture from similar landforms in different source images often show significantly different visual features, and these differences disturb the corresponding point extraction in the following image registration process. Considering that edges between heterogeneous images can provide homogeneous information and more consistent features can be extracted based on image edges, an edge consistency radiation-change insensitive feature transform (EC-RIFT) method is proposed in this paper. Firstly, the noise and texture interference are reduced by preprocessing according to the image characteristics. Secondly, image edges are extracted based on phase congruency, and an orthogonal Log-Gabor filter is performed to replace the global algorithm. Finally, the descriptors are built with logarithmic partition of the feature point neighborhood, which improves the robustness of the descriptors. Comparative experiments on datasets containing multi-source remote sensing image pairs show that the proposed EC-RIFT method outperforms other registration methods in terms of precision and effectiveness. Full article
Show Figures

Figure 1

Figure 1
<p>Texture and edge differences between multi-source images. The red box represents textured terrain, and the green box contains significant edges.</p>
Full article ">Figure 2
<p>Implementation flow chart of EC-RIFT algorithm.</p>
Full article ">Figure 3
<p>Results of preprocessing. (<b>a</b>) Original SAR image. (<b>b</b>) Denoised SAR image. (<b>c</b>) Original optical image. (<b>d</b>) Enhanced optical image.</p>
Full article ">Figure 4
<p>Features detected with LG/OLG. (<b>a</b>) LG-SAR1. (<b>b</b>) LG-OPT1. (<b>c</b>) LG-SAR2. (<b>d</b>) LG-OPT2. (<b>e</b>) OLG-SAR1. (<b>f</b>) OLG-OPT1. (<b>g</b>) OLG-SAR2. (<b>h</b>) OLG-OPT2.</p>
Full article ">Figure 5
<p>The descriptors of EC-RIFT method.</p>
Full article ">Figure 6
<p>Dataset presentation. The (<b>a</b>–<b>g</b>) denote sar-optical image pairs of different landforms. And the (<b>h</b>–<b>j</b>) represent infrared-optical, day-night, depth-optical landforms, respectively.</p>
Full article ">Figure 7
<p>Registration results with/without preprocessing. (<b>a</b>) Without preprocessing. (<b>b</b>) SAR image is denoised. (<b>c</b>) Optical image is enhanced. (<b>d</b>) Both images are preprocessed.</p>
Full article ">Figure 8
<p>Repeatability of OLG and LG. The bold numbers represent the higher scores.</p>
Full article ">Figure 9
<p>The similarity of MIM.</p>
Full article ">Figure 10
<p>Comparison of OLG and LG. The bold numbers represent the best results.</p>
Full article ">Figure 11
<p>Matching results for SAR–optical images. (<b>a</b>) Pair 1. (<b>b</b>) Pair 2. (<b>c</b>) Pair 3. (<b>d</b>) Pair 4. (<b>e</b>) Pair 5.</p>
Full article ">Figure 12
<p>Average running time of registration methods.</p>
Full article ">Figure 13
<p>Matching results for multi-source images. (<b>a</b>) Infrared–optical. (<b>b</b>) Day–night. (<b>c</b>) Depth–optical.</p>
Full article ">Figure 14
<p>MIMs of images. (<b>a</b>) MIM of original image pairs. (<b>b</b>) MIM of processed image pairs.</p>
Full article ">Figure 15
<p>Matching results with scale and rotation differences. (<b>a</b>) Scale difference. (<b>b</b>) Rotation difference. (<b>c</b>) Rotation difference.</p>
Full article ">
14 pages, 8937 KiB  
Article
A Fabric Defect Segmentation Model Based on Improved Swin-Unet with Gabor Filter
by Haitao Xu, Chengming Liu, Shuya Duan, Liangpin Ren, Guozhen Cheng and Bing Hao
Appl. Sci. 2023, 13(20), 11386; https://doi.org/10.3390/app132011386 - 17 Oct 2023
Viewed by 1255
Abstract
Fabric inspection is critical in fabric manufacturing. Automatic detection of fabric defects in the textile industry has always been an important research field. Previously, manual visual inspection was commonly used; however, there were drawbacks such as high labor costs, slow detection speed, and [...] Read more.
Fabric inspection is critical in fabric manufacturing. Automatic detection of fabric defects in the textile industry has always been an important research field. Previously, manual visual inspection was commonly used; however, there were drawbacks such as high labor costs, slow detection speed, and high error rates. Recently, many defect detection methods based on deep learning have been proposed. However, problems need to be solved in the existing methods, such as detection accuracy and interference of complex background textures. In this paper, we propose an efficient segmentation algorithm that combines traditional operators with deep learning networks to alleviate the existing problems. Specifically, we introduce a Gabor filter into the model, which provides the unique advantage of extracting low-level texture features to solve the problem of texture interference and enable the algorithm to converge quickly in the early stages of training. Furthermore, we design a U-shaped architecture that is not completely symmetrical, making model training easier. Meanwhile, multi-stage result fusion is proposed for precise location of defects. The design of this framework significantly improves the detection accuracy and effectively breaks through the limitations of transformer-based models. Experimental results show that on a dataset with one class, a small amount of data, and complex sample background texture, our method achieved 90.03% and 33.70% in ACC and IoU, respectively, which is almost 10% higher than other previous state of the art models. Experimental results based on three different fabric datasets consistently show that the proposed model has excellent performance and great application potential in the industrial field. Full article
Show Figures

Figure 1

Figure 1
<p>Swin Transformer block.</p>
Full article ">Figure 2
<p>The architecture of our model.</p>
Full article ">Figure 3
<p>Gabor kernel under the condition of <math display="inline"><semantics> <mrow> <mi>θ</mi> <mo>=</mo> <mo>[</mo> <mn>0</mn> <mo>,</mo> <mn>45</mn> <mo>,</mo> <mn>90</mn> <mo>,</mo> <mn>135</mn> <mo>]</mo> <mo> </mo> <mo>,</mo> <mi>k</mi> <mi>e</mi> <mi>r</mi> <mi>n</mi> <mi>e</mi> <mi>l</mi> <mo>=</mo> <mn>256</mn> <mo>,</mo> <mo> </mo> <mi>λ</mi> <mo>=</mo> <mn>15</mn> <mo>,</mo> <mo> </mo> <mi>γ</mi> <mo>=</mo> <mn>0.5</mn> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>σ</mi> <mo>=</mo> <mn>25</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 4
<p>Typical fabric samples in Cropped AITEX dataset.</p>
Full article ">Figure 5
<p>Representative fabric samples in Colored Fabric dataset.</p>
Full article ">Figure 6
<p>Representative fabric samples in One Class Fabric dataset.</p>
Full article ">Figure 7
<p>IoU.</p>
Full article ">Figure 8
<p>The proposed method compared with other state of the art methods. Analysis of defect detection results: first column, original defect image; second column, ground truth; third column, UNet segmentation results [<a href="#B27-applsci-13-11386" class="html-bibr">27</a>]; fourth column, AttuNet segmentation results [<a href="#B28-applsci-13-11386" class="html-bibr">28</a>]; fifth column, NestedUNet segmentation results [<a href="#B29-applsci-13-11386" class="html-bibr">29</a>]; sixth column, segmentation results of the method proposed by Tabernik et al. [<a href="#B3-applsci-13-11386" class="html-bibr">3</a>]; seventh column, segmentation results of the method proposed in this article.</p>
Full article ">Figure 9
<p>Evaluation index curves of different combinations. Swin-Unet refers to the original Swin-Unet using the BCE–Dice loss; Our(0) refers to our proposed model using the BCE-Dice loss; Our(1) represents the proposed model using FDL without the Gabor filter; Our(2) represents the proposed model trained with fixed parameters of the Gabor filter layer and FDL.</p>
Full article ">
33 pages, 24854 KiB  
Article
A New Region-Based Minimal Path Selection Algorithm for Crack Detection and Ground Truth Labeling Exploiting Gabor Filters
by Gonzalo de León, Nicholas Fiorentini, Pietro Leandri and Massimo Losa
Remote Sens. 2023, 15(11), 2722; https://doi.org/10.3390/rs15112722 - 24 May 2023
Cited by 6 | Viewed by 1418
Abstract
Cracks are fractures or breaks that occur in materials such as concrete, metals, rocks, and other solids. Various methods are used to detect and monitor cracks; among many of them, image-based methodologies allow fast identification of the distress and easy quantification of the [...] Read more.
Cracks are fractures or breaks that occur in materials such as concrete, metals, rocks, and other solids. Various methods are used to detect and monitor cracks; among many of them, image-based methodologies allow fast identification of the distress and easy quantification of the percentage of cracks in the scene. Two main categories can be identified: classical and deep learning approaches. In the last decade, the tendency has moved towards the use of the latter. Even though they have proven their outstanding predicting performance, they suffer some drawbacks: a “black-box” nature leaves the user blind and without the possibility of modifying any parameters, a huge amount of labeled data is generally needed, a process that requires expert judgment is always required, and, finally, they tend to be time-consuming. Accordingly, the present study details the methodology for a new algorithm for crack segmentation based on the theory of minimal path selection combined with a region-based approach obtained through the segmentation of texture features extracted using Gabor filters. A pre-processing step is described, enabling the equalization of brightness and shadows, which results in better detection of local minima. These local minimal are constrained by a minimum distance between adjacent points, enabling a better coverage of the cracks. Afterward, a region-based segmentation technique is introduced to determine two areas that are used to determine threshold values used for rejection. This step is critical to generalize the algorithm to images presenting close-up scenes or wide cracks. Finally, a geometrical thresholding step is presented, allowing the exclusion of rounded areas and small isolated cracks. The results showed a very competitive F1-score (0.839), close to state-of-the-art values achieved with deep learning techniques. The main advantage of this approach is the transparency of the workflow, contrary to what happens with deep learning frameworks. In the proposed approach, no prior information is required; however, the statistical parameters may have to be adjusted to the particular case and requirements of the situation. The proposed algorithm results in a useful tool for researchers and practitioners needing to validate their results against some reference or needing labeled data for their models. Moreover, the current study could establish the grounds to standardize the procedure for crack segmentation with a lower human bias and faster results. The direct application of the methodology to images obtained with any low-cost sensor makes the proposed algorithm an operational support tool for authorities needing crack detection systems in order to monitor and evaluate the current state of the infrastructures, such as roads, tunnels, or bridges. Full article
(This article belongs to the Special Issue Road Detection, Monitoring and Maintenance Using Remotely Sensed Data)
Show Figures

Figure 1

Figure 1
<p>Flat-field underestimation of <math display="inline"><semantics> <mrow> <mi>σ</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math>. <b>Up left:</b> image. <b>Up right:</b> color map. <b>Down left:</b> shading component. <b>Down right:</b> flat-fielded image.</p>
Full article ">Figure 2
<p>Flat-field overestimation with <math display="inline"><semantics> <mrow> <mi>σ</mi> <mo>=</mo> <mn>50</mn> </mrow> </semantics></math>. <b>Up left:</b> image. <b>Up right:</b> color map. <b>Down left:</b> shading component. <b>Down right:</b> flat-fielded image.</p>
Full article ">Figure 3
<p>Gaussian bank.</p>
Full article ">Figure 4
<p>Flat-fielded image with PCA Gaussian. <b>Up left:</b> image. <b>Up right:</b> color map. <b>Down left:</b> shading component. <b>Down right:</b> flat-fielded image.</p>
Full article ">Figure 5
<p>Gabor filter banks. Kernel size at top of each column. X-axis wavelengths vs. Y-axis orientation.</p>
Full article ">Figure 6
<p>Gabor filter real part at <math display="inline"><semantics> <mrow> <mi>λ</mi> <mo>=</mo> <mn>90.51</mn> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>θ</mi> <mo>=</mo> <msup> <mn>120</mn> <mo>∘</mo> </msup> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>γ</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 7
<p>Filtered Gabor magnitude response. Kernel size at top of each column. X-axis wavelengths vs. Y-axis orientation.</p>
Full article ">Figure 8
<p>Gabor textural feature. <b>Left:</b> flat-field image. <b>Right:</b> Gabor texture features.</p>
Full article ">Figure 9
<p>Gabor textural features. <b>Left:</b> without flat-field correction. <b>Right:</b> with flat-field correction.</p>
Full article ">Figure 10
<p>Local minima original. <b>Left:</b> full image. <b>Right:</b> zoomed-in.</p>
Full article ">Figure 11
<p>Original local minima. <b>Left:</b> synthetic image. <b>Right:</b> local minima.</p>
Full article ">Figure 12
<p>Local minima with repulsion. <b>Left:</b> synthetic image. <b>Center:</b> <span class="html-italic">d<sub>min</sub> =</span>1. <b>Right:</b> <span class="html-italic">d<sub>min</sub></span> = 2.</p>
Full article ">Figure 13
<p>Local minima with repulsion. Zoomed-in. <b>Left:</b> original local minima. <b>Right:</b> repulsion local minima.</p>
Full article ">Figure 14
<p>Decisional local minima flow chart.</p>
Full article ">Figure 15
<p>Example. <b>Left:</b> flat-fielded image. <b>Right:</b> Gabor features.</p>
Full article ">Figure 16
<p>Region segmentation. <b>Left:</b> Gabor features. <b>Right:</b> segmentation.</p>
Full article ">Figure 17
<p>Regional local minima. <b>Left:</b> Region 1. <b>Right:</b> Region 2.</p>
Full article ">Figure 18
<p>Local minima original vs. region-based. <b>Left:</b> original. <b>Right:</b> region-based.</p>
Full article ">Figure 19
<p>Influence from flat-field on the segmentation.</p>
Full article ">Figure 20
<p>Amount of local minima points. <b>Left:</b> global. Total points: 1768. <b>Right:</b> region-based. Total points: 931.</p>
Full article ">Figure 21
<p>Local minima. <b>Left:</b> Full size image. <b>Right:</b> Zoomed-in portion.</p>
Full article ">Figure 22
<p>Local minima. <b>Left:</b> zoomed-in local minima. <b>Right:</b> zoomed-in minimal paths.</p>
Full article ">Figure 23
<p>Zoomed-in minimal paths. <b>Left:</b> minimal paths. <b>Right:</b> number of times traveled.</p>
Full article ">Figure 24
<p>Rejection original. <b>Left:</b> all minimal paths. <b>Right:</b> accepted minimal paths.</p>
Full article ">Figure 25
<p>Histograms. <b>Left:</b> threshold original. <b>Right:</b> threshold with Gaussian fit.</p>
Full article ">Figure 26
<p>Accepted paths. <b>Left:</b> threshold original. <b>Right:</b> threshold with Gaussian fit.</p>
Full article ">Figure 27
<p>Histograms. <b>Left:</b> costs. <b>Right:</b> intensities.</p>
Full article ">Figure 28
<p>Accepted paths. <b>Left:</b> threshold original. <b>Right:</b> threshold with global.</p>
Full article ">Figure 29
<p>Regions. <b>Left:</b> image. <b>Right:</b> regions.</p>
Full article ">Figure 30
<p>Regions histogram. <b>Up:</b> global. <b>Bottom:</b> region-based.</p>
Full article ">Figure 31
<p>Thresholds. <b>Left:</b> original. <b>Center:</b> global. <b>Right:</b> region-based.</p>
Full article ">Figure 32
<p>Results. (<b>A</b>) Thin cracks in perspective and secondary cracks. (<b>B</b>) Wide crack and watermark. (<b>C</b>) Medium width crack with secondary cracks. (<b>D</b>) View from the top, thin cracks, secondary cracks, and oil stain.</p>
Full article ">Figure 33
<p>Filtered results. (<b>A</b>) Thin cracks in perspective and secondary cracks. (<b>B</b>) Wide crack and watermark. (<b>C</b>) Medium width crack with secondary cracks. (<b>D</b>) View from the top, thin cracks, secondary cracks, and oil stain.</p>
Full article ">Figure 34
<p>Results for RB-MPS (<b>A</b>,<b>C</b>,<b>E</b>,<b>G</b>) vs. MPS (<b>B</b>,<b>D</b>,<b>F</b>,<b>H</b>).</p>
Full article ">Figure 35
<p>Results. (<b>A</b>) Original images. (<b>B</b>) GT. (<b>C</b>) DC. (<b>D</b>) MPS. (<b>E</b>) RB-MPS.</p>
Full article ">
19 pages, 3302 KiB  
Article
Stain Defect Classification by Gabor Filter and Dual-Stream Convolutional Neural Network
by Min-Ho Ha, Young-Gyu Kim and Tae-Hyoung Park
Appl. Sci. 2023, 13(7), 4540; https://doi.org/10.3390/app13074540 - 3 Apr 2023
Cited by 2 | Viewed by 1523
Abstract
A stain defect is difficult to detect with the human eye because of its characteristic of having a very minimal difference in brightness with the local area of the surface. Recently, with the development of Deep learning, the Convolutional Neural Network (CNN) based [...] Read more.
A stain defect is difficult to detect with the human eye because of its characteristic of having a very minimal difference in brightness with the local area of the surface. Recently, with the development of Deep learning, the Convolutional Neural Network (CNN) based stain defect classification method has been proposed. This paper proposes a Dual-stream CNN for stain defect classification using a Gabor filter image. Using Dual-stream structure CNN, Gabor filter images and Gray image (Original) preserve their respective features. The experiment based on the Magnetic Tile (MT) stain data set and the Compact Camera Module (CCM) stain dataset confirms that the proposed method has an improved performance based on the precision, recall, and F1-score in comparison to the Single-stream extraction-based method. Gabor filter images have an advantage in image texture analysis and can be used as an input to the CNNs. The Dual-stream structure better extracts the features needed for classification. Full article
(This article belongs to the Section Robotics and Automation)
Show Figures

Figure 1

Figure 1
<p>Stain defect classification system.</p>
Full article ">Figure 2
<p>Gabor image according to phase in MT stain image. (<b>a</b>) gray-scale inspection image; from (<b>b</b>–<b>i</b>) Gabor filter images according to phase.</p>
Full article ">Figure 3
<p>Gabor image according to phase in CCM stain image. (<b>a</b>) gray-scale inspection image; from (<b>b</b>–<b>i</b>) Gabor filter images according to phase.</p>
Full article ">Figure 4
<p>Single-stream network structure based on ResNet50.</p>
Full article ">Figure 5
<p>Dual-stream network structure based on ResNet50.</p>
Full article ">Figure 6
<p>Magnetic Tile Dataset (<b>Left</b>: image, <b>Right</b>: Stain Defect segmentation image).</p>
Full article ">Figure 7
<p>Compact Camera Module Dataset (<b>Left</b>: image, <b>Right</b>: Stain Defect segmentation image).</p>
Full article ">Figure 8
<p>CAM image result for CCM Stain image in a Single-stream Network.</p>
Full article ">Figure 9
<p>CAM image result for CCM Stain image in a Dual-stream Network.</p>
Full article ">Figure 10
<p>CAM image result for CCM OK image in a Single-stream Network.</p>
Full article ">Figure 11
<p>CAM image result for CCM OK image in a Dual-stream Network.</p>
Full article ">
21 pages, 3830 KiB  
Article
GabROP: Gabor Wavelets-Based CAD for Retinopathy of Prematurity Diagnosis via Convolutional Neural Networks
by Omneya Attallah
Diagnostics 2023, 13(2), 171; https://doi.org/10.3390/diagnostics13020171 - 4 Jan 2023
Cited by 22 | Viewed by 3479
Abstract
One of the most serious and dangerous ocular problems in premature infants is retinopathy of prematurity (ROP), a proliferative vascular disease. Ophthalmologists can use automatic computer-assisted diagnostic (CAD) tools to help them make a safe, accurate, and low-cost diagnosis of ROP. All previous [...] Read more.
One of the most serious and dangerous ocular problems in premature infants is retinopathy of prematurity (ROP), a proliferative vascular disease. Ophthalmologists can use automatic computer-assisted diagnostic (CAD) tools to help them make a safe, accurate, and low-cost diagnosis of ROP. All previous CAD tools for ROP diagnosis use the original fundus images. Unfortunately, learning the discriminative representation from ROP-related fundus images is difficult. Textural analysis techniques, such as Gabor wavelets (GW), can demonstrate significant texture information that can help artificial intelligence (AI) based models to improve diagnostic accuracy. In this paper, an effective and automated CAD tool, namely GabROP, based on GW and multiple deep learning (DL) models is proposed. Initially, GabROP analyzes fundus images using GW and generates several sets of GW images. Next, these sets of images are used to train three convolutional neural networks (CNNs) models independently. Additionally, the actual fundus pictures are used to build these networks. Using the discrete wavelet transform (DWT), texture features retrieved from every CNN trained with various sets of GW images are combined to create a textural-spectral-temporal demonstration. Afterward, for each CNN, these features are concatenated with spatial deep features obtained from the original fundus images. Finally, the previous concatenated features of all three CNN are incorporated using the discrete cosine transform (DCT) to lessen the size of features caused by the fusion process. The outcomes of GabROP show that it is accurate and efficient for ophthalmologists. Additionally, the effectiveness of GabROP is compared to recently developed ROP diagnostic techniques. Due to GabROP’s superior performance compared to competing tools, ophthalmologists may be able to identify ROP more reliably and precisely, which could result in a reduction in diagnostic effort and examination time. Full article
(This article belongs to the Special Issue Advances in Retinopathy)
Show Figures

Figure 1

Figure 1
<p>Examples of the dataset’s images, (<b>a</b>) diseased and (<b>b</b>), not diseased.</p>
Full article ">Figure 2
<p>The stages of the proposed GabROP CAD tool.</p>
Full article ">Figure 3
<p>Samples of the generated GW images for both classes of the ROP dataset, (<b>a</b>) Diseased, (<b>b</b>) Not Diseased.</p>
Full article ">Figure 4
<p>Diagnostic accuracy of the three classifiers trained with the features extracted from ResNet-50 learned using the individual GW images compared the fused DWT features.</p>
Full article ">Figure 5
<p>Diagnostic accuracy of the three classifiers trained with the features extracted from DarkNet-53 learned using the individual GW images compared to the fused DWT features.</p>
Full article ">Figure 6
<p>Diagnostic accuracy of the three classifiers trained with the features extracted from MobileNet learned using the individual GW images compared to the fused DWT features.</p>
Full article ">Figure 7
<p>Diagnostic accuracy of the three classifiers trained with spatial features extracted from ResNet-50 learned using the original fundus images compared the fused DWT features obtained by CNNs learned with GW images and the combination of the two.</p>
Full article ">Figure 8
<p>Diagnostic accuracy of the three classifiers trained with spatial features extracted from DarkNet-53 learned using the original fundus images compared the fused DWT features obtained by CNNs learned with GW images and the combination of the two.</p>
Full article ">Figure 9
<p>Diagnostic accuracy of the three classifiers trained with spatial features extracted from MobileNet learned using the original fundus images compared the fused DWT features obtained by CNNs learned with GW images and the combination of the two.</p>
Full article ">Figure 10
<p>Diagnostic accuracy of the three classifiers trained with integrated features of the third fusion stage (fusing of the second fusion stage features of the three CNNs using DCT) versus the number of DCT features.</p>
Full article ">Figure 11
<p>ROC curve and the AUC of the SVM classifier trained with integrated features of the third fusion stage (fusing of the second fusion stage features of the three CNNs using DCT (2000 features)).</p>
Full article ">Figure 12
<p>Comparison among the highest accuracy attained in each fusion stage of GabROP.</p>
Full article ">
19 pages, 3260 KiB  
Article
Binary and Multi-Class Malware Threads Classification
by Ismail Taha Ahmed, Norziana Jamil, Marina Md. Din and Baraa Tareq Hammad
Appl. Sci. 2022, 12(24), 12528; https://doi.org/10.3390/app122412528 - 7 Dec 2022
Cited by 5 | Viewed by 1734
Abstract
The security of a computer system can be harmed by specific applications, such as malware. Malware comprises unwanted, dangerous enemies that aim to compromise the security and generate significant loss. Consequently, Malware Detection (MD) and Malware Classification (MC) has emerged as a key [...] Read more.
The security of a computer system can be harmed by specific applications, such as malware. Malware comprises unwanted, dangerous enemies that aim to compromise the security and generate significant loss. Consequently, Malware Detection (MD) and Malware Classification (MC) has emerged as a key issue for the cybersecurity society. MD only involves locating malware without determining what kind of malware it is, but MC comprises assigning a class of malware to a particular sample. Recently, a few techniques for analyzing malware quickly have been put out. However, there remain numerous difficulties, such as the low classification accuracy of samples from related malware families, the computational complexity, and consumption of resources. These difficulties make detecting and classifying malware very challenging. Therefore, in this paper, we proposed an efficient malware detection and classification technique that combines Segmentation-based Fractal Texture Analysis (SFTA) and Gaussian Discriminant Analysis (GDA). The outcomes of the experiment demonstrate that the SFTA-GDA produces a high classification rate. There are three main steps involved in our malware analysis, namely: (i) malware conversion; (ii) feature extraction; and (iii) classification. We initially convert the RGB malware images into grayscale malware images for effective malware analysis. The SFTA and Gabor features are then extracted from gray-scale images in the feature extraction step. Finally, the classification is carried out by GDA and Naive Bayes (NB). The proposed method is evaluated on a common MaleVis dataset. The proposed SFTA-GDA is the effective choice since it produces the highest accuracy rate across all families of the MaleVis Database. Experimental findings indicate that the accuracy rate was 98%, which is higher than the overall accuracy from the existing state-of-the-art methods. Full article
Show Figures

Figure 1

Figure 1
<p>Existing Taxonomy of Malware Analysis.</p>
Full article ">Figure 2
<p>SFTA Extraction process.</p>
Full article ">Figure 3
<p>A two-dimensional Gabor filter with eight orientations and five scales [<a href="#B31-applsci-12-12528" class="html-bibr">31</a>].</p>
Full article ">Figure 4
<p>Proposed method Flowchart.</p>
Full article ">Figure 5
<p>The Conversion Process Diagram.</p>
Full article ">Figure 6
<p>The distribution of MaleVis datasets samples among different malware classes.</p>
Full article ">Figure 7
<p>Various Samples collected from the MaleVis Dataset [<a href="#B36-applsci-12-12528" class="html-bibr">36</a>].</p>
Full article ">Figure 8
<p>Detection Accuracy Rate of Proposed Method.</p>
Full article ">Figure 9
<p>The classification Accuracy-based SFTA and Gabor features across NB Classifier.</p>
Full article ">Figure 10
<p>The classification Accuracy-based SFTA and Gabor features across GDA Classifier.</p>
Full article ">
15 pages, 1899 KiB  
Article
Obscurant Segmentation in Long Wave Infrared Images Using GLCM Textures
by Mohammed Abuhussein and Aaron Robinson
J. Imaging 2022, 8(10), 266; https://doi.org/10.3390/jimaging8100266 - 30 Sep 2022
Cited by 4 | Viewed by 1675
Abstract
The benefits of autonomous image segmentation are readily apparent in many applications and garners interest from stakeholders in many fields. The wide range of benefits encompass applications ranging from medical diagnosis, where the shape of the grouped pixels increases diagnosis accuracy, to autonomous [...] Read more.
The benefits of autonomous image segmentation are readily apparent in many applications and garners interest from stakeholders in many fields. The wide range of benefits encompass applications ranging from medical diagnosis, where the shape of the grouped pixels increases diagnosis accuracy, to autonomous vehicles where the grouping of pixels defines roadways, traffic signs, other vehicles, etc. It even proves beneficial in many phases of machine learning, where the resulting segmentation can be used as inputs to the network or as labels for training. The majority of the available image segmentation algorithmic development and results focus on visible image modalities. Therefore, in this treatment, the authors present the results of a study designed to identify and improve current semantic methods for infrared scene segmentation. Specifically, the goal is to propose a novel approach to provide tile-based segmentation of occlusion clouds in Long Wave Infrared images. This work complements the collection of well-known semantic segmentation algorithms applicable to thermal images but requires a vast dataset to provide accurate performance. We document performance in applications where the distinction between dust cloud tiles and clear tiles enables conditional processing. Therefore, the authors propose a Gray Level Co-Occurrence Matrix (GLCM) based method for infrared image segmentation. The main idea of our approach is that GLCM features are extracted from local tiles in the image and used to train a binary classifier to provide indication of tile occlusions. Our method introduces a new texture analysis scheme that is more suitable for image segmentation than the solitary Gabor segmentation or Markov Random Field (MRF) scheme. Our experimental results show that our algorithm performs well in terms of accuracy and a better inter-region homogeneity than the pixel-based infrared image segmentation algorithms. Full article
(This article belongs to the Section Computer Vision and Pattern Recognition)
Show Figures

Figure 1

Figure 1
<p>The GLCM matrix is calculated for center pixel P(x,y) using the pixels at <math display="inline"><semantics> <mrow> <mi>d</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 2
<p>Correlations in the initial GLCM feature vector.</p>
Full article ">Figure 3
<p>Sample images from our dataset (<b>top row</b>) along with corresponding occlusion segmentation ground truth (<b>bottom row</b>).</p>
Full article ">Figure 4
<p>In (<b>A</b>), the features are extracted by picking <math display="inline"><semantics> <mrow> <mn>128</mn> <mo>×</mo> <mn>128</mn> </mrow> </semantics></math> tiles which are either completely occluded, or clear. Then we extract the GLCM features from tiles with varying sizes. Next, we train the classifier using the feature vectors and the labels from the tiles collected (<b>B</b>). To segment the obscuring cloud (<b>C</b>), a larger sliding window passes along the image and generates the GLCM features for the window then classifies it. Then a smaller window will only process the tiles that include obscurants. The scaled sliding window process will continue until the stopping size is reached.</p>
Full article ">Figure 5
<p>Classification performance evaluation for different radii to generate the GLCM matrix.</p>
Full article ">Figure 6
<p>Classification performance evaluation. The normalized confusion matrix demonstrates the performance of the proposed model in classifying tiles containing obscurant clouds. The model tends to miss-classify tiles with tiles near the edge of the obscurant clouds since the human labels considered the edges clear although containing very light occlusion.</p>
Full article ">Figure 7
<p>Comparing performance when including several feature vectors per tile calculated at different distances from reference pixel. The negligible improvement in performance demonstrates the irrelevance in the calculated features at distances larger than 3 tiles. However, we notice an increase in the performance, although trivial, when including two vectors (at <math display="inline"><semantics> <mrow> <mi>d</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math>, and <math display="inline"><semantics> <mrow> <mi>d</mi> <mo>=</mo> <mn>4</mn> </mrow> </semantics></math>).</p>
Full article ">Figure 8
<p>Qualitative results from running tests with varying the window size. The top row is the input image, the middle row shows the results with window size of <math display="inline"><semantics> <mrow> <mn>64</mn> <mo>×</mo> <mn>64</mn> </mrow> </semantics></math>, and the bottom row shows the masks generated from <math display="inline"><semantics> <mrow> <mn>32</mn> <mo>×</mo> <mn>32</mn> </mrow> </semantics></math> tiles.</p>
Full article ">Figure 9
<p>Sample results with the corresponding ground truth comparing the results from the proposed approach, with results from Gabor and MRF segmentation. As displayed in the results above, the proposed approach provides accurate results compared to the other two methods for texture analysis. Although texture analysis is very accurate in RGB and grayscale images, it can produce far less accurate results when tested in thermal images.</p>
Full article ">
22 pages, 8813 KiB  
Article
Multiplicative Long Short-Term Memory with Improved Mayfly Optimization for LULC Classification
by Andrzej Stateczny, Shanthi Mandekolu Bolugallu, Parameshachari Bidare Divakarachari, Kavithaa Ganesan and Jamuna Rani Muthu
Remote Sens. 2022, 14(19), 4837; https://doi.org/10.3390/rs14194837 - 28 Sep 2022
Cited by 13 | Viewed by 1851
Abstract
Land Use and Land Cover (LULC) monitoring is crucial for global transformation, sustainable land control, urban planning, urban growth prediction, and the establishment of climate regulations for long-term development. Remote sensing images have become increasingly important in many environmental planning and land use [...] Read more.
Land Use and Land Cover (LULC) monitoring is crucial for global transformation, sustainable land control, urban planning, urban growth prediction, and the establishment of climate regulations for long-term development. Remote sensing images have become increasingly important in many environmental planning and land use surveys in recent times. LULC is evaluated in this research using the Sat 4, Sat 6, and Eurosat datasets. Various spectral feature bands are involved, but unexpectedly little consideration has been given to these characteristics in deep learning models. Due to the wide availability of RGB models in computer vision, this research mainly utilized RGB bands. Once the pre-processing is carried out for the images of the selected dataset, the hybrid feature extraction is performed using Haralick texture features, an oriented gradient histogram, a local Gabor binary pattern histogram sequence, and Harris Corner Detection to extract features from the images. After that, the Improved Mayfly Optimization (IMO) method is used to choose the optimal features. IMO-based feature selection algorithms have several advantages that include features such as a high learning rate and computational efficiency. After obtaining the optimal feature selection, the LULC classes are classified using a multi-class classifier known as the Multiplicative Long Short-Term Memory (mLSTM) network. The main functionality of the multiplicative LSTM classifier is to recall appropriate information for a comprehensive duration. In order to accomplish an improved result in LULC classification, a higher amount of remote sensing data should be processed. So, the simulation outcomes demonstrated that the proposed IMO-mLSTM efficiently classifies the LULC classes in terms of classification accuracy, recall, and precision. When compared with ConvNet and Alexnet, the proposed IMO-mLSTM method accomplished accuracies of 99.99% on Sat 4, 99.98% on Sat 6, and 98.52% on the Eurosat datasets. Full article
(This article belongs to the Special Issue New Advancements in Remote Sensing Image Processing)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Overview of satellite image classification.</p>
Full article ">Figure 2
<p>Sat 4 and Sat 6 databases.</p>
Full article ">Figure 3
<p>Eurosat database.</p>
Full article ">Figure 4
<p>Graphical depiction of multiplicative LSTM.</p>
Full article ">Figure 5
<p>Performance analysis of precision and recall on Sat 4.</p>
Full article ">Figure 6
<p>Performance analysis of accuracy on Sat 4.</p>
Full article ">Figure 7
<p>Performance analysis of precision and recall on the Sat 4 dataset.</p>
Full article ">Figure 8
<p>Performance analysis of accuracy on various classes.</p>
Full article ">Figure 9
<p>Visual results of the Sat 4 dataset.</p>
Full article ">Figure 10
<p>Confusion matrix for the Sat 4 dataset.</p>
Full article ">Figure 11
<p>Performance analysis of precision and recall on the Sat 6 dataset.</p>
Full article ">Figure 12
<p>Performance analysis of accuracy on Sat 6.</p>
Full article ">Figure 13
<p>Visual results of the Sat 6 dataset.</p>
Full article ">Figure 14
<p>Confusion Matrix for the Sat 4 dataset.</p>
Full article ">Figure 15
<p>Performance of precision, recall, and accuracy on Eurosat.</p>
Full article ">Figure 16
<p>Visual analysis of Eurosat.</p>
Full article ">Figure 17
<p>Confusion Matrix for the Eurosat dataset.</p>
Full article ">Figure 18
<p>Comparative analysis of accuracy with existing classes [<a href="#B22-remotesensing-14-04837" class="html-bibr">22</a>,<a href="#B23-remotesensing-14-04837" class="html-bibr">23</a>,<a href="#B24-remotesensing-14-04837" class="html-bibr">24</a>,<a href="#B29-remotesensing-14-04837" class="html-bibr">29</a>].</p>
Full article ">Figure 19
<p>Comparative analysis of accuracy with existing DBN [<a href="#B28-remotesensing-14-04837" class="html-bibr">28</a>].</p>
Full article ">
Back to TopTop