Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (864)

Search Parameters:
Keywords = underwater image

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 1701 KiB  
Article
Deep Learning-Based Feature Matching Algorithm for Multi-Beam and Side-Scan Images
by Yu Fu, Xiaowen Luo, Xiaoming Qin, Hongyang Wan, Jiaxin Cui and Zepeng Huang
Remote Sens. 2025, 17(4), 675; https://doi.org/10.3390/rs17040675 (registering DOI) - 16 Feb 2025
Abstract
Side-scan sonar and multi-beam echo sounder (MBES) are the most widely used underwater surveying tools in marine mapping today. The MBES offers high accuracy in depth measurement but is limited by low imaging resolution due to beam density constraints. Conversely, side-scan sonar provides [...] Read more.
Side-scan sonar and multi-beam echo sounder (MBES) are the most widely used underwater surveying tools in marine mapping today. The MBES offers high accuracy in depth measurement but is limited by low imaging resolution due to beam density constraints. Conversely, side-scan sonar provides high-resolution backscatter intensity images but lacks precise positional information and often suffers from distortions. Thus, MBES and side-scan images complement each other in depth accuracy and imaging resolution. To obtain high-quality seafloor topography images in practice, matching between MBES and side-scan images is necessary. However, due to the significant differences in content and resolution between MBES depth images and side-scan backscatter images, they represent a typical example of heterogeneous images, making feature matching difficult with traditional image matching methods. To address this issue, this paper proposes a feature matching network based on the LoFTR algorithm, utilizing the intermediate layers of the ResNet-50 network to extract shared features between the two types of images. By leveraging self-attention and cross-attention mechanisms, the features of the MBES and side-scan images are combined, and a similarity matrix of the two modalities is calculated to achieve mutual matching. Experimental results show that, compared to traditional methods, the proposed model exhibits greater robustness to noise interference and effectively reduces noise. It also overcomes challenges, such as large nonlinear differences, significant geometric distortions, and high matching difficulty between the MBES and side-scan images, significantly improving the optimized image matching results. The matching error RMSE has been reduced to within six pixels, enabling the accurate matching of multi-beam and side-scan images. Full article
14 pages, 2796 KiB  
Article
Multi-View and Multi-Type Feature Fusion Rotor Biofouling Recognition Method for Tidal Stream Turbine
by Haoran Xu, Dingding Yang, Tianzhen Wang and Mohamed Benbouzid
J. Mar. Sci. Eng. 2025, 13(2), 356; https://doi.org/10.3390/jmse13020356 (registering DOI) - 15 Feb 2025
Abstract
Power generation is affected and structural instability may occur when biofouling attaches to the rotor of tidal stream turbines (TSTs). Image signals are used to identify biofouling for biofouling recognition, thus achieving on-demand maintenance, optimizing power generation efficiency, and minimizing maintenance costs. However, [...] Read more.
Power generation is affected and structural instability may occur when biofouling attaches to the rotor of tidal stream turbines (TSTs). Image signals are used to identify biofouling for biofouling recognition, thus achieving on-demand maintenance, optimizing power generation efficiency, and minimizing maintenance costs. However, image signals are sensitive to background interferences, and underwater targets blend with the water background, making it difficult to extract target features. Changes in water turbidity can affect the effectiveness of image signal biofouling recognition, which can lead to reduced recognition accuracy. In order to solve these problems, a multi-view and multi-type feature fusion (MVTFF) method is proposed to recognize rotor biofouling on TSTs for applications in TST operation and maintenance. (1) Key boundary and semantic information are captured to solve the problem of background feature interference by comparing and fusing the extracted multi-view features. (2) The local geometric description and dependency are obtained by integrating contour features into multi-view features to address the issue of the target mixing with water. The mIoU, mPA, Precision, and Recall of the experimental results show that the method achieves superior recognition performance on TST datasets with different turbidity levels. Full article
Show Figures

Figure 1

Figure 1
<p>Comparison of clear and turbid TST segmentation images.</p>
Full article ">Figure 2
<p>MVTFF network for recognizing TST rotor biofouling.</p>
Full article ">Figure 3
<p>Examples of three types of segmentation artifacts.</p>
Full article ">Figure 4
<p>Comparison before and after introduction of different views.</p>
Full article ">Figure 5
<p>Comparison before and after introduction of OB.</p>
Full article ">Figure 6
<p>Results of comparative experiments. (<b>a</b>) Input. (<b>b</b>) MVTFF. (<b>c</b>) Unet. (<b>d</b>) Swin-Unet. (<b>e</b>) deeplabV3+. (<b>f</b>) SETR.</p>
Full article ">
16 pages, 6730 KiB  
Article
Restoration of Turbid Underwater Images of Cobalt Crusts Using Combined Homomorphic Filtering and a Polarization Imaging System
by Enzu Peng, Chengyi Liu and Haiming Zhao
Sensors 2025, 25(4), 1088; https://doi.org/10.3390/s25041088 - 11 Feb 2025
Abstract
Marine cobalt-rich crusts, extensively used in industries such as aerospace, automotive, and electronics, are crucial mineral resources located on the ocean floor. To effectively exploit these valuable resources, underwater imaging is essential for real-time detection and distribution mapping in mining areas. However, the [...] Read more.
Marine cobalt-rich crusts, extensively used in industries such as aerospace, automotive, and electronics, are crucial mineral resources located on the ocean floor. To effectively exploit these valuable resources, underwater imaging is essential for real-time detection and distribution mapping in mining areas. However, the presence of suspended particles in the seabed mining environment severely degrades image quality due to light scattering and absorption, hindering the effective identification of the target objects. Traditional image processing techniques—including spatial and frequency domain methods—are ineffective in addressing the interference caused by suspended particles and offer only limited enhancement effects. This paper proposes a novel underwater image restoration method that combines polarization imaging and homomorphic filtering. By exploiting the differences in polarization characteristics between suspended particles and target objects, polarization imaging is used to separate backscattered light from the target signal, enhancing the clarity of the cobalt crust images. Homomorphic filtering is then applied to improve the intensity distribution and contrast of the orthogonal polarization images. To optimize the parameters, a genetic algorithm is used with image quality evaluation indices as the fitness function. The proposed method was compared with traditional image processing techniques and classical polarization imaging methods. Experimental results demonstrate that the proposed approach more effectively suppresses backscattered light, enhancing the clarity of target object features. With significant improvements in image quality confirmed by several no-reference quality metrics, the method shows promise as a solution for high-quality underwater imaging in turbid environments, particularly for deep-sea mining of cobalt-rich crusts. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>Flowchart of the proposed method in this paper.</p>
Full article ">Figure 2
<p>Schematic Diagram of the Experimental Setup. (<b>a</b>) is taken from Figure 1a of the article A Volterra series-based method for extracting target echoes in the seafloor mining environment, published in the journal Ultrasonic by Elsevier, and used with permission. Copyright © [<a href="#B18-sensors-25-01088" class="html-bibr">18</a>].</p>
Full article ">Figure 3
<p>Underwater camera.</p>
Full article ">Figure 4
<p>Cobalt crust sample.</p>
Full article ">Figure 5
<p>Original orthogonal polarization images and filtered images.</p>
Full article ">Figure 5 Cont.
<p>Original orthogonal polarization images and filtered images.</p>
Full article ">Figure 6
<p>Restoration image comparison.</p>
Full article ">Figure 7
<p>Comparison of Different Methods for Restoring Images. (<b>a</b>) Original intensity image method. (<b>b</b>) Our method. (<b>c</b>) Schechner. (<b>d</b>) CLAHE. (<b>e</b>) Retinex. (<b>f</b>) HF.</p>
Full article ">Figure 8
<p>Comparison of restored images with different turbidity levels.</p>
Full article ">
19 pages, 32782 KiB  
Article
Artificial Fish Reef Site Evaluation Based on Multi-Source High-Resolution Acoustic Images
by Fangqi Wang, Yikai Feng, Senbo Liu, Yilan Chen and Jisheng Ding
J. Mar. Sci. Eng. 2025, 13(2), 309; https://doi.org/10.3390/jmse13020309 - 7 Feb 2025
Abstract
Marine geophysical and geological investigations are crucial for evaluating the construction suitability of artificial fish reefs (AFRs). Key factors such as seabed topography, geomorphology, sub-bottom structure, and sediment type significantly influence AFR design and site selection. Challenges such as material sinking, sediment instability, [...] Read more.
Marine geophysical and geological investigations are crucial for evaluating the construction suitability of artificial fish reefs (AFRs). Key factors such as seabed topography, geomorphology, sub-bottom structure, and sediment type significantly influence AFR design and site selection. Challenges such as material sinking, sediment instability, and scouring effects should be critically considered and addressed in the construction of AFR, particularly in areas with soft mud or dynamic environments. In this study, detailed investigations were conducted approximately seven months after the deployment of reef materials in the AFR experimental zones around Xiaoguan Island, located in the western South Yellow Sea, China. Based on morphological factors, using data from multibeam echosounders and side-scan sonar, the study area was divided into three geomorphic zones, namely, the tidal flat (TF), underwater erosion-accumulation slope (UEABS), and inclined erosion-accumulation shelf plain (IEASP) zones. The focus of this study was on the UEABS and IEASP experimental zones, where reef materials (concrete or stone blocks) were deployed seven months earlier. The comprehensive interpretation results of multi-source high-resolution acoustic images showed that the average settlement of individual reefs in the UEABS experimental zone was 0.49 m, and their surrounding seabed experienced little to no scouring. This suggested the formation of an effective range and height, making the zone suitable for AFR construction. However, in the IEASP experimental zone, the seabed sediment consisted of soft mud, causing the reef materials to sink into the seabed after deployment, preventing the formation of an effective range and height, and rendering the area unsuitable for AFR construction. These findings provided valuable scientific guidance for AFR construction in the study area and other similar coastal regions. Full article
(This article belongs to the Section Coastal Engineering)
Show Figures

Figure 1

Figure 1
<p>Geographical location of the study area. The red box indicates the scope of the study area.</p>
Full article ">Figure 2
<p>Research framework of this study.</p>
Full article ">Figure 3
<p>Overview of the multibeam bathymetric dataset for the study area, highlighting the locations of Figures 8a and 10a. The seabed topography in the study area was relatively simple, with water depths ranging from 0 to 15 m. The three types of submarine geomorphology in the surrounding waters of Xiaoguan Island were as follows: tidal flat (TF), underwater erosion-accumulation bank slope (UEABS), and inclined erosion-accumulation shelf plain (IEASP).</p>
Full article ">Figure 4
<p>Overview of the SSS mosaic for the study area with the highlighted locations of Figures 6a–f, 8b and 10b. The six microgeomorphology types were as follows: seabed disturbed zone, micro sand wave, tidal sand ridge, sand dune, natural reef, and AFR.</p>
Full article ">Figure 5
<p>Three-dimensional SBP display map highlighting the location of Figure 7a. The thickness of the unconsolidated sedimentary layer near the islands was controlled by the underlying bedrock. The Holocene sequence was stable, with a bottom boundary determined by a continuous unconformity interface.</p>
Full article ">Figure 6
<p>Typical SSS records of the six types of submarine microgeomorphology: (<b>a</b>) branch-like disturbed zone; (<b>b</b>) submarine sand wave; (<b>c</b>) tidal sand ridge; (<b>d</b>) sand dune; (<b>e</b>) natural reef; (<b>f</b>) AFR; and (<b>g</b>) photo of the AFR material.</p>
Full article ">Figure 7
<p>Typical (<b>a</b>) original image of the SBP (location shown in <a href="#jmse-13-00309-f005" class="html-fig">Figure 5</a>) and (<b>b</b>) its interpretation profile. TWT stands for two-way travel time of acoustic waves.</p>
Full article ">Figure 8
<p>(<b>a</b>) MBE image showing the outlines of artificial reef deposits, micro pits, and pockmarks on the seabed in the Xiaodao Bay experimental zone; (<b>b</b>) SSS mosaic highlighting the individual artificial reef structures as square-shaped areas with strong reflections and shaded regions; (<b>c</b>) photo of a sediment sample classified as mixed mud and sand (Ms) with visible gravel inclusions; (<b>d</b>) high-resolution SSS image of individual artificial reefs showing clear acoustic shadows, indicating minimal settlement and scouring effects.</p>
Full article ">Figure 9
<p>Geometric relationship between the SSS towfish, target, and acoustic shadow.</p>
Full article ">Figure 10
<p>(<b>a</b>) MBE image showing no obvious topographic anomalies on the seafloor, and not even the outline of AFR can be identified; (<b>b</b>) SSS image revealing a clear outline of AFR, but no obvious shadow area was observed; (<b>c</b>) photo of a sediment sample classified as soft mud (Si); (<b>d</b>,<b>e</b>) SBP images across the AFR showing acoustic shielding zones spanning 36 and 132 m due to the hard AFR materials in the surficial of seabed.</p>
Full article ">
21 pages, 3281 KiB  
Article
Multi-Space Feature Fusion and Entropy-Based Metrics for Underwater Image Quality Assessment
by Baozhen Du, Hongwei Ying, Jiahao Zhang and Qunxin Chen
Entropy 2025, 27(2), 173; https://doi.org/10.3390/e27020173 - 6 Feb 2025
Abstract
In marine remote sensing, underwater images play an indispensable role in ocean exploration, owing to their richness in information and intuitiveness. However, underwater images often encounter issues such as color shifts, loss of detail, and reduced clarity, leading to the decline of image [...] Read more.
In marine remote sensing, underwater images play an indispensable role in ocean exploration, owing to their richness in information and intuitiveness. However, underwater images often encounter issues such as color shifts, loss of detail, and reduced clarity, leading to the decline of image quality. Therefore, it is critical to study precise and efficient methods for assessing underwater image quality. A no-reference multi-space feature fusion and entropy-based metrics for underwater image quality assessment (MFEM-UIQA) are proposed in this paper. Considering the color shifts of underwater images, the chrominance difference map is created from the chrominance space and statistical features are extracted. Moreover, considering the information representation capability of entropy, entropy-based multi-channel mutual information features are extracted to further characterize chrominance features. For the luminance space features, contrast features from luminance images based on gamma correction and luminance uniformity features are extracted. In addition, logarithmic Gabor filtering is applied to the luminance space images for subband decomposition and entropy-based mutual information of subbands is captured. Furthermore, underwater image noise features, multi-channel dispersion information, and visibility features are extracted to jointly represent the perceptual features. The experiments demonstrate that the proposed MFEM-UIQA surpasses the state-of-the-art methods. Full article
(This article belongs to the Collection Entropy in Image Analysis)
Show Figures

Figure 1

Figure 1
<p>The Framework of MFEM-UIQA.</p>
Full article ">Figure 2
<p>Underwater images and corresponding UCD maps. (<b>a</b>) Underwater images of different quality levels; (<b>b</b>) corresponding UCD maps.</p>
Full article ">Figure 3
<p>Comparison of statistical distribution of MSCN coefficients for original underwater images and corresponding <span class="html-italic">Ψ<sub>D</sub></span>. (<b>a</b>) The statistical distribution of MSCN coefficients for the original underwater images, and (<b>b</b>) the statistical distribution of MSCN coefficients for <span class="html-italic">Ψ<sub>D</sub></span>.</p>
Full article ">Figure 4
<p>Underwater images of different quality levels and the corresponding fitting Rayleigh distribution shape parameter. (<b>a</b>) Underwater images of different quality levels; (<b>b</b>) fitting Rayleigh distribution shape parameters corresponding to three channel histograms of the OC space.</p>
Full article ">Figure 5
<p>Non-uniform brightness image and its block map. (<b>a</b>) Non-uniform brightness underwater image; (<b>b</b>) block map of (<b>a</b>).</p>
Full article ">Figure 6
<p>Underwater images with differing quality and corresponding K-L divergence distribution. (<b>a</b>) Underwater images of different quality levels; (<b>b</b>) the K-L divergence distribution of three channels in the OC space.</p>
Full article ">Figure 7
<p>Different quality underwater images and corresponding visibility values.</p>
Full article ">
14 pages, 5025 KiB  
Article
The Method for Storing a Seabed Photo Map of the During Surveys Conducted by an Autonomous Underwater Vehicle
by Chang Liu, Vladimir Filaretov, Eduard Mursalimov, Alexander Timoshenko and Alexander Zuev
Drones 2025, 9(2), 114; https://doi.org/10.3390/drones9020114 - 4 Feb 2025
Abstract
The paper introduces a novel method for creating a photographic map of the seabed using images captured by the on-board photo and video systems of autonomous underwater vehicles (AUVs) during various missions, while incorporating navigation parameters. Additionally, it presents a new approach for [...] Read more.
The paper introduces a novel method for creating a photographic map of the seabed using images captured by the on-board photo and video systems of autonomous underwater vehicles (AUVs) during various missions, while incorporating navigation parameters. Additionally, it presents a new approach for storing this photo map on the on-board device in a mosaic format (tiles), which significantly accelerates operational visual inspection by enabling the automatic search and recognition of underwater objects that may exceed the coverage area of a single photograph. This capability is achieved by organizing the photo map into layers with varying zoom levels. Semi-natural experiments were conducted with data from actual missions using the real underwater vehicle demonstrate the high efficiency of the proposed method and algorithm. Unlike existing methods that form photo maps after the underwater vehicle has taken pictures of the bottom using special high-performance computers, the developed method forms a photo map directly during the movement of the vehicle, using only the computing power of the on-board computer. In addition, in the event of accidents, when it is necessary to detect objects of interest on the seabed as quickly as possible, it is necessary to provide a quick visual inspection of the generated photo map. For this purpose, we have developed an algorithm for saving a photo map in the form of a mosaic, which is widely used in interactive geographic maps, such as Google Maps. This algorithm differs from existing methods in that it selectively saves data to the on-board storage device to reduce the number of read and write operations, thus ensuring the timely operation of the entire process of creating a photo map at a given frequency of photography. After the generated map has been stored as a mosaic and a high-speed connection with the vehicle has appeared, the operator can immediately view the entire generated map using a regular web browser. Full article
(This article belongs to the Special Issue Advances in Autonomous Underwater Drones)
Show Figures

Figure 1

Figure 1
<p>A general overview of the created photo map.</p>
Full article ">Figure 2
<p>The general structure of the developed algorithm.</p>
Full article ">Figure 3
<p>Photo image transformation to split into tiles.</p>
Full article ">Figure 4
<p>The principle of storing tiles on the on-board storage of AUV.</p>
Full article ">Figure 5
<p>Formation of tiles on the upper layers.</p>
Full article ">Figure 6
<p>The scheme of operation of the developed program.</p>
Full article ">Figure 7
<p>Program class diagram.</p>
Full article ">Figure 8
<p>The fragments of the generated photo map.</p>
Full article ">
20 pages, 4829 KiB  
Article
Study on Sound Field Properties of Parametric Array Under the Influence of Underwater Waveguide Interface Scattering Based on Non-Paraxial Model—Theory and Experiment
by Yuan Cao, Jie Shi, Jiangyi Zhang, Yuezhu Cheng and Haokang Shi
J. Mar. Sci. Eng. 2025, 13(2), 286; https://doi.org/10.3390/jmse13020286 - 4 Feb 2025
Abstract
This paper theoretically and experimentally studies the effect of underwater waveguide interface scattering on the nonlinear sound field characteristics of parametric array (PA) radiation. Based on the image source method, the components of the sound field in the waveguide are first analyzed. Then, [...] Read more.
This paper theoretically and experimentally studies the effect of underwater waveguide interface scattering on the nonlinear sound field characteristics of parametric array (PA) radiation. Based on the image source method, the components of the sound field in the waveguide are first analyzed. Then, a non-paraxial model is developed to account for the influence of interface scattering. This model enables accurate calculation of the wide-angle sound field. The impact of the sound source depth and the interface reflection coefficient on the distribution of the difference-frequency wave (DFW) sound field in the waveguide is studied. The interface alters the phase distribution of the DFW’s virtual source density function, thereby affecting the sound field accumulation process. Waveguide interfaces with different absorption coefficients influence the amplitude oscillation caused by interface reflection and change the sidelobe size of the DFW beam. The DFW sound field distribution is measured at three typical frequencies. Simulation and experimental results show that the attenuation of the DFW’s axial sound pressure level in the waveguide oscillates, and the DFW’s beamwidth gradually widens as the frequency decreases. The calculated results from the proposed model agree well with the measured data, with average errors along the sound axis and depth being less than 3 dB and 6 dB, respectively. This demonstrates the model’s superior applicability compared to the existing free-field model. Full article
(This article belongs to the Topic Advances in Underwater Acoustics and Aeroacoustics)
Show Figures

Figure 1

Figure 1
<p>Sketch of PA sound field in the waveguide.</p>
Full article ">Figure 2
<p>The SPL distribution on two-dimensional <math display="inline"><semantics> <mrow> <mi>y</mi> <mi>o</mi> <mi>z</mi> </mrow> </semantics></math> cross-section calculated by (<b>a</b>) RI model. (<b>b</b>) Conventional GBE model. (<b>c</b>) Non-paraxial model.</p>
Full article ">Figure 3
<p>Comparison of sound field and calculation error distribution. (<b>a</b>) Axial SPL. (<b>b</b>) Sound vertical directivity. (<b>c</b>) Error of conventional GBE model. (<b>d</b>) Error of Non-paraxial model.</p>
Full article ">Figure 4
<p>The linear sound field and DFW sound field results. (<b>a</b>) SPL distribution for linear sound field on two-dimensional <math display="inline"><semantics> <mrow> <mi>y</mi> <mi>o</mi> <mi>z</mi> </mrow> </semantics></math> cross-section. (<b>b</b>) Phase distribution for linear sound field on two-dimensional <math display="inline"><semantics> <mrow> <mi>y</mi> <mi>o</mi> <mi>z</mi> </mrow> </semantics></math> cross-section. (<b>c</b>) Vertical directivity for linear sound field. (<b>d</b>) SPL distribution for DFW sound field on two-dimensional <math display="inline"><semantics> <mrow> <mi>y</mi> <mi>o</mi> <mi>z</mi> </mrow> </semantics></math> cross-section. (<b>e</b>) Phase distribution for DFW sound field on two-dimensional <math display="inline"><semantics> <mrow> <mi>y</mi> <mi>o</mi> <mi>z</mi> </mrow> </semantics></math> cross-section. (<b>f</b>) Vertical directivity for DFW sound field.</p>
Full article ">Figure 5
<p>The DFW sound field for PA placed at different depths. (<b>a</b>) SPL distribution for PA placed at 10 m depth. (<b>b</b>) Phase distribution for PA placed at 10 m depth. (<b>c</b>) Vertical directivity for PA placed at 10 m depth. (<b>d</b>) SPL distribution for PA placed at 90 m depth. (<b>e</b>) Phase distribution for PA placed at 90 m depth. (<b>f</b>) Vertical directivity for PA placed at 90 m depth.</p>
Full article ">Figure 6
<p>The SPL and phase distribution of the integrand function for PA placed at different depths. (<b>a</b>) SPL distribution of the integrand function for PA placed at 10 m depth in the free field. (<b>b</b>) Phase distribution of the integrand function for PA placed at 10 m depth in the free field. (<b>c</b>) SPL distribution of the integrand function for PA placed at 10 m depth in the waveguide. (<b>d</b>) Phase distribution of the integrand function for PA placed at 10 m depth in the waveguide. (<b>e</b>) SPL distribution of the integrand function for PA placed at 90 m depth in the free field. (<b>f</b>) Phase distribution of the integrand function for PA placed at 90 m depth in the free field. (<b>g</b>) SPL distribution of the integrand function for PA placed at 90 m depth in the waveguide. (<b>h</b>) Phase distribution of the integrand function for PA placed at 90 m depth in the waveguide.</p>
Full article ">Figure 7
<p>The DFW sound field distribution for different waveguide interfaces. (<b>a</b>) Axial SPL. (<b>b</b>) Vertical directivity.</p>
Full article ">Figure 8
<p>Experimental environment and location distribution of measuring points (“A” represents an experimental platform).</p>
Full article ">Figure 9
<p>Diagram of the experimental equipment connection.</p>
Full article ">Figure 10
<p>The data recorded by the logger.</p>
Full article ">Figure 11
<p>Comparison of simulation and experimental results and error shadow distribution for axial SPL. (<b>a</b>) Axial SPL of DFW at a frequency 7.98 kHz. (<b>b</b>) Axial SPL of DFW at a frequency 3.98 kHz. (<b>c</b>) Axial SPL of DFW at a frequency 1.98 kHz. (<b>d</b>) The error shadow distribution for axial SPL at a frequency 7.98 kHz. (<b>e</b>) The error shadow distribution for axial SPL at a frequency 3.98 kHz. (<b>f</b>) The error shadow distribution for axial SPL at a frequency 1.98 kHz.</p>
Full article ">Figure 12
<p>Comparison between simulation and experimental distribution results for DFW sound field in the depth direction at the measuring point 5 m. (<b>a</b>) DFW sound field in the depth direction at a frequency 7.98 kHz. (<b>b</b>) DFW sound field in the depth direction at a frequency 3.98 kHz. (<b>c</b>) DFW sound field in the depth direction at a frequency 1.98 kHz.</p>
Full article ">
18 pages, 19258 KiB  
Article
Feature Fusion Network with Local Information Exchange for Underwater Object Detection
by Xiaopeng Liu, Pengwei Ma and Long Chen
Electronics 2025, 14(3), 587; https://doi.org/10.3390/electronics14030587 - 1 Feb 2025
Abstract
When using enhanced images for underwater object detection, issues such as detail loss and increased noise often arise, leading to decreased detection efficiency. To address this issue, we propose the Feature Fusion Network with Local Information Exchange (FFNLIE) for underwater object detection. We [...] Read more.
When using enhanced images for underwater object detection, issues such as detail loss and increased noise often arise, leading to decreased detection efficiency. To address this issue, we propose the Feature Fusion Network with Local Information Exchange (FFNLIE) for underwater object detection. We input raw and enhanced images into the Swin Transformer in parallel for feature extraction. Then, we propose a local information exchange module to enhance the feature extraction capability of the Swin Transformer. In order to fully utilize the complementary information of the two images, our feature fusion module consists of two core components: the Discrepancy Information Addition Block (DIAB) and the Common Information Addition Block (CIAB). The DIAB and CIAB are designed by utilizing and modifying cross-attention mechanisms, which can easily extract image discrepancy information and common information. Finally, the fused features are fed into the object detector to perform object detection tasks. The experimental findings demonstrate that the FFNLIE exhibited exceptional performance across four underwater datasets. Full article
Show Figures

Figure 1

Figure 1
<p>The influence of various preprocessing methods on object detection results for the UODD dataset.</p>
Full article ">Figure 2
<p>The overall framework of the FFNLIE.</p>
Full article ">Figure 3
<p>The network of the Swin Transformer block.</p>
Full article ">Figure 4
<p>The workflow of the window partition and shifted window partition.</p>
Full article ">Figure 5
<p>Visual comparison of feature maps using Grad-CAM with and without (second row) the LIE module. The darker the color in the image, the stronger the model’s attention to the current area.</p>
Full article ">Figure 6
<p>Qualitative comparison results on the UODD dataset.</p>
Full article ">Figure 7
<p>Qualitative comparison results on the UTDAC2020 dataset.</p>
Full article ">Figure 8
<p>Qualitative comparison results based on the RUOD dataset.</p>
Full article ">Figure 9
<p>Error analysis plots comparing the base method and the proposed FFNLIE on the UODD dataset. The first figure presents the plot for the basic method (not using feature fusion module and local information exchange module) and the second figure presents that for our method.</p>
Full article ">Figure 10
<p>Error analysis plots comparing the base method and the proposed FFNLIE on the RUOD dataset. The first figure presents the plot for the basic method (not using feature fusion module and local information exchange module) and the second figure presents that for our method.</p>
Full article ">
24 pages, 14510 KiB  
Article
PIC-GAN: Symmetry-Driven Underwater Image Enhancement with Partial Instance Normalisation and Colour Detail Modulation
by Kui You, Xinghua Li, Pengfei Yi, Yihe Zhang, Jingkang Xu, Jiajun Ren, Heyang Bai and Caihua Ma
Symmetry 2025, 17(2), 201; https://doi.org/10.3390/sym17020201 - 27 Jan 2025
Abstract
This model solves the problems of insufficient global feature attention, colour distortion, low contrast, and blurred details in previous methods by using a symmetric U-Net architecture and two new modules: the partial instance normalisation (PIN) module and the colour detail modulation (CDM) module. [...] Read more.
This model solves the problems of insufficient global feature attention, colour distortion, low contrast, and blurred details in previous methods by using a symmetric U-Net architecture and two new modules: the partial instance normalisation (PIN) module and the colour detail modulation (CDM) module. PIC-GAN effectively balances texture enhancement and feature retention while restoring colour details, ensuring excellent image quality. Experimental results on several publicly available underwater image datasets demonstrate the effectiveness of PIC-GAN, which improves PSNR by 3.23 dB and SSIM by 0.06 compared to WaterGAN, highlighting the potential of PIC-GAN as a powerful solution for underwater image enhancement tasks. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

Figure 1
<p>The simplified underwater optical imaging model.</p>
Full article ">Figure 2
<p>PINet structure.</p>
Full article ">Figure 3
<p>(<b>a</b>) shows the residual structure of the decoder section, and (<b>b</b>) shows the Discriminator network structure.</p>
Full article ">Figure 4
<p>Visualisation of process characteristics.</p>
Full article ">Figure 5
<p>The network structure of the PIN module.</p>
Full article ">Figure 6
<p>Network structure diagram with four division ratios.</p>
Full article ">Figure 7
<p>CDM module network structure.</p>
Full article ">Figure 8
<p>Visualisation results of different comparison algorithms.</p>
Full article ">Figure 9
<p>Visualisation results of the input image and the image before and after PIC-GAN augmentation.</p>
Full article ">Figure 10
<p>Visualisation of different algorithms in low-visibility environments.</p>
Full article ">Figure 11
<p>Visualisation results of different ablation experiments.</p>
Full article ">Figure 12
<p>Visualisation results for different instantiation ratios.</p>
Full article ">
16 pages, 3921 KiB  
Article
Effect of Seabed Type on Image Segmentation of an Underwater Object Obtained from a Side Scan Sonar Using a Deep Learning Approach
by Jungyong Park and Ho Seuk Bae
J. Mar. Sci. Eng. 2025, 13(2), 242; https://doi.org/10.3390/jmse13020242 - 26 Jan 2025
Abstract
This study examines the impact of seabed conditions on image segmentation for seabed target images acquired via side-scan sonar during sea experiments. The dataset comprised cylindrical target images overlying on two seabed types, mud and sand, categorized accordingly. The deep learning algorithm (U-NET) [...] Read more.
This study examines the impact of seabed conditions on image segmentation for seabed target images acquired via side-scan sonar during sea experiments. The dataset comprised cylindrical target images overlying on two seabed types, mud and sand, categorized accordingly. The deep learning algorithm (U-NET) was utilized for image segmentation. The analysis focused on two key factors influencing segmentation performance: the weighting method of the cross-entropy loss function and the combination of datasets categorized by seabed type for training, validation, and testing. The results revealed three key findings. First, applying equal weights to the loss function yielded better segmentation performance compared to pixel-frequency-based weighting. This improvement is indicated by Intersection over Union (IoU) for the highlight class in dataset 2 (0.41 compared to 0.37). Second, images from the mud area were easier to segment than those from the sand area. This was due to the clearer intensity contrast between the target highlight and background. This difference is indicated by the IoU for the highlight class (0.63 compared to 0.41). Finally, a network trained on a combined dataset from both seabed types improved segmentation performance. This improvement was observed in challenging conditions, such as sand areas. In comparison, a network trained on a single-seabed dataset showed lower performance. The IoU values for the highlight class in sand area images are as follows: 0.34 for training on mud, 0.41 for training on sand, and 0.45 for training on both. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) tow fish side scan sonar (SeaView400S), and (<b>b</b>) mock-up target (closed-end cylinder) from Kim et al. [<a href="#B15-jmse-13-00242" class="html-bibr">15</a>].</p>
Full article ">Figure 2
<p>Schematic diagram of survey line (redrawn from Kim et al. [<a href="#B15-jmse-13-00242" class="html-bibr">15</a>]).</p>
Full article ">Figure 3
<p>The examples of target images are from (<b>a</b>) Experiment 1, (<b>b</b>) Experiment 2, (<b>c</b>) Experiment 3, and (<b>d</b>) Experiment 4.</p>
Full article ">Figure 4
<p>Examples of U-NET structure.</p>
Full article ">Figure 5
<p>The examples of image segmentation results. The left figure in the panel shows the original image. The middle figure displays the overlap between the true highlight (HL) label region (white) and the predicted label region (red). The right figure illustrates the overlap between the true shadow label region (white) and the predicted label region (red). These comparisons include (<b>a</b>) test data from dataset 1 (mud) and (<b>b</b>) test data from dataset 2 (sand).</p>
Full article ">Figure 6
<p>Conceptual diagram for the performance metrics.</p>
Full article ">Figure 7
<p>Average of k-fold metrics (Precision, Recall and IoU) for the target highlight (HL) class and shadow class when the datasets for training, validation and test are the same. The color of the bar graph represents the dataset type and weight type for the loss function.</p>
Full article ">Figure 8
<p>Average of k-fold metrics (Precision, Recall and IoU) for the target highlight (HL) class. The color of the bar graph represents the different types of datasets for training, validation and testing.</p>
Full article ">Figure 9
<p>Average of k-fold metrics (Precision, Recall and IoU) for the target shadow class. The color of the bar graph represents the different types of datasets for training, validation and testing.</p>
Full article ">
17 pages, 4965 KiB  
Article
Neural Network for Underwater Fish Image Segmentation Using an Enhanced Feature Pyramid Convolutional Architecture
by Guang Yang, Junyi Yang, Wenyao Fan and Donghe Yang
J. Mar. Sci. Eng. 2025, 13(2), 238; https://doi.org/10.3390/jmse13020238 - 26 Jan 2025
Abstract
Underwater fish image segmentation is a crucial technique in marine fish monitoring. However, typical underwater fish images often suffer from issues such as color distortion, low contrast, and blurriness, primarily due to the complex and dynamic nature of the marine environment. To enhance [...] Read more.
Underwater fish image segmentation is a crucial technique in marine fish monitoring. However, typical underwater fish images often suffer from issues such as color distortion, low contrast, and blurriness, primarily due to the complex and dynamic nature of the marine environment. To enhance the accuracy of underwater fish image segmentation, this paper introduces an innovative neural network model that combines the attention mechanism with a feature pyramid module. After the backbone network processes the input image through convolution, the data pass through the enhanced feature pyramid module, where it is iteratively processed by multiple weighted branches. Unlike conventional methods, the multi-scale feature extraction module that we designed not only improves the extraction of high-level semantic features but also optimizes the distribution of low-level shape feature weights through the synergistic interactions of the branches, all while preserving the inherent properties of the image. This novel architecture significantly boosts segmentation accuracy, offering a new solution for fish image segmentation tasks. To further enhance the model’s robustness, the Mix-up and CutMix data augmentation techniques were employed. The model was validated using the Fish4Knowledge dataset, and the experimental results demonstrate that the model achieves a Mean Intersection over Union (MIoU) of 95.1%, with improvements of 1.3%, 1.5%, and 1.7% in the MIoU, Mean Pixel Accuracy (PA), and F1 score, respectively, compared to traditional segmentation methods. Additionally, a real fish image dataset captured in deep-sea environments was constructed to verify the practical applicability of the proposed algorithm. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

Figure 1
<p>Network Architecture.</p>
Full article ">Figure 2
<p>PAFE.</p>
Full article ">Figure 3
<p>Fish dataset after data augmentation.</p>
Full article ">Figure 4
<p>Data Comparison Chart.</p>
Full article ">Figure 5
<p>Results Display.</p>
Full article ">Figure 6
<p>Black Sea Spart.</p>
Full article ">Figure 7
<p>Ablation Experiment III.</p>
Full article ">Figure 8
<p>Presentation of Experimental Results for Deep-Sea Fish Images.</p>
Full article ">Figure 8 Cont.
<p>Presentation of Experimental Results for Deep-Sea Fish Images.</p>
Full article ">Figure 9
<p>Image enhancement effect.</p>
Full article ">
17 pages, 13467 KiB  
Article
An Improved YOLOv9s Algorithm for Underwater Object Detection
by Shize Zhou, Long Wang, Zhuoqun Chen, Hao Zheng, Zhihui Lin and Li He
J. Mar. Sci. Eng. 2025, 13(2), 230; https://doi.org/10.3390/jmse13020230 - 25 Jan 2025
Viewed by 162
Abstract
Monitoring marine life through underwater object detection technology serves as a primary means of understanding biodiversity and ecosystem health. However, the complex marine environment, poor resolution, color distortion in underwater optical imaging, and limited computational resources all affect the accuracy and efficiency of [...] Read more.
Monitoring marine life through underwater object detection technology serves as a primary means of understanding biodiversity and ecosystem health. However, the complex marine environment, poor resolution, color distortion in underwater optical imaging, and limited computational resources all affect the accuracy and efficiency of underwater object detection. To solve these problems, the YOLOv9s-SD underwater target detection algorithm is proposed to improve the detection performance in underwater environments. We combine the inverted residual structure of MobileNetV2 with Simple Attention Module (SimAM) and Squeeze-and-Excitation Attention (SE) to form the Simple Enhancement attention Module (SME) and optimize AConv, improving the sensitivity of the model to object details. Furthermore, we introduce the lightweight DySample operator to optimize feature recovery, enabling better adaptation to the complex characteristics of underwater targets. Finally, we employ Wise-IoU version 3 (WIoU v3) as the loss function to balance the loss weights for targets of different sizes. In comparison with the YOLOv9s model, according to the experiments conducted on the UPRC and Brackish underwater datasets, YOLOv9s-SD achieves an improvement of 1.3% and 1.2% in the mean Average Precision (mAP), reaching 83.0% and 94.3% on the respective datasets and demonstrating better adaptability to intricate underwater environments. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

Figure 1
<p>Theimproved YOLOv9s-SD network structure.</p>
Full article ">Figure 2
<p>The structural framework of each module: (<b>a</b>) structure of the SME attention module; (<b>b</b>) AConv; (<b>c</b>) SMEAConv.</p>
Full article ">Figure 3
<p>DySample module flowchart.</p>
Full article ">Figure 4
<p>Dynamic point sampling set generation process.</p>
Full article ">Figure 5
<p>The statistical data of the quantities of four types of organisms.</p>
Full article ">Figure 6
<p>Experimental detection results before and after YOLOv9s improvement; (<b>a</b>) real label; (<b>b</b>) YOLOv9s detection results; (<b>c</b>) YOLOv9s-SD detection results.</p>
Full article ">Figure 7
<p>Displaying heatmap visualization results of different attention mechanisms using LayerCAM algorithm: (<b>a</b>) raw image; (<b>b</b>) YOLOv9s; (<b>c</b>) YOLOv9s+SE; (<b>d</b>) YOLOv9s+SimAM; (<b>e</b>) YOLOv9s+SME.</p>
Full article ">Figure 8
<p>The mAP50% curve for each model.</p>
Full article ">Figure 9
<p>The statistical data of the quantities of six types of organisms.</p>
Full article ">Figure 10
<p>Experimental detection results before and after YOLOv9s improvement; (<b>a</b>) real label; (<b>b</b>) YOLOv9s detection results; (<b>c</b>) YOLOv9s-SD detection results.</p>
Full article ">
13 pages, 3352 KiB  
Article
Dual-CycleGANs with Dynamic Guidance for Robust Underwater Image Restoration
by Yu-Yang Lin, Wan-Jen Huang and Chia-Hung Yeh
J. Mar. Sci. Eng. 2025, 13(2), 231; https://doi.org/10.3390/jmse13020231 - 25 Jan 2025
Viewed by 225
Abstract
The field of underwater image processing has gained significant attention recently, offering great potential for enhanced exploration of underwater environments, including applications such as underwater terrain scanning and autonomous underwater vehicles. However, underwater images frequently face challenges such as light attenuation, color distortion, [...] Read more.
The field of underwater image processing has gained significant attention recently, offering great potential for enhanced exploration of underwater environments, including applications such as underwater terrain scanning and autonomous underwater vehicles. However, underwater images frequently face challenges such as light attenuation, color distortion, and noise introduced by artificial light sources. These degradations not only affect image quality but also hinder the effectiveness of related application tasks. To address these issues, this paper presents a novel deep network model for single under-water image restoration. Our model does not rely on paired training images and incorporates two cycle-consistent generative adversarial network (CycleGAN) structures, forming a dual-CycleGAN architecture. This enables the simultaneous conversion of an underwater image to its in-air (atmospheric) counterpart while learning a light field image to guide the underwater image towards its in-air version. Experimental results indicate that the proposed method provides superior (or at least comparable) image restoration performance, both in terms of quantitative measures and visual quality, when compared to existing state-of-the-art techniques. Our model significantly reduces computational complexity, resulting in a more efficient approach that maintains superior restoration capabilities, ensuring faster processing times and lower memory usage, making it highly suitable for real-world applications. Full article
(This article belongs to the Special Issue Application of Deep Learning in Underwater Image Processing)
Show Figures

Figure 1

Figure 1
<p>Illustration of the output from the light field module.</p>
Full article ">Figure 2
<p>The proposed unsupervised adversarial learning framework consisting of dual-CycleGAN with unpaired training images.</p>
Full article ">Figure 3
<p>The architecture of the generator in the proposed deep single underwater image restoration network.</p>
Full article ">Figure 4
<p>Qualitative evaluation results on the UFO-120 dataset.</p>
Full article ">
15 pages, 19344 KiB  
Article
An Underwater Polarization Imaging Technique Based on the Construction and Decomposition of the Low-Rank and Sparse Matrix in Stokes Space for Polarization State Imaging
by Pengfeng Liu, Yuxiang Zhai, Hongjin Zhu, Zijian Ye, Qinyu He, Zhilie Tang and Peijun Tang
Sensors 2025, 25(3), 704; https://doi.org/10.3390/s25030704 - 24 Jan 2025
Viewed by 234
Abstract
Traditional underwater polarization imaging methods can only provide clear degree of polarization (DOP) and intensity images of the object but cannot provide images of the polarization state of the object. This paper proposes a method to extract clear object information from turbid water [...] Read more.
Traditional underwater polarization imaging methods can only provide clear degree of polarization (DOP) and intensity images of the object but cannot provide images of the polarization state of the object. This paper proposes a method to extract clear object information from turbid water in all four Stokes parameter (I, Q, U, and V) channels by using the full Stokes camera, enabling clear polarization state image reconstruction. The method utilizes multiple images from different angles to construct a low-rank and sparse matrix. Then, by decomposing this matrix into sparse and low-rank components, clear Q, U, and V images (i.e., the full polarization state) can be obtained. Unlike traditional methods that assume the circularly polarized component (V component) to be zero, this method retains V channel information, allowing for circular polarization component measurement. The study successfully reconstructed clear underwater images of samples with inhomogeneous DOP distribution and obtained the clear polarization states of polarizers and fish in the turbid water. The results show that the proposed method can visualize and analyze the object’s polarization state quantitatively with high accuracy in turbid water for the first time, potentially extending the applicability of polarization underwater imaging in ocean exploration. Full article
(This article belongs to the Special Issue Underwater Vision Sensing System)
Show Figures

Figure 1

Figure 1
<p>A flowchart of the underwater polarization state imaging algorithm.</p>
Full article ">Figure 2
<p>(<b>a</b>) Equipment diagram. (<b>b</b>) Physical diagram of the experimental device.</p>
Full article ">Figure 3
<p>(<b>a</b>) A group of intensity images (<span class="html-italic">I</span>) of the ambient light. (<b>b</b>) Separated low-rank components obtained by the proposed method. (<b>c</b>) Separated sparse components obtained by the proposed method.</p>
Full article ">Figure 4
<p>Polarization underwater imaging of the plastic doll (low DOP) in different concentrations of the milk solution. (<b>a</b>–<b>c</b>) are the original intensity images with the milk concentrations of 6, 8, and 10 mL/L, respectively. (<b>d</b>–<b>f</b>) are recovered intensity images of the plastic doll with the milk concentrations of 6, 8, and 10 mL/L, respectively, using the proposed LRSD-based method.</p>
Full article ">Figure 5
<p>Polarization underwater imaging of the iron ruler (high DOP) and tape (low DOP) in a milk solution with the concentration of 6 mL/L. (<b>a</b>) Original underwater intensity image. (<b>b</b>) Image reconstructed by Treibitz’s method by setting the DOP of the sample as 0 (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>S</mi> </msub> </semantics></math> = 0). (<b>c</b>) Image reconstructed by Treibitz’s method by setting the DOP of the sample as 1 (<math display="inline"><semantics> <msub> <mi>P</mi> <mi>S</mi> </msub> </semantics></math> = 1). (<b>d</b>) Image reconstructed by CLAHE’s method. (<b>e</b>) Image reconstructed by ICA’s method. (<b>f</b>) Intensity image reconstructed by the proposed method. (<b>g</b>) Color-encoded polarization state image reconstructed by the proposed method.</p>
Full article ">Figure 6
<p>Curves of the grayscale values. (<b>a</b>) Gray value of the dotted line in <a href="#sensors-25-00704-f005" class="html-fig">Figure 5</a>. (<b>b</b>) Gray value of the solid line in <a href="#sensors-25-00704-f005" class="html-fig">Figure 5</a>.</p>
Full article ">Figure 7
<p>Polarization underwater imaging of three polarizers with different orientations. (<b>a</b>–<b>d</b>) are original underwater Stokes parameter images <math display="inline"><semantics> <mrow> <mo>[</mo> <msub> <mi>I</mi> <mrow> <mi>t</mi> <mi>o</mi> <mi>t</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>Q</mi> <mrow> <mi>t</mi> <mi>o</mi> <mi>t</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>U</mi> <mrow> <mi>t</mi> <mi>o</mi> <mi>t</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>V</mi> <mrow> <mi>t</mi> <mi>o</mi> <mi>t</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mo>]</mo> </mrow> </semantics></math>. (<b>e</b>–<b>h</b>) are Stokes parameter images reconstructed by the proposed method <math display="inline"><semantics> <mrow> <mo>[</mo> <msub> <mi>I</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>r</mi> <mi>g</mi> <mi>e</mi> <mi>t</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>Q</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>r</mi> <mi>g</mi> <mi>e</mi> <mi>t</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>U</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>r</mi> <mi>g</mi> <mi>e</mi> <mi>t</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>V</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>r</mi> <mi>g</mi> <mi>e</mi> <mi>t</mi> </mrow> </msub> <mo>]</mo> </mrow> </semantics></math>.</p>
Full article ">Figure 8
<p>Color-encoded polarization state image of the 3 polarizers with different orientations. (<b>a</b>) Original polarization state of the sample in the clean water. (<b>b</b>) Original underwater polarization state image of the sample. (<b>c</b>) Clear color-encoded polarization state image of the sample reconstructed by the proposed method. (<b>d</b>) Colormap for the linearly polarized light.</p>
Full article ">Figure 9
<p>Underwater images of the fish in the turbid water with milk concentration of 8 mL captured by the full Stokes camera. (<b>a</b>–<b>d</b>) are the raw Stokes images of the fish <math display="inline"><semantics> <mrow> <mo>[</mo> <msub> <mi>I</mi> <mrow> <mi>t</mi> <mi>o</mi> <mi>t</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>Q</mi> <mrow> <mi>t</mi> <mi>o</mi> <mi>t</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>U</mi> <mrow> <mi>t</mi> <mi>o</mi> <mi>t</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>V</mi> <mrow> <mi>t</mi> <mi>o</mi> <mi>t</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mo>]</mo> </mrow> </semantics></math>. (<b>e</b>–<b>h</b>) are the recovered clear images of the fish reconstructed by the proposed method <math display="inline"><semantics> <mrow> <mo>[</mo> <msub> <mi>I</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>r</mi> <mi>g</mi> <mi>e</mi> <mi>t</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>Q</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>r</mi> <mi>g</mi> <mi>e</mi> <mi>t</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>U</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>r</mi> <mi>g</mi> <mi>e</mi> <mi>t</mi> </mrow> </msub> <mspace width="4pt"/> <msub> <mi>V</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>r</mi> <mi>g</mi> <mi>e</mi> <mi>t</mi> </mrow> </msub> <mo>]</mo> </mrow> </semantics></math>.</p>
Full article ">Figure 10
<p>Color-encoded polarization state image of the fish. (<b>a</b>) Original polarization state of the sample in the clean water. (<b>b</b>) Original underwater polarization state image of the sample. (<b>c</b>) Clear color-encoded polarization state image of the sample reconstructed by the proposed method.</p>
Full article ">
23 pages, 11152 KiB  
Article
Self-Training Can Reduce Detection False Alarm Rate of High-Resolution Imaging Sonar
by Jingqi Han, Yue Fan, Zheng He, Zhenhang You, Peng Zhang and Zhengliang Hu
Appl. Sci. 2025, 15(3), 1189; https://doi.org/10.3390/app15031189 - 24 Jan 2025
Viewed by 298
Abstract
Imaging sonar is a primary means of underwater detection, but it faces challenges of high false alarm rates in sonar image target detection due to factors such as reverberation, noise, and resolution. This paper proposes a method to improve the false alarm rate [...] Read more.
Imaging sonar is a primary means of underwater detection, but it faces challenges of high false alarm rates in sonar image target detection due to factors such as reverberation, noise, and resolution. This paper proposes a method to improve the false alarm rate by self-training a deep learning detector on sonar images. Self-training automatically generates proxy classification tasks based on the sonar image target detection dataset, and pre-trains the deep learning detector through these proxy classification tasks to enhance its learning effectiveness of target and background features. This, in turn, improves the detector’s ability to distinguish between targets and backgrounds, thereby reducing the false alarm rate. For the first time, this paper conducts target detection experiments based on deep learning using high-resolution synthetic aperture sonar images at two frequencies. The results show that, under the conditions of equal or higher recall rates, this method can reduce the false alarm rate by 3.91% and 18.50% on 240 kHz and 450 kHz sonar images, respectively, compared to traditional transfer learning methods. Full article
Show Figures

Figure 1

Figure 1
<p>Pre-trained detection paradigm for sonar image target detection based on large-scale optical image datasets.</p>
Full article ">Figure 2
<p>Comparison diagram of this paper’s training framework and traditional training frameworks.</p>
Full article ">Figure 3
<p>The training strategy of Detectors based on proxy classification tasks.</p>
Full article ">Figure 4
<p>The creation of derived classification datasets.</p>
Full article ">Figure 5
<p>Targets and background in datasets SAS240.</p>
Full article ">Figure 6
<p>Targets and background in datasets SAS450.</p>
Full article ">Figure 7
<p>Distribution of target category quantities in the dataset. (<b>a</b>) Result of the SAS240 training set; (<b>b</b>) Result of the SAS240 validation set; (<b>c</b>) Result of the SAS450 training set; (<b>d</b>) Result of the SAS450 validation set.</p>
Full article ">Figure 8
<p>The relationship among different datasets used in the experiments.</p>
Full article ">Figure 9
<p>The detection results of Detectors A, B, and C on the SAS240 test set. (The yellow bounding box indicates a false alarm, and the red bounding box indicates a correct detection).</p>
Full article ">Figure 10
<p>The detection results of Detectors D, E, and F on the SAS450 test set. (The yellow bounding box indicates a false alarm, and the red bounding box indicates a correct detection).</p>
Full article ">Figure 11
<p>The performance of Detector A, Detector B and Detector C on the SAS240 test set: (<b>a</b>) The comparison of false alarm rates; (<b>b</b>) The comparison of miss rates.</p>
Full article ">Figure 12
<p>Missing alarm rate–false alarm rate curve among Detectors A, B and C.</p>
Full article ">Figure 13
<p>The training results of the classification task: (<b>a</b>) results of SAS240; (<b>b</b>) results of SAS450.</p>
Full article ">Figure 14
<p>The detection results of Detectors F and G on the SAS450 test set. (The yellow bounding box indicates a false alarm, and the red bounding box indicates a correct detection).</p>
Full article ">Figure 15
<p>The performance of Detector D, Detector F and Detector G on the SAS450 test set: (<b>a</b>) The comparison of false alarm rates; (<b>b</b>) The comparison of miss rates.</p>
Full article ">Figure 16
<p>Missing alarm rate–false alarm rate curve among Detectors D, F and G.</p>
Full article ">Figure 17
<p>The detection results of Detectors G and H on the SAS450 test set. (The yellow bounding box indicates a false alarm, and the red bounding box indicates a correct detection).</p>
Full article ">
Back to TopTop