Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (17)

Search Parameters:
Keywords = vision-based underwater measurements

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
16 pages, 4851 KiB  
Article
Underwater Refractive Stereo Vision Measurement and Simulation Imaging Model Based on Optical Path
by Guanqing Li, Shengxiang Huang, Zhi Yin, Jun Li and Kefei Zhang
J. Mar. Sci. Eng. 2024, 12(11), 1955; https://doi.org/10.3390/jmse12111955 - 1 Nov 2024
Viewed by 539
Abstract
When light passes through air–glass and glass–water interfaces, refraction occurs, which affects the accuracy of stereo vision three-dimensional measurements of underwater targets. To eliminate the impact of refraction, we developed a refractive stereo vision measurement model based on light propagation paths, utilizing the [...] Read more.
When light passes through air–glass and glass–water interfaces, refraction occurs, which affects the accuracy of stereo vision three-dimensional measurements of underwater targets. To eliminate the impact of refraction, we developed a refractive stereo vision measurement model based on light propagation paths, utilizing the normalized coordinate of the underwater target. This model is rigorous in theory, and easy to understand and apply. Additionally, we established an underwater simulation imaging model based on the principle that light travels the shortest time between two points. Simulation experiments conducted using this imaging model verified the performance of the underwater stereo vision measurement model. The results demonstrate that the accuracy achieved by the new measurement model is comparable to that of the stereo vision measurement model in the air and significantly higher than that of the existing refractive measurement model. This is because the light rays from the camera’s optical center to the refraction point at the air–glass interface do not always intersect. The experiments also indicate that the deviation in the refractive index of water lead to corresponding systematic errors in the measurement results. Therefore, in real underwater measurements, it is crucial to carefully calibrate the refractive index of water and maintain the validity of the calibration results. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

Figure 1
<p>Refracted light rays with two flat interfaces under air–glass–water condition.</p>
Full article ">Figure 2
<p>Propagation path of the light from the target to the camera.</p>
Full article ">Figure 3
<p>Simulation experimental design. (<b>a</b>) Experimental scenario; (<b>b</b>) distribution of 496 points in the target plane.</p>
Full article ">Figure 4
<p>Contour of the shortest distance between the two refracted rays from cameras.</p>
Full article ">Figure 5
<p>Discrepancy between the measured and true coordinates obtained by Su’s method under ideal conditions. (<b>a</b>) <math display="inline"><semantics> <mrow> <mi>X</mi> </mrow> </semantics></math> coordinate discrepancy. (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>Y</mi> </mrow> </semantics></math> coordinate discrepancy. (<b>c</b>) <math display="inline"><semantics> <mrow> <mi>Z</mi> </mrow> </semantics></math> coordinate discrepancy. (<b>d</b>) Total coordinate discrepancy. <math display="inline"><semantics> <mrow> <mi>Z</mi> <mo>=</mo> <mn>1</mn> <mo> </mo> <mi mathvariant="normal">m</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 6
<p>Discrepancy between the measured and true coordinates obtained by Su’s method under ideal conditions. (<b>a</b>) <math display="inline"><semantics> <mrow> <mi>X</mi> </mrow> </semantics></math> coordinate discrepancy. (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>Y</mi> </mrow> </semantics></math> coordinate discrepancy. (<b>c</b>) <math display="inline"><semantics> <mrow> <mi>Z</mi> </mrow> </semantics></math> coordinate discrepancy. (<b>d</b>) Total coordinate discrepancy. <math display="inline"><semantics> <mrow> <mi>Z</mi> <mo>=</mo> <mn>2</mn> <mo> </mo> <mi mathvariant="normal">m</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 7
<p>Discrepancy between the measured and true coordinates obtained under the condition pixel coordinates contain errors. (<b>a</b>–<b>d</b>) Our method, (<b>e</b>–<b>h</b>) Su’s method, (<b>i</b>–<b>l</b>) the regular method. <math display="inline"><semantics> <mrow> <mi>Z</mi> <mo>=</mo> <mn>1</mn> <mo> </mo> <mi mathvariant="normal">m</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 8
<p>Discrepancy normalized histogram and fitted Gaussian curve obtained under the condition pixel coordinates contain errors. (<b>a</b>–<b>c</b>) Our method, (<b>d</b>–<b>f</b>) Su’s method. <math display="inline"><semantics> <mrow> <mi>Z</mi> <mo>=</mo> <mn>1</mn> <mo> </mo> <mi mathvariant="normal">m</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 9
<p>Discrepancy between the measured and true coordinates obtained under the condition pixel coordinates contain errors and water refractive index contains deviation. (<b>a</b>–<b>d</b>) Our method, (<b>e</b>–<b>h</b>) Su’s method. <math display="inline"><semantics> <mrow> <mi>Z</mi> <mo>=</mo> <mn>1</mn> <mo> </mo> <mi mathvariant="normal">m</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 10
<p>Discrepancy normalized histogram and fitted Gaussian curve obtained under the condition pixel coordinates contain errors and water refractive index contains deviation. (<b>a</b>–<b>c</b>) Our method, (<b>d</b>–<b>f</b>) Su’s method. <math display="inline"><semantics> <mrow> <mi>Z</mi> <mo>=</mo> <mn>1</mn> <mo> </mo> <mi mathvariant="normal">m</mi> </mrow> </semantics></math>.</p>
Full article ">
19 pages, 15208 KiB  
Article
Analysis of the Influence of Refraction-Parameter Deviation on Underwater Stereo-Vision Measurement with Flat Refraction Interface
by Guanqing Li, Shengxiang Huang, Zhi Yin, Nanshan Zheng and Kefei Zhang
Remote Sens. 2024, 16(17), 3286; https://doi.org/10.3390/rs16173286 - 4 Sep 2024
Viewed by 587
Abstract
There has been substantial research on multi-medium visual measurement in fields such as underwater three-dimensional reconstruction and underwater structure monitoring. Addressing the issue where traditional air-based visual-measurement models fail due to refraction when light passes through different media, numerous studies have established refraction-imaging [...] Read more.
There has been substantial research on multi-medium visual measurement in fields such as underwater three-dimensional reconstruction and underwater structure monitoring. Addressing the issue where traditional air-based visual-measurement models fail due to refraction when light passes through different media, numerous studies have established refraction-imaging models based on the actual geometry of light refraction to compensate for the effects of refraction on cross-media imaging. However, the calibration of refraction parameters inevitably contains errors, leading to deviations in these parameters. To analyze the impact of refraction-parameter deviations on measurements in underwater structure visual navigation, this paper develops a dual-media stereo-vision measurement simulation model and conducts comprehensive simulation experiments. The results indicate that to achieve high-precision underwater-measurement outcomes, the calibration method for refraction parameters, the distribution of the targets in the field of view, and the distance of the target from the camera must all be meticulously designed. These findings provide guidance for the construction of underwater stereo-vision measurement systems, the calibration of refraction parameters, underwater experiments, and practical applications. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Refractive geometry for air–glass–water medium with two flat interfaces.</p>
Full article ">Figure 2
<p>Refracted light in the dual-medium scenario.</p>
Full article ">Figure 3
<p>Air–water dual-medium visual-measurement scenario with one flat interface.</p>
Full article ">Figure 4
<p>Target area and four feature points.</p>
Full article ">Figure 5
<p>Simulation process.</p>
Full article ">Figure 6
<p>Coordinate deviation of the target area at <math display="inline"><semantics> <mrow> <mi>Z</mi> <mo>=</mo> <mn>5</mn> <mo> </mo> <mi mathvariant="normal">m</mi> </mrow> </semantics></math> when the relative refraction is 0.99843<math display="inline"><semantics> <mrow> <msub> <mi>n</mi> <mi>r</mi> </msub> </mrow> </semantics></math>. <math display="inline"><semantics> <mrow> <mrow> <mi>max</mi> <mo> </mo> </mrow> <mfenced> <mrow> <msub> <mi>d</mi> <mi>p</mi> </msub> </mrow> </mfenced> <mo>=</mo> <mn>10</mn> <mrow> <mo> </mo> <mi>mm</mi> </mrow> </mrow> </semantics></math>.</p>
Full article ">Figure 7
<p>Coordinate deviation of the target area at <math display="inline"><semantics> <mrow> <mi>Z</mi> <mo>=</mo> <mn>5</mn> <mo> </mo> <mi mathvariant="normal">m</mi> </mrow> </semantics></math> when the distance is <math display="inline"><semantics> <mrow> <mn>0.805</mn> <mi>D</mi> </mrow> </semantics></math>. <math display="inline"><semantics> <mrow> <mrow> <mi>max</mi> <mo> </mo> </mrow> <mfenced> <mrow> <msub> <mi>d</mi> <mi>p</mi> </msub> </mrow> </mfenced> <mo>=</mo> <mn>10</mn> <mrow> <mo> </mo> <mi>mm</mi> </mrow> </mrow> </semantics></math>.</p>
Full article ">Figure 8
<p>Coordinate deviation of the target area at <math display="inline"><semantics> <mrow> <mi>Z</mi> <mo>=</mo> <mn>5</mn> <mo> </mo> <mi mathvariant="normal">m</mi> </mrow> </semantics></math> when the rotation angle of the normal direction of the refractive surface is 0.195°. <math display="inline"><semantics> <mrow> <mrow> <mi>max</mi> <mo> </mo> </mrow> <mfenced> <mrow> <msub> <mi>d</mi> <mi>p</mi> </msub> </mrow> </mfenced> <mo>=</mo> <mn>10</mn> <mrow> <mo> </mo> <mi>mm</mi> </mrow> </mrow> </semantics></math>.</p>
Full article ">Figure 9
<p><math display="inline"><semantics> <mi>X</mi> </semantics></math> coordinate deviation of the target area at different distances from the camera when the relative refraction is 0.99843<math display="inline"><semantics> <mrow> <msub> <mi>n</mi> <mi>r</mi> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 10
<p><math display="inline"><semantics> <mi>Y</mi> </semantics></math> coordinate deviation of the target area at different distances from the camera when the relative refraction is 0.99843<math display="inline"><semantics> <mrow> <msub> <mi>n</mi> <mi>r</mi> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 11
<p><math display="inline"><semantics> <mi>Z</mi> </semantics></math> coordinate deviation of the target area at different distances from the camera when the relative refraction is 0.99843<math display="inline"><semantics> <mrow> <msub> <mi>n</mi> <mi>r</mi> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 12
<p><math display="inline"><semantics> <mi>X</mi> </semantics></math> coordinate deviation of the target area at different distances from the camera when the distance is <math display="inline"><semantics> <mrow> <mn>0.805</mn> <mi>D</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 13
<p><math display="inline"><semantics> <mi>Y</mi> </semantics></math> coordinate deviation of the target area at different distances from the camera when the distance is <math display="inline"><semantics> <mrow> <mn>0.805</mn> <mi>D</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 14
<p><math display="inline"><semantics> <mi>Z</mi> </semantics></math> coordinate deviation of the target area at different distances from the camera when the distance is <math display="inline"><semantics> <mrow> <mn>0.805</mn> <mi>D</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 15
<p><math display="inline"><semantics> <mi>X</mi> </semantics></math> coordinate deviation of the target area at different distances from the camera when the rotation angle of the normal direction of the refractive surface is 0.195°.</p>
Full article ">Figure 16
<p><math display="inline"><semantics> <mi>Y</mi> </semantics></math> coordinate deviation of the target area at different distances from the camera when the rotation angle of the normal direction of the refractive surface is 0.195°.</p>
Full article ">Figure 17
<p><math display="inline"><semantics> <mi>Z</mi> </semantics></math> coordinate deviation of the target area at different distances from the camera when the rotation angle of the normal direction of the refractive surface is 0.195°.</p>
Full article ">Figure 18
<p>Relationship between the coordinate components of the four target points and <math display="inline"><semantics> <mi>Z</mi> </semantics></math> when the relative refraction is 0.99843<math display="inline"><semantics> <mrow> <msub> <mi>n</mi> <mi>r</mi> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 19
<p>Relationship between the coordinate components of the four target points and <math display="inline"><semantics> <mi>Z</mi> </semantics></math> when the distance is <math display="inline"><semantics> <mrow> <mn>0.805</mn> <mi>D</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 20
<p>Relationship between the coordinate components of the four target points and <math display="inline"><semantics> <mi>Z</mi> </semantics></math> when the rotation angle of the normal direction of the refractive surface is 0.195°.</p>
Full article ">Figure 21
<p>Relationship between target-coordinate deviation and relative refractive index deviation.</p>
Full article ">Figure 22
<p>Relationship between target-coordinate deviation and camera-to-refractive-interface distance deviation.</p>
Full article ">Figure 23
<p>Relationship between target-coordinate deviation and normal direction of refraction interface.</p>
Full article ">
18 pages, 5728 KiB  
Article
NUAM-Net: A Novel Underwater Image Enhancement Attention Mechanism Network
by Zhang Wen, Yikang Zhao, Feng Gao, Hao Su, Yuan Rao and Junyu Dong
J. Mar. Sci. Eng. 2024, 12(7), 1216; https://doi.org/10.3390/jmse12071216 - 19 Jul 2024
Viewed by 825
Abstract
Vision-based underwater exploration is crucial for marine research. However, the degradation of underwater images due to light attenuation and scattering poses a significant challenge. This results in the poor visual quality of underwater images and impedes the development of vision-based underwater exploration systems. [...] Read more.
Vision-based underwater exploration is crucial for marine research. However, the degradation of underwater images due to light attenuation and scattering poses a significant challenge. This results in the poor visual quality of underwater images and impedes the development of vision-based underwater exploration systems. Recent popular learning-based Underwater Image Enhancement (UIE) methods address this challenge by training enhancement networks with annotated image pairs, where the label image is manually selected from the reference images of existing UIE methods since the groundtruth of underwater images do not exist. Nevertheless, these methods encounter uncertainty issues stemming from ambiguous multiple-candidate references. Moreover, they often suffer from local perception and color perception limitations, which hinder the effective mitigation of wide-range underwater degradation. This paper proposes a novel NUAM-Net (Novel Underwater Image Enhancement Attention Mechanism Network) that addresses these limitations. NUAM-Net leverages a probabilistic training framework, measuring enhancement uncertainty to learn the UIE mapping from a set of ambiguous reference images. By extracting features from both the RGB and LAB color spaces, our method fully exploits the fine-grained color degradation clues of underwater images. Additionally, we enhance underwater feature extraction by incorporating a novel Adaptive Underwater Image Enhancement Module (AUEM) that incorporates both local and long-range receptive fields. Experimental results on the well-known UIEBD benchmark demonstrate that our method significantly outperforms popular UIE methods in terms of PSNR while maintaining a favorable Mean Opinion Score. The ablation study also validates the effectiveness of our proposed method. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

Figure 1
<p>Illustration of uncertainty issue in UIE learning. We show examples of UIEBD datasets, i.e., the original image, (<b>a</b>) selected reference, (<b>b</b>) contrast adjustment result, (<b>c</b>) saturation adjustment result, and (<b>d</b>) gamma correction result. Multiple potential solutions can be ambiguous in reference selection since different people might choose different labels as the reference.</p>
Full article ">Figure 2
<p>The network architecture of NUAM-Net. It consists of the feature extractor, PAdaIN, AUEM, and the output blocks. The extractor’s architecture is similar to the U-Net.</p>
Full article ">Figure 3
<p>The overview of the AUEM. It consisted of a conv block and AIEM block. In the AIEM block, we try to combine and enhance the probabilistic feature. AIEM includes PConv, DWConv, LKA, SG, and IMAConv, which are five types of convolution blocks.</p>
Full article ">Figure 4
<p>Structures of LKA, SG, and CA used in our AUEM module. (<b>a</b>) Large-Kernel Attention (LKA), (<b>b</b>) Simple Gate (SG), and (<b>c</b>) Channel Attention (CA).</p>
Full article ">Figure 5
<p>Structures of IMAConv used in our AUEM module.</p>
Full article ">Figure 6
<p>Examples of the extended UIEBD dataset, including 4 labels. Label-1 denotes the manually selected label in the original UIEBD dataset, label-2 is the contrast adjustment result, label-3 is the saturation adjustment result, and label-4 is the gamma correction result.</p>
Full article ">Figure 7
<p>Qualitative results of the UIEBD test dataset. (<b>a</b>) DCP, (<b>b</b>) GC, (<b>c</b>) Retinex, (<b>d</b>) SESR, (<b>e</b>) Water-Net, (<b>f</b>) Ucolor, (<b>g</b>) PUIE-MC, (<b>h</b>) PUIE-MP, (<b>i</b>) Ours.</p>
Full article ">Figure 8
<p>Enhancement examples of our ablation studies. We show the enhanced images of backbone, backbone+LAB, and backbone+LAB+AUEM on a subset of the UIEBD test data. It is evident from the image that our network demonstrates significant improvement in enhancement effectiveness.</p>
Full article ">Figure 9
<p>Pictures show two extracted results of backbone and our network. (<b>a</b>) represents the feature extracted by our network and (<b>b</b>) represents the feature extracted by backbone network.</p>
Full article ">
22 pages, 7148 KiB  
Article
An Improved YOLOv8n Used for Fish Detection in Natural Water Environments
by Zehao Zhang, Yi Qu, Tan Wang, Yuan Rao, Dan Jiang, Shaowen Li and Yating Wang
Animals 2024, 14(14), 2022; https://doi.org/10.3390/ani14142022 - 9 Jul 2024
Cited by 2 | Viewed by 1210
Abstract
To improve detection efficiency and reduce cost consumption in fishery surveys, target detection methods based on computer vision have become a new method for fishery resource surveys. However, the specialty and complexity of underwater photography result in low detection accuracy, limiting its use [...] Read more.
To improve detection efficiency and reduce cost consumption in fishery surveys, target detection methods based on computer vision have become a new method for fishery resource surveys. However, the specialty and complexity of underwater photography result in low detection accuracy, limiting its use in fishery resource surveys. To solve these problems, this study proposed an accurate method named BSSFISH-YOLOv8 for fish detection in natural underwater environments. First, replacing the original convolutional module with the SPD-Conv module allows the model to lose less fine-grained information. Next, the backbone network is supplemented with a dynamic sparse attention technique, BiFormer, which enhances the model’s attention to crucial information in the input features while also optimizing detection efficiency. Finally, adding a 160 × 160 small target detection layer (STDL) improves sensitivity for smaller targets. The model scored 88.3% and 58.3% in the two indicators of mAP@50 and mAP@50:95, respectively, which is 2.0% and 3.3% higher than the YOLOv8n model. The results of this research can be applied to fishery resource surveys, reducing measurement costs, improving detection efficiency, and bringing environmental and economic benefits. Full article
(This article belongs to the Section Aquatic Animals)
Show Figures

Figure 1

Figure 1
<p>YOLOv8 network structure.</p>
Full article ">Figure 2
<p>SPD-Conv module structure.</p>
Full article ">Figure 3
<p>BiFormer attention structure.</p>
Full article ">Figure 4
<p>Target size distribution of the dataset (Darker colours represent a greater number of distributions).</p>
Full article ">Figure 5
<p>BSSFISH-YOLOv8 network structure.</p>
Full article ">Figure 6
<p>Examples of enhanced images: (<b>a</b>) vertical flip; (<b>b</b>) horizontal flip; (<b>c</b>) brightness adjustment; (<b>d</b>) Gaussian blur; (<b>e</b>) affine transformation translation; (<b>f</b>) affine transformation scaling; (<b>g</b>) channel addition; (<b>h</b>) rotate; (<b>i</b>) Gaussian noise.</p>
Full article ">Figure 7
<p>Typical images in the dataset: (<b>a</b>) blur; (<b>b</b>) occlusion; (<b>c</b>) small targets.</p>
Full article ">Figure 8
<p>Comparison of heat maps: (<b>a</b>) YOLOv8n; (<b>b</b>) BSSFISH-YOLOv8.</p>
Full article ">Figure 9
<p>Feature maps of different scales: (<b>a</b>) 160 × 160; (<b>b</b>) 80 × 80; (<b>c</b>) 40 × 40; (<b>d</b>) 20 × 20.</p>
Full article ">Figure 10
<p>Comparison of improvement effects. From top to bottom: original images; YOLOv8n; BSSFISH-YOLOv8. (<b>a</b>) blur; (<b>b</b>) occlusion; (<b>c</b>) small targets.</p>
Full article ">Figure 11
<p>mAP@50 curve.</p>
Full article ">Figure 12
<p>Confusion matrix for identification of different fish species. (<b>a</b>) quantitative information (<b>b</b>) ratio information.</p>
Full article ">Figure 13
<p>Demonstration of cases with higher and lower fish detection accuracy: (<b>a</b>) Blue Catfish; (<b>b</b>) Yellowfin Bream; (<b>c</b>) Eastern Striped Grunter.</p>
Full article ">
16 pages, 7318 KiB  
Article
Non-Contact Tilapia Mass Estimation Method Based on Underwater Binocular Vision
by Guofu Feng, Bo Pan and Ming Chen
Appl. Sci. 2024, 14(10), 4009; https://doi.org/10.3390/app14104009 - 8 May 2024
Cited by 1 | Viewed by 836
Abstract
The non-destructive measurement of fish is an important link in intelligent aquaculture, and realizing the accurate estimation of fish mass is the key to the stable operation of this link. Taking tilapia as the object, this study proposes an underwater tilapia mass estimation [...] Read more.
The non-destructive measurement of fish is an important link in intelligent aquaculture, and realizing the accurate estimation of fish mass is the key to the stable operation of this link. Taking tilapia as the object, this study proposes an underwater tilapia mass estimation method, which can accurately estimate the mass of free-swimming tilapia under non-contact conditions. First, image enhancement is performed on the original image, and the depth image is obtained by correcting and stereo matching the enhanced image using binocular stereo vision technology. And the fish body is segmented by an SAM model. Then, the segmented fish body is labeled with key points, thus realizing the 3D reconstruction of tilapia. Five mass estimation models are established based on the relationship between the body length and the mass of tilapia, so as to realize the mass estimation of tilapia. The results showed that the average relative errors of the method models were 5.34%~7.25%. The coefficient of determination of the final tilapia mass estimation with manual measurement was 0.99, and the average relative error was 5.90%. The improvement over existing deep learning methods is about 1.54%. This study will provide key technical support for the non-destructive measurement of tilapia, which is of great significance to the information management of aquaculture, the assessment of fish growth condition, and baiting control. Full article
(This article belongs to the Special Issue Engineering of Smart Agriculture—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Experimental equipment and materials: 1, binocular camera; 2, breeding water; 3, calibration board; 4, breeding box; 5, cables; 6, laptop; 7, electronic scale.</p>
Full article ">Figure 2
<p>Flowchart for mass estimation of tilapia based on underwater binocular vision.</p>
Full article ">Figure 3
<p>Relative position of binocular camera to calibration board.</p>
Full article ">Figure 4
<p>Calibration board corner point detection.</p>
Full article ">Figure 5
<p>Comparison of reprojection error before and after processing: (<b>a</b>) histogram of reprojection error for first calibration; (<b>b</b>) histogram of reprojection error after removal of large errors.</p>
Full article ">Figure 6
<p>Underwater image enhancement results based on Retinex: (<b>a</b>) original image; (<b>b</b>) SSR; (<b>c</b>) MSRCR; (<b>d</b>) MSR.</p>
Full article ">Figure 7
<p>Binocular camera ranging model.</p>
Full article ">Figure 8
<p>Maps generated by the SGBM algorithm: (<b>a</b>) disparity map; (<b>b</b>) depth map.</p>
Full article ">Figure 9
<p>Image segmentation platform.</p>
Full article ">Figure 10
<p>SAM platform segmentation results: (<b>a</b>) fish body layering; (<b>b</b>) full segmentation result.</p>
Full article ">Figure 11
<p>Key point pixel labeling.</p>
Full article ">Figure 12
<p>Effect of fish distance and angle variation on pixel body length.</p>
Full article ">Figure 13
<p>Quadratic term fit plots for different mass groups of tilapia.</p>
Full article ">Figure 14
<p>Comparison between estimated and manually measured fish mass.</p>
Full article ">
19 pages, 11331 KiB  
Article
Advanced Underwater Measurement System for ROVs: Integrating Sonar and Stereo Vision for Enhanced Subsea Infrastructure Maintenance
by Jiawei Zhang, Fenglei Han, Duanfeng Han, Jianfeng Yang, Wangyuan Zhao and Hansheng Li
J. Mar. Sci. Eng. 2024, 12(2), 306; https://doi.org/10.3390/jmse12020306 - 9 Feb 2024
Viewed by 1942
Abstract
In the realm of ocean engineering and maintenance of subsea structures, accurate underwater distance quantification plays a crucial role. However, the precision of such measurements is often compromised in underwater environments due to backward scattering and feature degradation, adversely affecting the accuracy of [...] Read more.
In the realm of ocean engineering and maintenance of subsea structures, accurate underwater distance quantification plays a crucial role. However, the precision of such measurements is often compromised in underwater environments due to backward scattering and feature degradation, adversely affecting the accuracy of visual techniques. Addressing this challenge, our study introduces a groundbreaking method for underwater object measurement, innovatively combining image sonar with stereo vision. This approach aims to supplement the gaps in underwater visual feature detection with sonar data while leveraging the distance information from sonar for enhanced visual matching. Our methodology seamlessly integrates sonar data into the Semi-Global Block Matching (SGBM) algorithm used in stereo vision. This integration involves introducing a novel sonar-based cost term and refining the cost aggregation process, thereby both elevating the precision in depth estimations and enriching the texture details within the depth maps. This represents a substantial enhancement over existing methodologies, particularly in the texture augmentation of depth maps tailored for subaquatic environments. Through extensive comparative analyses, our approach demonstrates a substantial reduction in measurement errors by 1.6%, showing significant promise in challenging underwater scenarios. The adaptability and accuracy of our algorithm in generating detailed depth maps make it particularly relevant for underwater infrastructure maintenance, exploration, and inspection. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

Figure 1
<p>Comparative visualization of an underwater object. (<b>a</b>) captured through optical imaging, showing the object’s appearance with ambient light, and (<b>b</b>) captured through image sonar, depicting acoustic reflections used for object detection and mapping in low-visibility underwater conditions.</p>
Full article ">Figure 2
<p>Framework of the proposed method. This framework shows a multi-stage approach to underwater stereo matching with a stereo camera and image sonar, starting with PART I: Vision-Based Cost Calculation utilizing stereo image optimization and semantic segmentation. PART II: Sonar-Based Cost Calculation, where sonar imagery informs cost metrics. PART III: Aggregation of costs from both visual and sonar data to compute disparity, illustrating the synergy between acoustic and optical sensing modalities for enhanced underwater perception.</p>
Full article ">Figure 3
<p>Joint coordinate system for left camera and sonar.</p>
Full article ">Figure 4
<p>Cost volume. The left volume is the origin cost volume, and the right volume is the cost volume after improvement, where C means the cost.</p>
Full article ">Figure 5
<p>The upper portion of the illustration depicts the traditional process for calculating the SAD cost. In contrast, the lower portion illustrates the cost calculation process utilizing a dilated SAD window. <math display="inline"><semantics> <msub> <mi>N</mi> <mi>P</mi> </msub> </semantics></math> is the reference support window, and <math display="inline"><semantics> <msub> <mi>N</mi> <mrow> <mi>p</mi> <mi>d</mi> </mrow> </msub> </semantics></math> is the target support window.</p>
Full article ">Figure 6
<p>Target objects used in the experiment: (<b>a</b>) frame-type structure; (<b>b</b>) pressure tank; (<b>c</b>) sphere structure; (<b>d</b>) platform structure.</p>
Full article ">Figure 7
<p>ROV system with stereo camera and image sonar.</p>
Full article ">Figure 8
<p>Comparisons of different stereo matching methods: (<b>a</b>) sonar image, (<b>b</b>) raw image, (<b>c</b>) IGEV, (<b>d</b>) GA-Net, (<b>e</b>) SGBM, (<b>f</b>) BM, (<b>g</b>) ours.</p>
Full article ">Figure 9
<p>Algorithm performance (<b>a</b>) without sonar data and (<b>b</b>) with sonar data.</p>
Full article ">Figure 10
<p>Histograms of the measurement accuracy.</p>
Full article ">Figure 11
<p>Results with different weights: (<b>a</b>) <math display="inline"><semantics> <mrow> <mi>w</mi> <mo>=</mo> <mn>0.3</mn> </mrow> </semantics></math>, (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>w</mi> <mo>=</mo> <mn>0.6</mn> </mrow> </semantics></math>, (<b>c</b>) <math display="inline"><semantics> <mrow> <mi>w</mi> <mo>=</mo> <mn>0.9</mn> </mrow> </semantics></math>.</p>
Full article ">
17 pages, 16728 KiB  
Article
Seaweed Growth Monitoring with a Low-Cost Vision-Based System
by Jeroen Gerlo, Dennis G. Kooijman, Ivo W. Wieling, Ritchie Heirmans and Steve Vanlanduit
Sensors 2023, 23(22), 9197; https://doi.org/10.3390/s23229197 - 15 Nov 2023
Cited by 3 | Viewed by 2048
Abstract
In this paper, we introduce a method for automated seaweed growth monitoring by combining a low-cost RGB and stereo vision camera. While current vision-based seaweed growth monitoring techniques focus on laboratory measurements or above-ground seaweed, we investigate the feasibility of the underwater imaging [...] Read more.
In this paper, we introduce a method for automated seaweed growth monitoring by combining a low-cost RGB and stereo vision camera. While current vision-based seaweed growth monitoring techniques focus on laboratory measurements or above-ground seaweed, we investigate the feasibility of the underwater imaging of a vertical seaweed farm. We use deep learning-based image segmentation (DeeplabV3+) to determine the size of the seaweed in pixels from recorded RGB images. We convert this pixel size to meters squared by using the distance information from the stereo camera. We demonstrate the performance of our monitoring system using measurements in a seaweed farm in the River Scheldt estuary (in The Netherlands). Notwithstanding the poor visibility of the seaweed in the images, we are able to segment the seaweed with an intersection of the union (IoU) of 0.9, and we reach a repeatability of 6% and a precision of the seaweed size of 18%. Full article
(This article belongs to the Special Issue Intelligent Sensing and Machine Vision in Precision Agriculture)
Show Figures

Figure 1

Figure 1
<p>Common seaweed farm typology: Floaters or buoys (3) are held in position on the water surface using anchors (1) and tethers (2). A long rope line (4) is suspended between the fixed buoys. The seaweed (6) is seeded on this suspended line, growing downwards and flowing freely in the water current. Attached buoys and/or floaters (3, 5) keep the seaweed in the upper water column in order to catch sunlight.</p>
Full article ">Figure 2
<p>Seaweed farm schematic during growth. (7) Dense seaweed bundles; (8) delineated surface area of the seaweed, approximating it as a continuous 2D sheet.</p>
Full article ">Figure 3
<p>Sequence of the automated monitoring process. The separate components are designed to be integrated in the UTOPIA framework.</p>
Full article ">Figure 4
<p>(<b>Left</b>) Area within the red dashed line shows the waterproof housing and mounting of the Realsense camera. (<b>Right</b>) The camera setup submerged and recording during one of the measurement campaigns at Neeltje Jans.</p>
Full article ">Figure 5
<p>(<b>Left</b>) Sharp images of seaweed in the foreground, but too close to determine plant size. (<b>Right</b>) The seaweed plant is in frame, but is too far away to be sufficiently sharp due to the haze created by turbidity and lighting conditions at the moment of measurement.</p>
Full article ">Figure 6
<p>Example of underwater images taken with the Realsense RGB camera. Seaweed plants can take on different colors and shapes depending on current lighting conditions, depth and position within the seaweed bundles.</p>
Full article ">Figure 7
<p>Example of manual annotation on a seaweed RGB image. (<b>Left</b>): original image. (<b>Right</b>): original image with the annotation superimposed in blue with a black border. Two bundles of seaweed are visible in the foreground, both growing on the closest rope line. Another bundle of seaweed is visible in the background, much more obscured by the water haze.</p>
Full article ">Figure 8
<p>DeeplabV3+ architecture. The structure of the Atruous Spatial Pyramid Pooling (ASSP) decoder module is shown, followed by a simple decoder to acquire image prediction. Adapted from [<a href="#B31-sensors-23-09197" class="html-bibr">31</a>].</p>
Full article ">Figure 9
<p>Segmentation results of the DeeplabV3+ model for different hyperparameters. Segmentation masks are shown as a yellow overlay, with the base image in the bottom right.</p>
Full article ">Figure 10
<p>More examples of seaweed segmentation with the DeeplabV3+ model using our optimal hyperparameters. The model correctly distinguishes a gap between the seaweed plants in the foreground, and ignores the seaweed on a second line in the background. Finer details of trailing seaweed at the bottom of the plants is not detected due to limitations on pixel resolution for the segmentation.</p>
Full article ">Figure 11
<p>Evolution of the intersection over union (IoU) metric of the training images using different hyperparameters. A perfect model would reach an IoU of 1. The 576 px model converges towards an IoU of 0.9 after 100 iterations (epochs).</p>
Full article ">Figure 12
<p>Intersection over union metric for the DeeplabV3+ model for seaweed segmentation. Training and validation IoU are approximately identical after 100 epochs.</p>
Full article ">Figure 13
<p>Spread of the segmented surface area as a percentile of the total image, for two one-minute steady camera positions. There are no large outliers in either test.</p>
Full article ">Figure 14
<p>Construction of a disparity map. (<b>top</b>) Original left and right stereo images. (<b>bottom</b>) Normalized disparity map, calculated via the Semi-Global Block Matching algorithm.</p>
Full article ">Figure 15
<p>Spread of the distance measurement 2D median with a steady plant target, calculated for 100 images. The red line denotes the mean.</p>
Full article ">Figure 16
<p>Segmented images, with seaweed crop yield in square meters as calculated by our algorithm. The pixel area of seaweed on the left image is larger than the right, but the left image was captured at a shorter distance from the seaweed (1.99 m vs. 2.97 m). Our method returns a smaller surface area in m<math display="inline"><semantics> <msup> <mrow/> <mn>2</mn> </msup> </semantics></math> for the left image.</p>
Full article ">Figure 17
<p>Additional examples of segmented images, with seaweed crop yield in square meters as calculated by our algorithm.</p>
Full article ">
17 pages, 25768 KiB  
Article
SwimmerNET: Underwater 2D Swimmer Pose Estimation Exploiting Fully Convolutional Neural Networks
by Nicola Giulietti, Alessia Caputo, Paolo Chiariotti and Paolo Castellini
Sensors 2023, 23(4), 2364; https://doi.org/10.3390/s23042364 - 20 Feb 2023
Cited by 14 | Viewed by 3436
Abstract
Professional swimming coaches make use of videos to evaluate their athletes’ performances. Specifically, the videos are manually analyzed in order to observe the movements of all parts of the swimmer’s body during the exercise and to give indications for improving swimming technique. This [...] Read more.
Professional swimming coaches make use of videos to evaluate their athletes’ performances. Specifically, the videos are manually analyzed in order to observe the movements of all parts of the swimmer’s body during the exercise and to give indications for improving swimming technique. This operation is time-consuming, laborious and error prone. In recent years, alternative technologies have been introduced in the literature, but they still have severe limitations that make their correct and effective use impossible. In fact, the currently available techniques based on image analysis only apply to certain swimming styles; moreover, they are strongly influenced by disturbing elements (i.e., the presence of bubbles, splashes and reflections), resulting in poor measurement accuracy. The use of wearable sensors (accelerometers or photoplethysmographic sensors) or optical markers, although they can guarantee high reliability and accuracy, disturb the performance of the athletes, who tend to dislike these solutions. In this work we introduce swimmerNET, a new marker-less 2D swimmer pose estimation approach based on the combined use of computer vision algorithms and fully convolutional neural networks. By using a single 8 Mpixel wide-angle camera, the proposed system is able to estimate the pose of a swimmer during exercise while guaranteeing adequate measurement accuracy. The method has been successfully tested on several athletes (i.e., different physical characteristics and different swimming technique), obtaining an average error and a standard deviation (worst case scenario for the dataset analyzed) of approximately 1 mm and 10 mm, respectively. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>A single camera was fixed sideways underwater, and only the submerged athlete’s body is framed.</p>
Full article ">Figure 2
<p>Skeleton that represents a model of a human body. The black dots represent the manually annotated targets.</p>
Full article ">Figure 3
<p>Example of wrist trajectories. (<b>a</b>): Curves originally labeled by the models. Labels are highly mixed due to the symmetry of human body with respect to the sagittal plane. (<b>b</b>): Curves corrected by the proposed algorithm. In the example, the red curve represents the right wrist’s trajectory and the blue curve represents the left wrist’s trajectory.</p>
Full article ">Figure 4
<p>The athlete is identified by the model and used to define a fixed-size ROI of 1024 × 576 pixel in order to obtain small images with the swimmer always in the center of the frame. If the identified ROI falls outside of the input image, the frame is discarded and the next one is used.</p>
Full article ">Figure 5
<p>The binary mask identified by the semantic segmentation model for target 1 (i.e., the swimmer’s head) is superimposed on the input image. The position of the target, in pixel, is taken as the position of the center of gravity from the area identified by the model.</p>
Full article ">Figure 6
<p>Through iterative application of the developed semantic segmentation models, a coordinate is assigned to each targeted body part that is visible within the frame.</p>
Full article ">Figure 7
<p>Examples of the different videos used for testing the SwimmerNET method: one athlete swimming freestyle for training (<b>a</b>) and two new athletes in different pools for the test phase—specifically, a female athlete performing freestyle (<b>c</b>) and a male athlete performing dolphin (<b>b</b>) and backstroke (<b>d</b>). Finally, there are frames extrapolated from videos gathered from the public Sports Videos in the Wild (SVW) repository [<a href="#B27-sensors-23-02364" class="html-bibr">27</a>] showing a male athlete during a freestyle swimming session (<b>e</b>,<b>f</b>).</p>
Full article ">Figure 8
<p>Examples of trajectories of the right parts of the body during a swimming stroke: a 1.90 m tall male athlete during freestyle (<b>a</b>) and a 1.80 m tall male athlete during dolphin style (<b>b</b>).</p>
Full article ">Figure 9
<p>Percentage of targets not recognized by the proposed method divided by body part.</p>
Full article ">Figure 10
<p>The presence of outliers in the target location increases the standard deviation of the error. Once eliminated, the standard deviation remains below 5 pixels for each target.</p>
Full article ">Figure 11
<p>Mean error in locating body parts.</p>
Full article ">Figure 12
<p>Standard deviation of mean error in locating body parts.</p>
Full article ">
14 pages, 2941 KiB  
Article
An Automatic Recognition Method for Fish Species and Length Using an Underwater Stereo Vision System
by Yuxuan Deng, Hequn Tan, Minghang Tong, Dianzhuo Zhou, Yuxiang Li and Ming Zhu
Fishes 2022, 7(6), 326; https://doi.org/10.3390/fishes7060326 - 10 Nov 2022
Cited by 15 | Viewed by 2980
Abstract
Developing new methods to detect biomass information on freshwater fish in farm conditions enables the creation of decision bases for precision feeding. In this study, an approach based on Keypoints R-CNN is presented to identify species and measure length automatically using an underwater [...] Read more.
Developing new methods to detect biomass information on freshwater fish in farm conditions enables the creation of decision bases for precision feeding. In this study, an approach based on Keypoints R-CNN is presented to identify species and measure length automatically using an underwater stereo vision system. To enhance the model’s robustness, stochastic enhancement is performed on image datasets. For further promotion of the features extraction capability of the backbone network, an attention module is integrated into the ResNeXt50 network. Concurrently, the feature pyramid network (FPN) is replaced by an improved path aggregation network (I-PANet) to achieve a greater fusion of effective feature maps. Compared to the original model, the mAP of the improved one in object and key point detection tasks increases by 4.55% and 2.38%, respectively, with a small increase in the number of model parameters. In addition, a new algorithm is introduced for matching the detection results of neural networks. On the foundation of the above contents, coordinates of head and tail points in stereo images as well as fish species can be obtained rapidly and accurately. A 3D reconstruction of the fish head and tail points is performed utilizing the calibration parameters and projection matrix of the stereo camera. The estimated length of the fish is acquired by calculating the Euclidean distance between two points. Finally, the precision of the proposed approach proved to be acceptable for five kinds of common freshwater fish. The accuracy of species identification exceeds 94%, and the relative errors of length measurement are less than 10%. In summary, this method can be utilized to help aquaculture farmers efficiently collect real-time information about fish length. Full article
(This article belongs to the Section Fishery Facilities, Equipment, and Information Technology)
Show Figures

Figure 1

Figure 1
<p>Schematic diagram of the image acquisition platform, (<b>a</b>) Structure of the platform, (<b>b</b>) Photo of the actual device.</p>
Full article ">Figure 2
<p>Schematic diagram of the improved Keypoints RCNN model, (<b>a</b>) General structure of the model, (<b>b</b>) Main process of the RPN and RCNN part.</p>
Full article ">Figure 3
<p>Structure schematic of the CBAM module. Note: <math display="inline"><semantics> <mrow> <msub> <mi>w</mi> <mi>c</mi> </msub> </mrow> </semantics></math> is the channel attention weight, <math display="inline"><semantics> <mrow> <msub> <mi>w</mi> <mi>s</mi> </msub> </mrow> </semantics></math> is the spatial attention weight, <span class="html-italic">F</span> is the original feature map, <span class="html-italic">F′</span> is the channel weighted feature map, and <span class="html-italic">F″</span> is the spatial weighted feature map.</p>
Full article ">Figure 4
<p>Schematic diagram of the I-PANet, (<b>a</b>) Structure of the I-PANet, (<b>b</b>) Structure of the “Conv block”.</p>
Full article ">Figure 5
<p>Training process of ResNeXt with I-PANet.</p>
Full article ">Figure 6
<p>Confusion matrix of recognition results.</p>
Full article ">Figure 7
<p>Detection and stereo matching results of the left and right image, (<b>a</b>) the left image, (<b>b</b>) the right image, (<b>c</b>) the MS matrix.</p>
Full article ">Figure 7 Cont.
<p>Detection and stereo matching results of the left and right image, (<b>a</b>) the left image, (<b>b</b>) the right image, (<b>c</b>) the MS matrix.</p>
Full article ">Figure 8
<p>The results of the fish length measurement experiment, (<b>a</b>) result mark on the original image, (<b>b</b>) relative errors of the length measuring.</p>
Full article ">
23 pages, 30385 KiB  
Article
Binocular-Vision-Based Obstacle Avoidance Design and Experiments Verification for Underwater Quadrocopter Vehicle
by Meiyan Zhang, Wenyu Cai, Qinan Xie and Shenyang Xu
J. Mar. Sci. Eng. 2022, 10(8), 1050; https://doi.org/10.3390/jmse10081050 - 30 Jul 2022
Cited by 5 | Viewed by 2301
Abstract
As we know, for autonomous robots working in a complex underwater region, obstacle avoidance design will play an important role in underwater tasks. In this paper, a binocular-vision-based underwater obstacle avoidance mechanism is discussed and verified with our self-made Underwater Quadrocopter Vehicle. The [...] Read more.
As we know, for autonomous robots working in a complex underwater region, obstacle avoidance design will play an important role in underwater tasks. In this paper, a binocular-vision-based underwater obstacle avoidance mechanism is discussed and verified with our self-made Underwater Quadrocopter Vehicle. The proposed Underwater Quadrocopter Vehicle (UQV for short), like a quadrocopter drone working underwater, is a new kind of Autonomous Underwater Vehicle (AUV), which is equipped with four propellers along the vertical direction of the robotic body to adjust its body posture and two propellers arranged at the sides of the robotic body to provide propulsive and turning force. Moreover, an underwater binocular-vision-based obstacle positioning method is studied to measure an underwater spherical obstacle’s radius and its distance from the UQV. Due to its perfect ability of full-freedom underwater actions, the proposed UQV has obvious advantages such as a zero turning radius compared with existing torpedo-shaped AUVs. Therefore, one semicircle-curve-based obstacle avoidance path is planned on the basis of an obstacle’s coordinates. Practical pool experiments show that the proposed binocular vision can locate an underwater obstacle accurately, and the designed UQV has the ability to effectively avoid multiple obstacles along the predefined trajectory. Full article
(This article belongs to the Special Issue Advances in Marine Vehicles, Automation and Robotics)
Show Figures

Figure 1

Figure 1
<p>The mechanical structure of the underwater quadrocopter vehicle.</p>
Full article ">Figure 2
<p>Hardware diagram.</p>
Full article ">Figure 3
<p>Dynamic model of the UQV.</p>
Full article ">Figure 4
<p>Rotational model in the body coordinate system.</p>
Full article ">Figure 5
<p>Underwater binocular-vision-based obstacle positioning.</p>
Full article ">Figure 6
<p>Camera imaging model.</p>
Full article ">Figure 7
<p>Camera calibration images.</p>
Full article ">Figure 8
<p>Calibration modeling results.</p>
Full article ">Figure 9
<p>Imaging pixels of target point between two cameras.</p>
Full article ">Figure 10
<p>Block stereo-matching algorithm.</p>
Full article ">Figure 11
<p>Obstacle depth computing model.</p>
Full article ">Figure 12
<p>Obstacle -avoidance-based path planning framework.</p>
Full article ">Figure 13
<p>Obstacle avoidance path-planning framework.</p>
Full article ">Figure 14
<p>Actual and ideal path.</p>
Full article ">Figure 15
<p>Projected paths in the <math display="inline"><semantics> <mrow> <mi>X</mi> <mo>−</mo> <mi>a</mi> <mi>x</mi> <mi>i</mi> <mi>s</mi> </mrow> </semantics></math>, the <math display="inline"><semantics> <mrow> <mi>Y</mi> <mo>−</mo> <mi>a</mi> <mi>x</mi> <mi>i</mi> <mi>s</mi> </mrow> </semantics></math> and the <math display="inline"><semantics> <mrow> <mi>Z</mi> <mo>−</mo> <mi>a</mi> <mi>x</mi> <mi>i</mi> <mi>s</mi> </mrow> </semantics></math>, respectively.</p>
Full article ">Figure 15 Cont.
<p>Projected paths in the <math display="inline"><semantics> <mrow> <mi>X</mi> <mo>−</mo> <mi>a</mi> <mi>x</mi> <mi>i</mi> <mi>s</mi> </mrow> </semantics></math>, the <math display="inline"><semantics> <mrow> <mi>Y</mi> <mo>−</mo> <mi>a</mi> <mi>x</mi> <mi>i</mi> <mi>s</mi> </mrow> </semantics></math> and the <math display="inline"><semantics> <mrow> <mi>Z</mi> <mo>−</mo> <mi>a</mi> <mi>x</mi> <mi>i</mi> <mi>s</mi> </mrow> </semantics></math>, respectively.</p>
Full article ">Figure 16
<p>The photograph of the UQV.</p>
Full article ">Figure 17
<p>Web page control interfaces.</p>
Full article ">Figure 18
<p>Obstacle-free trajectory.</p>
Full article ">Figure 19
<p>X-axis and Y-axis curves of the obstacle-free trajectory.</p>
Full article ">Figure 20
<p>Experimental environment.</p>
Full article ">Figure 21
<p>The distribution of three spherical obstacles.</p>
Full article ">Figure 22
<p>Underwater obstacle image processing results.</p>
Full article ">Figure 23
<p>Key steps of obstacle avoidance trajectory.</p>
Full article ">
25 pages, 6991 KiB  
Article
Stereo Vision System for Vision-Based Control of Inspection-Class ROVs
by Stanisław Hożyń and Bogdan Żak
Remote Sens. 2021, 13(24), 5075; https://doi.org/10.3390/rs13245075 - 14 Dec 2021
Cited by 9 | Viewed by 3111
Abstract
The inspection-class Remotely Operated Vehicles (ROVs) are crucial in underwater inspections. Their prime function is to allow the replacing of humans during risky subaquatic operations. These vehicles gather videos from underwater scenes that are sent online to a human operator who provides control. [...] Read more.
The inspection-class Remotely Operated Vehicles (ROVs) are crucial in underwater inspections. Their prime function is to allow the replacing of humans during risky subaquatic operations. These vehicles gather videos from underwater scenes that are sent online to a human operator who provides control. Furthermore, these videos are used for analysis. This demands an RGB camera operating at a close distance to the observed objects. Thus, to obtain a detailed depiction, the vehicle should move with a constant speed and a measured distance from the bottom. As very few inspection-class ROVs possess navigation systems that facilitate these requirements, this study had the objective of designing a vision-based control method to compensate for this limitation. To this end, a stereo vision system and image-feature matching and tracking techniques were employed. As these tasks are challenging in the underwater environment, we carried out analyses aimed at finding fast and reliable image-processing techniques. The analyses, through a sequence of experiments designed to test effectiveness, were carried out in a swimming pool using a VideoRay Pro 4 vehicle. The results indicate that the method under consideration enables automatic control of the vehicle, given that the image features are present in stereo-pair images as well as in consecutive frames captured by the left camera. Full article
Show Figures

Figure 1

Figure 1
<p>Summary of the developed method.</p>
Full article ">Figure 2
<p>Stereo image rectification.</p>
Full article ">Figure 3
<p>Canonical stereo setup.</p>
Full article ">Figure 4
<p>Distance reconstruction using a stereo vision system.</p>
Full article ">Figure 5
<p>Corresponding region of visibility for different distances from bottom and baselines.</p>
Full article ">Figure 6
<p>Corresponding region of visibility for different distances from bottom and the vehicle’s velocities.</p>
Full article ">Figure 7
<p>An underwater vehicle with cameras.</p>
Full article ">Figure 8
<p>Corresponding region of visibility for different values of <math display="inline"><semantics> <mi>α</mi> </semantics></math>.</p>
Full article ">Figure 9
<p>Feature points matching for velocity calculation.</p>
Full article ">Figure 10
<p>Body-fixed and inertial frames.</p>
Full article ">Figure 11
<p>Control system.</p>
Full article ">Figure 12
<p>Surge measurement with the Kalman filter.</p>
Full article ">Figure 13
<p>(<b>a</b>) VideoRay Pro 4 with stereo vision cameras; (<b>b</b>) swimming pool tests.</p>
Full article ">Figure 14
<p>Membership functions: (<b>a</b>) error; (<b>b</b>) derivative of error; and (<b>c</b>) output.</p>
Full article ">Figure 15
<p>Heading control for velocity equals 0.1 m/s: (<b>a</b>) stereo vision; (<b>b</b>) SLAM lasers; (<b>c</b>) SLAM stereo; (<b>d</b>) SLAM IMU.</p>
Full article ">Figure 16
<p>Heading control for velocity equals 0.2 m/s: (<b>a</b>) stereo vision; (<b>b</b>) SLAM lasers; (<b>c</b>) SLAM stereo; (<b>d</b>) SLAM IMU.</p>
Full article ">Figure 17
<p>Surge control at variable speed: (<b>a</b>) stereo vision; (<b>b</b>) SLAM lasers; (<b>c</b>) SLAM stereo; (<b>d</b>) SLAM IMU.</p>
Full article ">Figure 18
<p>Surge control at constant speed equals 0.2 m/s: (<b>a</b>) stereo vision; (<b>b</b>) SLAM lasers; (<b>c</b>) SLAM stereo; (<b>d</b>) SLAM IMU.</p>
Full article ">Figure 19
<p>Distance-from-the-bottom control at a speed of 0.1 m/s: (<b>a</b>) stereo vision; (<b>b</b>) SLAM lasers; (<b>c</b>) SLAM stereo; (<b>d</b>) SLAM IMU.</p>
Full article ">Figure 20
<p>Distance-from-the-bottom control at a speed of 0.2 m/s: (<b>a</b>) stereo vision; (<b>b</b>) SLAM lasers; (<b>c</b>) SLAM stereo; (<b>d</b>) SLAM IMU.</p>
Full article ">
18 pages, 3983 KiB  
Article
Method for the Coordination of Referencing of Autonomous Underwater Vehicles to Man-Made Objects Using Stereo Images
by Valery Bobkov, Alexey Kudryashov and Alexander Inzartsev
J. Mar. Sci. Eng. 2021, 9(9), 1038; https://doi.org/10.3390/jmse9091038 - 21 Sep 2021
Cited by 9 | Viewed by 2505
Abstract
The use of an autonomous underwater vehicle (AUV) to inspect underwater industrial infrastructure requires the precise, coordinated movement of the AUV relative to subsea objects. One significant underwater infrastructure system is the subsea production system (SPS), which includes wells for oil and gas [...] Read more.
The use of an autonomous underwater vehicle (AUV) to inspect underwater industrial infrastructure requires the precise, coordinated movement of the AUV relative to subsea objects. One significant underwater infrastructure system is the subsea production system (SPS), which includes wells for oil and gas production, located on the seabed. The present paper suggests a method for the accurate navigation of AUVs in a distributed SPS to coordinate space using video information. This method is based on the object recognition and computation of the AUV coordinate references to SPS objects. Stable high accuracy during the continuous movement of the AUV in SPS space is realized through the regular updating of the coordinate references to SPS objects. Stereo images, a predefined geometric SPS model, and measurements of the absolute coordinates of a limited number of feature points of objects are used as initial data. The matrix of AUV coordinate references to the SPS object coordinate system is computed using 3D object points matched with the model. The effectiveness of the proposed method is estimated based on the results of computational experiments with virtual scenes generated in the simulator for AUV, and with real data obtained by the Karmin2 stereo camera (Nerian Vision, Stuttgart, Germany) in laboratory conditions. Full article
(This article belongs to the Special Issue Maritime Autonomous Vessels)
Show Figures

Figure 1

Figure 1
<p>Subsea production system (SPS).</p>
Full article ">Figure 2
<p>Construction of an object coordinate system based on points P<sub>1</sub>, P<sub>2</sub>, and P<sub>3</sub> specified in the <span class="html-italic">WCS</span>: the X axis is determined by points P<sub>1</sub>P<sub>3</sub>, the Z axis is normal to the plane of the P<sub>1</sub>P<sub>2</sub> and P<sub>1</sub>P<sub>3</sub> vectors, and the Y axis is normal to the ZX plane.</p>
Full article ">Figure 3
<p>Coordinate systems (<span class="html-italic">WCS</span>, CS SPS, object CS, CS associated with an AUV in the initial and current positions) used and geometric transformations between them.</p>
Full article ">Figure 4
<p>Points in min–max-shells in the 3D cloud, marked in black, are identified by the search algorithm as belonging to the SPS object.</p>
Full article ">Figure 5
<p>Calculation of the geometric transformation matrix from the AUV CS (<span class="html-italic">CS<sup>AUV</sup></span>) to the CS of the SPS object (<span class="html-italic">CS<sup>ob_k</sup></span>) using the auxiliary CS (<span class="html-italic">CS<sup>ad</sup></span>).</p>
Full article ">Figure 6
<p>Virtual stage: AUV performs SPS inspection.</p>
Full article ">Figure 7
<p>Identification of points belonging to the SPS object: (<b>a</b>) extraction of characteristic points by the Harris angle detector in the image taken by the camera; (<b>b</b>) a set of 3D points constructed from 2D images. All selected points in the scene are marked in white. Points belonging to SPS objects are marked in black. Six points belonging to SPS have been identified.</p>
Full article ">Figure 8
<p>In the photo taken by the Karmin2 camera, the desired objects are marked: A–F. Points belonging to the desired objects (model) are marked in black. Their number is 48, as indicated by the operator showing eight on each box. Of these, 33 points fell into the camera’s field of view: on object A—7, on B—7, on C—6, on D—2, on E—4, on F—7. The coordinate system (CS), in which all points of objects (model) were set, was built on three corner points of object A. The points built by the Harris detector are marked with white circles. There are 89 of them in the scene.</p>
Full article ">Figure 9
<p>The figure shows the points identified in the 3D cloud (marked in black) as belonging to the sought objects. Their number was 13: on object A—2, on B—3, on C—3, on D—1, on E—1, on F—3. The matrix connecting the CS of objects with the CS of the camera was calculated by 3 points (they are marked with numbers 1, 2, 3), which were selected by the algorithm from the found points.</p>
Full article ">
33 pages, 13767 KiB  
Article
Elevation Angle Estimations of Wide-Beam Acoustic Sonar Measurements for Autonomous Underwater Karst Exploration
by Yohan Breux and Lionel Lapierre
Sensors 2020, 20(14), 4028; https://doi.org/10.3390/s20144028 - 20 Jul 2020
Cited by 4 | Viewed by 3374
Abstract
This paper proposes a solution for merging the measurements from two perpendicular profiling sonars with different beam-widths, in the context of underwater karst (cave) exploration and mapping. This work is a key step towards the development of a full 6D pose SLAM framework [...] Read more.
This paper proposes a solution for merging the measurements from two perpendicular profiling sonars with different beam-widths, in the context of underwater karst (cave) exploration and mapping. This work is a key step towards the development of a full 6D pose SLAM framework adapted to karst aquifer, where potential water turbidity disqualifies vision-based methods, hence relying on acoustic sonar measurements. Those environments have complex geometries which require 3D sensing. Wide-beam sonars are mandatory to cover previously seen surfaces but do not provide 3D measurements as the elevation angles are unknown. The approach proposed in this paper leverages the narrow-beam sonar measurements to estimate local karst surface with Gaussian process regression. The estimated surface is then further exploited to infer scaled-beta distributions of elevation angles from a wide-beam sonar. The pertinence of the method was validated through experiments on simulated environments. As a result, this approach allows one to benefit from the high coverage provided by wide-beam sonars without the drawback of loosing 3D information. Full article
(This article belongs to the Special Issue Sensors and System for Vehicle Navigation)
Show Figures

Figure 1

Figure 1
<p>Karstic environment (extracted from [<a href="#B5-sensors-20-04028" class="html-bibr">5</a>]).</p>
Full article ">Figure 2
<p>(<b>a</b>) Model of the underwater robot Telemaque, equipped for karstic exploration, and (<b>b</b>) frames definition.</p>
Full article ">Figure 3
<p>The vertical and horizontal profiling sonars mounted on Telemaque while scanning the environment’s surface <math display="inline"><semantics> <mo>Ω</mo> </semantics></math>.</p>
Full article ">Figure 4
<p>A single measure from the horizontal sonar can correspond to several 3D points and in consequence to several elevation angles.</p>
Full article ">Figure 5
<p>Elliptic cylinder fitting of vertical sonar points for the prior shape illustrated from Top (<b>a</b>) and Side (<b>b</b>) views. The RPCA results are represented in orange and the final results after Levendberg–Marquardt optimization in purple. The RPCA base is represented by a frame in the center of the cylinder.</p>
Full article ">Figure 6
<p>Illustration of elevation angle maximum likelihood estimation on a vertical slice. The green curve is the mean surface (cofounded here with the true surface for visibility) and blue curves are the confidence interval at ±3<math display="inline"><semantics> <mi>σ</mi> </semantics></math>. Note that in this particular configuration <math display="inline"><semantics> <mrow> <mi>ρ</mi> <mo>=</mo> <msup> <mi>ρ</mi> <mi>h</mi> </msup> <mo>=</mo> <msup> <mover accent="true"> <mi>ρ</mi> <mo stretchy="false">^</mo> </mover> <mi>v</mi> </msup> </mrow> </semantics></math>.</p>
Full article ">Figure 7
<p>Elevation angle estimation illustration where the vertical sonar plane is parallel to the ZY plane of the robot.</p>
Full article ">Figure 8
<p>Segment selection for the <span class="html-italic">k</span>-th horizontal sonar measurement and <span class="html-italic">q</span>-th sample <math display="inline"><semantics> <mrow> <mi mathvariant="bold">x</mi> <mo stretchy="false">(</mo> <msub> <mi>θ</mi> <mi>q</mi> </msub> <mo stretchy="false">)</mo> </mrow> </semantics></math>. We have represented the vertical sonar planes and rotation axis <math display="inline"><semantics> <msubsup> <mi mathvariant="bold">n</mi> <mrow> <mo>|</mo> <mo stretchy="false">{</mo> <mn>0</mn> <mo stretchy="false">}</mo> <mo>,</mo> <mi>i</mi> </mrow> <mi>v</mi> </msubsup> </semantics></math> along the trajectory from a top view. The blue (resp. red) vector corresponds to <math display="inline"><semantics> <mrow> <mfenced separators="" open="&#x2329;" close="&#x232A;"> <msubsup> <mi mathvariant="bold">n</mi> <mrow> <mo>|</mo> <mo stretchy="false">{</mo> <mn>0</mn> <mo stretchy="false">}</mo> <mo>,</mo> <mi>i</mi> </mrow> <mi>v</mi> </msubsup> <mo>,</mo> <mi mathvariant="bold">x</mi> <mrow> <mo stretchy="false">(</mo> <msub> <mi>θ</mi> <mi>q</mi> </msub> <mo stretchy="false">)</mo> </mrow> <mo>−</mo> <msub> <mi mathvariant="bold-italic">τ</mi> <mi>i</mi> </msub> </mfenced> <msubsup> <mi mathvariant="bold">n</mi> <mrow> <mo>|</mo> <mo stretchy="false">{</mo> <mn>0</mn> <mo stretchy="false">}</mo> <mo>,</mo> <mi>i</mi> </mrow> <mi>v</mi> </msubsup> </mrow> </semantics></math> (resp. <math display="inline"><semantics> <mrow> <mfenced separators="" open="&#x2329;" close="&#x232A;"> <msubsup> <mi mathvariant="bold">n</mi> <mrow> <mo>|</mo> <mo stretchy="false">{</mo> <mn>0</mn> <mo stretchy="false">}</mo> <mo>,</mo> <mi>i</mi> <mo>+</mo> <mn>1</mn> </mrow> <mi>v</mi> </msubsup> <mo>,</mo> <mi mathvariant="bold">x</mi> <mrow> <mo stretchy="false">(</mo> <msub> <mi>θ</mi> <mi>q</mi> </msub> <mo stretchy="false">)</mo> </mrow> <mo>−</mo> <msub> <mi mathvariant="bold-italic">τ</mi> <mrow> <mi>i</mi> <mo>+</mo> <mn>1</mn> </mrow> </msub> </mfenced> <msubsup> <mi mathvariant="bold">n</mi> <mrow> <mo>|</mo> <mo stretchy="false">{</mo> <mn>0</mn> <mo stretchy="false">}</mo> <mo>,</mo> <mi>i</mi> <mo>+</mo> <mn>1</mn> </mrow> <mi>v</mi> </msubsup> </mrow> </semantics></math>).</p>
Full article ">Figure 9
<p>Thresholds <math display="inline"><semantics> <msub> <mi>σ</mi> <mi>l</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>σ</mi> <mi>m</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>σ</mi> <mi>T</mi> </msub> </semantics></math> relative to <math display="inline"><semantics> <mover accent="true"> <mi>θ</mi> <mo>¯</mo> </mover> </semantics></math> with <math display="inline"><semantics> <mrow> <mi>b</mi> <mo>=</mo> <mn>0.61</mn> </mrow> </semantics></math> radians (35 deg).</p>
Full article ">Figure 10
<p>Uncertainty estimation. (<b>a</b>) Corresponding horizontal arc measurement with colors related to the beta pdf values. The top (resp. bottom) distribution correspond to the blue (resp. green) curve in the figure below. (<b>b</b>) Graphics for <math display="inline"><semantics> <mi>θ</mi> </semantics></math> likelihood (top) and inverse fisher information (bottom). (<b>c</b>) Beta distributions corresponding to the two local maxima.</p>
Full article ">Figure 11
<p>Models used in our simulations. Note that for generating sparse measurements, we used a scaled down version of the smooth karst model. (<b>a</b>) Outside view of the smooth karst model. (<b>b</b>) Inside view of the smooth karst model. (<b>c</b>) Outside view of the chaotic karst model. (<b>d</b>) Inside view of the chaotic karst model.</p>
Full article ">Figure 12
<p>Some 3D views of the resulting GP regressions. The first two lines correspond to the sparse case and the last one to the dense case. The mean surface is represented in green and the lower/upper bound at +/− 3<math display="inline"><semantics> <mi>σ</mi> </semantics></math> is in dark blue/blue. Red points correspond to the simulated vertical sonar measurements.</p>
Full article ">Figure 13
<p>Contours of estimated surface with <math display="inline"><semantics> <mrow> <msub> <mi>K</mi> <mi>ψ</mi> </msub> <mo>=</mo> <msubsup> <mi>K</mi> <mfrac> <mn>5</mn> <mn>2</mn> </mfrac> <mi>M</mi> </msubsup> <mo>∘</mo> <msub> <mi>d</mi> <mrow> <mi>c</mi> <mi>h</mi> <mi>o</mi> <mi>r</mi> <mi>d</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> </mrow> </semantics></math> and several <math display="inline"><semantics> <msub> <mi>K</mi> <mi>s</mi> </msub> </semantics></math> with different lengthscale <math display="inline"><semantics> <msub> <mi>l</mi> <mi>ψ</mi> </msub> </semantics></math>.</p>
Full article ">Figure 14
<p>Contours of estimated surface with fixed <math display="inline"><semantics> <mrow> <msub> <mi>K</mi> <mi>s</mi> </msub> <mo>=</mo> <msubsup> <mi>K</mi> <mfrac> <mn>5</mn> <mn>2</mn> </mfrac> <mi>M</mi> </msubsup> </mrow> </semantics></math>.</p>
Full article ">Figure 14 Cont.
<p>Contours of estimated surface with fixed <math display="inline"><semantics> <mrow> <msub> <mi>K</mi> <mi>s</mi> </msub> <mo>=</mo> <msubsup> <mi>K</mi> <mfrac> <mn>5</mn> <mn>2</mn> </mfrac> <mi>M</mi> </msubsup> </mrow> </semantics></math>.</p>
Full article ">Figure 15
<p>Surface estimation for a chaotic environment.</p>
Full article ">Figure 16
<p>Surface estimation with a random trajectory.</p>
Full article ">Figure 17
<p>Distributions obtained in the experiment with <math display="inline"><semantics> <mrow> <msub> <mi>K</mi> <mi>s</mi> </msub> <mo>=</mo> <msubsup> <mi>K</mi> <mfrac> <mn>5</mn> <mn>2</mn> </mfrac> <mi>M</mi> </msubsup> </mrow> </semantics></math>. For better visibility, uniform distribution are not displayed.</p>
Full article ">Figure 18
<p>Diagrams of error distributions. The box-plot center box represents the interquartile range, IQR = <math display="inline"><semantics> <mrow> <mi>Q</mi> <mn>1</mn> <mo>−</mo> <mi>Q</mi> <mn>3</mn> </mrow> </semantics></math>. The filled (resp. dot) line corresponds to the median (resp. mean). The upper (resp. lower) whisker corresponds to the largest (resp. lowest) value at a distance below (resp. above) 1.5 times the IQR. Values above and below the whiskers are outliers drawn as circles.</p>
Full article ">
25 pages, 5282 KiB  
Article
A Self-triggered Position Based Visual Servoing Model Predictive Control Scheme for Underwater Robotic Vehicles
by Shahab Heshmati-alamdari, Alina Eqtami, George C. Karras, Dimos V. Dimarogonas and Kostas J. Kyriakopoulos
Machines 2020, 8(2), 33; https://doi.org/10.3390/machines8020033 - 11 Jun 2020
Cited by 29 | Viewed by 3906
Abstract
An efficient position based visual sevroing control approach for Autonomous Underwater Vehicles (AUVs) by employing Non-linear Model Predictive Control (N-MPC) is designed and presented in this work. In the proposed scheme, a mechanism is incorporated within the vision-based controller that determines when the [...] Read more.
An efficient position based visual sevroing control approach for Autonomous Underwater Vehicles (AUVs) by employing Non-linear Model Predictive Control (N-MPC) is designed and presented in this work. In the proposed scheme, a mechanism is incorporated within the vision-based controller that determines when the Visual Tracking Algorithm (VTA) should be activated and new control inputs should be calculated. More specifically, the control loop does not close periodically, i.e., between two consecutive activations (triggering instants), the control inputs calculated by the N-MPC at the previous triggering time instant are applied to the underwater robot in an open-loop mode. This results in a significantly smaller number of requested measurements from the vision tracking algorithm, as well as less frequent computations of the non-linear predictive control law. This results in a reduction in processing time as well as energy consumption and, therefore, increases the accuracy and autonomy of the Autonomous Underwater Vehicle. The latter is of paramount importance for persistent underwater inspection tasks. Moreover, the Field of View constraints (FoV), control input saturation, the kinematic limitations due to the underactuated degree of freedom in sway direction, and the effect of the model uncertainties as well as external disturbances have been considered during the control design. In addition, the stability and convergence of the closed-loop system has been guaranteed analytically. Finally, the efficiency and performance of the proposed vision-based control framework is demonstrated through a comparative real-time experimental study while using a small underwater vehicle. Full article
(This article belongs to the Special Issue Intelligent Mechatronics Systems)
Show Figures

Figure 1

Figure 1
<p>The classic periodic time-triggered framework is depicted in the top block diagram. The bottom diagram represents the self-triggered control.</p>
Full article ">Figure 2
<p>Navigation and stabilization in front of a visual target while maintaining the visual target within the camera’s Field of View (FoV), ©2014 IEEE [<a href="#B58-machines-08-00033" class="html-bibr">58</a>].</p>
Full article ">Figure 3
<p>System coordination. The under-actuated, as well as the actuated, degrees of freedom are indicated with red and green color, respectively, ©2014 IEEE [<a href="#B58-machines-08-00033" class="html-bibr">58</a>].</p>
Full article ">Figure 4
<p>Visibility constraints formulation Equations (<a href="#FD8b-machines-08-00033" class="html-disp-formula">8b</a>)–(<a href="#FD8e-machines-08-00033" class="html-disp-formula">8e</a>) and modeling of the external disturbance Equation (<a href="#FD5-machines-08-00033" class="html-disp-formula">5</a>), ©2014 IEEE [<a href="#B58-machines-08-00033" class="html-bibr">58</a>].</p>
Full article ">Figure 5
<p>Experimental setup. The underwater robot at the initial and desired configuration with respect to a visual marker. Vehicle’s view at initial and the desired position, respectively, ©2014 IEEE [<a href="#B58-machines-08-00033" class="html-bibr">58</a>].</p>
Full article ">Figure 6
<p>The evaluation of the underwater robot coordinates regarding the visual target. (<b>Left</b>) Proposed self-Triggered N-MPC. (<b>Right</b>) Classic N-MPC, ©2014 IEEE [<a href="#B58-machines-08-00033" class="html-bibr">58</a>].</p>
Full article ">Figure 7
<p>Camera view during the experiment. From initial view (top and left) to the final view (bottom and right). The target remains within the field of view of the camera.</p>
Full article ">Figure 8
<p>Image coordinates of the visual target center during the experiment.</p>
Full article ">Figure 9
<p>The triggering instants in Self triggered NMPC, ©2014 IEEE [<a href="#B58-machines-08-00033" class="html-bibr">58</a>].</p>
Full article ">Figure 10
<p>Control inputs. (<b>Left</b>) Proposed self-Triggered N-MPC. (<b>Right</b>) Classic N-MPC , ©2014 IEEE [<a href="#B58-machines-08-00033" class="html-bibr">58</a>].</p>
Full article ">
28 pages, 9973 KiB  
Article
Data Gathering from a Multimodal Dense Underwater Acoustic Sensor Network Deployed in Shallow Fresh Water Scenarios
by Alberto Signori, Filippo Campagnaro, Fabian Steinmetz, Bernd-Christian Renner and Michele Zorzi
J. Sens. Actuator Netw. 2019, 8(4), 55; https://doi.org/10.3390/jsan8040055 - 30 Nov 2019
Cited by 24 | Viewed by 7202
Abstract
The Robotic Vessels as-a-Service (RoboVaaS) project intends to exploit the most advanced communication and marine vehicle technologies to revolutionize shipping and near-shore operations, offering on-demand and cost-effective robotic-aided services. In particular, the RoboVaaS vision includes a ship hull inspection service, a quay walls [...] Read more.
The Robotic Vessels as-a-Service (RoboVaaS) project intends to exploit the most advanced communication and marine vehicle technologies to revolutionize shipping and near-shore operations, offering on-demand and cost-effective robotic-aided services. In particular, the RoboVaaS vision includes a ship hull inspection service, a quay walls inspection service, an antigrounding service, and an environmental and bathymetry data collection service. In this paper, we present a study of the underwater environmental data collection service, performed by a low-cost autonomous vehicle equipped with both a commercial modem and a very low-cost acoustic modem prototype, the smartPORT Acoustic Underwater Modem (AHOI). The vehicle mules the data from a network of low cost submerged acoustic sensor nodes to a surface sink. To this end, an underwater acoustic network composed by both static and moving nodes has been implemented and simulated with the DESERT Underwater Framework, where the performance of the AHOI modem has been mapped in the form of lookup tables. The performance of the AHOI modem has been measured near the Port of Hamburg, where the RoboVaaS concept will be demonstrated with a real field evaluation. The transmission with the commercial modem, instead, has been simulated with the Bellhop ray tracer thanks to the World Ocean Simulation System (WOSS), by considering both the bathymetry and the sound speed profile of the Port of Hamburg. The set up of the polling-based MAC protocol parameters, such as the maximum backoff time of the sensor nodes, appears to be crucial for the network performance, in particular for the low-cost low-rate modems. In this work, to tune the maximum backoff time during the data collection mission, an adaptive mechanism has been implemented. Specifically, the maximum backoff time is updated based on the network density. This adaptive mechanism results in an 8% improvement of the network throughput. Full article
(This article belongs to the Section Network Services and Applications)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Robotic Vessels-as-a-Service (RoboVaaS) envisioned example scenario showing ship-hull inspection and anti-grounding services enabled through a fleet of autonomous vessels connected with acoustic underwater and surface radio communication.</p>
Full article ">Figure 2
<p>RoboVaaS underwater data collection service, where both an autonomous surface vessel (ASV) and an autonomous underwater vehicle (AUV) collect data from static underwater sensor nodes. The figure also portrays the inclusion of the control center and cloud integration.</p>
Full article ">Figure 3
<p>AHOI acoustic underwater modem with an AS-1 hydrophone (and 1-euro coin).</p>
Full article ">Figure 4
<p>Nodes during the field test with five receivers (<math display="inline"><semantics> <mi mathvariant="script">A</mi> </semantics></math> to <math display="inline"><semantics> <mi mathvariant="script">E</mi> </semantics></math>) and a single transmitter (<math display="inline"><semantics> <mi mathvariant="script">TX</mi> </semantics></math>).</p>
Full article ">Figure 5
<p>Power spectral densities (PSDs) of the simulated modem signals and additional shipping noise. The PSDs correspond to received signals (packets or noise) with <math display="inline"><semantics> <mrow> <msub> <mi>d</mi> <mi mathvariant="normal">M</mi> </msub> <mo>,</mo> <msub> <mi>d</mi> <mrow> <mi>ship</mi> </mrow> </msub> <mo>∈</mo> <mfenced separators="" open="{" close="}"> <mn>25</mn> <mo> </mo> <mi mathvariant="normal">m</mi> <mo>,</mo> <mo> </mo> <mn>50</mn> <mo> </mo> <mi mathvariant="normal">m</mi> <mo>,</mo> <mo> </mo> <mn>75</mn> <mo> </mo> <mi mathvariant="normal">m</mi> <mo>,</mo> <mo> </mo> <mn>100</mn> <mo> </mo> <mi mathvariant="normal">m</mi> </mfenced> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>d</mi> <mrow> <mi>AUV</mi> </mrow> </msub> <mo>∈</mo> <mfenced separators="" open="{" close="}"> <mn>1</mn> <mo> </mo> <mi mathvariant="normal">m</mi> <mo>,</mo> <mo> </mo> <mn>2.5</mn> <mo> </mo> <mi mathvariant="normal">m</mi> <mo>,</mo> <mo> </mo> <mn>5</mn> <mo> </mo> <mi mathvariant="normal">m</mi> <mo>,</mo> <mo> </mo> <mn>10</mn> <mo> </mo> <mi mathvariant="normal">m</mi> </mfenced> </mrow> </semantics></math> distance to the transmitter or noise source.</p>
Full article ">Figure 6
<p>Packet delivery ratio vs. Signal to Interference Ratio of the AHOI modem.</p>
Full article ">Figure 7
<p>From left to right: state machine of the UW-POLLING protocol for an AUV (<b>a</b>), for a node (<b>b</b>) and for a sink (<b>c</b>), respectively.</p>
Full article ">Figure 8
<p>Protocol stacks used by the sensor nodes (<b>a</b>) and the AUV (<b>b</b>), respectively, during the single mode scenario with the EvoLogics S2CM HS modem.</p>
Full article ">Figure 9
<p>Examples of scenarios with the high speed modem: node deployment with a fixed node density <math display="inline"><semantics> <mi>λ</mi> </semantics></math>= 100 nodes/km<sup>2</sup> (<b>a</b>) and node deployment with variable <math display="inline"><semantics> <mi>λ</mi> </semantics></math> ranging from 10 to 200 nodes/km<sup>2</sup> (<b>b</b>).</p>
Full article ">Figure 10
<p>Simulation results in the first scenario with high speed modem: overall throughput as a function of the maximum backoff time for different values of <math display="inline"><semantics> <mi>λ</mi> </semantics></math> (<b>a</b>), throughput as a function of <math display="inline"><semantics> <mi>λ</mi> </semantics></math> comparing adaptive backoff approaches and fixed backoff case (<b>b</b>).</p>
Full article ">Figure 11
<p>Throughput in the variable density scenario with the adaptive backoff approaches (Full knowledge case and realistic case (FC and RC) and the fixed backoff case (AVG) (<b>a</b>). Optimal backoff time as a function of the network density obtained via simulation considering the first scenario with fixed node density (<b>b</b>).</p>
Full article ">Figure 12
<p>Protocol stacks used by the sensor nodes (<b>a</b>) and the AUV (<b>b</b>), respectively, during the single mode scenario with the AHOI modem.</p>
Full article ">Figure 13
<p>Examples of the two scenarios considered in the low rate modem simulations: node deployment with a fixed node density <math display="inline"><semantics> <mi>λ</mi> </semantics></math> = 300 nodes/km<sup>2</sup> (<b>a</b>) and node deployment with variable <math display="inline"><semantics> <mi>λ</mi> </semantics></math> ranging from 50 to 400 nodes/km<sup>2</sup> (<b>b</b>).</p>
Full article ">Figure 14
<p>Simulation results in the first scenario with low rate modem: overall throughput as a function of the maximum backoff time for different values of <math display="inline"><semantics> <mi>λ</mi> </semantics></math> (<b>a</b>), throughput as a function of <math display="inline"><semantics> <mi>λ</mi> </semantics></math> comparing adaptive backoff approaches and fixed backoff case (<b>b</b>).</p>
Full article ">Figure 15
<p>Throughput in the variable density scenario with the adaptive backoff approaches (FC and RC) and the fixed backoff case (AVG).</p>
Full article ">Figure 16
<p>Protocol stack used by the sensor nodes (<b>a</b>), the AUV (<b>b</b>) and the sink (<b>c</b>), during the complete multimodal scenario, where the AUV delivers the collected data to the sink.</p>
Full article ">Figure 17
<p>Node deployment along the Elbe river in the port of Hamburg (<b>a</b>). Three clusters of nodes are identified with the letters A, B and C. (<b>b</b>) represents the port of Hamburg bathymetry related to the zone depicted in (<b>a</b>).</p>
Full article ">Figure 18
<p>Overall throughput in the port of Hamburg scenario as a function of the offered traffic.</p>
Full article ">Figure 19
<p>Jain’s Fairness Index for the whole network (<b>a</b>), cluster A (<b>b</b>), cluster B (<b>c</b>) and cluster C (<b>d</b>).</p>
Full article ">
Back to TopTop