Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (634)

Search Parameters:
Keywords = multi-target resolution

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
26 pages, 3046 KiB  
Review
Polymerase Chain Reaction Chips for Biomarker Discovery and Validation in Drug Development
by Dang-Khoa Vo and Kieu The Loan Trinh
Micromachines 2025, 16(3), 243; https://doi.org/10.3390/mi16030243 - 20 Feb 2025
Abstract
Polymerase chain reaction (PCR) chips are advanced, microfluidic platforms that have revolutionized biomarker discovery and validation because of their high sensitivity, specificity, and throughput levels. These chips miniaturize traditional PCR processes for the speed and precision of nucleic acid biomarker detection relevant to [...] Read more.
Polymerase chain reaction (PCR) chips are advanced, microfluidic platforms that have revolutionized biomarker discovery and validation because of their high sensitivity, specificity, and throughput levels. These chips miniaturize traditional PCR processes for the speed and precision of nucleic acid biomarker detection relevant to advancing drug development. Biomarkers, which are useful in helping to explain disease mechanisms, patient stratification, and therapeutic monitoring, are hard to identify and validate due to the complexity of biological systems and the limitations of traditional techniques. The challenges to which PCR chips respond include high-throughput capabilities coupled with real-time quantitative analysis, enabling researchers to identify novel biomarkers with greater accuracy and reproducibility. More recent design improvements of PCR chips have further expanded their functionality to also include digital and multiplex PCR technologies. Digital PCR chips are ideal for quantifying rare biomarkers, which is essential in oncology and infectious disease research. In contrast, multiplex PCR chips enable simultaneous analysis of multiple targets, therefore simplifying biomarker validation. Furthermore, single-cell PCR chips have made it possible to detect biomarkers at unprecedented resolution, hence revealing heterogeneity within cell populations. PCR chips are transforming drug development, enabling target identification, patient stratification, and therapeutic efficacy assessment. They play a major role in the development of companion diagnostics and, therefore, pave the way for personalized medicine, ensuring that the right patient receives the right treatment. While this tremendously promising technology has exhibited many challenges regarding its scalability, integration with other omics technologies, and conformity with regulatory requirements, many still prevail. Future breakthroughs in chip manufacturing, the integration of artificial intelligence, and multi-omics applications will further expand PCR chip capabilities. PCR chips will not only be important for the acceleration of drug discovery and development but also in raising the bar in improving patient outcomes and, hence, global health care as these technologies continue to mature. Full article
(This article belongs to the Special Issue PCR Chips for Biomarker Discovery and Validation in Drug Development)
Show Figures

Figure 1

Figure 1
<p>An example of a microchamber array digital PCR chip using silicon–glass material for SARS-CoV-2 virus and ultra-early-stage lung cancer marker quantitative detection. MEMS: microelectromechanical systems. Copyright ACS publisher (2012) [<a href="#B43-micromachines-16-00243" class="html-bibr">43</a>].</p>
Full article ">Figure 2
<p>An overall schematic of assembly and operation of a nanoliter PCR SlipChip for <span class="html-italic">Staphylococcus aureus</span> detection using SlipChip. Copyright ACS (2010) [<a href="#B66-micromachines-16-00243" class="html-bibr">66</a>].</p>
Full article ">Figure 3
<p>An overall illustration schematic of circulating tumor DNA methylation markers offers insights into early detection, prognosis, minimal residual disease, and therapeutic response. The primary analytical methodologies are founded either on PCR after sodium bisulfite (SB) conversion or on a comprehensive omic approach. For a clearer view of the figure details, please refer to the original source. Copyright Wiley (2021) [<a href="#B87-micromachines-16-00243" class="html-bibr">87</a>].</p>
Full article ">Figure 4
<p>Schematic diagram of the HIV integrase-specific DNA biosensing platform based on the rolling circle amplification (RCA) technique for multiple amplifications for HIV detection. LTR: long terminal repeat sequences. Copyright ACS (2024) [<a href="#B140-micromachines-16-00243" class="html-bibr">140</a>].</p>
Full article ">Figure 5
<p>An overall schematic representation of biomarker detection and identification platforms by combining point-of-care microdevice systems and intelligent technologies. Copyright Wiley (2024) [<a href="#B145-micromachines-16-00243" class="html-bibr">145</a>].</p>
Full article ">Figure 6
<p>An example of smartphone-based platforms integrating microfluidic detection with image-based artificial intelligence in point-of-care testing applications. Copyright <span class="html-italic">Nature</span> (2023) [<a href="#B216-micromachines-16-00243" class="html-bibr">216</a>].</p>
Full article ">
32 pages, 13857 KiB  
Article
SPDC-YOLO: An Efficient Small Target Detection Network Based on Improved YOLOv8 for Drone Aerial Image
by Jingxin Bi, Keda Li, Xiangyue Zheng, Gang Zhang and Tao Lei
Remote Sens. 2025, 17(4), 685; https://doi.org/10.3390/rs17040685 - 17 Feb 2025
Abstract
Target detection in UAV images is of great significance in fields such as traffic safety, emergency rescue, and environmental monitoring. However, images captured by UAVs usually have multi-scale features, complex backgrounds, uneven illumination, and low target resolution, which makes target detection in UAV [...] Read more.
Target detection in UAV images is of great significance in fields such as traffic safety, emergency rescue, and environmental monitoring. However, images captured by UAVs usually have multi-scale features, complex backgrounds, uneven illumination, and low target resolution, which makes target detection in UAV images very challenging. To tackle these challenges, this paper introduces SPDC-YOLO, a novel model built upon YOLOv8. In the backbone, the model eliminates the last C2f module and the final downsampling module, thus avoiding the loss of small target features. In the neck, this paper proposes a novel feature pyramid, SPC-FPN, which employs the SBA (Selective Boundary Aggregation) module to fuse features from two distinct scales. In the head, the P5 detection head is eliminated, and a new detection head, Dyhead-DCNv4, is proposed, replacing DCNv2 in the original Dyhead with DCNv4 and utilizing three attention mechanisms for dynamic feature weighting. In addition, the model uses the CGB (Context Guided Block) module for downsampling, which can learn and fuse local features with surrounding contextual information, and the PPA (Parallelized Patch-Aware Attention) module replacing the original C2f module to further improve feature expression capability. Finally, SPDC-YOLO adopts EIoU as the loss function to optimize target localization accuracy. On the public dataset VisDrone2019, the experimental results show that SPDC-YOLO improves mAP50 by 3.4% compared to YOLOv8n while reducing the parameters count by 1.03 M. Compared with other related methods, SPDC-YOLO demonstrates better performance. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Figure 1

Figure 1
<p>YOLOv8 structure diagram.</p>
Full article ">Figure 2
<p>SPDC-YOLO structure diagram and its differences from Yolov8.</p>
Full article ">Figure 3
<p>SPC-FPN structure diagram.</p>
Full article ">Figure 4
<p>SBA module structure diagram.</p>
Full article ">Figure 5
<p>PPA module structure diagram.</p>
Full article ">Figure 6
<p>Context Guided Block flowchart.</p>
Full article ">Figure 7
<p>Dyhead Structure Diagram.</p>
Full article ">Figure 8
<p>EIoU Calculation Diagram.</p>
Full article ">Figure 9
<p>VisDrone2019 data distribution.</p>
Full article ">Figure 10
<p>Typical images from the VisDrone2019 dataset.</p>
Full article ">Figure 11
<p>Comparison of different models in the validation set for each category mAP<sub>50</sub>.</p>
Full article ">Figure 12
<p>Comparison of different models in the test set for each category mAP<sub>50</sub>.</p>
Full article ">Figure 13
<p>Aerial images taken at a tilted angle at the intersection in the evening.</p>
Full article ">Figure 14
<p>Road images taken under strong light.</p>
Full article ">Figure 15
<p>Images captured by drones over the road at night.</p>
Full article ">Figure 16
<p>Comparison of heat maps before and after adding PPA modules.</p>
Full article ">Figure 17
<p>Comparison of heat maps.</p>
Full article ">
35 pages, 37221 KiB  
Article
Target Ship Recognition and Tracking with Data Fusion Based on Bi-YOLO and OC-SORT Algorithms for Enhancing Ship Navigation Assistance
by Shuai Chen, Miao Gao, Peiru Shi, Xi Zeng and Anmin Zhang
J. Mar. Sci. Eng. 2025, 13(2), 366; https://doi.org/10.3390/jmse13020366 - 16 Feb 2025
Abstract
With the ever-increasing volume of maritime traffic, the risks of ship navigation are becoming more significant, making the use of advanced multi-source perception strategies and AI technologies indispensable for obtaining information about ship navigation status. In this paper, first, the ship tracking system [...] Read more.
With the ever-increasing volume of maritime traffic, the risks of ship navigation are becoming more significant, making the use of advanced multi-source perception strategies and AI technologies indispensable for obtaining information about ship navigation status. In this paper, first, the ship tracking system was optimized using the Bi-YOLO network based on the C2f_BiFormer module and the OC-SORT algorithms. Second, to extract the visual trajectory of the target ship without a reference object, an absolute position estimation method based on binocular stereo vision attitude information was proposed. Then, a perception data fusion framework based on ship spatio-temporal trajectory features (ST-TF) was proposed to match GPS-based ship information with corresponding visual target information. Finally, AR technology was integrated to fuse multi-source perceptual information into the real-world navigation view. Experimental results demonstrate that the proposed method achieves a mAP0.5:0.95 of 79.6% under challenging scenarios such as low resolution, noise interference, and low-light conditions. Moreover, in the presence of the nonlinear motion of the own ship, the average relative position error of target ship visual measurements is maintained below 8%, achieving accurate absolute position estimation without reference objects. Compared to existing navigation assistance, the AR-based navigation assistance system, which utilizes ship ST-TF-based perception data fusion mechanism, enhances ship traffic situational awareness and provides reliable decision-making support to further ensure the safety of ship navigation. Full article
Show Figures

Figure 1

Figure 1
<p>Organization diagram of the sections of this paper.</p>
Full article ">Figure 2
<p>A perception data fusion framework based on ship ST-TF for ship AR navigation assistance.</p>
Full article ">Figure 3
<p>The structure of the Bi-YOLO network.</p>
Full article ">Figure 4
<p>(<b>a</b>) Details of a BiFormer block; (<b>b</b>) Structure of the BRA.</p>
Full article ">Figure 5
<p>(<b>a</b>) to (<b>b</b>) illustrate the Driving-Leaves binocular camera before and after calibration and stereo rectification, and (<b>c</b>) to (<b>d</b>) illustrate the Baymax binocular camera before and after calibration and stereo rectification.</p>
Full article ">Figure 6
<p>Conceptual diagram of the binocular imaging process.</p>
Full article ">Figure 7
<p>Illustration of coordinate system conversion.</p>
Full article ">Figure 8
<p>Synchronization process of different sensor frequencies.</p>
Full article ">Figure 9
<p>Asynchronous nonlinear ship trajectory sequence association based on the DTW algorithm.</p>
Full article ">Figure 10
<p>Asynchronous ship trajectory association and joint data storage method.</p>
Full article ">Figure 11
<p>The MASSs used in the experimental process.</p>
Full article ">Figure 12
<p>The data samples from the FLShip dataset.</p>
Full article ">Figure 13
<p>Training mAP@0.5 curves for Bi-YOLO and various object detection algorithms.</p>
Full article ">Figure 14
<p>(<b>a</b>–<b>f</b>) respectively show the comparison of detection effects between YOLO11s and Bi-YOLO.</p>
Full article ">Figure 15
<p>Tracking performance comparison of four state-of-the-art object trackers in Scene-2.</p>
Full article ">Figure 16
<p>Tracking performance comparison of four state-of-the-art object trackers in Scene-4.</p>
Full article ">Figure 17
<p>Tracking performance comparison of four state-of-the-art object trackers in Scene-6.</p>
Full article ">Figure 18
<p>The visual position estimation results of the ‘Roaring-Flame’ MASS in Scene-1.</p>
Full article ">Figure 19
<p>The visual position estimation result of the ‘Baymax’ MASS in scene-2.</p>
Full article ">Figure 20
<p>The visual position estimation results of the ‘Baymax’ MASS in Scene-3.</p>
Full article ">Figure 21
<p>The AR navigation assistance effects of ships constructed at different timestamps in multiple scenes.</p>
Full article ">
21 pages, 14440 KiB  
Article
Spectral Super-Resolution Technology Based on Fabry–Perot Interferometer for Temporally and Spatially Modulated Fourier Transform Imaging Spectrometer
by Yu Zhang, Qunbo Lv, Jianwei Wang, Yinhui Tang, Jia Si, Xinwen Chen and Yangyang Liu
Sensors 2025, 25(4), 1201; https://doi.org/10.3390/s25041201 - 16 Feb 2025
Abstract
A new spectral super-resolution technique was proposed by combining the Fabry–Perot interferometer (FPI) with Temporally and Spatially Modulated Fourier Transform Imaging Spectrometer (TSMFTIS). This study uses the multi-beam interference of the FPI to modulate the target spectrum periodically, and it acquires the modulated [...] Read more.
A new spectral super-resolution technique was proposed by combining the Fabry–Perot interferometer (FPI) with Temporally and Spatially Modulated Fourier Transform Imaging Spectrometer (TSMFTIS). This study uses the multi-beam interference of the FPI to modulate the target spectrum periodically, and it acquires the modulated interferogram through TSMFTIS. The combined interference of the two techniques overcomes the limitations of the maximum optical path difference (OPD) on spectral resolution. FPI is used to encode high-frequency interference information into low-frequency interference information, proposing an inversion algorithm to recover high-frequency information, studying the impact of FPI optical defects on the system, and proposing targeted improvement algorithms. The simulation results indicate that this method can achieve multi-component joint interference imaging, improving spectral resolution by twofold. This technology offers advantages such as high throughput, stability, simple and compact structure, straightforward principles, high robustness, and low cost. It provides new insights into TSMFTIS spectral super-resolution research. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>FPI principle schematic diagram.</p>
Full article ">Figure 2
<p>Schematic diagram of system structure.</p>
Full article ">Figure 3
<p>Schematic diagram of teh TSMFTIS interferogram scanning method: (<b>a</b>–<b>c</b>) interference images captured by TSMFTIS at times <math display="inline"><semantics> <msub> <mi>t</mi> <mn>1</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>t</mi> <mi>i</mi> </msub> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>t</mi> <mi>j</mi> </msub> </semantics></math>, where the position of a specific target point in the image is <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>x</mi> <mrow> <mo>(</mo> <msub> <mi>t</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>x</mi> <mrow> <mo>(</mo> <msub> <mi>t</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </semantics></math>, and <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>x</mi> <mrow> <mo>(</mo> <msub> <mi>t</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </semantics></math>, respectively; (<b>d</b>) pushing the scan along the X direction, interference data are assembled based on the actual spatial positions and time sequence; (<b>e</b>) interference patterns formed by the Y dimension of space and the T dimension of time; (<b>f</b>) The interferogram of a specific target point formed by assembling data in time sequence.</p>
Full article ">Figure 4
<p>(<b>a</b>) Optical equivalent model of LSI; (<b>b</b>) Optical equivalent model of FPI; (<b>c</b>) optical equivalent model of SMJI between the LSI and FPI.</p>
Full article ">Figure 5
<p>Schematic diagram of the spectrum mixing principle. (<b>a</b>) Object spectrum and (<b>b</b>) its interferogram; (<b>c</b>) cosine-modulated spectrum and (<b>d</b>) its interferogram.</p>
Full article ">Figure 6
<p>(<b>a</b>) Spectrum after FPI modulation; (<b>b</b>) schematic diagram of its interferogram.</p>
Full article ">Figure 7
<p>(<b>a</b>) Overall flowchart of the restoration algorithm; (<b>b</b>) flowchart of the interpolation algorithm.</p>
Full article ">Figure 8
<p>Input spectrum: (<b>a</b>) Gaussian function; (<b>b</b>) spectral data.</p>
Full article ">Figure 9
<p>Interferograms <math display="inline"><semantics> <msub> <mi>I</mi> <mn>1</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>I</mi> <mn>2</mn> </msub> </semantics></math>.</p>
Full article ">Figure 10
<p>Interferogram of Gaussian input spectrum: (<b>a</b>) overall restoration result; (<b>b</b>) partial magnification of <span class="html-italic">L</span>∼<math display="inline"><semantics> <mrow> <mn>2</mn> <mi>L</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 11
<p>(<b>a</b>) Overall super-resolution spectral result of the Gaussian input spectrum; (<b>b</b>) partial magnification and the fitting and <span class="html-italic">FWHM</span> calculation of <math display="inline"><semantics> <msub> <mi>B</mi> <mrow> <mi>s</mi> <mi>u</mi> <mi>p</mi> </mrow> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>B</mi> <mrow> <mi>c</mi> <mi>l</mi> <mi>a</mi> <mi>s</mi> </mrow> </msub> </semantics></math>.</p>
Full article ">Figure 12
<p>Interferogram of spectral data input spectrum: (<b>a</b>) overall restoration result; (<b>b</b>) partial magnification of <span class="html-italic">L</span>∼<math display="inline"><semantics> <mrow> <mn>2</mn> <mi>L</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 13
<p>(<b>a</b>) Overall super-resolution spectral result of spectral data input spectrum; (<b>b</b>–<b>d</b>) partial magnifications.</p>
Full article ">Figure 14
<p>(<b>a</b>) Super-resolution interferogram and (<b>b</b>) super-resolution spectrum for each row of the field of view and its partial magnification.</p>
Full article ">Figure 15
<p>Frequency shift broadening caused by different degrees of non-parallelism for (<b>a</b>) <math display="inline"><semantics> <mrow> <mi>m</mi> <mo>=</mo> <mn>0</mn> </mrow> </semantics></math>, (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>m</mi> <mo>=</mo> <mn>1</mn> </mrow> </semantics></math>, (<b>c</b>) <math display="inline"><semantics> <mrow> <mi>m</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math>, (<b>d</b>) <math display="inline"><semantics> <mrow> <mi>m</mi> <mo>=</mo> <mn>3</mn> </mrow> </semantics></math>; frequency shift broadening caused by different degrees of non-flatness for (<b>e</b>) <math display="inline"><semantics> <mrow> <mi>m</mi> <mo>=</mo> <mn>0</mn> </mrow> </semantics></math>, (<b>f</b>) <math display="inline"><semantics> <mrow> <mi>m</mi> <mo>=</mo> <mn>1</mn> </mrow> </semantics></math>, (<b>g</b>) <math display="inline"><semantics> <mrow> <mi>m</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math>, (<b>h</b>) <math display="inline"><semantics> <mrow> <mi>m</mi> <mo>=</mo> <mn>3</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 16
<p>Interferogram restoration results for Defect1 with <span class="html-italic">FWH</span><math display="inline"><semantics> <msub> <mi>M</mi> <mi>D</mi> </msub> </semantics></math> values of 0.125 <math display="inline"><semantics> <mi mathvariant="sans-serif">μ</mi> </semantics></math>m, 0.25 <math display="inline"><semantics> <mi mathvariant="sans-serif">μ</mi> </semantics></math>m, and 0.5 <math display="inline"><semantics> <mi mathvariant="sans-serif">μ</mi> </semantics></math>m.</p>
Full article ">
20 pages, 7029 KiB  
Article
Tracking of Low Radar Cross-Section Super-Sonic Objects Using Millimeter Wavelength Doppler Radar and Adaptive Digital Signal Processing
by Yair Richter, Shlomo Zach, Maxi Y. Blum, Gad A. Pinhasi and Yosef Pinhasi
Remote Sens. 2025, 17(4), 650; https://doi.org/10.3390/rs17040650 - 14 Feb 2025
Abstract
Small targets with low radar cross-section (RCS) and high velocities are very hard to track by radar as long as the frequent variations in speed and location demand shorten the integration temporal window. In this paper, we propose a technique for tracking evasive [...] Read more.
Small targets with low radar cross-section (RCS) and high velocities are very hard to track by radar as long as the frequent variations in speed and location demand shorten the integration temporal window. In this paper, we propose a technique for tracking evasive targets using a continuous wave (CW) radar array of multiple transmitters operating in the millimeter wavelength (MMW). The scheme is demonstrated to detect supersonic moving objects, such as rifle projectiles, with extremely short integration times while utilizing an adaptive processing algorithm of the received signal. Operation at extremely high frequencies qualifies spatial discrimination, leading to resolution improvement over radars operating in commonly used lower frequencies. CW transmissions result in efficient average power utilization and consumption of narrow bandwidths. It is shown that although CW radars are not naturally designed to estimate distances, the array arrangement can track the instantaneous location and velocity of even supersonic targets. Since a CW radar measures the target velocity via the Doppler frequency shift, it is resistant to the detection of undesired immovable objects in multi-scattering scenarios; thus, the tracking ability is not impaired in a stationary, cluttered environment. Using the presented radar scheme is shown to enable the processing of extremely weak signals that are reflected from objects with a low RCS. In the presented approach, the significant improvement in resolution is beneficial for the reduction in the required detection time. In addition, in relation to reducing the target recording time for processing, the presented scheme stimulates the detection and tracking of objects that make frequent changes in their velocity and position. Full article
Show Figures

Figure 1

Figure 1
<p>Doppler radar system scheme.</p>
Full article ">Figure 2
<p>Multi-transmitters—single-receiver Doppler radar system.</p>
Full article ">Figure 3
<p>A scale model for a radar system to detect velocity components.</p>
Full article ">Figure 4
<p>Velocity resolution <math display="inline"><semantics> <mrow> <mo>∆</mo> <msub> <mrow> <mi mathvariant="normal">v</mi> </mrow> <mrow> <mi mathvariant="normal">r</mi> </mrow> </msub> </mrow> </semantics></math> vs. integration time <math display="inline"><semantics> <mrow> <mi mathvariant="normal">T</mi> <mo>/</mo> <msub> <mrow> <mi mathvariant="normal">T</mi> </mrow> <mrow> <mi mathvariant="normal">o</mi> <mi mathvariant="normal">p</mi> <mi mathvariant="normal">t</mi> </mrow> </msub> </mrow> </semantics></math> as a result of the trade-off between measurement resolution improvement and frequency broadening by data overabundance.</p>
Full article ">Figure 5
<p>Resolution vs. integration time over different transmission frequencies of the radar, with the same acceleration <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>a</mi> </mrow> <mrow> <mi mathvariant="normal">r</mi> </mrow> </msub> <mo>=</mo> <mn>100</mn> <mstyle scriptlevel="0" displaystyle="true"> <mfrac> <mrow> <mi mathvariant="normal">m</mi> </mrow> <mrow> <msup> <mrow> <mi mathvariant="normal">s</mi> </mrow> <mrow> <mn>2</mn> </mrow> </msup> </mrow> </mfrac> </mstyle> </mrow> </semantics></math>.</p>
Full article ">Figure 6
<p>Locations of the systems used during the experiment. The radar system was placed alongside the bullet’s expected trajectory.</p>
Full article ">Figure 7
<p>The radar system used in the experiment; (<b>a</b>) the master unit and (<b>b</b>) the slave unit. *Multiplier: QMM-9940615060.</p>
Full article ">Figure 8
<p>Diagram of the measurement performed in the range. The red line is the illustrated trajectory of the measured object. The blue beam and the green beam are the illustrated beams of the antennas.</p>
Full article ">Figure 9
<p>Spectral-time representation of the recording from the radar. Two frequencies are obtained at any time slot and transferred to velocity by <math display="inline"><semantics> <mrow> <mi mathvariant="normal">v</mi> <mo>=</mo> <mstyle scriptlevel="0" displaystyle="true"> <mfrac> <mrow> <mo>λ</mo> </mrow> <mrow> <mn>2</mn> </mrow> </mfrac> </mstyle> <mi mathvariant="normal">f</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 10
<p>Spectral-time representation of the recording from the radar as presented in <a href="#remotesensing-17-00650-f009" class="html-fig">Figure 9</a>, with additional analysis of velocity components extracting.</p>
Full article ">Figure 11
<p>Optimal integration times for each velocity component separately over time.</p>
Full article ">Figure 12
<p>Spectrogram calculations optimized for the acceleration of the target and extraction of high-quality velocity components. The spectrogram in (<b>a</b>) shows an optimal spectrogram creation process for the upper-velocity component; compared with (<b>b</b>)<b>,</b> it provides earlier data regarding the target that moves towards the radar. The spectrogram in (<b>b</b>) shows an optimal result for the lower velocity component when the graph is narrower and, therefore, shows greater accuracy than the spectrogram from (<b>a</b>). The velocity components were extracted after generating optimal spectrograms, with the upper-velocity component in (<b>c</b>) extracted from the spectrogram in (<b>a</b>) and the velocity component in (<b>d</b>) extracted from the spectrogram in (<b>b</b>).</p>
Full article ">Figure 13
<p>Distance calculation from the radar measurements and the estimated distance from the moment of receiving the trigger at 40 ms.</p>
Full article ">
25 pages, 115458 KiB  
Article
RSAM-Seg: A SAM-Based Model with Prior Knowledge Integration for Remote Sensing Image Semantic Segmentation
by Jie Zhang, Yunxin Li, Xubing Yang, Rui Jiang and Li Zhang
Remote Sens. 2025, 17(4), 590; https://doi.org/10.3390/rs17040590 - 8 Feb 2025
Abstract
High-resolution remote sensing satellites have revolutionized remote sensing research, yet accurately segmenting specific targets from complex satellite imagery remains challenging. While the Segment Anything Model (SAM) has emerged as a promising universal segmentation model, its direct application to remote sensing imagery yields suboptimal [...] Read more.
High-resolution remote sensing satellites have revolutionized remote sensing research, yet accurately segmenting specific targets from complex satellite imagery remains challenging. While the Segment Anything Model (SAM) has emerged as a promising universal segmentation model, its direct application to remote sensing imagery yields suboptimal results. To address these limitations, we propose RSAM-Seg, a novel deep learning model adapted from SAM specifically designed for remote sensing applications. Our model incorporates two key components: Adapter-Scale and Adapter-Feature modules. The Adapter-Scale modules, integrated within Vision Transformer (ViT) blocks, enhance model adaptability through learnable transformations, while the Adapter-Feature modules, positioned between ViT blocks, generate image-informed prompts by incorporating task-specific information. Extensive experiments across four binary and two multi-class segmentation scenarios demonstrate the superior performance of RSAM-Seg, achieving an F1 score of 0.815 in cloud detection, 0.834 in building segmentation, and 0.755 in road extraction, consistently outperforming established architectures like U-Net, DeepLabV3+, and Segformer. Moreover, RSAM-Seg shows significant improvements of up to 56.5% in F1 score compared to the original SAM. In addition, RSAM-Seg maintains robust performance in few-shot learning scenarios, achieving an F1 score of 0.656 with only 1% of the training data and increasing to 0.815 with full data availability. Furthermore, RSAM-Seg exhibits the capability to detect missing areas within the ground truth of certain datasets, highlighting its capability for completion. Full article
(This article belongs to the Special Issue Advanced AI Technology for Remote Sensing Analysis)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Comparison of segmentation results on various scenarios using point and box prompt modes, as well as the segment everything mode. Green squares indicate box prompt positions, green points show point prompt coordinates, and red areas represent predicted segmentation results. In Segment Everything results, each color represents a distinct segmented instance.</p>
Full article ">Figure 2
<p>The architecture and key components of RSAM-Seg. (<b>A</b>) The overall structure of RSAM-Seg. The encoder’s ViT blocks are modified to include the internal Adapter-Scale components and interleaved Adapter-Feature layers for enhanced image information extraction (⊕ denotes feature fusion), while the mask decoder remains unchanged and promptless. Each stage’s dimensional changes are marked in the figure, where ‘B’ represents batch size. (<b>B</b>) The structure of the modified transformer block and Adapter-Scale, where “Learned Embedding from (i − 1)-th ViT Block” represents features processed by the previous transformer layer, and “Reshaped Prompts” are dimension-adjusted prompts generated by Adapter-Feature for the ViT Block. (<b>C</b>) The structure of Adapter-Feature, which extracts HFC features through Fourier transform for prompt generation.</p>
Full article ">Figure 3
<p>Visual examples from different datasets depict various scenes, including clouds, buildings, fields, and roads. The image above shows remote sensing images with the corresponding mask. GT represents ground truth.</p>
Full article ">Figure 4
<p>Visual examples from the CSWV and GID datasets, illustrating cloud-like scenes in mountainous regions and diverse land-cover scenarios.</p>
Full article ">Figure 5
<p><math display="inline"><semantics> <mrow> <mi>P</mi> <mi>r</mi> <mi>e</mi> <mi>c</mi> <mi>i</mi> <mi>s</mi> <mi>i</mi> <mi>o</mi> <mi>n</mi> </mrow> </semantics></math> vs. <math display="inline"><semantics> <mrow> <mi>R</mi> <mi>e</mi> <mi>c</mi> <mi>a</mi> <mi>l</mi> <mi>l</mi> </mrow> </semantics></math> across various datasets, showing the performance of different segmentation models (U-Net, DeepLabV3+, Segformer, SAM(center +), SAM(center−), SAM(manual), and RSAM-Seg in different types of scenarios.</p>
Full article ">Figure 6
<p>Comparison of cloud segmentation results using the 38-Cloud dataset for RSAM-Seg, SAM, U-Net, DeepLabV3+, and Segformer. For SAM(manual), the green points indicate manually annotated positive samples, the blue points represent negative samples, and the red masks show the segmentation results. For SAM(center), + and − denote the segmentation results where the center points are marked as positive and negative classes, respectively.</p>
Full article ">Figure 7
<p>Comparison of field segmentation results for the Sentinel-2 dataset with RSAM-Seg, SAM, U-Net, DeepLabV3+, and Segformer. For SAM(manual), the green points indicate manually annotated positive samples, the blue points represent negative samples, and the red masks show the segmentation results. For SAM(center), + and − denote the segmentation results where the center points are marked as positive and negative classes, respectively.</p>
Full article ">Figure 8
<p>Comparison of building segmentation results for the Inria dataset with RSAM-Seg, SAM, U-Net, DeepLabV3+, and Segformer. In SAM(manual), the green points indicate manually annotated positive samples, the blue points represent negative samples, and the red masks show the segmentation results. For SAM(center), + and − denote the segmentation results where the center points are marked as positive and negative classes, respectively.</p>
Full article ">Figure 9
<p>Comparison of the road segmentation results for the DG-Road dataset using RSAM-Seg, SAM, U-Net, DeepLabV3+, and Segformer. For SAM(manual), the green points indicate manually annotated positive samples, the blue points represent negative samples, and the red masks show the segmentation results. For SAM(center), + and − denote the segmentation results where the center points are marked as positive and negative classes, respectively.</p>
Full article ">Figure 10
<p>Comparison of cloud and snow segmentation results for the CSWV dataset with original imagery, ground truth, U-Net, SAM (in the point and segment everything modes), and RSAM-Seg.</p>
Full article ">Figure 11
<p>Comparison of the land-cover multi-class segmentation results for the GID dataset with ground truth, U-Net, SAM, and RSAM-Seg.</p>
Full article ">Figure 12
<p>Visualization results of the ablation study for RSAM-Seg using the 38-Cloud dataset, where the complete RSAM-Seg model shows the best performance compared to its ablated versions.</p>
Full article ">Figure 13
<p>Comparison of images, ground truth, HFC features, and RSAM-Seg results across four distinct scenarios of clouds, fields, buildings, and roads, demonstrating segmentation through HFC feature extraction.</p>
Full article ">Figure 14
<p>Examples of few-shot segmentation results using the 38-Cloud dataset.</p>
Full article ">Figure 15
<p>Examples of completion results on the DG-Road dataset.</p>
Full article ">
26 pages, 13025 KiB  
Article
Unified Spatial-Frequency Modeling and Alignment for Multi-Scale Small Object Detection
by Jing Liu, Ying Wang, Yanyan Cao, Chaoping Guo, Peijun Shi and Pan Li
Symmetry 2025, 17(2), 242; https://doi.org/10.3390/sym17020242 - 6 Feb 2025
Abstract
Small object detection in aerial imagery remains challenging due to sparse feature representation, limited spatial resolution, and complex background interference. Current deep learning approaches enhance detection performance through multi-scale feature fusion, leveraging convolutional operations to expand the receptive field or self-attention mechanisms for [...] Read more.
Small object detection in aerial imagery remains challenging due to sparse feature representation, limited spatial resolution, and complex background interference. Current deep learning approaches enhance detection performance through multi-scale feature fusion, leveraging convolutional operations to expand the receptive field or self-attention mechanisms for global context modeling. However, these methods primarily rely on spatial-domain features, while self-attention introduces high computational costs, and conventional fusion strategies (e.g., concatenation or addition) often result in weak feature correlation or boundary misalignment. To address these challenges, we propose a unified spatial-frequency modeling and multi-scale alignment fusion framework, termed USF-DETR, for small object detection. The framework comprises three key modules: the Spatial-Frequency Interaction Backbone (SFIB), the Dual Alignment and Balance Fusion FPN (DABF-FPN), and the Efficient Attention-AIFI (EA-AIFI). The SFIB integrates the Scharr operator for spatial edge and detail extraction and FFT/IFFT for capturing frequency-domain patterns, achieving a balanced fusion of global semantics and local details. The DABF-FPN employs bidirectional geometric alignment and adaptive attention to enhance the significance expression of the target area, suppress background noise, and improve feature asymmetry across scales. The EA-AIFI streamlines the Transformer attention mechanism by removing key-value interactions and encoding query relationships via linear projections, significantly boosting inference speed and contextual modeling. Experiments on the VisDrone and TinyPerson datasets demonstrate the effectiveness of USF-DETR, achieving improvements of 2.3% and 1.4% mAP over baselines, respectively, while balancing accuracy and computational efficiency. The framework outperforms state-of-the-art methods in small object detection. Full article
(This article belongs to the Special Issue Symmetry and Asymmetry Study in Object Detection)
Show Figures

Figure 1

Figure 1
<p>Comparison between RT-DETR and the proposed USF-DETR method. The feature maps generated by USF-DETR (bottom row) exhibit sharper edges and richer details due to the SFIB and EA-AIFI modules. After multi-scale alignment fusion through the DABF-FPN Encoder, USF-DETR produces more accurate heatmaps, effectively highlighting small objects and improving detection results with fewer missed detections and false positives, as demonstrated by the red bounding boxes.</p>
Full article ">Figure 2
<p>Architecture of the proposed USF-DETR, which includes three modules: SFIB, EA-AIFI, and DABF-FPN. The top part illustrates the pipeline of USF-DETR, while the bottom part presents the module flowchart.</p>
Full article ">Figure 3
<p>The pipeline of SFIB consists of four stages, with each stage including a Conv layer and a SFI block. The SFI block, shown in the lower left figure, is connected across layers using the CSP concept; As depicted in the lower right image, the SFI extracts spatial and frequency domain features of the image and then fuses them.</p>
Full article ">Figure 4
<p>The overall structure of the DABF-FPN integrates bidirectional feature fusion to enhance small object detection and outputs multi-scale features (P2, N3, N4, and N5) for further processing.</p>
Full article ">Figure 5
<p>The structure of the DABF module involves high-level semantic features and low-level detailed features being adaptively processed to extract mutual representations. Two DABF blocks facilitate comprehensive information exchange and enhance feature fusion quality.</p>
Full article ">Figure 6
<p>EA-AIFI module. (<b>a</b>) Through input embedding and positional encoding, combined with enhanced representation of contextual information, further internal feature interaction and optimization are carried out through a FFN. (<b>b</b>) Efficient Additive Attention eliminates key value interactions and relies solely on linear projections.</p>
Full article ">Figure 7
<p>Bounding box distribution. (<b>a</b>) VisDrone2019-DET Dataset. (<b>b</b>) TinyPerson Dataset. The vertical axis represents the categories of annotated bounding boxes, while the horizontal axis depicts the square root of the bounding box area, measured in pixels.</p>
Full article ">Figure 8
<p>Visualization of feature maps. (<b>a</b>) Input image. (<b>b</b>) Feature map generated without using the SFI module in the baseline model. (<b>c</b>) Feature map generated with the SFI module in USF-DETR.</p>
Full article ">Figure 9
<p>Visualizing the detection results and heatmap on TinyPerson. The highlighted area represents the region of network attention, demonstrating the outstanding performance of USF-DETR in detecting small objects.</p>
Full article ">Figure 10
<p>Detection results of the USF-DETR on the VisDrone dataset. Boxes of different colors represent different target categories.</p>
Full article ">Figure 11
<p>A comparison of detection results between USF-DETR and the baseline model is presented. Green boxes indicate correct detections, blue boxes represent false positives, and red boxes denote missed detections.</p>
Full article ">Figure 12
<p>A comparison of detection performance between the two methods. The first row represents USF-DETR, while the second row shows the baseline method. USF-DETR significantly reduces false positives (blue) and false negatives (red).</p>
Full article ">Figure 13
<p>Comparison of detection performance between USF-DETR and popular methods. The yellow circle shows the outstanding detection effect of USF-DETR.</p>
Full article ">
26 pages, 5439 KiB  
Article
Particle Filter Tracking System Based on Digital Zoom and Regional Image Measure
by Qisen Zhao, Liquan Dong, Xuhong Chu, Ming Liu, Lingqin Kong and Yuejin Zhao
Sensors 2025, 25(3), 880; https://doi.org/10.3390/s25030880 - 31 Jan 2025
Abstract
To address the challenges of low accuracy and the difficulty in balancing a large field of view and long distance when tracking high-speed moving targets with a single sensor, an ROI adaptive digital zoom tracking method is proposed. In this paper, we discuss [...] Read more.
To address the challenges of low accuracy and the difficulty in balancing a large field of view and long distance when tracking high-speed moving targets with a single sensor, an ROI adaptive digital zoom tracking method is proposed. In this paper, we discuss the impact of ROI on image processing and describe the design of the ROI adaptive digital zoom tracking system. Additionally, we construct an adaptive ROI update model based on normalized target information. To capture target changes effectively, we introduce the multi-scale regional measure and propose an improved particle filter algorithm, referred to as the improved multi-scale regional measure resampling particle filter (IMR-PF). This method enables high temporal resolution processing efficiency within a high-resolution large field of view, which is particularly beneficial for high-resolution videos. The IMR-PF can maintain high temporal resolution within a wide field of view with high resolution. Simulation results demonstrate that the improved target tracking method effectively improves tracking robustness to target motion changes and reduces the tracking center error by 20%, as compared to other state-of-the-art methods. The IMR-PF still maintains good performance even when confronted with various interference factors and in real-world scenario applications. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>The overall structure of the proposed method.</p>
Full article ">Figure 2
<p>MENS filter diagram.</p>
Full article ">Figure 3
<p>Tracking results of the top six methods on the U-skier dataset (different video clips from top to bottom).</p>
Full article ">Figure 4
<p>Evaluation result on the U-skin dataset. (<b>a</b>), Success rate of different methods on the U-skier dataset; (<b>b</b>), Precision of different methods on the U-skier dataset.</p>
Full article ">Figure 5
<p>System schematic: large FOV camera is used to capture object images, PC processor is used to target detection and tracking on the captured images, and the tracking results are used to drive the Pan-tilt Platform to rotate. Video camera is a zoom camera used to capture close-up images of the target and evaluate the results. <math display="inline"><semantics> <mi>L</mi> </semantics></math> is the depth range of the entire system.</p>
Full article ">Figure 6
<p>Tracking results of the top five tracking methods on the U-skier dataset in an indoor simulation environment (The following line shows the tracking results on a high-resolution image with a large field of view. The top line shows the close-up camera captured images corresponding to each tracking moment. The trajectory curve in the figure was formed by manually registering the front and rear frames and connecting the predicted center positions of different methods. The images in column (<b>a</b>) depict a scenario where the athlete’s target leaps in the distance, with interference from similar targets in the background. The images in column (<b>b</b>) show the athlete’s target descending and encountering a sudden change in trajectory due to the ski slope. The images in column (<b>c</b>) illustrate the athlete’s target leaping into the air from a close distance. The images in column (<b>d</b>) capture the athlete’s target descending at a close distance, experiencing a sudden change in trajectory upon encountering the ski slope).</p>
Full article ">Figure 7
<p>Tracking results on the U-skier dataset in an indoor simulation environment. (<b>a</b>) Ground truth of the target trajectory; (<b>b</b>) Changes in the target trajectory and target area measure; (<b>c</b>,<b>d</b>) Tracking error curves of different methods as the video frame changes.</p>
Full article ">Figure 8
<p>Results of the top six different methods on the VOT2021 dataset for tracking different data segments on the screen. (<b>a</b>) represents the success rate of these methods on the graduate_set of the VOT2021; (<b>b</b>) represents the success rate of these methods on the matrix_set of the VOT2021; (<b>c</b>) represents the success rate of these methods on the pedestrian_set of the VOT2021; (<b>d</b>) represents the success rate of these methods on the road_set of the VOT2021; and (<b>e</b>) represents the success rate of these methods on the shaking_set of the VOT2021.</p>
Full article ">Figure 9
<p>Success rate of different methods on the different sub-datasets of the VOT2021 dataset. (<b>a</b>) represents the Success rate of these methods at graduate_set in the VOT2021 dataset; (<b>b</b>) represents the Success rate of these methods at matrix_set in the VOT2021 dataset; (<b>c</b>) represents the Success rate of these methods at soccer_set in the VOT2021 dataset; (<b>d</b>) represents the Success rate of these methods at nature_set in the VOT2021 dataset; (<b>e</b>) represents the Success rate of these methods at road_set in the VOT2021 dataset; (<b>f</b>) represents the Success rate of these methods at racing_set in the VOT2021 dataset; (<b>g</b>) represents the Success rate of these methods at pedestrian_set in the VOT2021 dataset; and (<b>h</b>) represents the Success rate of these methods at shaking_set in the VOT2021 dataset.</p>
Full article ">Figure 10
<p>Precision of different methods on data with different properties in the VOT2021 dataset. (<b>a</b>) represents the precision of these methods at size variations properties; (<b>b</b>) represents the precision of these methods at occlusion properties; (<b>c</b>) represents the precision of these methods at background clutter properties; (<b>d</b>) represents the precision of these methods at camera movement properties; (<b>e</b>) represents the precision of these methods at target fast motion properties; and (<b>f</b>) represents the precision of these methods at illumination variable properties.</p>
Full article ">Figure 11
<p>Tracking error and success rate curves of different module improvement methods on the U-skier dataset. (<b>a</b>) Representing the central error of methods with different structures at Uskier sequence; (<b>b</b>) Repeat the success rate of methods with different structures at Uskier sequence.</p>
Full article ">
49 pages, 68388 KiB  
Article
Improved Stereophotogrammetric and Multi-View Shape-from-Shading DTMs of Occator Crater and Its Interior Cryovolcanism-Related Bright Spots
by Alicia Neesemann, Stephan van Gasselt, Ralf Jaumann, Julie C. Castillo-Rogez, Carol A. Raymond, Sebastian H. G. Walter and Frank Postberg
Remote Sens. 2025, 17(3), 437; https://doi.org/10.3390/rs17030437 - 27 Jan 2025
Abstract
Over the course of NASA’s Dawn Discovery mission, the onboard framing camera mapped Ceres across a wide wavelength spectrum at varying polar science orbits and altitudes. With increasing resolution, the uniqueness of the 92 km wide, young Occator crater became evident. Its central [...] Read more.
Over the course of NASA’s Dawn Discovery mission, the onboard framing camera mapped Ceres across a wide wavelength spectrum at varying polar science orbits and altitudes. With increasing resolution, the uniqueness of the 92 km wide, young Occator crater became evident. Its central cryovolcanic dome, Cerealia Tholus, and especially the associated bright carbonate and ammonium chloride deposits—named Cerealia Facula and the thinner, more dispersed Vinalia Faculae—are the surface expressions of a deep brine reservoir beneath Occator. Understandably, this made this crater the target for future sample return mission studies. The planning and preparation for this kind of mission require the characterization of potential landing sites based on the most accurate topography and orthorectified image data. In this work, we demonstrate the capabilities of the freely available and open-source USGS Integrated Software for Imagers and Spectrometers (ISIS 3) and Ames Stereo Pipeline (ASP 2.7) in creating high-quality image data products as well as stereophotogrammetric (SPG) and multi-view shape-from-shading (SfS) digital terrain models (DTMs) of the aforementioned spectroscopically challenging features. The main data products of our work are four new DTMs, including one SPG and one SfS DTM based on High-Altitude Mapping Orbit (HAMO) (CSH/CXJ) and one SPG and one SfS DTM based on Low-Altitude Mapping Orbit (LAMO) (CSL/CXL), along with selected Extended Mission Orbit 7 (XMO7) framing camera (FC) data. The SPG and SfS DTMs were calculated to a GSD of 1 and 0.5 px, corresponding to 136 m (HAMO SPG), 68 m (HAMO SfS), 34 m (LAMO SPG), and 17 m (LAMO SfS). Finally, we show that the SPG and SfS approaches we used yield consistent results even in the presence of high albedo differences and highlight how our new DTMs differ from those previously created and published by the German Aerospace Center (DLR) and the Jet Propulsion Laboratory (JPL). Full article
Show Figures

Figure 1

Figure 1
<p>CSH CYCLE 1 photometrically corrected RGB orthomosaic of the Occator crater (F5IR, F2GREEN, F8BLUE). The map is a stereographic projection with the projection and image center at 22.879°N<sub><math display="inline"><semantics> <mi>φ</mi> </semantics></math></sub>/239.429°E (19.865°N<sub><math display="inline"><semantics> <mi>ψ</mi> </semantics></math></sub>). RGB values were limited to R: 0.0257–0.0339; G: 0.0555–0.0869; and B: 0.0552–0.0815. (<b>a</b>) Due to the high albedo difference, bright deposits within Occator appear overexposed in the applied histogram stretch. To get an idea about their shape and distribution, we compiled a CSL/CXL RGB orthomosaic with a histogram stretch optimized for the bright deposits, shown in <a href="#remotesensing-17-00437-f002" class="html-fig">Figure 2</a>. (<b>b</b>) Extent of our CSL/CXL ASP SPG DTM. (<b>c</b>,<b>d</b>) Areas for which we calculated the highest resolution DTM (CSL/CXL ASP SfS DTMs).</p>
Full article ">Figure 2
<p>Photometrically corrected CSL/CXL RGB orthomosaic of Occator’s interior. The mosaic was compiled from 13 individual images per channel (see <a href="#remotesensing-17-00437-t0A2" class="html-table">Table A2</a>) using F5IR, F2GREEN, and F8BLUE as the three RGB bands. In contrast to <a href="#remotesensing-17-00437-f001" class="html-fig">Figure 1</a>, images were photometrically corrected based on the 482 × 446 km ellipsoid, not on the DTM, to preserve the topography-related brightness variations and morphology, respectively. RGB values were limited to R: 0.0016–0.6634; G: 0.0003–1.5579; and B: 0.0000–1.1953. The figure is a stereographic projection centered at 23.05°N<sub><math display="inline"><semantics> <mi>φ</mi> </semantics></math></sub>/241.05°E (20.02°N<sub><math display="inline"><semantics> <mi>ψ</mi> </semantics></math></sub>). To the left, we see Cerealia Facula with its fractured dome, Cerealia Tholus, and the orange-colored exposed hydrated sodium chloride [<a href="#B24-remotesensing-17-00437" class="html-bibr">24</a>]. To the right, we see various smaller and thinner bright deposits collectively named Vinalia Faculae (<b>a</b>–<b>e</b>).</p>
Full article ">Figure 3
<p>Number of acquired FC2 images and geometries of polar mapping orbits during Dawn’s mission at Ceres. Left: Number of images acquired by the FC2 during Dawn’s mission to Ceres. Left: During the course of Dawn’s mission at Ceres, 67,969 images in total were acquired by the FC2. Except for images acquired for the purpose of the camera calibration and orientation of the spacecraft, as well as for the search of moons orbiting Ceres, by far the most images mapping Ceres directly were taken during the HAMO and LAMO. Ancillary image acquisition was carried out by the FC1 in order to increase spatial coverage during Dawn’s time-limited final mission phase, XM2, but is not included in this figure (see <a href="#remotesensing-17-00437-t001" class="html-table">Table 1</a>). Right: Indicated orbits correspond to the median distance to Ceres’ center during the different orbit phases. 360° corresponds to the period between 2015/01/01 and 2018/12/31. Geometries of Dawn’s highly elliptical 2nd extended mission orbits (XMO5–XMO7) flown during the final mission, extended mission XM2, are not shown in this figure for scale reasons.</p>
Full article ">Figure 4
<p>Detection of erroneous pixels. In the upper two rows, we present average images created for the eight individual FC filters and two average flat-field images for the F1 and F7 filters taken while the front door was closed and the calibration lamp (callamp) on. Note that images used in this context were not photometrically but only radiometrically corrected. The east–west shading therefore stems from the illumination conditions during image acquisition and not from camera shading. The five static erroneous pixel clusters recognized in all the average images are marked by an ‘x’ in the upper left subfigure. In the bottom row, we present magnifications of the five erroneous pixel clusters to illustrate their extent, marked by the dashed outlines. A 3 sigma histogram stretch was applied to each average image.</p>
Full article ">Figure 5
<p>Basic ISIS 3 pre-processing workflow. Gray: functions; red: temporary textfiles (here, GCP.net files); green: temporary raster files; blue: final photometrically corrected, orthorectified F1CLEAR image; yellow: F1CLEAR orthomosaic, HAMO-based SPG DTM, and reconstructed SPICE kernels. The asterisk in campt* stands for a script we wrote that reads out the lat/lon values at a specific sample/line position, converts them into cartesian coordinates (the ApprioryX, Y, and Z values) and creates a <span class="html-italic">qtie</span>-readable GCP netfile. A detailed description of the flowchart is given in <a href="#sec4dot4-remotesensing-17-00437" class="html-sec">Section 4.4</a>.</p>
Full article ">Figure 6
<p>ASP SPG processing workflow. This figure is a version of Figure 14.1 from the ASP 2.7 documentation (<a href="https://stereopipeline.readthedocs.io/en/latest/correlation.html" target="_blank">https://stereopipeline.readthedocs.io/en/latest/correlation.html</a>, accessed on 23 Decemeber 2024) that we have modified. The parameters specified in the stereo.txt file and passed to the <span class="html-italic">stereo</span> command are listed in Appendix <a href="#remotesensing-17-00437-t0A8" class="html-table">Table A8</a>. Other parameters passed to the <span class="html-italic">point2dem</span> command to remove additional erroneous points from the point cloud during the DTM generation are described in <a href="#sec4dot5-remotesensing-17-00437" class="html-sec">Section 4.5</a>.</p>
Full article ">Figure 7
<p>Latitude- and longitude-dependent deviation between the CSH/CXJ and CSL/CXL ASP SPG and SfS DTMs created in our study. For the upper two plots, the mean latitude and longitude values were calculated from the extent of area 1, while they were calculated for area 4 in the four lower plots.</p>
Full article ">Figure 8
<p>Latitude- and longitude-dependent deviation between the HAMO and LAMO SPG DTMs. Mean latitude and longitude values were calculated for the extent of area 3.</p>
Full article ">Figure 9
<p>Latitude- and longitude-dependent deviation of our 4 new CSH/CXJ and CSL/CXL SPG and SfS DTMs and the JPL HAMO/LAMO SPC DTM [<a href="#B58-remotesensing-17-00437" class="html-bibr">58</a>,<a href="#B59-remotesensing-17-00437" class="html-bibr">59</a>] from the DLR HAMO SPG DTM [<a href="#B55-remotesensing-17-00437" class="html-bibr">55</a>,<a href="#B56-remotesensing-17-00437" class="html-bibr">56</a>].</p>
Full article ">Figure 10
<p>Visual comparison of the seven DTMs for the example of the fresh crater located at 14.281°N<sub><math display="inline"><semantics> <mi>φ</mi> </semantics></math></sub>/233.489°E (12.295°N<sub><math display="inline"><semantics> <mi>ψ</mi> </semantics></math></sub>). Images used to create the orthomosaics are listed in <a href="#remotesensing-17-00437-t0A5" class="html-table">Table A5</a>.</p>
Full article ">Figure 11
<p>Topographic profiles of the fresh crater. Subfigures (<b>a</b>,<b>b</b>) (left and middle panel) are stereographic projections centered at 14.281°N<sub><math display="inline"><semantics> <mi>φ</mi> </semantics></math></sub>/233.489°E (12.295°N<sub><math display="inline"><semantics> <mi>ψ</mi> </semantics></math></sub>). (<b>a</b>) Photometrically corrected FC2 F1CLEAR close-up view of the fresh crater. (<b>b</b>) Photometrically corrected RGB (F5IR, F2GREEN, F8BLUE) orthomosaic of the fresh crater. In total, we derived 52 profiles at 3 degree intervals between 24–90°, 144–177°, and 240–288° for each of the seven DTMs extending from the crater center. (<b>c</b>) (right panel) Average topographic profiles for each of the seven DTMs. <sup>a</sup> [<a href="#B59-remotesensing-17-00437" class="html-bibr">59</a>], <sup>b</sup> [<a href="#B58-remotesensing-17-00437" class="html-bibr">58</a>], <sup>c</sup> [<a href="#B106-remotesensing-17-00437" class="html-bibr">106</a>], <sup>d</sup> [<a href="#B104-remotesensing-17-00437" class="html-bibr">104</a>], <sup>e</sup> [<a href="#B56-remotesensing-17-00437" class="html-bibr">56</a>], <sup>f</sup> [<a href="#B55-remotesensing-17-00437" class="html-bibr">55</a>].</p>
Full article ">Figure 12
<p>Detailed comparison of the topographic profiles of the fresh crater. As a reference (black solid line), we used our new CSL/CXL ASP SfS DTM, as it has the highest effective resolution and the highest d/D ratio of 0.255 and plotted it together with the profiles extracted from the other six DTMs (<b>a</b>–<b>f</b>). Black and gray triangles mark the inflection points (the highest elevation of the rim crest) of our reference profile and the other profiles. Note that the highest congruence exists between profiles taken from our CSL/CXL ASP SfS, our CSL/CXL ASP SPG, and the DLR CSL/CXL SPG DTMs.</p>
Full article ">Figure 13
<p>Visual comparison of the seven DTMs for the example of the Cerealia Tholus at 22.626°N<sub><math display="inline"><semantics> <mi>φ</mi> </semantics></math></sub>/239.581°E (19.648°N<sub><math display="inline"><semantics> <mi>ψ</mi> </semantics></math></sub>). Images used to create the orthomosaics are listed in <a href="#remotesensing-17-00437-t0A6" class="html-table">Table A6</a>.</p>
Full article ">Figure 14
<p>Overview of the location of the topographic profile across the Cerealia Tholus. We created three different maps, each with their advantages and disadvantages, in order to show which specific surface features are covered by our topographic profile. Our 25,498 m long profile goes from west to east while crossing the highest elevations (the Lohri, Cerealia, and Kekri Tholus) within Occator. (<b>a</b>) Semi-transparent, color-coded CSL/CXL/XMO7 ASP SfS DTM overlaid on the corresponding hillshade model. Topography contour lines are plotted in 100 m intervals. (<b>b</b>) CSL/CXL RGB color composite of Cerealia Facula. (<b>c</b>) Generated slope map overlaid on a curvature map. This combined map highlights aspects of the surface’s shape or features, such as the circular and other rather subparallel fault systems as well as numerous little mounds, at a detailed level.</p>
Full article ">Figure 15
<p>Topographic profiles across the Cerealia Tholus.</p>
Full article ">Figure 16
<p>Topographic profiles across Vinalia Faculae. (<b>a</b>) Topographic profiles across Vinalia Faculae, extracted from our CSL/CXL SPG (black line) and SfS (dark grey line) DTMs, as well as the HAMO/LAMO-based SPC DTM (blue line) from the JPL. Additionally, the albedo along the profile line was extracted based on the photometrically corrected CSL/CXL F1CLEAR orthomosaic included in this study. (<b>b</b>) Deviations between the lower resolution yet more robust CSL/CXL ASP SPG DTM, the CSL/CXL ASP SfS DTM, and the HAMO/LAMO-based SPC DTM from the JPL. (<b>c</b>) The CSL/CXL ASP SfS DTM of Vinalia Faculae, represented as elevations above the 482 × 446 km ellipsoid. (<b>d</b>) Photometrically corrected CSL/CXL F1CLEAR orthomosaic of Vinalia Faculae. Both maps are stereographic projections, centered at 23.292°N<sub><math display="inline"><semantics> <mi>φ</mi> </semantics></math></sub>/242.487°E (20.234°N<sub><math display="inline"><semantics> <mi>ψ</mi> </semantics></math></sub>). The course of the topographic profiles shown in panel (<b>a</b>) is indicated by a black-and-green dashed line.</p>
Full article ">
25 pages, 8829 KiB  
Article
Novel Surveillance View: A Novel Benchmark and View-Optimized Framework for Pedestrian Detection from UAV Perspectives
by Chenglizhao Chen, Shengran Gao, Hongjuan Pei, Ning Chen, Lei Shi and Peiying Zhang
Sensors 2025, 25(3), 772; https://doi.org/10.3390/s25030772 - 27 Jan 2025
Abstract
To address the issues of insufficient samples, limited scene diversity, missing perspectives, and low resolution in existing UAV-based pedestrian detection datasets, this paper proposes a novel UAV-based pedestrian detection benchmark dataset named the Novel Surveillance View (NSV). This dataset encompasses diverse scenes and [...] Read more.
To address the issues of insufficient samples, limited scene diversity, missing perspectives, and low resolution in existing UAV-based pedestrian detection datasets, this paper proposes a novel UAV-based pedestrian detection benchmark dataset named the Novel Surveillance View (NSV). This dataset encompasses diverse scenes and pedestrian information captured from multiple perspectives, and introduces an innovative data mining approach that leverages tracking and optical flow information. This approach significantly improves data acquisition efficiency while ensuring annotation quality. Furthermore, an improved pedestrian detection method is proposed to overcome the performance degradation caused by significant perspective changes in top-down UAV views. Firstly, the View-Agnostic Decomposition (VAD) module decouples features into perspective-dependent and perspective-independent branches to enhance the model’s generalization ability to perspective variations. Secondly, the Deformable Conv-BN-SiLU (DCBS) module dynamically adjusts the receptive field shape to better adapt to the geometric deformations of pedestrians. Finally, the Context-Aware Pyramid Spatial Attention (CPSA) module integrates multi-scale features with attention mechanisms to address the challenge of drastic target scale variations. The experimental results demonstrate that the proposed method improves the mean Average Precision (mAP) by 9% on the NSV dataset, thereby validating that the approach effectively enhances pedestrian detection accuracy from UAV perspectives by optimizing perspective features. Full article
Show Figures

Figure 1

Figure 1
<p>Mainstream pedestrian detectors are unable to detect pedestrians effectively from a UAV perspective.</p>
Full article ">Figure 2
<p>Overall architecture diagram of the TAHA framework. The framework consists of two main modules: (<b>a</b>) the target tracking module based on SMILEtrack, which processes target annotations from a normal viewpoint, and (<b>b</b>) the optical flow-assisted detection module, designed specifically for target annotations from a top-down viewpoint. The outputs of the two modules (<math display="inline"><semantics> <msub> <mi mathvariant="italic">Labels</mi> <mi mathvariant="italic">Track</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi mathvariant="italic">Labels</mi> <mi mathvariant="italic">Flow</mi> </msub> </semantics></math>) are fused to achieve robust annotation across different viewpoint scenarios.</p>
Full article ">Figure 3
<p>Architecture diagram of the optical flow feature transfer in the Flow module. The framework consists of two key stages: (<b>a</b>) the transfer learning stage based on standard perspective data, where the GMFlow algorithm is used to extract optical flow features and feature mapping is performed using detection results from SMILEtrack; (<b>b</b>) the vertical perspective detection stage, where a category consistency filtering strategy is applied to achieve accurate detection of pedestrian optical flow features.</p>
Full article ">Figure 4
<p>Filtering non-pedestrian labels using initial frames.</p>
Full article ">Figure 5
<p>Overall architecture of the Pedestrian-DVC network framework is based on the YOLOv7 architecture and integrates three innovative modules: View-Agnostic Decomposition (VAD) <a href="#sec4dot1-sensors-25-00772" class="html-sec">Section 4.1</a>, Deformable Conv-BN-SiLU (DCBS) <a href="#sec4dot2-sensors-25-00772" class="html-sec">Section 4.2</a>, and Context-Aware Pyramid Spatial Attention (CPSA) <a href="#sec4dot3-sensors-25-00772" class="html-sec">Section 4.3</a>. The innovative modules are highlighted with red dashed circles, and the black dashed boxes represent the specific content or meaning of each module, while the other modules are part of the YOLOv7 architecture.</p>
Full article ">Figure 6
<p>Architecture of the View-Agnostic Detection (VAD) module for extracting stable viewpoint-invariant feature representations.</p>
Full article ">Figure 7
<p>Context-Aware Pyramid Spatial Attention (CPSA) module adaptively processes pedestrian targets at different scales and suppresses background interference.</p>
Full article ">Figure 8
<p>Comparison of visualization information of attention heatmaps. (<b>a</b>–<b>h</b>) Each group shows the attention heatmap results for the same image using different methods. In each heatmap, areas with redder tones indicate higher levels of attention. Red boxes denote the ground truth locations of pedestrians.</p>
Full article ">Figure 9
<p>Performance of the Pedestrian-DVC model on the NSV testset.</p>
Full article ">
21 pages, 10436 KiB  
Technical Note
Rapid Micro-Motion Feature Extraction of Multiple Space Targets Based on Improved IRT
by Jing Wu, Xiaofeng Ai, Zhiming Xu, Yiqi Zhu and Qihua Wu
Remote Sens. 2025, 17(3), 434; https://doi.org/10.3390/rs17030434 - 27 Jan 2025
Abstract
Micro-motion feature extraction is of great significance for target recognition. However, traditional methods mostly focus on single target and struggle to correctly separate the severely overlapping micro-motion curves of multiple targets. In this paper, a rapid micro-motion feature extraction algorithm of multiple space [...] Read more.
Micro-motion feature extraction is of great significance for target recognition. However, traditional methods mostly focus on single target and struggle to correctly separate the severely overlapping micro-motion curves of multiple targets. In this paper, a rapid micro-motion feature extraction algorithm of multiple space targets based on inverse radon transform (IRT) with a modified model is proposed. First, the high-resolution range profile (HRRP) generated from echo is subject to binarization to improve the unstable estimation caused by noise. Then, the micro-motion period in a complicated multi-target scenario is obtained by a period estimation method based on the autocorrelation coefficients of binarized HRRP. To further improve the extraction accuracy, the IRT model of the micro-range curve is modified from the sine function to second-order sine function. By searching for the remaining unknown parameters in the model in conjunction with the period, the precise micro-range curves are quickly separated. Each time the curves of a target are extracted, they are removed, and the next extraction is carried out until all the targets have been searched. Finally, simulation and experimental results indicate that the proposed algorithm can not only correctly separate the micro-motion feature curves of multiple space targets under low signal-to-noise ratio (SNR) conditions but also significantly outperforms the original IRT in terms of extraction speed. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Micro-motion model of a single target.</p>
Full article ">Figure 2
<p>Transformation between the reference coordinate system and global coordinate system.</p>
Full article ">Figure 3
<p>Micro-motion model of multiple targets.</p>
Full article ">Figure 4
<p>Algorithm flowchart.</p>
Full article ">Figure 5
<p>Comparison figure of the fast and slow time results.</p>
Full article ">Figure 6
<p>The autocorrelation coefficients before and after binarization.</p>
Full article ">Figure 7
<p>The results of different autocorrelation coefficient treatments.</p>
Full article ">Figure 8
<p>The autocorrelation coefficient with a period of 0.333 s.</p>
Full article ">Figure 9
<p>Period estimation result.</p>
Full article ">Figure 9 Cont.
<p>Period estimation result.</p>
Full article ">Figure 10
<p>IRT result of simulation.</p>
Full article ">Figure 11
<p>Curve extraction result.</p>
Full article ">Figure 12
<p>The number of successfully estimated periods.</p>
Full article ">Figure 13
<p>The period estimation errors with different SNRs.</p>
Full article ">Figure 14
<p>Curve extraction error.</p>
Full article ">Figure 15
<p>Design of experimental scenario.</p>
Full article ">Figure 16
<p>Experimental setup.</p>
Full article ">Figure 17
<p>Theoretical curves. (<b>a</b>) The first group. (<b>b</b>) The second group.</p>
Full article ">Figure 18
<p>IRT result of experiment.</p>
Full article ">Figure 19
<p>The extracted result of the first group.</p>
Full article ">Figure 20
<p>The extracted result of the second group.</p>
Full article ">
23 pages, 44092 KiB  
Article
A Global-Scale Overlapping Pixels Calculation Method for Whisk-Broom Payloads with Multi-Module-Staggered Longlinear-Array Detectors
by Xinwang Du, Chao Wu, Quan Liang, Lixing Zhao, Yixuan Xu, Junhong Guo, Xiaoyan Li and Fansheng Chen
Remote Sens. 2025, 17(3), 433; https://doi.org/10.3390/rs17030433 - 27 Jan 2025
Abstract
A multi-module staggered (MMS) long-linear-array (LLA) detector is presently recognized as an effective and widely adopted means of improving the field of view (FOV) of in-orbit optical line-array cameras. In particular, in terms of low-orbit whisk-broom payloads, the MMS LLA detector combined with [...] Read more.
A multi-module staggered (MMS) long-linear-array (LLA) detector is presently recognized as an effective and widely adopted means of improving the field of view (FOV) of in-orbit optical line-array cameras. In particular, in terms of low-orbit whisk-broom payloads, the MMS LLA detector combined with the one-dimensional scanning mirror is capable of achieving both large-swath and high-resolution imaging. However, because of the complexity of the instantaneous relative motion model (IRMM) of the whisk-broom imaging mechanism, it is really difficult to determine and verify the actual numbers of overlapping pixels of adjacent detector sub-module images and consecutive images in the same and opposite scanning directions, which are exceedingly crucial to the instrument design pre-launch as well as the in-orbit geometric quantitative processing and application post-launch. Therefore, in this paper, aiming at addressing the problems above, we propose a global-scale overlapping pixels calculation method based on the IRMM and rigorous geometric positioning model (RGPM) of the whisk-broom payloads with an MMS LLA detector. First, in accordance with the imaging theory and the specific optical–mechanical structure, the RGPM of the whisk-broom payload is constructed and introduced elaborately. Then, we qualitatively analyze the variation tendency of the overlapping pixels of adjacent detector sub-module images with the IRMM of the imaging targets, and establish the associated overlapping pixels calculation model based on the RGPM. And subsequently, the global-scale overlapping pixels calculation models for consecutive images of the same and opposite scanning directions of the whisk-broom payload are also built. Finally, the corresponding verification method is presented in detail. The proposed method is validated using both simulation data and in-orbit payload data from the Thermal Infrared Spectrometer (TIS) of the Sustainable Development Goals Satellite-1 (SDGSAT-1), launched on 5 November 2021, demonstrating its effectiveness and accuracy with overlapping pixel errors of less than 0.3 pixels between sub-modules and less than 0.5 pixels between consecutive scanning images. Generally, this method is suitable and versatile for the other scanning cameras with a MMS LLA detector because of the similarity of the imaging mechanism. Full article
(This article belongs to the Special Issue Optical Remote Sensing Payloads, from Design to Flight Test)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Overall framework of the global-scale overlapping pixels calculation method for whisk-broom payloads with multi-module-staggered long-linear-array detectors.</p>
Full article ">Figure 2
<p>Schematic of Push-Broom and Whisk-Broom imaging systems. (<b>a</b>) Push-Broom imaging system; (<b>b</b>) Whisk-Broom imaging system.</p>
Full article ">Figure 3
<p>Diagram of rigorous geometric imaging model of whisk-broom camera. (<b>a</b>) In-orbit geometric imaging model; (<b>b</b>) Interior geometry of whisk-broom camera.</p>
Full article ">Figure 4
<p>Diagram of instantaneous relative motion for the ascending sub-module images. (<b>a</b>) Retrograde orbit (<math display="inline"><semantics> <mrow> <mi mathvariant="italic">θ</mi> </mrow> </semantics></math> &gt; 90°) forward-scanning; (<b>b</b>) Prograde orbit (<math display="inline"><semantics> <mrow> <mi mathvariant="italic">θ</mi> </mrow> </semantics></math> &lt; 90°) forward-scanning; (<b>c</b>) Relative motion of the pixel in stitching seam during forward-scanning; (<b>d</b>) Retrograde orbit (<math display="inline"><semantics> <mrow> <mi mathvariant="italic">θ</mi> </mrow> </semantics></math> &gt; 90°) backward-scanning; (<b>e</b>) Prograde orbit (<math display="inline"><semantics> <mrow> <mi mathvariant="italic">θ</mi> </mrow> </semantics></math> &lt; 90°) backward-scanning; (<b>f</b>) Relative motion of the pixel in stitching seam during backward-scanning.</p>
Full article ">Figure 5
<p>Diagram of instantaneous relative motion for the descending sub-module images. (<b>a</b>) Retrograde orbit (<math display="inline"><semantics> <mrow> <mi mathvariant="italic">θ</mi> </mrow> </semantics></math> &gt; 90°) forward-scanning; (<b>b</b>) Prograde orbit (<math display="inline"><semantics> <mrow> <mi mathvariant="italic">θ</mi> </mrow> </semantics></math> &lt; 90°) forward-scanning; (<b>c</b>) Relative motion of the pixel in stitching seam during forward-scanning; (<b>d</b>) Retrograde orbit (<math display="inline"><semantics> <mrow> <mi mathvariant="italic">θ</mi> </mrow> </semantics></math> &gt; 90°) backward-scanning; (<b>e</b>) Prograde orbit (<math display="inline"><semantics> <mrow> <mi mathvariant="italic">θ</mi> </mrow> </semantics></math> &lt; 90°) backward-scanning; (<b>f</b>) Relative motion of the pixel in stitching seam during backward-scanning.</p>
Full article ">Figure 6
<p>Diagram of the overlapping pixels calculation for adjacent detector sub-module images.</p>
Full article ">Figure 7
<p>Schematic diagram of the overlapping area of whisk-broom scanning imaging. (<b>a</b>) The overlap region between two consecutive scans in the same direction; (<b>b</b>) The overlap region between the forward scan and backward scan.</p>
Full article ">Figure 8
<p>Overlapping pixels change with orbital altitude and latitude. (<b>a</b>) Prograde, ascending, forward scanning (descending, backward-scanning); (<b>b</b>) Prograde, ascending, backward-scanning (descending, forward-scanning); (<b>c</b>) Retrograde, ascending, forward-scanning (descending, backward-scanning); (<b>d</b>) Retrograde, ascending, backward-scanning (descending, forward-scanning).</p>
Full article ">Figure 9
<p>Overlapping pixels vary with latitude at orbital altitude of 505 km.</p>
Full article ">Figure 10
<p>Overlapping pixels change with orbital altitude and the angular velocity of the scanning mirror. (<b>a</b>) Prograde, ascending, forward-scanning (descending, backward-scanning); (<b>b</b>) Prograde, ascending, backward-scanning (descending, forward-scanning); (<b>c</b>) Retrograde, ascending, forward-scanning (descending, backward-scanning); (<b>d</b>) Retrograde, ascending, backward-scanning (descending, forward-scanning).</p>
Full article ">Figure 11
<p>Overlapping pixels change with orbital altitude and orbital inclination. (<b>a</b>) Ascending, forward-scanning (descending, backward-scanning); (<b>b</b>) Ascending, backward-scanning (descending, forward-scanning).</p>
Full article ">Figure 12
<p>Overlapping pixels between the same scanning direction images change with orbital altitude and latitude. (<b>a</b>) Prograde; (<b>b</b>) Retrograde.</p>
Full article ">Figure 13
<p>Overlapping pixels between the same scanning direction images change with orbital altitude and scanning mirror velocity. (<b>a</b>) Prograde; (<b>b</b>) Retrograde.</p>
Full article ">Figure 14
<p>Overlapping pixels between the same scanning direction images change with orbital altitude and inclination.</p>
Full article ">Figure 15
<p>Overlapping pixels between the opposite scanning direction images change with orbital altitude and pixel location. (<b>a</b>) Prograde; (<b>b</b>) Retrograde.</p>
Full article ">Figure 16
<p>Overlapping pixels between the same scanning direction images change with scanning mirror velocity and pixel location. (<b>a</b>) Prograde; (<b>b</b>) Retrograde.</p>
Full article ">Figure 17
<p>Overlapping pixels of adjacent detector sub-module images during ascending. (<b>a</b>–<b>c</b>) are the overlapping positions of M1-M2, M2-M3, and M3-M4 during forward-scanning, respectively. (<b>d</b>–<b>f</b>) are the overlapping positions of M1-M2, M2-M3, and M3-M4 during backward-scanning, respectively.</p>
Full article ">Figure 18
<p>Overlapping pixels of adjacent detector sub-module images during descending. (<b>a</b>–<b>c</b>) are the overlapping positions of M1-M2, M2-M3, and M3-M4 during forward-scanning, respectively. (<b>d</b>–<b>f</b>) are the overlapping positions of M1-M2, M2-M3, and M3-M4 during backward-scanning, respectively.</p>
Full article ">Figure 19
<p>Overlapping pixels between the same scanning direction images. (<b>a</b>) Ascending, forward-scanning; (<b>b</b>) Ascending, backward-scanning; (<b>c</b>) Descending, forward-scanning; (<b>d</b>) Descending, backward-scanning.</p>
Full article ">Figure 20
<p>Overlapping pixels between the opposite scanning direction images. (<b>a</b>) Actual results; (<b>b</b>) Calculation results.</p>
Full article ">
18 pages, 3117 KiB  
Article
MonoDFNet: Monocular 3D Object Detection with Depth Fusion and Adaptive Optimization
by Yuhan Gao, Peng Wang, Xiaoyan Li, Mengyu Sun, Ruohai Di, Liangliang Li and Wei Hong
Sensors 2025, 25(3), 760; https://doi.org/10.3390/s25030760 - 27 Jan 2025
Abstract
Monocular 3D object detection refers to detecting 3D objects using a single camera. This approach offers low sensor costs, high resolution, and rich texture information, making it widely adopted. However, monocular sensors face challenges from environmental factors like occlusion and truncation, leading to [...] Read more.
Monocular 3D object detection refers to detecting 3D objects using a single camera. This approach offers low sensor costs, high resolution, and rich texture information, making it widely adopted. However, monocular sensors face challenges from environmental factors like occlusion and truncation, leading to reduced detection accuracy. Additionally, the lack of depth information poses significant challenges for predicting 3D positions. To address these issues, this paper presents a monocular 3D object detection method based on improvements to MonoCD, designed to enhance detection accuracy and robustness in complex environments. In order to effectively obtain and integrate depth information, this paper designs a multi-branch depth prediction with weight sharing module. Furthermore, an adaptive focus mechanism is proposed to emphasize target regions while minimizing interference from irrelevant areas. The experimental results demonstrate that MonoDFNet achieves significant improvements over existing methods, with AP3D gains of +4.09% (Easy), +2.78% (Moderate), and +1.63% (Hard), confirming its effectiveness in 3D object detection. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Figure 1
<p>Schematic of monocular 3D object detection.</p>
Full article ">Figure 2
<p>Workflow for parameter selection of <span class="html-italic">α</span>, <span class="html-italic">β</span>, and <span class="html-italic">γ</span>.</p>
Full article ">Figure 3
<p>MonoDFNet architecture overview.</p>
Full article ">Figure 4
<p>Architecture of the adaptive focus mechanism.</p>
Full article ">Figure 5
<p>Comparison of feature map processing effects.</p>
Full article ">Figure 6
<p>Architecture of the multi-branch depth prediction with weight sharing module.</p>
Full article ">Figure 7
<p>Feature visualization and depth map fusion in the multi-branch architecture.</p>
Full article ">Figure 8
<p>Detection visualization results. <math display="inline"><semantics> <mrow> <msub> <mi>z</mi> <mrow> <mi>s</mi> <mi>o</mi> <mi>f</mi> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math> : simple fusion based on SoftMax; <math display="inline"><semantics> <mrow> <msub> <mi>z</mi> <mrow> <mi>d</mi> <mi>i</mi> <mi>r</mi> </mrow> </msub> </mrow> </semantics></math> dynamic fusion strategy; <math display="inline"><semantics> <mrow> <msub> <mi>z</mi> <mrow> <mi>k</mi> <mi>e</mi> <mi>y</mi> </mrow> </msub> </mrow> </semantics></math>: depth fusion based on key features; <math display="inline"><semantics> <mrow> <msub> <mi>z</mi> <mrow> <mi>c</mi> <mi>o</mi> <mi>m</mi> <mi>p</mi> </mrow> </msub> </mrow> </semantics></math>: final fusion results combining multiple methods.</p>
Full article ">
22 pages, 19013 KiB  
Article
Exploring Inequality: A Multi-Scale Analysis of China’s Consumption Carbon Footprint
by Feng Xu, Xinqi Zheng, Minrui Zheng, Dongya Liu, Yin Ma, Jizong Peng, Ye Shen, Xu Han and Mengdi Zhang
ISPRS Int. J. Geo-Inf. 2025, 14(2), 49; https://doi.org/10.3390/ijgi14020049 - 26 Jan 2025
Abstract
Carbon emission inequality has become a critical factor constraining the coordinated development of socio-economic systems and the natural environment. This inequality exacerbates the disparity in carbon emissions across regions, hindering efforts to achieve sustainable development and environmental justice. Previous research has primarily focused [...] Read more.
Carbon emission inequality has become a critical factor constraining the coordinated development of socio-economic systems and the natural environment. This inequality exacerbates the disparity in carbon emissions across regions, hindering efforts to achieve sustainable development and environmental justice. Previous research has primarily focused on the structure of carbon footprints and their influencing factors, but there has been limited quantitative research on carbon emission inequality, particularly from a multi-scale perspective. This study constructs a 250 m-high-resolution consumption-based carbon footprint grid for China and uses the Theil index to reveal significant spatial inequalities in carbon footprints. The results indicate that smaller-scale analyses better reveal the spatiotemporal heterogeneity of carbon footprints within regions. At the county level, carbon footprints exhibit significant inequalities, with hotspots concentrated in regions such as Beijing–Tianjin–Hebei, the Yangtze River Delta, and the Pearl River Delta. The top 5% of areas with the highest carbon footprints (139 cities) contributed 19.6% of the national total, indicating a concentration in a few large cities. The decomposition of the Theil index shows that county-level cities contributed 55% of the national carbon inequality. The study also reveals the complex relationship between carbon footprints and income, as well as urban-rural disparities. The underdeveloped central and western regions exhibit a pronounced spatial lag effect, with the growth rate of carbon footprints in rural areas surpassing that of urban areas. Carbon footprints in impoverished areas and inter-provincial marginal areas overlap significantly with low-emission zones, demonstrating characteristics of “low-carbon growth”. To achieve carbon peak and carbon neutrality targets, China must adopt comprehensive measures to reduce carbon footprints and their inequalities, including strengthening multi-scale carbon inequality monitoring, implementing differentiated carbon reduction policies, and promoting coordinated emission reduction development at the county level. Full article
Show Figures

Figure 1

Figure 1
<p>Grid model of China’s fine-scale consumption-based carbon footprints.</p>
Full article ">Figure 2
<p>Technical process flowchart.</p>
Full article ">Figure 3
<p>High-resolution (250 m) Grid Distribution of 2017 Consumption-Based Carbon Footprints and Kernel Density Analysis of Carbon Emissions.</p>
Full article ">Figure 4
<p>Spatial distribution of county-level consumption-based carbon footprints and—Lorenz curve in 2017.</p>
Full article ">Figure 5
<p>Hotspot analysis of county-level total carbon emissions and per capita carbon footprints i-n 2017.</p>
Full article ">Figure 6
<p>Provincial consumption-based carbon emissions and emission structures—in 2015 and 2017.</p>
Full article ">Figure 7
<p>The Theil Index and Hotspot Analysis of County-level Carbon Footprints in 2017.</p>
Full article ">Figure 8
<p>Spatial lag analysis of carbon emissions and disposable income in county-level cities.</p>
Full article ">Figure 9
<p>Boxplot of urban and rural per capita carbon footprints by province in China.</p>
Full article ">Figure 10
<p>Distribution of national PSCs in China and scatter plot of linear regression between carbon emissions and income for PSCs and non-PSCs. (The blue line represents the trend of the correlation between carbon emissions and disposable income, the black dots indicate the disposable income and carbon emissions of individual county-level cities, the green bar chart shows the distribution of income data, and the orange bar chart illustrates the distribution of carbon emissions data).</p>
Full article ">
22 pages, 11926 KiB  
Article
PJ-YOLO: Prior-Knowledge and Joint-Feature-Extraction Based YOLO for Infrared Ship Detection
by Yongjie Liu, Chaofeng Li and Guanghua Fu
J. Mar. Sci. Eng. 2025, 13(2), 226; https://doi.org/10.3390/jmse13020226 - 25 Jan 2025
Viewed by 233
Abstract
Infrared ship images have low resolution and limited recognizable features, especially for small targets, leading to low accuracy and poor generalization of traditional detection methods. To address this, we design a prior knowledge auxiliary loss for leveraging the unique brightness distribution of infrared [...] Read more.
Infrared ship images have low resolution and limited recognizable features, especially for small targets, leading to low accuracy and poor generalization of traditional detection methods. To address this, we design a prior knowledge auxiliary loss for leveraging the unique brightness distribution of infrared ship images, we construct a joint feature extraction module that sufficiently captures context awareness, channel differentiation, and global information, and then we propose a prior-knowledge- and joint-feature-extraction-based YOLO (PJ-YOLO) for use in detecting infrared ships. Additionally, a residual deformable attention module is designed to integrate multi-scale information, enhancing detail capture. Experimental results on the SFISD and InfiRray Ships datasets demonstrate that the proposed PJ-YOLO achieves state-of-the-art detection performance for infrared ship targets. In particular, PJ-YOLO achieves improvements of 1.6%, 5.0%, and 2.8% in mAP50, mAP75, and mAP50:95 on the SFISD dataset, respectively. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

Figure 1
<p>Overall architecture of PJ-YOLO.</p>
Full article ">Figure 2
<p>Structure of the JFE module.</p>
Full article ">Figure 3
<p>Structure of the R-DA module.</p>
Full article ">Figure 4
<p>Brightness distribution of the infrared ship image.</p>
Full article ">Figure 5
<p>Mean brightness of the bounding box. In the right image, (<b>a</b>–<b>e</b>) are extracted from the left image, each containing the ground truth (GT) bounding box and the predicted bounding box. Green numbers represent the mean brightness values of pixels within the ground truth (GT) bounding boxes, while red numbers represent the mean brightness values of pixels within the predicted bounding boxes. Blue numbers indicate the difference in mean brightness between the GT and predicted bounding boxes. A larger difference value indicates a greater discrepancy in brightness distribution.</p>
Full article ">Figure 6
<p>Bounding box adjustment process in PKA Loss: green box (GT) and red box (predicted).</p>
Full article ">Figure 7
<p>The size distribution of targets in the datasets: (<b>a</b>) the SFISD dataset and (<b>b</b>) the InfiRray Ships dataset. The darker the dot color, the larger the quantity of targets with normalized width and height.</p>
Full article ">Figure 8
<p>Visual comparison of results for the SFISD dataset. False negative samples are highlighted with blue circles, and false positive samples are indicated with purple circles.</p>
Full article ">Figure 9
<p>Visual comparison of results for the InfiRray Ships dataset. False negative samples are highlighted with blue circles, and false positive samples are indicated with purple circles.</p>
Full article ">Figure 10
<p>Grad-CAM visualization results for different stages of the baseline network and PJ-YOLO across diverse scenarios in the InfiRray Ships dataset: (<b>a</b>) single-target scenario and (<b>b</b>) multi-target scenario.</p>
Full article ">Figure 11
<p>Visualization results of the ablation study from multiple scenarios.</p>
Full article ">Figure 12
<p>Comparison of latency, FLOPs, and mAP<sub>50:95</sub> (on the SFISD dataset) between PJ-YOLO and other competing methods. The number of FLOPs is represented by the radius of the circle.</p>
Full article ">Figure 13
<p>Visualization comparison of prediction results on real-world images.</p>
Full article ">
Back to TopTop