Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (201)

Search Parameters:
Keywords = Second-Generation Global Imager

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 12823 KiB  
Article
Remote Sensing Small Object Detection Network Based on Multi-Scale Feature Extraction and Information Fusion
by Junsuo Qu, Tong Liu, Zongbing Tang, Yifei Duan, Heng Yao and Jiyuan Hu
Remote Sens. 2025, 17(5), 913; https://doi.org/10.3390/rs17050913 - 5 Mar 2025
Viewed by 223
Abstract
Nowadays, object detection algorithms are widely used in various scenarios. However, there are further small object detection requirements in some special scenarios. Due to the problems related to small objects, such as their less available features, unbalanced samples, higher positioning accuracy requirements, and [...] Read more.
Nowadays, object detection algorithms are widely used in various scenarios. However, there are further small object detection requirements in some special scenarios. Due to the problems related to small objects, such as their less available features, unbalanced samples, higher positioning accuracy requirements, and fewer data sets, a small object detection algorithm is more complex than a general object detection algorithm. The detection effect of the model for small objects is not ideal. Therefore, this paper takes YOLOXs as the benchmark network and enhances the feature information on small objects by improving the network’s structure so as to improve the detection effect of the model for small objects. This specific research is presented as follows: Aiming at the problem of a neck network based on an FPN and its variants being prone to information loss in the feature fusion of non-adjacent layers, this paper proposes a feature fusion and distribution module, which replaces the information transmission path, from deep to shallow, in the neck network of YOLOXs. This method first fuses and extracts the feature layers used by the backbone network for prediction to obtain global feature information containing multiple-size objects. Then, the global feature information is distributed to each prediction branch to ensure that the high-level semantic and fine-grained information are more efficiently integrated so as to help the model effectively learn the discriminative information on small objects and classify them correctly. Finally, after testing on the VisDrone2021 dataset, which corresponds to a standard image size of 1080p (1920 × 1080), the resolution of each image is high and the video frame rate contained in the dataset is usually 30 frames/second (fps), with a high resolution in time, it can be used to detect objects of various sizes and for dynamic object detection tasks. And when we integrated the module into a YOLOXs network (named the FE-YOLO network) with the three improvement points of the feature layer, channel number, and maximum pool, the mAP and APs were increased by 1.0% and 0.8%, respectively. Compared with YOLOV5m, YOLOV7-Tiny, FCOS, and other advanced models, it can obtain the best performance. Full article
Show Figures

Figure 1

Figure 1
<p>Two common feature fusion diagrams.</p>
Full article ">Figure 2
<p>Convolution and deconvolution diagram.</p>
Full article ">Figure 3
<p>PANet information fusion diagram.</p>
Full article ">Figure 4
<p>The schematic diagram of the improved neck network.</p>
Full article ">Figure 5
<p>Schematic diagram of FFDN module.</p>
Full article ">Figure 6
<p>Map of mAP0.5 during the training of Fe-YOLO and FFDN-YOLO.</p>
Full article ">Figure 7
<p>Comparison between FE-YOLO and FFDN-YOLO models.</p>
Full article ">Figure 8
<p>Comparison diagram of model detection and manual labeling.</p>
Full article ">Figure 9
<p>Comparison chart of detection results of different models.</p>
Full article ">Figure 10
<p>Variation diagram of loss value during model training.</p>
Full article ">Figure 11
<p>Dataset test diagram.</p>
Full article ">
21 pages, 15325 KiB  
Article
Spatiotemporal Variations in Sea Ice Albedo: A Study of the Dynamics of Sea Ice Albedo in the Sea of Okhotsk
by Yingzhen Zhou, Wei Li, Nan Chen, Takenobu Toyota, Yongzhen Fan, Tomonori Tanikawa and Knut Stamnes
Remote Sens. 2025, 17(5), 772; https://doi.org/10.3390/rs17050772 - 23 Feb 2025
Viewed by 176
Abstract
This study utilizes a novel albedo retrieval framework combining radiative transfer modeling with scientific machine learning (RTM-SciML) to investigate sea ice dynamics in the Sea of Okhotsk. By validating albedo estimates derived from the MODIS sensor against in situ pyranometer measurements near the [...] Read more.
This study utilizes a novel albedo retrieval framework combining radiative transfer modeling with scientific machine learning (RTM-SciML) to investigate sea ice dynamics in the Sea of Okhotsk. By validating albedo estimates derived from the MODIS sensor against in situ pyranometer measurements near the Hokkaido coast, we achieved a robust Pearson coefficient of 0.86 and an RMSE of 0.089 for all sea ice types, with even higher correlations for specific surfaces like snow-covered ice (Pearson-r = 0.89) and meltwater/open water (Pearson-r = 0.90). This confirms the framework’s efficacy across varying surface conditions. Cross-sensor comparisons between MODIS and the Second-Generation Global Imager (SGLI) further demonstrated its consistency, achieving an overall Pearson-r of 0.883 and RMSE of 0.036. Integrating these albedo estimates with sea ice concentration data from the Advanced Microwave Scanning Radiometer 2 (AMSR-2), we analyzed the complex role of the Sea of Okhotsk’s polynya systems and ice interactions in regional climate processes. Our results highlight the synergistic advantage of pairing optical sensors, like MODIS and SGLI, with microwave sensors, offering a more comprehensive understanding of evolving sea ice conditions and paving the way for future climate and cryosphere studies. Full article
(This article belongs to the Special Issue Monitoring Sea Ice Loss with Remote Sensing Techniques)
Show Figures

Figure 1

Figure 1
<p>Map of the Sea of Okhotsk. The 200 m and 1000 m isobaths are indicated by red and blue dashed lines. The red box delineates the specific region where the PV Soya Icebreaker predominantly operated between 2002 and 2015 (see <a href="#sec2dot2-remotesensing-17-00772" class="html-sec">Section 2.2</a>). The orange box (i) represents the Tartar Strait region, and the green box (ii) corresponds to the Northern Polynya, indicating areas selected for specific studies (see <a href="#sec5dot1-remotesensing-17-00772" class="html-sec">Section 5.1</a>, <a href="#sec5dot2-remotesensing-17-00772" class="html-sec">Section 5.2</a> and <a href="#sec5dot3-remotesensing-17-00772" class="html-sec">Section 5.3</a>).</p>
Full article ">Figure 2
<p>(<b>a</b>) Close-up map of the region in the Sea of Okhotsk (the red box in <a href="#remotesensing-17-00772-f001" class="html-fig">Figure 1</a>). The trace colors indicate the different navigational paths of the Soya Icebreaker in the years between 2002 and 2015, with the corresponding voyage years indicated in the colorbar placed at the bottom. (<b>b</b>) A detailed view of the PV Soya navigating through the Sea of Okhotsk, surrounded by sea ice and polynyas. The red circle on the tip of the ship highlights the location of the EKO MR-40 pyranometer that was used for the irradiance measurements.</p>
Full article ">Figure 3
<p>Correlation between shortwave albedo measurements from the Soya Icebreaker and RTM-SciML retrievals. Panels (<b>a</b>,<b>b</b>) display the results with maximum time differences of three hours and one hour, respectively, between the measurement time and the MODIS overpass time. The color indicates the time interval between pyranometer measurements and satellite overpass. On the top left, the correlation equation (<math display="inline"><semantics> <mrow> <mi>y</mi> <mo>=</mo> <mi>a</mi> <mo>·</mo> <mi>x</mi> <mo>+</mo> <mi>b</mi> </mrow> </semantics></math>), Pearson <span class="html-italic">r</span> coefficient, root mean square error (RMSE), and the number of pixels (<span class="html-italic">N</span>) used to calculate the statistics are provided. The solid black lines represent (0,0)–(1,1), and the dashed black lines represent the 15% error range.</p>
Full article ">Figure 4
<p>Representative visualizations from 10 February 2004 (<b>a</b>) and 9 February 2008 (<b>b</b>). From top to bottom, each column displays: (<b>1</b>) true color RGB maps constructed using MODIS channels 645 nm, 555 nm, and 469 nm as the (R, G, B) bands, respectively; (<b>2</b>) surface classification maps with the spatial overlay of the Soya voyage shown in purple; (<b>3</b>) albedo retrieval maps with the spatial overlay of pyranometer measured values on RTM-SciML retrieved albedo; and (<b>4</b>) scatter–dot comparisons between the measurements and retrievals. On (<b>a-3</b>,<b>b-3</b>), the boxed A and B indicate the start and end points of the ship.</p>
Full article ">Figure 5
<p>Representative visualizations from 14 February 2002 (<b>a</b>) and 11 February 2008 (<b>b</b>). From top to bottom, each column displays: (<b>1</b>) true color RGB maps constructed using MODIS channels 645 nm, 555 nm, and 469 nm as the (R, G, B) bands, respectively; (<b>2</b>) surface classification maps with the spatial overlay of the Soya voyage shown in purple; (<b>3</b>) albedo retrieval maps with the spatial overlay of pyranometer measured values on RTM-SciML retrieved albedo; and (<b>4</b>) scatter–dot comparisons between the measurements and retrievals. On (<b>a-3</b>,<b>b-3</b>), the boxed A and B indicate the start and end points of the ship.</p>
Full article ">Figure 6
<p>Comparison of shortwave albedo measurements between the Soya Icebreaker and RTM-SciML retrievals represented as scatter plots. Each panel indicates the Pearson <span class="html-italic">r</span> coefficient and the number of pixels (<span class="html-italic">N</span>) at the top left. The dotted black lines delineate the MODIS overpass time. The <span class="html-italic">x</span>-axis across all panels displays the pyranometer measurement time (UTC).</p>
Full article ">Figure 7
<p>Density plots illustrating the correlation between MODIS and SGLI albedo retrievals. Top two rows (subfigures 1): Bare sea ice; bottom two rows (subfigures 2): Melt-water/Water. (<b>a</b>–<b>h</b>) represents the eight time periods discussed in the main text. Each plot provides the Pearson correlation coefficient (r) and the root mean square error (RMSE) for the respective time periods on the top left.</p>
Full article ">Figure 8
<p>Density plots comparing albedo retrievals from MODIS and SGLI sensors for different surface types over the total observation period from January to May 2021. (<b>a</b>) from bare ice surface, (<b>b</b>) from snow-covered sea ice surface, (<b>c</b>) from meltwater or open water and (<b>d</b>) from all valid sea-ice/water surfaces combined.</p>
Full article ">Figure 9
<p>Comprehensive visualization of various parameters over the Sea of Okhotsk. From top to bottom, the rows depict (<b>a</b>) shortwave albedo, (<b>b</b>) surface classification, (<b>c</b>) brightness temperature as captured by SGLI’s <math display="inline"><semantics> <mrow> <mn>10.8</mn> <mspace width="0.166667em"/> <mi mathvariant="sans-serif">μ</mi> </mrow> </semantics></math>m channel, and (<b>d</b>) high-resolution sea ice concentration from AMSR-2. Each column (numbers 1–7) corresponds to distinct retrieval periods as detailed in <a href="#remotesensing-17-00772-t004" class="html-table">Table 4</a>, showcasing the evolution of sea ice conditions.</p>
Full article ">Figure 10
<p>Average surface albedo of sea ice (left axis) and pixel percentage of the sea ice with different subtypes (right axis) during 1 April 2021∼7 April 2021. The colored bar plots and the labelled texts show the composition of sea ice. The black line is the relation between the average albedo (of all subtypes) and the SIC level. Error bars are the standard deviations of albedo.</p>
Full article ">Figure 11
<p>(<b>a</b>) Snow pixel albedo values and their coverage percentages. (<b>b</b>) Sea ice (bare ice and ice with melt-water coverage) pixel albedo values and their percentages. The blue line and shadings are the relation and RMSE of 0.062 derived by [<a href="#B4-remotesensing-17-00772" class="html-bibr">4</a>].</p>
Full article ">Figure 12
<p>Probability density curves of albedo in (<b>a</b>) Tatar Strait and the NWP adjoining the Japan Sea and (<b>b</b>) northern polynya. The number of pixels used to generate the curves in each panel are labelled at the top.</p>
Full article ">Figure 13
<p>A detailed examination of the Tatar Strait region, derived from the broader overview provided in <a href="#remotesensing-17-00772-f009" class="html-fig">Figure 9</a>. The rows from top to bottom feature (<b>a</b>–<b>f</b>) shortwave albedo, (<b>g</b>–<b>l</b>) surface classification, and (<b>m</b>–<b>r</b>) AMSR-2 sea ice concentration at 15 km resolution, alongside (<b>s</b>–<b>x</b>) brightness temperature data from SGLI’s 10.8 µm channel. Color-bars for each parameter are included for reference at the bottom of their respective rows.</p>
Full article ">
28 pages, 1296 KiB  
Article
Fidex and FidexGlo: From Local Explanations to Global Explanations of Deep Models
by Guido Bologna, Jean-Marc Boutay, Damian Boquete, Quentin Leblanc, Deniz Köprülü and Ludovic Pfeiffer
Algorithms 2025, 18(3), 120; https://doi.org/10.3390/a18030120 - 20 Feb 2025
Viewed by 194
Abstract
Deep connectionist models are characterized by many neurons grouped together in many successive layers. As a result, their data classifications are difficult to understand. We present two novel algorithms which explain the responses of several black-box machine learning models. The first is Fidex, [...] Read more.
Deep connectionist models are characterized by many neurons grouped together in many successive layers. As a result, their data classifications are difficult to understand. We present two novel algorithms which explain the responses of several black-box machine learning models. The first is Fidex, which is local and thus applied to a single sample. The second, called FidexGlo, is global and uses Fidex. Both algorithms generate explanations by means of propositional rules. In our framework, the discriminative boundaries are parallel to the input variables and their location is precisely determined. Fidex is a heuristic algorithm that, at each step, establishes where the best hyperplane is that has increased fidelity the most. The algorithmic complexity of Fidex is proportional to the maximum number of steps, the number of possible hyperplanes, which is finite, and the number of samples. We first used FidexGlo with ensembles and support vector machines (SVMs) to show that its performance on three benchmark problems is competitive in terms of complexity, fidelity and accuracy. The most challenging part was then to apply it to convolutional neural networks. We achieved this with three classification problems based on images. We obtained accurate results and described the characteristics of the rules generated, as well as several examples of explanations illustrated with their corresponding images. To the best of our knowledge, this is one of the few works showing a global rule extraction technique applied to both ensembles, SVMs and deep neural networks. Full article
(This article belongs to the Special Issue Machine Learning Algorithms for Image Understanding and Analysis)
Show Figures

Figure 1

Figure 1
<p>A DIMLP network that discriminates circles and triangles. The step activation function for neuron <span class="html-italic">h</span> creates a hyperplane at <math display="inline"><semantics> <mrow> <mi>x</mi> <mo>=</mo> <mn>3</mn> </mrow> </semantics></math>. Generally, the presence or absence of this hyperplane depends on the values of <math display="inline"><semantics> <msup> <mi>w</mi> <mo>′</mo> </msup> </semantics></math> and <math display="inline"><semantics> <msup> <mi>b</mi> <mo>′</mo> </msup> </semantics></math>.</p>
Full article ">Figure 2
<p>Schematic representation of the global explainability for a DIMLP model. From a trained DIMLP, the hyperplanes are precisely located from the weights of the first hidden layer; then, FidexGlo runs Fidex for all the training samples. Finally, by removing as many rules as possible, FidexGlo simplifies the rules generated by Fidex.</p>
Full article ">Figure 3
<p>An example of ruleset generated by FidexGlo applied to an ensemble of 25 DIMLPs with a single hidden layer.</p>
Full article ">Figure 4
<p>An example of ruleset generated by FidexGlo applied to an ensemble of 100 DTs of depth three trained by gradient boosting.</p>
Full article ">Figure 5
<p>Variation of the dropout parameters (<span class="html-italic">p</span> and <span class="html-italic">q</span>) for the “Ionosphere” classification problem trained with DIMLP ensembles. The average number of rules is shown on the <b>top left</b>, the average number of antecedents per rule on the <b>top right</b>, the average fidelity on the <b>bottom left</b>, and the average prediction accuracy on the <b>bottom right</b>.</p>
Full article ">Figure 6
<p>Variation of the dropout parameters (<span class="html-italic">p</span> and <span class="html-italic">q</span>) for the “Ionosphere” classification problem trained with random forests.</p>
Full article ">Figure 7
<p>Samples covered by a rule of class ‘1’; rule antecedents are represented by colored dots. The top left shows the centroid of the 1426 covered samples in the training set. The other pictures show three of the two hundred forty-two samples in the test set that activate this rule.</p>
Full article ">Figure 8
<p>Samples covered by a rule of class ‘0’; rule antecedents are represented by colored dots. The top left shows the centroid of the 1114 covered samples in the training set. The other pictures show three of the one hundred eighty-three samples in the test set that activate this rule.</p>
Full article ">Figure 9
<p>Samples covered by a rule of class ‘4’; rule antecedents are represented by colored dots. The top left shows the centroid of the 565 covered samples in the training set. The other pictures show three of the one hundred three samples in the test set that activate this rule.</p>
Full article ">Figure 10
<p>Samples covered by a rule of class ‘Happy’; rule antecedents are represented by colored dots. The top left shows the centroid of the twenty-seven covered samples in the training set, while the other pictures show three of the eight samples in the test set that activate this rule.</p>
Full article ">Figure 11
<p>Samples covered by a rule of class ‘Happy’; rule antecedents are represented by colored dots. The top left shows the centroid of the thirty covered samples in the training set, while the other pictures show three of the six samples in the test set that activate this rule.</p>
Full article ">Figure 12
<p>Samples covered by a rule of class ‘Happy’; rule antecedents are represented by colored dots. The top left shows the centroid of the twenty-nine covered samples in the training set, while the other pictures show three of the eight samples in the test set that activate this rule.</p>
Full article ">Figure 13
<p>Samples covered by a rule of class ‘Cracks’; rule antecedents are represented by colored dots. The top left shows the centroid of the 4549 covered samples in the training set. The other pictures show three of the one thousand one hundred twenty samples in the test set that activate this rule.</p>
Full article ">Figure 13 Cont.
<p>Samples covered by a rule of class ‘Cracks’; rule antecedents are represented by colored dots. The top left shows the centroid of the 4549 covered samples in the training set. The other pictures show three of the one thousand one hundred twenty samples in the test set that activate this rule.</p>
Full article ">Figure 14
<p>Samples covered by a rule of class ‘Cracks’; rule antecedents are represented by colored dots. The top left shows the centroid of the 4429 covered samples in the training set. The other pictures show three of the one thousand seventy samples in the test set that activate this rule.</p>
Full article ">
20 pages, 2628 KiB  
Review
Confocal Laser Endomicroscopy: Enhancing Intraoperative Decision Making in Neurosurgery
by Francesco Carbone, Nicola Pio Fochi, Giuseppe Di Perna, Arthur Wagner, Jürgen Schlegel, Elena Ranieri, Uwe Spetzger, Daniele Armocida, Fabio Cofano, Diego Garbossa, Augusto Leone and Antonio Colamaria
Diagnostics 2025, 15(4), 499; https://doi.org/10.3390/diagnostics15040499 - 19 Feb 2025
Viewed by 338
Abstract
Brain tumors, both primary and metastatic, represent a significant global health burden due to their high incidence, mortality, and the severe neurological deficits they frequently cause. Gliomas, especially high-grade gliomas (HGGs), rank among the most aggressive and lethal neoplasms, with only modest gains [...] Read more.
Brain tumors, both primary and metastatic, represent a significant global health burden due to their high incidence, mortality, and the severe neurological deficits they frequently cause. Gliomas, especially high-grade gliomas (HGGs), rank among the most aggressive and lethal neoplasms, with only modest gains in long-term survival despite extensive molecular research and established standard therapies. In neurosurgical practice, maximizing the extent of safe resection is a principal strategy for improving clinical outcomes. Yet, the infiltrative nature of gliomas often complicates the accurate delineation of tumor margins. Confocal laser endomicroscopy (CLE), originally introduced in gastroenterology, has recently gained prominence in neuro-oncology by enabling real-time, high-resolution cellular imaging during surgery. This technique allows for intraoperative tumor characterization and reduces dependence on time-consuming frozen-section analyses. Recent technological advances, including device miniaturization and second-generation CLE systems, have substantially improved image quality and diagnostic utility. Furthermore, integration with deep learning algorithms and telepathology platforms fosters automated image interpretation and remote expert consultations, thereby accelerating surgical decision making and enhancing diagnostic consistency. Future work should address remaining challenges, such as mitigating motion artifacts, refining training protocols, and broadening the range of applicable fluorescent probes, to solidify CLE’s role as a critical intraoperative adjunct in neurosurgical oncology. Full article
(This article belongs to the Special Issue Confocal Microscopy: Clinical Impacts and Innovation—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Time map of the publications investigating the roles and limitations of CLE in neuro-oncology over the years.</p>
Full article ">Figure 2
<p>(<b>A</b>) Glioblastoma. High cellularity of pleomorphic tumor cells, with fluorescence enriched in tumor matrix. (<b>B</b>) Meningioma. Prominent psammoma bodies typical for psammomatous meningiomas. (<b>C</b>) Schwannoma. Pseudo-palisadal arrangement of tumor cells with parallel fibrous structures. (<b>D</b>) Metastasis of Squamous Cell Carcinoma. Squamous growth pattern of tumor cells, with fluorescence enhancement in tumor cells’ cytoplasms.</p>
Full article ">
23 pages, 12422 KiB  
Article
Mapping Coastal Marine Habitats Using UAV and Multispectral Satellite Imagery in the NEOM Region, Northern Red Sea
by Emma Sullivan, Nikolaos Papagiannopoulos, Daniel Clewley, Steve Groom, Dionysios E. Raitsos and Ibrahim Hoteit
Remote Sens. 2025, 17(3), 485; https://doi.org/10.3390/rs17030485 - 30 Jan 2025
Viewed by 813
Abstract
Effective management to conserve marine environments requires up-to-date information on the location, distribution, and extent of major benthic habitats. Remote sensing is a key tool for such assessments, enabling consistent, repeated measurements over large areas. There is particular interest in using freely available [...] Read more.
Effective management to conserve marine environments requires up-to-date information on the location, distribution, and extent of major benthic habitats. Remote sensing is a key tool for such assessments, enabling consistent, repeated measurements over large areas. There is particular interest in using freely available satellite images such as from the Copernicus Sentinel-2 series for accessible repeat assessments. In this study, an area of 438 km2 of the northern Red Sea coastline, adjacent to the NEOM development was mapped using Sentinel-2 imagery. A hierarchical Random Forest classification method was used, where the initial level classified pixels into a geomorphological class, followed by a second level of benthic cover classification. Uncrewed Aerial Vehicle (UAV) surveys were carried out in 12 locations in the NEOM area to collect field data on benthic cover for training and validation. The overall accuracy of the geomorphic and benthic classifications was 84.15% and 72.97%, respectively. Approximately 12% (26.26 km2) of the shallow Red Sea study area was classified as coral or dense algae and 16% (36.12 km2) was classified as rubble. These reef environments offer crucial ecosystem services and are believed to be internationally important as a global warming refugium. Seagrass meadows, covering an estimated 29.17 km2 of the study area, play a regionally significant role in carbon sequestration and are estimated to store 200 tonnes of carbon annually, emphasising the importance of their conservation for meeting the environmental goals of the NEOM megaproject. This is the first map of this region generated using Sentinel-2 data and demonstrates the feasibility of using an open source and reproducible methodology for monitoring coastal habitats in the region. The use of training data derived from UAV imagery provides a low-cost and time-efficient alternative to traditional methods of boat or snorkel surveys for covering large areas in remote sites. Full article
(This article belongs to the Topic Conservation and Management of Marine Ecosystems)
Show Figures

Figure 1

Figure 1
<p>Map of Uncrewed Aerial Vehicle sampling stations in NEOM. Sites 2, 4, 6, 7, 9, and 11 were used for training, while sites 1, 3, 5, 8, 10, and 12 were used for validation. Satellite image ESRI basemap is from Earthstar Geographics.</p>
Full article ">Figure 2
<p>Flowchart diagram summarising the methodology adopted for the supervised classification of Red Sea benthic habitats in the NEOM area.</p>
Full article ">Figure 3
<p>(<b>a</b>) True colour composite image of Tiran Island, northern Red Sea, after atmospheric correction and compositing scenes. (<b>b</b>) False colour image of the same area using depth invariant indices. The red channel of the image is the green–red DII, the green channel is the blue–red DII, and the blue channel is the green–blue DII. Land and deep water are masked in black.</p>
Full article ">Figure 4
<p>Geomorphic classification map in Sindalah/Strait of Tiran in the NEOM region, Red Sea. The six classes mapped are reef crest, inner and outer reef flat, sand/mud flat, shallow lagoon, and deep water.</p>
Full article ">Figure 5
<p>Benthic habitat classification map in Sindalah/Strait of Tiran in the NEOM region, Red Sea. The six benthic classes mapped are coral and dense algae, rubble, seagrass, lagoonal sands, sand/mud flats, and rock (or other hard substrates).</p>
Full article ">Figure 6
<p>Geomorphic classification map in Sharma, a semi-enclosed tropical lagoon in the NEOM region, Red Sea. The six mapped classes are reef crest, inner and outer reef flat, sand/mud flat, shallow lagoon, and deep water.</p>
Full article ">Figure 7
<p>Benthic habitat classification map in Sharma, a semi-enclosed tropical lagoon in the NEOM region, Red Sea. The six benthic classes mapped are coral and dense algae, rubble, seagrass, lagoonal sands, sand/mud flats, and rock (or other hard substrates).</p>
Full article ">Figure 8
<p>Comparison of three different benthic classification maps covering Sanafir Island, in the northern Red Sea: (<b>a</b>) this study, (<b>b</b>) Allen Coral Atlas, and (<b>c</b>) JICA and NCWCD 2000.</p>
Full article ">Figure 9
<p>Comparison of three different benthic classification maps for a coastal stretch near Gayal in the NEOM region, Red Sea, showing sills and depressions of reticulated reef structures: (<b>a</b>) this study, (<b>b</b>) Allen Coral Atlas, and (<b>c</b>) JICA and NCWCD 2000.</p>
Full article ">
57 pages, 13137 KiB  
Article
Compositional and Numerical Geomorphology Along a Basement–Foreland Transition, SE Germany, with Special Reference to Landscape-Forming Indices and Parameters in Genetic and Applied Terrain Analyses
by Harald G. Dill, Andrei Buzatu, Sorin-Ionut Balaban and Christopher Kleyer
Geosciences 2025, 15(2), 37; https://doi.org/10.3390/geosciences15020037 - 23 Jan 2025
Viewed by 558
Abstract
The Münchberg Gneiss Complex (Central European Variscides, Germany) is separated by a deep-seated lineamentary fault zone, the Franconian Lineamentary Fault Zone, from its Mesozoic foreland. The study area offers insight into a great variety of landforms created by fluvial and mass wasting processes [...] Read more.
The Münchberg Gneiss Complex (Central European Variscides, Germany) is separated by a deep-seated lineamentary fault zone, the Franconian Lineamentary Fault Zone, from its Mesozoic foreland. The study area offers insight into a great variety of landforms created by fluvial and mass wasting processes together with their bedrocks, covering the full range from unmetamorphosed sediments to high-grade regionally metamorphic rocks. It renders the region an ideal place to conduct a study of compositional and numerical geomorphology and their landscape-forming indices and parameters. The landforms under consideration are sculpted out of the bedrocks (erosional landforms) and overlain by depositional landforms which are discussed by means of numerical landform indices (LFIs), all of which are coined for the first time in the current paper. They are designed to be suitable for applied geosciences such as extractive/economic geology as well as environmental geology. The erosional landform series are subdivided into three categories: (1) The landscape roughness indices, e.g., VeSival (vertical sinuosity—valley of landform series) and the VaSlAnalti (variation in slope angle altitude), which are used for a first order classification of landscapes into relief generations. The second order classification LFIs are devoted to the material properties of the landforms’ bedrocks, such as the rock strength (VeSilith) and the bedrock anisotropy (VaSlAnnorm). The third order scheme describes the hydrography as to its vertical changes by the inclination of the talweg and the different types of knickpoints (IncTallith/grad) and horizontal sinuosity (HoSilith/grad). The study area is subjected to a tripartite zonation into the headwater zone, synonymous with the paleoplain which undergoes some dissection at its edge, the step-fault plain representative of the track zone which undergoes widespread fluvial piracy, and the foreland plains which act as an intermediate sedimentary trap named the deposition zone. The area can be described in space and time with these landform indices reflecting fluvial and mass wasting processes operative in four different stages (around 17 Ma, 6 to 4 Ma, <1.7 Ma, and <0.4 Ma). The various groups of LFIs are a function of landscape maturity (pre-mature, mature, and super-mature). The depositional landforms are numerically defined in the same way and only differ from each other by their subscripts. Their set of LFIs is a mirror image of the composition of depositional landforms in relation to their grain size. The leading part of the acronym, such as QuantSanheav and QuantGravlith, refers to the process of quantification, the second part to the grain size, such as sand and gravel, and the subscript to the material, such as heavy minerals or lithological fragments. The three numerical indices applicable to depositional landforms are a direct measurement of the hydrodynamic and gravity-driven conditions of the fluvial and mass wasting processes using granulometry, grain morphology, and situmetry (clast orientation). Together with the previous compositional indices, the latter directly translate into the provenance analysis which can be used for environmental analyses and as a tool for mineral exploration. It creates a network between numerical geomorphology, geomorphometry, and the E&E issue disciplines (economic/extractive geology vs. environmental geology). The linguistics of the LFIs adopted in this publication are designed so as to be open for individual amendments by the reader. An easy adaptation to different landform suites worldwide, irrespective of their climatic conditions, geodynamic setting, and age of formation, is feasible due to the use of a software and a database available on a global basis. Full article
(This article belongs to the Section Sedimentology, Stratigraphy and Palaeontology)
Show Figures

Figure 1

Figure 1
<p>Geodynamic overview of the NE Bavarian basement and the study area at the western edge of the Münchberg Gneiss Complex, SE Germany. (<b>a</b>) The position of the study area in Germany. (<b>b</b>) The geological setting of the study area in SE Germany and its neighboring geodynamic units of the Frankenwald and Fichtelgebirge Mts. (modified from Emmert et al. [<a href="#B26-geosciences-15-00037" class="html-bibr">26</a>]. (<b>c</b>) Legend for the map in <a href="#geosciences-15-00037-f001" class="html-fig">Figure 1</a>b. The area with the dashed line denotes the close-up view of the geological setting in <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>.</p>
Full article ">Figure 2
<p>Geological overview and the bedrock lithologies of the landform series under consideration. (<b>a</b>) The geological map of the study area with the Cenozoic overburden and the fluvial drainage network and sampling sites. The geological basis is the geological maps published by Emmert and Weinelt [<a href="#B36-geosciences-15-00037" class="html-bibr">36</a>], Emmert et al. [<a href="#B35-geosciences-15-00037" class="html-bibr">35</a>], Stettner [<a href="#B39-geosciences-15-00037" class="html-bibr">39</a>] and Stettner [<a href="#B40-geosciences-15-00037" class="html-bibr">40</a>] which, in places, have been updated during the current investigation. (<b>b</b>) Lithological units shown in <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a and symbols used in the cross-sections through the landforms (see <a href="#geosciences-15-00037-f004" class="html-fig">Figure 4</a>).</p>
Full article ">Figure 2 Cont.
<p>Geological overview and the bedrock lithologies of the landform series under consideration. (<b>a</b>) The geological map of the study area with the Cenozoic overburden and the fluvial drainage network and sampling sites. The geological basis is the geological maps published by Emmert and Weinelt [<a href="#B36-geosciences-15-00037" class="html-bibr">36</a>], Emmert et al. [<a href="#B35-geosciences-15-00037" class="html-bibr">35</a>], Stettner [<a href="#B39-geosciences-15-00037" class="html-bibr">39</a>] and Stettner [<a href="#B40-geosciences-15-00037" class="html-bibr">40</a>] which, in places, have been updated during the current investigation. (<b>b</b>) Lithological units shown in <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a and symbols used in the cross-sections through the landforms (see <a href="#geosciences-15-00037-f004" class="html-fig">Figure 4</a>).</p>
Full article ">Figure 3
<p>Geomorphological overview of the washboard landscape and the study areas defined by the two paleosurfaces, I and II. (<b>a</b>) Cartoon showing two paleosurfaces. Paleosurface I is gently dipping off the FLFZ (Franconian Line Fault Zone) as a presumed architectural planar element covering the Franconian Scarpland. Paleosurface II is a presumed surface covering the basement and the immediate foreland affected by the FLFZ. It is a tripartite curved surface covering three plains [<a href="#B29-geosciences-15-00037" class="html-bibr">29</a>] (<b>b</b>) Cartoon providing an idealized cross-section of the tripartite paleosurface II. For geology, see key of <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a. Dotted line marks the modern-day surface and longitudinal profile of the talweg with its knickpoints. (<b>c</b>) Digital terrain model of the study area showing the controlling linear tectonic elements of the main anticline of the MGC. (<b>d</b>) Topographic map showing the altitude of the study area in meters a.m.s.l. (<b>e</b>) Thematic map showing the slope angle values of the various land forms under consideration in degrees. (<b>f</b>) Geomorphological index map showing the morphotectonic units currently on display: 1 = paleoplain undissected, 2 = paleoplain dissected, 3 = step-fault plain inclined, 4 = foreland plain inclined off the basement, and 5 = foreland plain towards the basement (<a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>b). The position of the reference cross-sections (<a href="#geosciences-15-00037-f004" class="html-fig">Figure 4</a>) and the maximum aerial extension of relief generations R1 to R4 (Rre = R2 landforms are patchily preserved as relics on R3) determined based upon the vertical sinuosity—valley (VeSi<sub>val</sub>) index are displayed by red full, stippled, and dashed-stippled lines. (<b>g</b>) Geological index map (for legend, see <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a). The position of the reference cross-sections (<a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>g) and the maximum aerial extension of relief generations R1 to R4 (Rre = R2 landforms are patchily preserved as relics on R3) determined based upon the vertical sinuosity—valley (VeSi<sub>val</sub>) index are displayed by red full, stippled, and dashed-stippled line.</p>
Full article ">Figure 3 Cont.
<p>Geomorphological overview of the washboard landscape and the study areas defined by the two paleosurfaces, I and II. (<b>a</b>) Cartoon showing two paleosurfaces. Paleosurface I is gently dipping off the FLFZ (Franconian Line Fault Zone) as a presumed architectural planar element covering the Franconian Scarpland. Paleosurface II is a presumed surface covering the basement and the immediate foreland affected by the FLFZ. It is a tripartite curved surface covering three plains [<a href="#B29-geosciences-15-00037" class="html-bibr">29</a>] (<b>b</b>) Cartoon providing an idealized cross-section of the tripartite paleosurface II. For geology, see key of <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a. Dotted line marks the modern-day surface and longitudinal profile of the talweg with its knickpoints. (<b>c</b>) Digital terrain model of the study area showing the controlling linear tectonic elements of the main anticline of the MGC. (<b>d</b>) Topographic map showing the altitude of the study area in meters a.m.s.l. (<b>e</b>) Thematic map showing the slope angle values of the various land forms under consideration in degrees. (<b>f</b>) Geomorphological index map showing the morphotectonic units currently on display: 1 = paleoplain undissected, 2 = paleoplain dissected, 3 = step-fault plain inclined, 4 = foreland plain inclined off the basement, and 5 = foreland plain towards the basement (<a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>b). The position of the reference cross-sections (<a href="#geosciences-15-00037-f004" class="html-fig">Figure 4</a>) and the maximum aerial extension of relief generations R1 to R4 (Rre = R2 landforms are patchily preserved as relics on R3) determined based upon the vertical sinuosity—valley (VeSi<sub>val</sub>) index are displayed by red full, stippled, and dashed-stippled lines. (<b>g</b>) Geological index map (for legend, see <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a). The position of the reference cross-sections (<a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>g) and the maximum aerial extension of relief generations R1 to R4 (Rre = R2 landforms are patchily preserved as relics on R3) determined based upon the vertical sinuosity—valley (VeSi<sub>val</sub>) index are displayed by red full, stippled, and dashed-stippled line.</p>
Full article ">Figure 3 Cont.
<p>Geomorphological overview of the washboard landscape and the study areas defined by the two paleosurfaces, I and II. (<b>a</b>) Cartoon showing two paleosurfaces. Paleosurface I is gently dipping off the FLFZ (Franconian Line Fault Zone) as a presumed architectural planar element covering the Franconian Scarpland. Paleosurface II is a presumed surface covering the basement and the immediate foreland affected by the FLFZ. It is a tripartite curved surface covering three plains [<a href="#B29-geosciences-15-00037" class="html-bibr">29</a>] (<b>b</b>) Cartoon providing an idealized cross-section of the tripartite paleosurface II. For geology, see key of <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a. Dotted line marks the modern-day surface and longitudinal profile of the talweg with its knickpoints. (<b>c</b>) Digital terrain model of the study area showing the controlling linear tectonic elements of the main anticline of the MGC. (<b>d</b>) Topographic map showing the altitude of the study area in meters a.m.s.l. (<b>e</b>) Thematic map showing the slope angle values of the various land forms under consideration in degrees. (<b>f</b>) Geomorphological index map showing the morphotectonic units currently on display: 1 = paleoplain undissected, 2 = paleoplain dissected, 3 = step-fault plain inclined, 4 = foreland plain inclined off the basement, and 5 = foreland plain towards the basement (<a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>b). The position of the reference cross-sections (<a href="#geosciences-15-00037-f004" class="html-fig">Figure 4</a>) and the maximum aerial extension of relief generations R1 to R4 (Rre = R2 landforms are patchily preserved as relics on R3) determined based upon the vertical sinuosity—valley (VeSi<sub>val</sub>) index are displayed by red full, stippled, and dashed-stippled lines. (<b>g</b>) Geological index map (for legend, see <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a). The position of the reference cross-sections (<a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>g) and the maximum aerial extension of relief generations R1 to R4 (Rre = R2 landforms are patchily preserved as relics on R3) determined based upon the vertical sinuosity—valley (VeSi<sub>val</sub>) index are displayed by red full, stippled, and dashed-stippled line.</p>
Full article ">Figure 4
<p>The reference cross-sections provide the link between the landscape and the lithological composition. For lithology, see key of <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>b, and for their position, see <a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>d. In the rectangles, the vertical sinuosity—valley (VeSi<sub>val</sub>) index of the landform series portrayed by the reference cross-section is given. It is the landscape roughness index of regional scale (for more information see text). (<b>a</b>) M1-M2, (<b>b</b>) L1-L2, (<b>c</b>) Pe1-Pe2, (<b>d</b>) K1-K2, (<b>e</b>) Ku1-Ku2, (<b>f</b>) We1-We, (<b>g</b>) J3-J4, (<b>h</b>) H1-I2, (<b>i</b>) H1-H2, (<b>j</b>) G1-G2, (<b>k</b>) F1-E2, (<b>l</b>) E1-E2, (<b>m</b>) D1-D2, (<b>n</b>) C1-C2, (<b>o</b>) B1-B2, (<b>p</b>) A1-A2, (<b>q</b>) Lu1-Lu2, (<b>r</b>) Wü1-Wü2, (<b>s</b>) A3-A4, (<b>t</b>) A3-A4 FW, (<b>u</b>) A1-A2 FW, (<b>v</b>) J1-J2, and (<b>w</b>) J5-J6.</p>
Full article ">Figure 4 Cont.
<p>The reference cross-sections provide the link between the landscape and the lithological composition. For lithology, see key of <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>b, and for their position, see <a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>d. In the rectangles, the vertical sinuosity—valley (VeSi<sub>val</sub>) index of the landform series portrayed by the reference cross-section is given. It is the landscape roughness index of regional scale (for more information see text). (<b>a</b>) M1-M2, (<b>b</b>) L1-L2, (<b>c</b>) Pe1-Pe2, (<b>d</b>) K1-K2, (<b>e</b>) Ku1-Ku2, (<b>f</b>) We1-We, (<b>g</b>) J3-J4, (<b>h</b>) H1-I2, (<b>i</b>) H1-H2, (<b>j</b>) G1-G2, (<b>k</b>) F1-E2, (<b>l</b>) E1-E2, (<b>m</b>) D1-D2, (<b>n</b>) C1-C2, (<b>o</b>) B1-B2, (<b>p</b>) A1-A2, (<b>q</b>) Lu1-Lu2, (<b>r</b>) Wü1-Wü2, (<b>s</b>) A3-A4, (<b>t</b>) A3-A4 FW, (<b>u</b>) A1-A2 FW, (<b>v</b>) J1-J2, and (<b>w</b>) J5-J6.</p>
Full article ">Figure 4 Cont.
<p>The reference cross-sections provide the link between the landscape and the lithological composition. For lithology, see key of <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>b, and for their position, see <a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>d. In the rectangles, the vertical sinuosity—valley (VeSi<sub>val</sub>) index of the landform series portrayed by the reference cross-section is given. It is the landscape roughness index of regional scale (for more information see text). (<b>a</b>) M1-M2, (<b>b</b>) L1-L2, (<b>c</b>) Pe1-Pe2, (<b>d</b>) K1-K2, (<b>e</b>) Ku1-Ku2, (<b>f</b>) We1-We, (<b>g</b>) J3-J4, (<b>h</b>) H1-I2, (<b>i</b>) H1-H2, (<b>j</b>) G1-G2, (<b>k</b>) F1-E2, (<b>l</b>) E1-E2, (<b>m</b>) D1-D2, (<b>n</b>) C1-C2, (<b>o</b>) B1-B2, (<b>p</b>) A1-A2, (<b>q</b>) Lu1-Lu2, (<b>r</b>) Wü1-Wü2, (<b>s</b>) A3-A4, (<b>t</b>) A3-A4 FW, (<b>u</b>) A1-A2 FW, (<b>v</b>) J1-J2, and (<b>w</b>) J5-J6.</p>
Full article ">Figure 5
<p>X–Y diagrams of the variation in slope angle altitude (VaSlAn<sub>alti</sub>) index with the X-axis giving the mean slope angle in degrees and the Y-axis giving the altitude in meters above mean sea level. R2 = correlation coefficient. (<b>a</b>) M1-M2, (<b>b</b>) L1-L2, (<b>c</b>) Pe1-Pe2, (<b>d</b>) K1-K2, (<b>e</b>) Ku1-Ku2, (<b>f</b>) We1-We, (<b>g</b>) J3-J4, (<b>h</b>) H1-I2, (<b>i</b>) H1-H2, (<b>j</b>) G1-G2, (<b>k</b>) F1-E2, (<b>l</b>) E1-E2, (<b>m</b>) D1-D2, (<b>n</b>) C1-C2, (<b>o</b>) B1-B2, (<b>p</b>) A1-A2, (<b>q</b>) Lu1-Lu2, (<b>r</b>) Wü1-Wü2, (<b>s</b>) A3-A4, (<b>t</b>) A3-A4 FW, (<b>u</b>) A1-A2 FW, (<b>v</b>) J1-J2, and (<b>w</b>) J5-J6.</p>
Full article ">Figure 5 Cont.
<p>X–Y diagrams of the variation in slope angle altitude (VaSlAn<sub>alti</sub>) index with the X-axis giving the mean slope angle in degrees and the Y-axis giving the altitude in meters above mean sea level. R2 = correlation coefficient. (<b>a</b>) M1-M2, (<b>b</b>) L1-L2, (<b>c</b>) Pe1-Pe2, (<b>d</b>) K1-K2, (<b>e</b>) Ku1-Ku2, (<b>f</b>) We1-We, (<b>g</b>) J3-J4, (<b>h</b>) H1-I2, (<b>i</b>) H1-H2, (<b>j</b>) G1-G2, (<b>k</b>) F1-E2, (<b>l</b>) E1-E2, (<b>m</b>) D1-D2, (<b>n</b>) C1-C2, (<b>o</b>) B1-B2, (<b>p</b>) A1-A2, (<b>q</b>) Lu1-Lu2, (<b>r</b>) Wü1-Wü2, (<b>s</b>) A3-A4, (<b>t</b>) A3-A4 FW, (<b>u</b>) A1-A2 FW, (<b>v</b>) J1-J2, and (<b>w</b>) J5-J6.</p>
Full article ">Figure 5 Cont.
<p>X–Y diagrams of the variation in slope angle altitude (VaSlAn<sub>alti</sub>) index with the X-axis giving the mean slope angle in degrees and the Y-axis giving the altitude in meters above mean sea level. R2 = correlation coefficient. (<b>a</b>) M1-M2, (<b>b</b>) L1-L2, (<b>c</b>) Pe1-Pe2, (<b>d</b>) K1-K2, (<b>e</b>) Ku1-Ku2, (<b>f</b>) We1-We, (<b>g</b>) J3-J4, (<b>h</b>) H1-I2, (<b>i</b>) H1-H2, (<b>j</b>) G1-G2, (<b>k</b>) F1-E2, (<b>l</b>) E1-E2, (<b>m</b>) D1-D2, (<b>n</b>) C1-C2, (<b>o</b>) B1-B2, (<b>p</b>) A1-A2, (<b>q</b>) Lu1-Lu2, (<b>r</b>) Wü1-Wü2, (<b>s</b>) A3-A4, (<b>t</b>) A3-A4 FW, (<b>u</b>) A1-A2 FW, (<b>v</b>) J1-J2, and (<b>w</b>) J5-J6.</p>
Full article ">Figure 5 Cont.
<p>X–Y diagrams of the variation in slope angle altitude (VaSlAn<sub>alti</sub>) index with the X-axis giving the mean slope angle in degrees and the Y-axis giving the altitude in meters above mean sea level. R2 = correlation coefficient. (<b>a</b>) M1-M2, (<b>b</b>) L1-L2, (<b>c</b>) Pe1-Pe2, (<b>d</b>) K1-K2, (<b>e</b>) Ku1-Ku2, (<b>f</b>) We1-We, (<b>g</b>) J3-J4, (<b>h</b>) H1-I2, (<b>i</b>) H1-H2, (<b>j</b>) G1-G2, (<b>k</b>) F1-E2, (<b>l</b>) E1-E2, (<b>m</b>) D1-D2, (<b>n</b>) C1-C2, (<b>o</b>) B1-B2, (<b>p</b>) A1-A2, (<b>q</b>) Lu1-Lu2, (<b>r</b>) Wü1-Wü2, (<b>s</b>) A3-A4, (<b>t</b>) A3-A4 FW, (<b>u</b>) A1-A2 FW, (<b>v</b>) J1-J2, and (<b>w</b>) J5-J6.</p>
Full article ">Figure 6
<p>Overview of the variation in slope angle altitude (VaSlAn<sub>alti</sub>) index as facies marker.</p>
Full article ">Figure 7
<p>Overview of the petrophysical–geomorphological parameters vertical sinuosity—lithology (VeSi<sub>lith</sub>) of the landform and variation in normalized slope angle (VaSlAn<sub>norm</sub>) of the landform.</p>
Full article ">Figure 8
<p>Meta-sedimentary, meta-intrusive, and meta-volcanic magmatic rocks and their landforms featuring different values of VaSlAn<sub>norm</sub> and VeSi<sub>lith</sub>. For numerical, compositional, topographic, and more detailed geomorphological data, see <a href="#geosciences-15-00037-t002" class="html-table">Table 2</a>. (<b>a</b>) Mica gneiss with subhorizontal jointing on top of a hillock of a large and shallow valley. The top slope is strewn with boulders undergoing creep and solifluction. (<b>b</b>) Close-up view of one of the boulders which displays a lens-shaped and strong foliation. (<b>c</b>) A well-rounded paragneiss-hornfels boulder similar in outward appearance and internal texture but of rock strength twice as much as the mica gneiss. (<b>d</b>) Layered phyllite exposed on the mid-slope of a V-shaped valley. (<b>e</b>) Tightly foliated and folded phyllite as an allochthonous block. See ignition key for scale. Dashed line highlights wrinkled folding. (<b>f</b>) Alternating beds of chert, forming ledges and slates with the beginning of disintegration into debris of flakes at the footslope of a V-shaped valley. (<b>g</b>) Plates of (roof)slate in the D horizon of the pedosphere. The argillaceous rocks are transformed into individual slaps of slate preserving the original siting of the rocks with the slaty cleavage. (<b>h</b>) Completely disintegrated pencil slates randomly scattered along the footslope of a V-shaped valley while forming a talus apron of flakey gravel. See hammer for scale. (<b>i</b>) Augengneiss ledges protruding out of the top slope of a V-shaped valley. The inset displays the tight arrangement of layers composed of quartz, K feldspar, and plagioclase with dark micaceous layers. (<b>j</b>) Meta-granite-to granodiorite showing a massive texture devoid of any strong foliation. (<b>k</b>) Steeply-dipping layers of tightly foliated epidote amphibolite (prasinite). (<b>l</b>) Layers of meta-basalt with narrowly-spaced joints near the escarpment of the inclined step-and-fault plain which is identical to the highland-boundary fault FLFZ (see <a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>b,c). The inset shows the disintegration of the meta-basalt (diabase) as a consequence of weathering. (<b>m</b>) Amphibolite with a vaguely expressed layering which is intruded by an alkaline feldspar pegmatoid rimmed by a stippled line. It constitutes the edge of a V-shaped valley (wide angle) passing into a large and shallow valley. See geologists for scale. (<b>n</b>) Monadnock with subrounded exposures of bronzite-serpentinite displaying typical rillen features of “silica karst”. (<b>o</b>) Disharmonic tight folding of alkaline feldspar—quartz mobilisates in massive layered amphibolite gneiss. (<b>p</b>) A monadnock made of massive eclogite and eclogite amphibolite surrounded by a blockmeer of the same lithology. Inset shows a slightly weathered massive eclogite with red Fe-Al-Mn garnet and green omphacitic pyroxene.</p>
Full article ">Figure 9
<p>Longitudinal sections along the talweg of drainage systems. The X-axis denotes the station points, and the Y-axis denotes the dip angle of the talweg in degree. The station points are characterized by Arabic numerals. The third variable is the wall rock or bedrock lithology of the host rocks exposed in the river banks and the river bed which, when different from each other on the left- and right-hand bank, are given by more than one numeral which refers to the notation in <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>b, e.g., profile X7-X9 13 + 12 + 14 = phyllite &gt; epidote amphibolite &gt; talc schist (for lithology, see <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>b). The correlation coefficient R2 between the two data sets is given in the upper right-hand corner of the diagram. (<b>a</b>) X–Y plot showing the inclination of the talweg (IncTal<sub>lith/grad</sub> index) in degrees. <span class="html-italic">Y</span>-axis versus the station point downstream of longitudinal profile X15-X16 FW. The knickpoints intensity can be directly assessed by the length of the various intervals of the graph and the type of knickpoint (see text) by its upward and downward directions. At station point 11, the longitudinal section is intersected by the cross-section A3-A4 FW (<a href="#geosciences-15-00037-f005" class="html-fig">Figure 5</a>t). The red rectangle marks the IncTal<sub>lith/grad</sub> index fluvial facies in the close-up view of <a href="#geosciences-15-00037-f009" class="html-fig">Figure 9</a>b. (<b>b</b>) Incision of an acute-angle single-channel non-alluvial V-shaped valley into the Devonian chert unit (slope angle 30° ⇒ 35°, talweg angle 2.7° ⇒ 0.7°). See reference profile with steps and pools in (stippled white line = strike of bedding). (<b>c</b>) Geological index map (for more detail and key, see <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a) with horizontal sinuosity—lithology plus gradient index (HoSi<sub>lith/grad</sub>) given in the white boxes; the knickpoint types 1 and 2 and the start and end points of longitudinal sections are displayed in <a href="#geosciences-15-00037-f009" class="html-fig">Figure 9</a>d–i by X–Y diagrams plotting the station points and inclinations data. The red dots mark mines of talc—(purple), pegmatoid—(dark blue), and Cu-(Au) deposits (yellow). (<b>d</b>) X1-X2, (<b>e</b>) X2-X3, (<b>f</b>) X5-X6, (<b>g</b>) X7-X8, (<b>h</b>) X9-X10, and (<b>i</b>) X11-X12 (for color symbols, see <a href="#geosciences-15-00037-f009" class="html-fig">Figure 9</a>c).</p>
Full article ">Figure 9 Cont.
<p>Longitudinal sections along the talweg of drainage systems. The X-axis denotes the station points, and the Y-axis denotes the dip angle of the talweg in degree. The station points are characterized by Arabic numerals. The third variable is the wall rock or bedrock lithology of the host rocks exposed in the river banks and the river bed which, when different from each other on the left- and right-hand bank, are given by more than one numeral which refers to the notation in <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>b, e.g., profile X7-X9 13 + 12 + 14 = phyllite &gt; epidote amphibolite &gt; talc schist (for lithology, see <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>b). The correlation coefficient R2 between the two data sets is given in the upper right-hand corner of the diagram. (<b>a</b>) X–Y plot showing the inclination of the talweg (IncTal<sub>lith/grad</sub> index) in degrees. <span class="html-italic">Y</span>-axis versus the station point downstream of longitudinal profile X15-X16 FW. The knickpoints intensity can be directly assessed by the length of the various intervals of the graph and the type of knickpoint (see text) by its upward and downward directions. At station point 11, the longitudinal section is intersected by the cross-section A3-A4 FW (<a href="#geosciences-15-00037-f005" class="html-fig">Figure 5</a>t). The red rectangle marks the IncTal<sub>lith/grad</sub> index fluvial facies in the close-up view of <a href="#geosciences-15-00037-f009" class="html-fig">Figure 9</a>b. (<b>b</b>) Incision of an acute-angle single-channel non-alluvial V-shaped valley into the Devonian chert unit (slope angle 30° ⇒ 35°, talweg angle 2.7° ⇒ 0.7°). See reference profile with steps and pools in (stippled white line = strike of bedding). (<b>c</b>) Geological index map (for more detail and key, see <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a) with horizontal sinuosity—lithology plus gradient index (HoSi<sub>lith/grad</sub>) given in the white boxes; the knickpoint types 1 and 2 and the start and end points of longitudinal sections are displayed in <a href="#geosciences-15-00037-f009" class="html-fig">Figure 9</a>d–i by X–Y diagrams plotting the station points and inclinations data. The red dots mark mines of talc—(purple), pegmatoid—(dark blue), and Cu-(Au) deposits (yellow). (<b>d</b>) X1-X2, (<b>e</b>) X2-X3, (<b>f</b>) X5-X6, (<b>g</b>) X7-X8, (<b>h</b>) X9-X10, and (<b>i</b>) X11-X12 (for color symbols, see <a href="#geosciences-15-00037-f009" class="html-fig">Figure 9</a>c).</p>
Full article ">Figure 10
<p>Quantification of fluvial and mass wasting deposits as well as their ratios (quantification of fluvial–mass wasting index (Quant<sub>flu/mas</sub>). For geomorphological background, see <a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>f. 1 + 2: Mass wasting deposits: 2.302 to 0.768 per km<sup>2</sup>, fluvial deposits: 0.888 to 0.135 per km<sup>2</sup>. 3: Mass wasting deposits: 0.457 to 0.061 per km<sup>2</sup>, fluvial deposits: 0.335 to 0.017 per km<sup>2</sup>. 4: Mixed type (mass wasting and fluvial): 3.443 to 0.393 per km<sup>2</sup>, mass wasting deposits 3.132 to 0.393 per km<sup>2</sup>, fluvial deposits: 2.798 to 0.028 per km<sup>2</sup>. 5: Mass wasting deposits 0.019 per km<sup>2</sup>, fluvial deposits: 0.076 per km<sup>2</sup>. In the case of very small quantities of the landform-related mass wasting and fluvial deposits, only the ratio of the deposits is presented as a sector diagram. In the case of very high quantities of these unconsolidated deposits, columnar diagram are used instead.</p>
Full article ">Figure 11
<p>Composition of siliciclastic deposits of the study area. For geology of the sampling sites, see <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a. The mineralogical and petrological composition is given by sector diagrams (100%). (<b>a</b>) Abundance of sand-sized light minerals (Quant<sub>san/ligh</sub>). (<b>b</b>) Abundance of sand-sized heavy minerals (Quant<sub>san/heav</sub>). (<b>c</b>) Abundance of gravel-sized debris (Quant<sub>grav/lith</sub>).</p>
Full article ">Figure 11 Cont.
<p>Composition of siliciclastic deposits of the study area. For geology of the sampling sites, see <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a. The mineralogical and petrological composition is given by sector diagrams (100%). (<b>a</b>) Abundance of sand-sized light minerals (Quant<sub>san/ligh</sub>). (<b>b</b>) Abundance of sand-sized heavy minerals (Quant<sub>san/heav</sub>). (<b>c</b>) Abundance of gravel-sized debris (Quant<sub>grav/lith</sub>).</p>
Full article ">Figure 12
<p>Landforms hosting gravel-sized debris accumulations subjected to GMS analyses (granulometry–morphometry–situmetry). For sampling sites, see the geological setting presented in <a href="#geosciences-15-00037-f011" class="html-fig">Figure 11</a> and the legend on display in <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a,b. (<b>a</b>) A V-shaped valley (acute angle 22 to 25°) with a small floodplain narrowing upstream towards a gorge (alluvial to non-alluvial). The inset situgram shows a bimodal clast orientation. Sampling site 7. (<b>b</b>) Non-alluvial V-shaped valley (acute angle 25 to 30°) chocked with gravel-sized clast and concentrated in side- and mid-channel longitudinal bars. Sampling site 15. (<b>c</b>) V-shaped valley with a small raised side bar on the slip bank (wide angle 5 to 11°) Sampling site 2. (<b>d</b>) Wide valley (angle 5 to 15°) showing a floodplain with gallery forests lined up along the meander belts, S = 1.407. Sampling site 12. (<b>e</b>) Two valleys telescoped into each other. The large and shallow valley (angle &lt;&lt; 10°) is cut by an acute V-shaped valley near the FLFZ. Sampling site 13 photography facing towards the W with the scarpland on the horizon. (<b>f</b>) A polymodal clast orientation representative of different landscape-forming processes superimposed on each other. Situgram of sampling site 15. (<b>g</b>) Unimodal clast orientation preserved on the raised sidebar of a slip bank. Situgram of sampling site 2.</p>
Full article ">Figure 12 Cont.
<p>Landforms hosting gravel-sized debris accumulations subjected to GMS analyses (granulometry–morphometry–situmetry). For sampling sites, see the geological setting presented in <a href="#geosciences-15-00037-f011" class="html-fig">Figure 11</a> and the legend on display in <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>a,b. (<b>a</b>) A V-shaped valley (acute angle 22 to 25°) with a small floodplain narrowing upstream towards a gorge (alluvial to non-alluvial). The inset situgram shows a bimodal clast orientation. Sampling site 7. (<b>b</b>) Non-alluvial V-shaped valley (acute angle 25 to 30°) chocked with gravel-sized clast and concentrated in side- and mid-channel longitudinal bars. Sampling site 15. (<b>c</b>) V-shaped valley with a small raised side bar on the slip bank (wide angle 5 to 11°) Sampling site 2. (<b>d</b>) Wide valley (angle 5 to 15°) showing a floodplain with gallery forests lined up along the meander belts, S = 1.407. Sampling site 12. (<b>e</b>) Two valleys telescoped into each other. The large and shallow valley (angle &lt;&lt; 10°) is cut by an acute V-shaped valley near the FLFZ. Sampling site 13 photography facing towards the W with the scarpland on the horizon. (<b>f</b>) A polymodal clast orientation representative of different landscape-forming processes superimposed on each other. Situgram of sampling site 15. (<b>g</b>) Unimodal clast orientation preserved on the raised sidebar of a slip bank. Situgram of sampling site 2.</p>
Full article ">Figure 13
<p>GMS indices (granulometry–morphology–situmetry) and their fluvial networks of the X1-X2 drainage system and its tributaries X3-X4 and X7-X8. For more details on the numerical parameters of the drainage systems, see <a href="#geosciences-15-00037-f009" class="html-fig">Figure 9</a>, and for geology, <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>. (<b>a</b>) Granulometry of gravel-sized debris illustrated by the numerical index QuantSed<sub>gran/sort</sub> with a cumulative frequency grain-size distribution of all samples from the study area above represented by the blue shaded area. (<b>b</b>) The regional variation in the minimum values of the QuantSed<sub>morp/roun</sub> of gravel-sized debris (map above) and a reference site showing the QuantSed<sub>morp/roun</sub> compared with the QuantSed<sub>morp/cycl</sub> numerically and visually for the most widespread lithology of the study area, the muscovite-biotite gneisses. (<b>c</b>) Situmetry of gravel-sized debris illustrated by 360° circle diagrams showing the true orientation of the river course and of various maxima of the longest axis of gravel clasts (<b>above</b>). The reference samples show a topographically non-oriented semi-circle rose diagram with a trimodal arrangement of gravel clasts with a sharpness of maximum as follows: first maximum 60.0, second maximum 19.4, and third maximum 19.0.</p>
Full article ">Figure 13 Cont.
<p>GMS indices (granulometry–morphology–situmetry) and their fluvial networks of the X1-X2 drainage system and its tributaries X3-X4 and X7-X8. For more details on the numerical parameters of the drainage systems, see <a href="#geosciences-15-00037-f009" class="html-fig">Figure 9</a>, and for geology, <a href="#geosciences-15-00037-f002" class="html-fig">Figure 2</a>. (<b>a</b>) Granulometry of gravel-sized debris illustrated by the numerical index QuantSed<sub>gran/sort</sub> with a cumulative frequency grain-size distribution of all samples from the study area above represented by the blue shaded area. (<b>b</b>) The regional variation in the minimum values of the QuantSed<sub>morp/roun</sub> of gravel-sized debris (map above) and a reference site showing the QuantSed<sub>morp/roun</sub> compared with the QuantSed<sub>morp/cycl</sub> numerically and visually for the most widespread lithology of the study area, the muscovite-biotite gneisses. (<b>c</b>) Situmetry of gravel-sized debris illustrated by 360° circle diagrams showing the true orientation of the river course and of various maxima of the longest axis of gravel clasts (<b>above</b>). The reference samples show a topographically non-oriented semi-circle rose diagram with a trimodal arrangement of gravel clasts with a sharpness of maximum as follows: first maximum 60.0, second maximum 19.4, and third maximum 19.0.</p>
Full article ">Figure 14
<p>The manual from fieldwork (geological, geomorphological, and lithological mapping) to numerical geomorphology &gt; geomorphometry (genetic geosciences) and economic and environmental geology (applied geosciences). The landform indices are the missing links. See also <a href="#geosciences-15-00037-t001" class="html-table">Table 1</a>.</p>
Full article ">Figure 15
<p>The evolution of landscape and re-orientation of the drainage system from the ancient Donau River to the modern Rhein River systems on display as a series of landscape contours true to scale as a function of altitude and distance based upon the VeSi<sub>val</sub>, VaSlAn<sub>alti</sub>, IncTal<sub>lith/grad</sub>, and geochronological data (for reference, see text). Periods correspond to the relief generations shown in plan view in <a href="#geosciences-15-00037-f003" class="html-fig">Figure 3</a>f,g. For the geology and landforms of each cross-section, see <a href="#geosciences-15-00037-f004" class="html-fig">Figure 4</a> and <a href="#geosciences-15-00037-f005" class="html-fig">Figure 5</a>. (<b>a</b>) Stage of peneplanation at full swing (Ro). (<b>b</b>) Stage of peneplanation (R1) transitioning into pediplanation (R2) (fossiliferous badlands). (<b>c</b>) Stage of the re-orientation of the paleogradientaccompanied by river piracy (R2e) and linear erosion (R3). (<b>d</b>) Stage of the re-direction of the fluvial regime from dip to strike stream and perched pedimentation (R4).</p>
Full article ">Figure 16
<p>Clay minerals (QuantClaSil), sand-sized light minerals (QuantSan<sub>/ligh</sub>), heavy minerals (QuantSan<sub>heav</sub>), and gravel (QuantGrav<sub>lith</sub>) of different lithologies represented by the range of dispersal off their source rocks.</p>
Full article ">Figure 17
<p>The “graphical conclusions” to underscore what the compositional terrain analysis is all about. The tripartite subdivision of the geoscientific disciplines involved: (<b>a</b>) A digital terrain model showing the interrelationship between morphotectonic linear architectural elements (fold axis), and hydrography (strike stream vs. dip stream). (<b>b</b>) The sedimentological GMS technology encompassing <b>g</b>ranulometry, morphometry, and <b>s</b>itumetry. (<b>c</b>) The pie-chart diagram commonly used in sediment petrography to quantify the lithological changes during transport.</p>
Full article ">
42 pages, 7150 KiB  
Article
LightweightUNet: Multimodal Deep Learning with GAN-Augmented Imaging Data for Efficient Breast Cancer Detection
by Hari Mohan Rai, Joon Yoo, Saurabh Agarwal and Neha Agarwal
Bioengineering 2025, 12(1), 73; https://doi.org/10.3390/bioengineering12010073 - 15 Jan 2025
Viewed by 1141
Abstract
Breast cancer ranks as the second most prevalent cancer globally and is the most frequently diagnosed cancer among women; therefore, early, automated, and precise detection is essential. Most AI-based techniques for breast cancer detection are complex and have high computational costs. Hence, to [...] Read more.
Breast cancer ranks as the second most prevalent cancer globally and is the most frequently diagnosed cancer among women; therefore, early, automated, and precise detection is essential. Most AI-based techniques for breast cancer detection are complex and have high computational costs. Hence, to overcome this challenge, we have presented the innovative LightweightUNet hybrid deep learning (DL) classifier for the accurate classification of breast cancer. The proposed model boasts a low computational cost due to its smaller number of layers in its architecture, and its adaptive nature stems from its use of depth-wise separable convolution. We have employed a multimodal approach to validate the model’s performance, using 13,000 images from two distinct modalities: mammogram imaging (MGI) and ultrasound imaging (USI). We collected the multimodal imaging datasets from seven different sources, including the benchmark datasets DDSM, MIAS, INbreast, BrEaST, BUSI, Thammasat, and HMSS. Since the datasets are from various sources, we have resized them to the uniform size of 256 × 256 pixels and normalized them using the Box-Cox transformation technique. Since the USI dataset is smaller, we have applied the StyleGAN3 model to generate 10,000 synthetic ultrasound images. In this work, we have performed two separate experiments: the first on a real dataset without augmentation and the second on a real + GAN-augmented dataset using our proposed method. During the experiments, we used a 5-fold cross-validation method, and our proposed model obtained good results on the real dataset (87.16% precision, 86.87% recall, 86.84% F1-score, and 86.87% accuracy) without adding any extra data. Similarly, the second experiment provides better performance on the real + GAN-augmented dataset (96.36% precision, 96.35% recall, 96.35% F1-score, and 96.35% accuracy). This multimodal approach, which utilizes LightweightUNet, enhances the performance by 9.20% in precision, 9.48% in recall, 9.51% in F1-score, and a 9.48% increase in accuracy on the combined dataset. The LightweightUNet model we proposed works very well thanks to a creative network design, adding fake images to the data, and a multimodal training method. These results show that the model has a lot of potential for use in clinical settings. Full article
(This article belongs to the Special Issue Application of Deep Learning in Medical Diagnosis)
Show Figures

Figure 1

Figure 1
<p>LightweightUNet-driven breast cancer detection methodology.</p>
Full article ">Figure 2
<p>Sample images of benign and malignant pathology types from USI modality.</p>
Full article ">Figure 3
<p>Sample images of benign and malignant pathology types from MGI modality.</p>
Full article ">Figure 4
<p>The structural view of the StyleGAN3 model [<a href="#B56-bioengineering-12-00073" class="html-bibr">56</a>].</p>
Full article ">Figure 5
<p>Architectural diagrams of the proposed LightweightUNet model for the detection of breast cancer.</p>
Full article ">Figure 6
<p>LightweightUNet training history on the real dataset (loss and accuracy).</p>
Full article ">Figure 6 Cont.
<p>LightweightUNet training history on the real dataset (loss and accuracy).</p>
Full article ">Figure 7
<p>Breast cancer classification results using LightweightUNet (5-Fold CV) in terms of confusion matrix and metrics on a real dataset.</p>
Full article ">Figure 8
<p>LightweightUNet training history on real + GAN dataset (loss and accuracy).</p>
Full article ">Figure 8 Cont.
<p>LightweightUNet training history on real + GAN dataset (loss and accuracy).</p>
Full article ">Figure 9
<p>Breast cancer classification results using LightweightUNet (5-Fold CV) in terms of confusion matrix and metrics on real + GAN dataset.</p>
Full article ">Figure 10
<p>Impact of GAN-generated data on breast cancer classification for the benign class using LightweightUNet model.</p>
Full article ">Figure 11
<p>Impact of GAN-generated data on breast cancer classification for the malignant class using LightweightUNet model.</p>
Full article ">Figure 12
<p>Impact of GAN-generated data on breast cancer classification metrics for breast cancer detection using LightweightUNet model.</p>
Full article ">
25 pages, 7245 KiB  
Article
Long-Term Evaluation of GCOM-C/SGLI Reflectance and Water Quality Products: Variability Among JAXA G-Portal and JASMES
by Salem Ibrahim Salem, Mitsuhiro Toratani, Hiroto Higa, SeungHyun Son, Eko Siswanto and Joji Ishizaka
Remote Sens. 2025, 17(2), 221; https://doi.org/10.3390/rs17020221 - 9 Jan 2025
Cited by 1 | Viewed by 640
Abstract
The Global Change Observation Mission-Climate (GCOM-C) satellite, launched in December 2017, is equipped with the Second-generation Global Imager (SGLI) sensor, featuring a moderate spatial resolution of 250 m and 19 spectral bands, including the unique 380 nm band. After six years in orbit, [...] Read more.
The Global Change Observation Mission-Climate (GCOM-C) satellite, launched in December 2017, is equipped with the Second-generation Global Imager (SGLI) sensor, featuring a moderate spatial resolution of 250 m and 19 spectral bands, including the unique 380 nm band. After six years in orbit, a comprehensive evaluation of SGLI products and their temporal consistency is needed. Remote sensing reflectance (Rrs) is the primary product for monitoring water quality, forming the basis for deriving key oceanic constituents such as chlorophyll-a (Chla) and total suspended matter (TSM). The Japan Aerospace Exploration Agency (JAXA) provides Rrs products through two platforms, G-Portal and JASMES, each employing different atmospheric correction methodologies and assumptions. This study aims to evaluate the SGLI full-resolution Rrs products from G-Portal and JASMES at regional scales (Japan and East Asia) and assess G-Portal Rrs products globally between January 2018 and December 2023. The evaluation employs in situ matchups from NASA’s Aerosol Robotic Network-Ocean Color (AERONET-OC) and cruise measurements. We also assess the retrieval accuracy of two water quality indices, Chla and TSM. The AERONET-OC data analysis reveals that JASMES systematically underestimates Rrs values at shorter wavelengths, particularly at 412 nm. While the Rrs accuracy at 412 nm is relatively low, G-Portal’s Rrs products perform better than JASMES at shorter wavelengths, showing lower errors and stronger correlations with AERONET-OC data. Both G-Portal and JASMES show lower agreement with AERONET-OC and cruise datasets at shorter wavelengths but demonstrate improved agreement at longer wavelengths (530 nm, 565 nm, and 670 nm). JASMES generates approximately 12% more matchup data points than G-Portal, likely due to G-Portal’s stricter atmospheric correction thresholds that exclude pixels with high reflectance. In situ measurements indicate that G-Portal provides better overall agreement, particularly at lower Rrs magnitudes and Chla concentrations below 5 mg/m3. This evaluation underscores the complexities and challenges of atmospheric correction, particularly in optically complex coastal waters (Case 2 waters), which may require tailored atmospheric correction methods different from the standard approach. The assessment of temporal consistency and seasonal variations in Rrs data shows that both platforms effectively capture interannual trends and maintain temporal stability, particularly from the 490 nm band onward, underscoring the potential of SGLI data for long-term monitoring of coastal and oceanic environments. Full article
Show Figures

Figure 1

Figure 1
<p>Global distribution of matchups between validation datasets and GCOM-C/SGLI. (<b>a</b>) Locations of 331 in situ <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> matchups (red cross), 753 in situ Chla and TSM matchups (green circle) of cruise measurements, and 22 sites for G-Portal and JASMES comparison (blue cross). (<b>b</b>) Distribution of 3704 AERONET-OC matchups, with red circles indicating locations and sizes proportional to the number of measurements at each site. The dashed box represents the footprint of the JASMES full-resolution <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> product over Japan.</p>
Full article ">Figure 2
<p>Schematic diagram of atmospheric correction processes for (<b>a</b>) SGLI G-Portal and (<b>b</b>) SGLI JASMES.</p>
Full article ">Figure 3
<p>Comparison of <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> spectra for various matchup scenarios. The first row represents <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> spectra for matchups between (<b>a</b>) AERONET-OC and both (<b>b</b>) SGLI G-Portal and (<b>c</b>) SGLI JASMES. The second row shows <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> spectra for matchups between (<b>d</b>) cruise measurements and both (<b>e</b>) SGLI G-Portal and (<b>f</b>) SGLI JASMES. Grey lines represent individual matchups, with black lines indicating the mean <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> values across different wavelengths. The legend at the top identifies the mean <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> values at various AERONET-OC sites, as shown in panel (<b>a</b>).</p>
Full article ">Figure 4
<p>Scatterplots (<b>a</b>–<b>f</b>) comparing AERONET-OC <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> with SGLI <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> of G-Portal (x symbols) and JASMES (• symbols) over Japan at four AERONET-OC sites: Ieodo_Station, Socheongcho, Ariake_Tower, and Kemigawa_Offshore. N refers to the number of matchups, R<sup>2</sup> to the coefficient of determination, β to bias, δ to mean absolute difference, ∆ to root mean square difference, and σ to mean absolute percentage difference in percent (%). For each wavelength, the evaluation metrics for G-Portal are listed first, followed by those for JASMES in parentheses. The dashed grey line represents the 1:1 line.</p>
Full article ">Figure 5
<p>Scatterplots (<b>a</b>–<b>f</b>) comparing AERONET-OC <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> values with SGLI G-Portal <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> values at 29 AERONET-OC sites, covering z global scale. N refers to the number of matchups, R<sup>2</sup> to the coefficient of determination, β to bias, δ to mean absolute difference, ∆ to root mean square difference, and σ to mean absolute percentage difference in percent (%). The dashed grey line represents the 1:1 line.</p>
Full article ">Figure 6
<p>Time series plots comparing <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> values from AERONET-OC (green circles), SGLI G-Portal (blue crosses), and SGLI JASMES (orange fork) across three observation sites: Kemigawa_Offshore, Ariake_Tower, and Socheongcho. Each row corresponds to specific SGLI bands (412 nm, 443 nm, 490 nm, 530 nm, 565 nm, and 670 nm). Parenthetical numbers adjacent to each SGLI band on the Y-axes indicate the closest corresponding AERONET-OC band. For instance, 670 (667) on the <span class="html-italic">Y</span>-axis of the last row represents the SGLI and AERONET-OC bands 670 nm and 667 nm, respectively.</p>
Full article ">Figure 7
<p>Scatterplots (<b>a</b>–<b>g</b>) compare cruise <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> with SGLI <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> of G-Portal (x symbols) and JASMES (• symbols). The number of matchups (N), coefficient of determination (R<sup>2</sup>), bias (β), mean absolute difference (δ), root mean square difference (∆), mean absolute percentage difference (σ) in percent (%), and chlorophyll-a (Chla) are shown. For each wavelength, the metrics for G-Portal are listed first, followed by those for JASMES in parentheses. The dashed grey line represents the 1:1 line.</p>
Full article ">Figure 8
<p>Difference in <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> values (Δ<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math>) between cruise measurement and SGLI of (<b>a</b>) G-Portal globally, (<b>b</b>) G-Portal over Japan, and (<b>c</b>) JASMES. N refers to the number of matchups. Each grey line represents individual observations, while the black line indicates the mean Δ<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math>, and the vertical bars represent the standard deviation of Δ<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> at each wavelength.</p>
Full article ">Figure 9
<p>Scatterplots comparing in situ measurements and SGLI-derived values from G-Portal and JASMES for two products: (<b>a</b>) Chla and (<b>b</b>) TSM. The number of matchups (N), coefficient of determination (R<sup>2</sup>), bias (β), mean absolute difference (δ), root mean square difference (∆), and mean absolute percentage difference (σ) in percent (%) are shown. The metrics for G-Portal are listed first, followed by those for JASMES in parentheses. The dashed grey line represents the 1:1 line.</p>
Full article ">Figure 10
<p>Density scatterplots (<b>a</b>–<b>g</b>) illustrate SGLI G-Portal <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> versus SGLI JASMES <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> over 22 sites (blue cross, <a href="#remotesensing-17-00221-f001" class="html-fig">Figure 1</a>a). N refers to the number of matchups, while R<sup>2</sup> and S denote the coefficient of determination and the slope, respectively. The dashed grey line represents the 1:1 line.</p>
Full article ">Figure 11
<p>Density scatterplots comparing SGLI G-Portal and SGLI JASMES for (<b>a</b>) Chla and (<b>b</b>) TSM over 22 sites covering Case 1 and Case 2 waters (blue cross, <a href="#remotesensing-17-00221-f001" class="html-fig">Figure 1</a>a). N refers to the number of matchups, while R<sup>2</sup> and S denote the coefficient of determination and the slope, respectively. The dashed grey line represents the 1:1 line.</p>
Full article ">Figure A1
<p>Trend and seasonality components of <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> for five key bands (<math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math>_412, <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math>_443, <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math>_490, <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math>_530, and <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>R</mi> </mrow> <mrow> <mi>r</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math>_670) at the Socheongcho AERONET-OC site. The trend plots (left column) and seasonality plots (right column) illustrate the temporal behavior of remote sensing reflectance across G-Portal, JASMES, and AERONET-OC data.</p>
Full article ">
19 pages, 13029 KiB  
Article
Object-Level Contrastive-Learning-Based Multi-Branch Network for Building Change Detection from Bi-Temporal Remote Sensing Images
by Shiming Li, Fengtao Yan, Cheng Liao, Qingfeng Hu, Kaifeng Ma, Wei Wang and Hui Zhang
Remote Sens. 2025, 17(2), 217; https://doi.org/10.3390/rs17020217 - 9 Jan 2025
Viewed by 553
Abstract
Buildings are fundamental elements of human environments, and detecting changes in them is crucial for land cover studies, urban expansion monitoring, and the detection of illegal construction activities. Existing methods primarily focus on pixel-level differences in bi-temporal remote sensing imagery. However, pseudo-changes, such [...] Read more.
Buildings are fundamental elements of human environments, and detecting changes in them is crucial for land cover studies, urban expansion monitoring, and the detection of illegal construction activities. Existing methods primarily focus on pixel-level differences in bi-temporal remote sensing imagery. However, pseudo-changes, such as variations in non-building areas caused by differences in illumination, seasonal changes, and other factors, pose significant challenges for reliable building change detection. To address these issues, we propose a novel object-level contrastive-learning-based multi-branch network (OCL-Net) for detecting building changes by integrating bi-temporal remote sensing images. First, we design a multi-head decoder to separately extract more distinguishable building change features and auxiliary semantic features from bi-temporal images, effectively leveraging building-specific priors. Second, an object-level contrastive learning loss is designed and jointly optimized with a pixel-level similarity loss to ensure the global consistency of buildings. Finally, an attention-based discriminative feature generation and fusion block is designed to enhance the representation of multi-scale change features. We validate the effectiveness of the proposed method through comparative experiments on the publicly available WHU-CD and S2Looking datasets. Our approach achieves IoU values of 88.54% and 51.94%, respectively, surpassing state-of-the-art methods for building change detection. Full article
Show Figures

Figure 1

Figure 1
<p>Overview of the proposed OCL-Net. Prior information about demolished and newly added building labels is used to supervise the model only during the training stage.</p>
Full article ">Figure 2
<p>Detailed architecture of the difference feature generation (DFG) and adaptive feature fusion (AFF) modules.</p>
Full article ">Figure 3
<p>Schematic diagram of positive and negative pair construction for the semi-supervised object-level contrastive loss.</p>
Full article ">Figure 4
<p>Sample examples from the WHU-CD and S2Looking datasets.</p>
Full article ">Figure 5
<p>Samples of building changes extracted from the WHU-CD test dataset, where (<b>a</b>) represents pre-temporal images, (<b>b</b>) represents post-temporal images, (<b>c</b>) represents change labels, and (<b>d</b>) shows the predicted building changes from OCL-Net.</p>
Full article ">Figure 6
<p>Samples of building changes extracted from the S2Looking test dataset, where (<b>a</b>) represents pre-temporal images, (<b>b</b>) represents post-temporal images, (<b>c</b>) represents change labels, and (<b>d</b>) shows the predicted building changes from OCL-Net.</p>
Full article ">Figure 7
<p>Comparison of the IoU of the accuracy over training epochs for each ablation experiment on the S2Looking test set.</p>
Full article ">Figure 8
<p>Comparison of the IoU of the accuracy over training epochs for each ablation experiment on the WHU-CD test set.</p>
Full article ">Figure 9
<p>Comparison of the proposed method with each introduced module on the S2Looking dataset.</p>
Full article ">Figure 10
<p>Comparison of the proposed method with each introduced module on the WHU-CD dataset.</p>
Full article ">Figure 11
<p>Comparison of models’ computational complexity and the numbers of parameters across different methods.</p>
Full article ">
20 pages, 4126 KiB  
Article
FD-YOLO: A YOLO Network Optimized for Fall Detection
by Hoseong Hwang, Donghyun Kim and Hochul Kim
Appl. Sci. 2025, 15(1), 453; https://doi.org/10.3390/app15010453 - 6 Jan 2025
Viewed by 807
Abstract
Falls are defined by the World Health Organization (WHO) as incidents in which an individual unintentionally falls to the ground or a lower level. Falls represent a serious public health issue, ranking as the second leading cause of death from unintentional injuries, following [...] Read more.
Falls are defined by the World Health Organization (WHO) as incidents in which an individual unintentionally falls to the ground or a lower level. Falls represent a serious public health issue, ranking as the second leading cause of death from unintentional injuries, following traffic accidents. While fall prevention is crucial, prompt intervention after a fall is equally necessary. Delayed responses can result in severe complications, reduced recovery potential, and a negative impact on quality of life. This study focuses on detecting fall situations using image-based methods. The fall images utilized in this research were created by combining three open-source datasets to enhance generalization and adaptability across diverse scenarios. Because falls must be detected promptly, the YOLO (You Only Look Once) network, known for its effectiveness in real-time detection, was applied. To better capture the complex body structures and interactions with the floor during a fall, two key techniques were integrated. First, a global attention module (GAM) based on the Convolutional Block Attention Module (CBAM) was employed to improve detection performance. Second, a Transformer-based Swin Transformer module was added to effectively learn global spatial information and enable a more detailed analysis of body movements. This study prioritized minimizing missed fall detections (false negatives, FN) as the key performance metric, since undetected falls pose greater risks than false detections. The proposed Fall Detection YOLO (FD-YOLO) network, developed by integrating the Swin Transformer and GAM into YOLOv9, achieved a high [email protected] score of 0.982 and recorded only 134 missed fall incidents, demonstrating optimal performance. When implemented in environments equipped with standard camera systems, the proposed FD-YOLO network is expected to enable real-time fall detection and prompt post-fall responses. This technology has the potential to significantly improve public health and safety by preventing fall-related injuries and facilitating rapid interventions. Full article
Show Figures

Figure 1

Figure 1
<p>Fall dataset.</p>
Full article ">Figure 2
<p>Attention block.</p>
Full article ">Figure 3
<p>Global attention mechanism.</p>
Full article ">Figure 4
<p>Swin Transformer.</p>
Full article ">Figure 5
<p>FD module.</p>
Full article ">Figure 6
<p>FD-YOLO structure.</p>
Full article ">Figure 7
<p>Detection samples.</p>
Full article ">Figure 8
<p>FD-YOLO PR curve.</p>
Full article ">Figure 9
<p>FD-YOLO training curve.</p>
Full article ">Figure 10
<p>FD-YOLO confusion matrix.</p>
Full article ">
24 pages, 46652 KiB  
Article
Hyperspectral Reconstruction Method Based on Global Gradient Information and Local Low-Rank Priors
by Chipeng Cao, Jie Li, Pan Wang, Weiqiang Jin, Runrun Zou and Chun Qi
Remote Sens. 2024, 16(24), 4759; https://doi.org/10.3390/rs16244759 - 20 Dec 2024
Viewed by 614
Abstract
Hyperspectral compressed imaging is a novel imaging detection technology based on compressed sensing theory that can quickly acquire spectral information of terrestrial objects in a single exposure. It combines reconstruction algorithms to recover hyperspectral data from low-dimensional measurement images. However, hyperspectral images from [...] Read more.
Hyperspectral compressed imaging is a novel imaging detection technology based on compressed sensing theory that can quickly acquire spectral information of terrestrial objects in a single exposure. It combines reconstruction algorithms to recover hyperspectral data from low-dimensional measurement images. However, hyperspectral images from different scenes often exhibit high-frequency data sparsity and existing deep reconstruction algorithms struggle to establish accurate mapping models, leading to issues with detail loss in the reconstruction results. To address this issue, we propose a hyperspectral reconstruction method based on global gradient information and local low-rank priors. First, to improve the prior model’s efficiency in utilizing information of different frequencies, we design a gradient sampling strategy and training framework based on decision trees, leveraging changes in the loss function gradient information to enhance the model’s predictive capability for data of varying frequencies. Second, utilizing the local low-rank prior characteristics of the representative coefficient matrix, we develop a sparse sensing denoising module to effectively improve the local smoothness of point predictions. Finally, by establishing a regularization term for the reconstruction process based on the semantic similarity between the denoised results and prior spectral data, we ensure spatial consistency and spectral fidelity in the reconstruction results. Experimental results indicate that the proposed method achieves better detail recovery across different scenes, demonstrates improved generalization performance for reconstructing information of various frequencies, and yields higher reconstruction quality. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Structural composition of the DCCHI system and data structure of SD-CASSI detector sampling.</p>
Full article ">Figure 2
<p>Reconstruction algorithm framework.</p>
Full article ">Figure 3
<p>RGB images from the KAIST, Harvard, and hyperspectral remote sensing datasets.</p>
Full article ">Figure 4
<p>Selected spectral curves of the pixel point with coordinates (180, 70), showing a visual comparison of different methods in the spectral dimension and comparing the pseudocolor images and local spatial detail information under different wavelengths.</p>
Full article ">Figure 5
<p>Selected spectral curves of the pixel point with coordinates (150, 100), showing a visual comparison of different methods in the spectral dimension and comparing the pseudocolor images and local spatial detail information of different wavelengths.</p>
Full article ">Figure 6
<p>Comparison of the spectral consistency of reconstruction results with different methods on the PaviaU hyperspectral remote sensing datast at sample point coordinates (180, 110), along with a comparison of the spatial detail information of the reconstruction results at different wavelengths.</p>
Full article ">Figure 7
<p>Comparison of the spectral consistency of reconstruction results with different methods on the PaviaC hyperspectral remote sensing dataset at sample point coordinates (190, 50), along with a comparison of the spatial detail information of the reconstruction results at different wavelengths.</p>
Full article ">Figure 8
<p>Comparison of spectral reconstruction results for different crops.</p>
Full article ">Figure 9
<p>Impact of hyperparameter settings on reconstruction quality.</p>
Full article ">Figure 10
<p>Comparison of pseudocolor images generated from the predictions of different prior models for the Harvard Scene 04 hyperspectral data at the 5th, 12th, and 25th bands, along with the spectral differences of the predictions at different wavelengths.</p>
Full article ">Figure 11
<p>Comparison of pseudocolor images generated from the predictions of different prior models for the PaviaU hyperspectral remote sensing data at the 3rd, 13th, and 26th bands, along with the spectral differences of the predictions at different wavelengths.</p>
Full article ">Figure 12
<p>Variation in reconstruction quality with increasing iteration count under the same solving framework for different regularization constraint methods.</p>
Full article ">
16 pages, 3705 KiB  
Article
Multimodal Contrastive Learning for Remote Sensing Image Feature Extraction Based on Relaxed Positive Samples
by Zhenshi Zhang, Qiujun Li, Wenxuan Jing, Guangjun He, Lili Zhu and Shijuan Gao
Sensors 2024, 24(23), 7719; https://doi.org/10.3390/s24237719 - 3 Dec 2024
Cited by 1 | Viewed by 801
Abstract
Traditional multimodal contrastive learning brings text and its corresponding image closer together as a positive pair, where the text typically consists of fixed sentence structures or specific descriptive statements, and the image features are generally global features (with some fine-grained work using local [...] Read more.
Traditional multimodal contrastive learning brings text and its corresponding image closer together as a positive pair, where the text typically consists of fixed sentence structures or specific descriptive statements, and the image features are generally global features (with some fine-grained work using local features). Similar to unimodal self-supervised contrastive learning, this approach can be seen as enforcing a strict identity constraint in a multimodal context. However, due to the inherent complexity of remote sensing images, which cannot be easily described in a single sentence, and the fact that remote sensing images contain rich ancillary information beyond just object features, this strict identity constraint may be insufficient. To fully leverage the characteristics of remote sensing images, we propose a multimodal contrastive learning method for remote sensing image feature extraction, based on positive sample tripartite relaxation, where the model is relaxed in three aspects. The first aspect of relaxation involves both the text and image inputs. By introducing learnable parameters in the language and image branches, instead of relying on fixed sentence structures and fixed image features, the network can achieve a more flexible description of remote sensing images in text and extract ancillary information from the image features, thereby relaxing the input constraints. Second relaxation is achieved through multimodal alignment of various features. By aligning semantic information with the corresponding semantic regions in the images, the method allows for the relaxation of local image features under semantic constraints. This approach addresses the issue of selecting image patches in unimodal settings, where there is no semantic constraint. The proposed method for remote sensing image feature extraction has been validated on four datasets. On the PatternNet dataset, it achieved a 91.1% accuracy with just one-shot. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

Figure 1
<p>Overall structure diagram of MRiSSNet.</p>
Full article ">Figure 2
<p>Language Prompt Diagram: the left side represents the previous prompt method, and the right side represents our prompt method.</p>
Full article ">Figure 3
<p>Visual Prompt Diagram: the left side represents the previous prompt method, and the right side represents our prompt method.</p>
Full article ">Figure 4
<p>Relaxed identity positive sample selection under semantic guidance, where the red parts represent the top k image patches most similar to the semantics.</p>
Full article ">Figure 5
<p>Dataset visualization results.</p>
Full article ">Figure 6
<p>Experimental results of various methods on different datasets for 2/4/8/16 shots.</p>
Full article ">Figure 7
<p>Visualization of results for relaxed identity sample selection under semantic guidance.</p>
Full article ">
22 pages, 96008 KiB  
Article
HSD2Former: Hybrid-Scale Dual-Domain Transformer with Crisscrossed Interaction for Hyperspectral Image Classification
by Binxin Luo, Meihui Li, Yuxing Wei, Haorui Zuo, Jianlin Zhang and Dongxu Liu
Remote Sens. 2024, 16(23), 4411; https://doi.org/10.3390/rs16234411 - 25 Nov 2024
Viewed by 653
Abstract
An unescapable trend of hyperspectral image (HSI) has been toward classification with high accuracy and splendid performance. In recent years, Transformers have made remarkable progress in the HSI classification task. However, Transformer-based methods still encounter two main challenges. First, they concentrate on extracting [...] Read more.
An unescapable trend of hyperspectral image (HSI) has been toward classification with high accuracy and splendid performance. In recent years, Transformers have made remarkable progress in the HSI classification task. However, Transformer-based methods still encounter two main challenges. First, they concentrate on extracting spectral information and are incapable of using spatial information to a great extent. Second, they lack the utilization of multiscale features and do not sufficiently combine the advantages of the Transformer’s global feature extraction and multiscale feature extraction. To tackle these challenges, this article proposes a new solution named the hybrid-scale dual-domain Transformer with crisscrossed interaction (HSD2Former) for HSI classification. HSD2Former consists of three functional modules: dual-dimension multiscale convolutional embedding (D2MSCE), mixed domainFormer (MDFormer), and pyramid scale fusion block (PSFB). D2MSCE supersedes conventional patch embedding to generate spectral and spatial tokens at different scales, effectively enriching the diversity of spectral-spatial features. MDFormer is designed to facilitate self-enhancement and information interaction between the spectral domain and spatial domain, alleviating the heterogeneity of the spatial domain and spectral domain. PSFB introduces a straightforward fusion manner to achieve advanced semantic information for classification. Extensive experiments conducted on four datasets demonstrate the robustness and significance of HSD2Former. The classification evaluation indicators of OA, AA, and Kappa on four datasets almost exceed 98%, reaching state-of-the-art performance. Full article
Show Figures

Figure 1

Figure 1
<p>Overall architecture of hybrid-scale dual-domain Transformer with crisscrossed interaction (HSD<sup>2</sup>Former).</p>
Full article ">Figure 2
<p>Structure of dual-dimension multiscale convolutional embedding (D<sup>2</sup>MSCE). (<b>a</b>) spectral multiscale convolutional embedding (SeMSCE), (<b>b</b>) spatial multiscale convolutional embedding (SaMSCE).</p>
Full article ">Figure 3
<p>Structure of mixed domainFormer (MDFormer).</p>
Full article ">Figure 4
<p>Structure of pyramid scale fusion block (PSFB).</p>
Full article ">Figure 5
<p>Visual comparison results on the S-A dataset.</p>
Full article ">Figure 6
<p>Visual comparison results on the UP dataset.</p>
Full article ">Figure 7
<p>Visual comparison results on the HU dataset.</p>
Full article ">Figure 8
<p>Visual comparison results on the IP dataset.</p>
Full article ">Figure 9
<p>t-SNE on the S-A dataset.</p>
Full article ">Figure 10
<p>t-SNE on the UP dataset.</p>
Full article ">Figure 11
<p>t-SNE on the HU dataset.</p>
Full article ">Figure 12
<p>t-SNE on the IP dataset.</p>
Full article ">Figure 13
<p>Sensitivity of principal component number on four datasets: (<b>a</b>) on the S-A dataset, (<b>b</b>) on the UP dataset, (<b>c</b>) on the HU dataset, (<b>d</b>) on the IP dataset.</p>
Full article ">Figure 14
<p>Sensitivity of space size on four datasets: (<b>a</b>) on the S-A dataset, (<b>b</b>) on the UP dataset, (<b>c</b>) on the HU dataset, (<b>d</b>) on the IP dataset.</p>
Full article ">Figure 15
<p>Sensitivity of convolutional kernel number on four datasets: (<b>a</b>) on the S-A dataset, (<b>b</b>) on the UP dataset, (<b>c</b>) on the HU dataset, (<b>d</b>) on the IP dataset.</p>
Full article ">Figure 16
<p>Sensitivity of multihead number on four datasets: (<b>a</b>) on the S-A dataset, (<b>b</b>) on the UP dataset, (<b>c</b>) on the HU dataset, (<b>d</b>) on the IP dataset.</p>
Full article ">Figure 17
<p>Sensitivity of pooling stride on four datasets: (<b>a</b>) on the S-A dataset, (<b>b</b>) on the UP dataset, (<b>c</b>) on the HU dataset, (<b>d</b>) on the IP dataset.</p>
Full article ">Figure 18
<p>Sensitivity of training sample ratio on four datasets: (<b>a</b>) on the S-A dataset, (<b>b</b>) on the UP dataset, (<b>c</b>) on the HU dataset, (<b>d</b>) on the IP dataset.</p>
Full article ">
23 pages, 32729 KiB  
Article
PLC-Fusion: Perspective-Based Hierarchical and Deep LiDAR Camera Fusion for 3D Object Detection in Autonomous Vehicles
by Husnain Mushtaq, Xiaoheng Deng, Fizza Azhar, Mubashir Ali and Hafiz Husnain Raza Sherazi
Information 2024, 15(11), 739; https://doi.org/10.3390/info15110739 - 19 Nov 2024
Cited by 1 | Viewed by 1483
Abstract
Accurate 3D object detection is essential for autonomous driving, yet traditional LiDAR models often struggle with sparse point clouds. We propose perspective-aware hierarchical vision transformer-based LiDAR-camera fusion (PLC-Fusion) for 3D object detection to address this. This efficient, multi-modal 3D object detection framework integrates [...] Read more.
Accurate 3D object detection is essential for autonomous driving, yet traditional LiDAR models often struggle with sparse point clouds. We propose perspective-aware hierarchical vision transformer-based LiDAR-camera fusion (PLC-Fusion) for 3D object detection to address this. This efficient, multi-modal 3D object detection framework integrates LiDAR and camera data for improved performance. First, our method enhances LiDAR data by projecting them onto a 2D plane, enabling the extraction of object perspective features from a probability map via the Object Perspective Sampling (OPS) module. It incorporates a lightweight perspective detector, consisting of interconnected 2D and monocular 3D sub-networks, to extract image features and generate object perspective proposals by predicting and refining top-scored 3D candidates. Second, it leverages two independent transformers—CamViT for 2D image features and LidViT for 3D point cloud features. These ViT-based representations are fused via the Cross-Fusion module for hierarchical and deep representation learning, improving performance and computational efficiency. These mechanisms enhance the utilization of semantic features in a region of interest (ROI) to obtain more representative point features, leading to a more effective fusion of information from both LiDAR and camera sources. PLC-Fusion outperforms existing methods, achieving a mean average precision (mAP) of 83.52% and 90.37% for 3D and BEV detection, respectively. Moreover, PLC-Fusion maintains a competitive inference time of 0.18 s. Our model addresses computational bottlenecks by eliminating the need for dense BEV searches and global attention mechanisms while improving detection range and precision. Full article
(This article belongs to the Special Issue Emerging Research in Object Tracking and Image Segmentation)
Show Figures

Figure 1

Figure 1
<p>The architecture of our PLC-Fusion model for 3D object detection using LiDAR and camera data. The raw point cloud from LiDAR and raw image data are processed by separate 3D and 2D backbones, respectively. Perspective-based sampling is applied to both modalities before passing through a vision transformer (ViT)-based model (LiDViT for LiDAR data and CamViT for image data) to establish 2D and 3D correspondence. The Cross-Fusion module integrates these features, followed by region of interest (RoI)-based 3D detection for generating 3D bounding box predictions.</p>
Full article ">Figure 2
<p>Graphical depiction of the object perspective sampling process for LiDAR and camera data within the multimodal fusion model.</p>
Full article ">Figure 3
<p>Illustration of our object perspective sampling and projection process for LiDAR and camera data within the multimodal fusion model. The sampled points from LiDAR and camera images are projected into their respective 3D and 2D coordinate systems. Sparse feature extraction is applied to both modalities before being passed into vision transformer (ViT)-based encoders (LiDAR-ViT for LiDAR features and camera-ViT for image features). These extracted features are then fused in the Cross-Fusion module to establish a 2D–3D correspondence for improved multimodal 3D object detection.</p>
Full article ">Figure 4
<p>The figure illustrates the vision transformer (ViT)-based cross-fusion approach for 3D object detection, combining camera and LiDAR data. Object perspective sampling extracts features from both sensors. The camera branch (CamViT) generates 3D and 2D feature maps <math display="inline"><semantics> <mrow> <msub> <mi mathvariant="bold">H</mi> <mi>c</mi> </msub> <mo>∈</mo> <msup> <mi mathvariant="double-struck">R</mi> <mrow> <msub> <mi>N</mi> <mi>c</mi> </msub> <mo>×</mo> <msub> <mi>D</mi> <mi>c</mi> </msub> </mrow> </msup> </mrow> </semantics></math> using multi-head attention (MH-Attention) and a feedforward neural network (FFN), while the LiDAR branch (LiDViT) processes 3D voxel features <math display="inline"><semantics> <mrow> <msub> <mi mathvariant="bold">H</mi> <mi>v</mi> </msub> <mo>∈</mo> <msup> <mi mathvariant="double-struck">R</mi> <mrow> <mi>V</mi> <mo>×</mo> <msub> <mi>D</mi> <mi>v</mi> </msub> </mrow> </msup> </mrow> </semantics></math> through a similar transformer architecture. The 2D and 3D feature maps from both modalities are concatenated <math display="inline"><semantics> <mrow> <msub> <mi mathvariant="bold">F</mi> <mrow> <mi>f</mi> <mi>u</mi> <mi>s</mi> <mi>i</mi> <mi>o</mi> <mi>n</mi> </mrow> </msub> <mo>=</mo> <mrow> <mo>[</mo> <msub> <mi mathvariant="bold">A</mi> <mrow> <mi>c</mi> <mi>v</mi> </mrow> </msub> <mo>;</mo> <msub> <mi mathvariant="bold">H</mi> <mi>c</mi> </msub> <mo>;</mo> <msub> <mi mathvariant="bold">A</mi> <mrow> <mi>v</mi> <mi>c</mi> </mrow> </msub> <mo>;</mo> <msub> <mi mathvariant="bold">H</mi> <mi>v</mi> </msub> <mo>]</mo> </mrow> </mrow> </semantics></math> and undergo cross-attention to align visual and geometric data. A final FFN refines the fused representation <math display="inline"><semantics> <mrow> <msub> <mi mathvariant="bold">F</mi> <mrow> <mi>f</mi> <mi>i</mi> <mi>n</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mo>=</mo> <mi>MLP</mi> <mrow> <mo>(</mo> <msub> <mi mathvariant="bold">F</mi> <mrow> <mi>f</mi> <mi>u</mi> <mi>s</mi> <mi>i</mi> <mi>o</mi> <mi>n</mi> </mrow> </msub> <mo>)</mo> </mrow> </mrow> </semantics></math>, providing deep multimodal features for accurate object detection in 3D space.</p>
Full article ">Figure 5
<p>Visual results of the proposed method on the KITTI validation dataset. For each case of sub-figures (<b>a</b>–<b>d</b>), the top row shows the visualization in the RGB image, and the bottom row displays the visualization in the LiDAR point cloud. Green represents the ground truth, and blue denotes the predicted outcomes.</p>
Full article ">Figure 6
<p>Visual results of the proposed method on the KITTI test and validation datasets. Row (<b>a</b>) presents the testing results, and row (<b>b</b>) displays the validation outcomes. The detection results demonstrate the effectiveness of our method, with the dotted circles highlighting the undetected instances caused by distance and heavy occlusion.</p>
Full article ">Figure 7
<p>Car class with Moderate condition: AP vs. IoU on KITII validation set.</p>
Full article ">Figure 8
<p>Comparative analysis of the runtime of our model with recent methods.</p>
Full article ">
24 pages, 392 KiB  
Article
Updated-Food Choice Questionnaire: Cultural Adaptation and Validation in a Spanish-Speaking Population from Mexico
by Miguel Amaury Salas-García, María Fernanda Bernal-Orozco, Andrés Díaz-López, Alejandra Betancourt-Núñez, Pablo Alejandro Nava-Amante, Ina Danquah, J. Alfredo Martínez, Daniel A. de Luis and Barbara Vizmanos
Nutrients 2024, 16(21), 3749; https://doi.org/10.3390/nu16213749 - 31 Oct 2024
Cited by 1 | Viewed by 1778
Abstract
Background: Determinants and motives related to food selection have evolved in a globalized and changing world. The traditional and useful Food Choice Questionnaire (FCQ), created in 1995, needs to be updated, adapted to new scenarios, and validated. Objectives: This study aimed to: (1) [...] Read more.
Background: Determinants and motives related to food selection have evolved in a globalized and changing world. The traditional and useful Food Choice Questionnaire (FCQ), created in 1995, needs to be updated, adapted to new scenarios, and validated. Objectives: This study aimed to: (1) assess face validity (FV) of the original 36-item FCQ, (2) generate an Updated-FCQ (U-FCQ) and assess its content validity (CV) (instrument suitability), and (3) evaluate its construct validity and reliability in a Spanish-speaking population from Mexico. Methods: FV involved a panel of nutrition professionals (NPs) rating the original items’ clarity, relevance, specificity, and representativeness. A literature review process updated the FCQ by adding new items. CV with a second NP panel allowed calculating content validity ratio (CVR). Construct validation was performed via exploratory and confirmatory factor analysis (EFA-CFA). Internal consistency through Cronbach’s alpha (CA) and test–retest reliability via intra-class correlation (ICC) were assessed. Results: The FV (n = 8) resulted in the modification of 11 original items. The literature review added 36 new items (15 from previous adaptations and 21 original items). The CV (n = 13) identified nine items (non-acceptable CVR), prompting reformulation of seven and removal of two. The NPs’ feedback added six new items. The EFA-CFA (n = 788) developed a 75-item U-FCQ with eight dimensions: sensory appeal, mood, health and nutritional content, price, food identity, environmental and wildlife awareness, convenience, and image management. CA ranged from 0.74–0.97 (good–excellent) and ICC from 0.51–0.78 (moderate–good). Conclusions: This study provides a useful instrument for the assessment of food choices and lays the groundwork for future cross-cultural comparisons, expanding its applicability in wider settings. Full article
(This article belongs to the Section Nutrition and Public Health)
Show Figures

Graphical abstract

Graphical abstract
Full article ">
Back to TopTop