Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (328)

Search Parameters:
Keywords = dual head

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 14147 KiB  
Article
Few-Shot Object Detection for Remote Sensing Imagery Using Segmentation Assistance and Triplet Head
by Jing Zhang, Zhaolong Hong, Xu Chen and Yunsong Li
Remote Sens. 2024, 16(19), 3630; https://doi.org/10.3390/rs16193630 - 29 Sep 2024
Viewed by 1398
Abstract
The emergence of few-shot object detection provides a new approach to address the challenge of poor generalization ability due to data scarcity. Currently, extensive research has been conducted on few-shot object detection in natural scene datasets, and notable progress has been made. However, [...] Read more.
The emergence of few-shot object detection provides a new approach to address the challenge of poor generalization ability due to data scarcity. Currently, extensive research has been conducted on few-shot object detection in natural scene datasets, and notable progress has been made. However, in the realm of remote sensing, this technology is still lagging behind. Furthermore, many established methods rely on two-stage detectors, prioritizing accuracy over speed, which hinders real-time applications. Considering both detection accuracy and speed, in this paper, we propose a simple few-shot object detection method based on the one-stage detector YOLOv5 with transfer learning. First, we propose a Segmentation Assistance (SA) module to guide the network’s attention toward foreground targets. This module assists in training and enhances detection accuracy without increasing inference time. Second, we design a novel detection head called the Triplet Head (Tri-Head), which employs a dual distillation mechanism to mitigate the issue of forgetting base-class knowledge. Finally, we optimize the classification loss function to emphasize challenging samples. Evaluations on the NWPUv2 and DIOR datasets showcase the method’s superiority. Full article
Show Figures

Figure 1

Figure 1
<p>Comparative visualization between the detection results of TFA (first row) versus the proposed method (second row) on the NWPUv2 test dataset under the 3-shot setting. There are many missed detections and false detections in the TFA results.</p>
Full article ">Figure 2
<p>Overview of the proposed framework. In the base training stage, the entire model is trained using the base set. During the fine-tuning stage, the backbone is frozen, while the remaining parts follow the training strategy outlined in <a href="#sec3dot6-remotesensing-16-03630" class="html-sec">Section 3.6</a>. SA enhances attention to foreground targets via binary map prediction and <math display="inline"><semantics> <mrow> <msub> <mi>L</mi> <mrow> <mi>S</mi> <mi>A</mi> </mrow> </msub> </mrow> </semantics></math> calculation against the real binary mask map. Tri-Head, a Triplet Head, adopts a dual knowledge distillation mechanism to mitigate the issue of forgetting base-class knowledge. <math display="inline"><semantics> <mrow> <msub> <mi>L</mi> <mrow> <mi>E</mi> <mi>n</mi> <mo>−</mo> <mi>C</mi> <mi>l</mi> <mi>s</mi> </mrow> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>L</mi> <mrow> <mi>E</mi> <mi>n</mi> <mo>−</mo> <mi>O</mi> <mi>b</mi> <mi>j</mi> </mrow> </msub> </mrow> </semantics></math> are enhanced classification loss functions, emphasizing challenging samples.</p>
Full article ">Figure 3
<p>The structure of SA.</p>
Full article ">Figure 4
<p>The process of obtaining labels required by SA. The figure shows the process of obtaining labels under the 3-shot setting of the NWPUv2 dataset.</p>
Full article ">Figure 5
<p>The structure of Triplet Head. <math display="inline"><semantics> <mrow> <msub> <mi>H</mi> <mrow> <mi>b</mi> <mi>a</mi> <mi>s</mi> <mi>e</mi> </mrow> </msub> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <msub> <mi>H</mi> <mrow> <mi>o</mi> <mi>u</mi> <mi>t</mi> <mi>p</mi> <mi>u</mi> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math>, and <math display="inline"><semantics> <mrow> <msub> <mi>H</mi> <mrow> <mi>n</mi> <mi>o</mi> <mi>v</mi> <mi>e</mi> <mi>l</mi> </mrow> </msub> </mrow> </semantics></math> refer to base-class detection head, final output head, and novel-class detection head, respectively. <math display="inline"><semantics> <mrow> <msub> <mi>L</mi> <mrow> <mi>D</mi> <mi>i</mi> <mi>s</mi> <mi>t</mi> <mi>i</mi> <mi>l</mi> <mi>l</mi> <mi>a</mi> <mi>t</mi> <mi>i</mi> <mi>o</mi> <mi>n</mi> <mo>-</mo> <mi>b</mi> <mi>a</mi> <mi>s</mi> <mi>e</mi> </mrow> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>L</mi> <mrow> <mrow> <mi mathvariant="italic">Distillation</mi> <mo mathvariant="italic">-</mo> <mi mathvariant="italic">novel</mi> </mrow> </mrow> </msub> </mrow> </semantics></math> are base-class distillation loss and novel-class distillation loss, respectively.</p>
Full article ">Figure 6
<p>Training sample statistics on the DIOR dataset under the 20-shot setting in Split1.</p>
Full article ">Figure 7
<p>Visualization of detection results of our method versus YOLOv5 and G-FSDet on the NWPUv2 dataset under the 3-shot setting.</p>
Full article ">Figure 8
<p>Results of each category in base classes on the NWPUv2 dataset under the 3-shot setting.</p>
Full article ">Figure 9
<p>Results of each category in novel classes on the NWPUv2 dataset under the 3-shot setting.</p>
Full article ">Figure 10
<p>Confusion matrix on NWPUv2 under the 3-shot setting.</p>
Full article ">Figure 11
<p>Visualization of detection results of our method versus YOLOv5 and G-FSDet on the DIOR dataset under the 10-shot setting in Split1.</p>
Full article ">
11 pages, 820 KiB  
Article
Graph Neural Network-Based Modeling with Subcategory Exploration for Drug Repositioning
by Rong Lu, Yong Liang, Jiatai Lin and Yuqiang Chen
Electronics 2024, 13(19), 3835; https://doi.org/10.3390/electronics13193835 - 28 Sep 2024
Viewed by 370
Abstract
Drug repositioning is a cost-effective approach to identifying new indications for existing drugs by predicting their associations with new diseases or symptoms. Recently, deep learning-based models have become the mainstream for drug repositioning. Existing methods typically regard the drug-repositioning task as a binary [...] Read more.
Drug repositioning is a cost-effective approach to identifying new indications for existing drugs by predicting their associations with new diseases or symptoms. Recently, deep learning-based models have become the mainstream for drug repositioning. Existing methods typically regard the drug-repositioning task as a binary classification problem to find the new drug–disease associations. However, drug–disease associations may encompass some potential subcategories that can be used to enhance the classification performance. In this paper, we propose a prototype-based subcategory exploration (PSCE) model to guide the model learned with the information of a potential subcategory for drug repositioning. To achieve this, we first propose a prototype-based feature-enhancement mechanism (PFEM) that uses clustering centroids as the attention to enhance the drug–disease features by introducing subcategory information to improve the association prediction. Second, we introduce the drug–disease dual-task classification head (D3TC) of the model, which consists of a traditional binary classification head and a subcategory-classification head to learn with subcategory exploration. It leverages finer-grained pseudo-labels of subcategories to introduce additional knowledge for precise drug–disease association classification. In this study, we conducted experiments on four public datasets to compare the proposed PSCE with existing state-of-the-art approaches and our PSCE achieved a better performance than the existing ones. Finally, the effectiveness of the PFEM and D3TC was demonstrated using ablation studies. Full article
(This article belongs to the Special Issue Network Security Management in Heterogeneous Networks)
Show Figures

Figure 1

Figure 1
<p>Illustration of the proposed PSCE pipeline. The middle part of this diagram shows the main process of the entire pipeline. (<b>a</b>) The proposed prototype-based feature enhancement mechanism (PFEM), (<b>b</b>) the feature concatenation and split steps, and (<b>c</b>) the proposed drug–disease dual-task classification head (D3TC).</p>
Full article ">Figure 2
<p>Bar chart of performance that compares our PSCE and six existing methods. The <span style="color: #0000FF">blue</span> and <span style="color: #00FF00">green</span> bars represent the performances according to the AUROC and AUPRC metrics, respectively.</p>
Full article ">Figure 3
<p>Visualization of performance generated by our PSCE and existing methods. Top and bottom represent the scatter plot of performances according to AUROC and AUPRC metrics, respectively.</p>
Full article ">Figure 4
<p>The effects of different combinations of the proposed PFEM and D3TC on four datasets. The top and bottom represent the line charts of performances with AUROC and AUPRC metrics, respectively.</p>
Full article ">
13 pages, 803 KiB  
Brief Report
Emulating the Delivery of Sawtooth Proton Arc Therapy Plans on a Cyclotron-Based Proton Beam Therapy System
by Samuel Burford-Eyre, Adam Aitkenhead, Jack D. Aylward, Nicholas T. Henthorn, Samuel P. Ingram, Ranald Mackay, Samuel Manger, Michael J. Merchant, Peter Sitch, John-William Warmenhoven and Robert B. Appleby
Cancers 2024, 16(19), 3315; https://doi.org/10.3390/cancers16193315 - 27 Sep 2024
Viewed by 267
Abstract
Purpose: To evaluate and compare the deliverability of ‘sawtooth’ proton arc therapy (PAT) plans relative to static intensity modulated proton therapy (IMPT) at a cyclotron-based clinical facility. Methods: The delivery of single and dual arc Sawtooth PAT plans for an abdominal [...] Read more.
Purpose: To evaluate and compare the deliverability of ‘sawtooth’ proton arc therapy (PAT) plans relative to static intensity modulated proton therapy (IMPT) at a cyclotron-based clinical facility. Methods: The delivery of single and dual arc Sawtooth PAT plans for an abdominal CT phantom and multiple clinical cases of brain, head and neck (H&N) and base of skull (BoS) targets was emulated under the step-and-shoot and continuous PAT delivery regimes and compared to that of a corresponding static IMPT plan. Results: Continuous PAT delivery increased the time associated with beam delivery and gantry movement in single/dual PAT plans by 4.86/7.34 min (brain), 7.51/12.40 min (BoS) and 6.59/10.57 min (H&N) on average relative to static IMPT. Step-and-shoot PAT increased this delivery time further by 4.79 min on average as the delivery was limited by gantry motion. Conclusions: The emulator can approximately model clinical sawtooth PAT delivery but requires experimental validation. No clear benefit was observed regarding beam-on time for sawtooth PAT relative to static IMPT. Full article
(This article belongs to the Special Issue The Advance of Pencil Beam Scanning Proton Beam Therapy in Cancers)
Show Figures

Figure 1

Figure 1
<p>Single-fraction dose distribution of the ependymoma E1 case under continuous single (<b>a</b>) and dual (<b>c</b>) arc PAT delivery. Dose differences of each plan relative to the planned dose distribution of step-and-shoot PAT are shown in the right-hand column (<b>b</b>,<b>d</b>). CTV and brainstem contours shown in yellow and red, respectively. Isocentre marked using black cross. Isodose lines shown in 10% and 5% intervals for 10–90% and 90–105% single-fraction dose, respectively.</p>
Full article ">
24 pages, 14553 KiB  
Article
Multiple-Point Metamaterial-Inspired Microwave Sensors for Early-Stage Brain Tumor Diagnosis
by Nantakan Wongkasem and Gabriel Cabrera
Sensors 2024, 24(18), 5953; https://doi.org/10.3390/s24185953 - 13 Sep 2024
Viewed by 497
Abstract
Simple, instantaneous, contactless, multiple-point metamaterial-inspired microwave sensors, composed of multi-band, low-profile metamaterial-inspired antennas, were developed to detect and identify meningioma tumors, the most common primary brain tumors. Based on a typical meningioma tumor size of 5–20 mm, a higher operating frequency, where the [...] Read more.
Simple, instantaneous, contactless, multiple-point metamaterial-inspired microwave sensors, composed of multi-band, low-profile metamaterial-inspired antennas, were developed to detect and identify meningioma tumors, the most common primary brain tumors. Based on a typical meningioma tumor size of 5–20 mm, a higher operating frequency, where the wavelength is similar or smaller than the tumor target, is crucial. The sensors, designed for the microwave Ku band range (12–18 GHz), where the electromagnetic property values of tumors are available, were implemented in this study. A seven-layered head phantom, including the meningioma tumors, was defined using actual electromagnetic parametric values in the frequency range of interest to mimic the actual human head. The reflection coefficients can be recorded and analyzed instantaneously, reducing high electromagnetic radiation consumption. It has been shown that a single-band detection point is not adequate to classify the nonlinear tumor and head model parameters. On the other hand, dual-band and tri-band metamaterial-inspired antennas, with additional detecting points, create a continuous function solution for the nonlinear problem by adding extra observation points using multiple-band excitation. The point mapping values can be used to enhance the tumor detection capability. Two-point mapping showed a consistent trend between the S11 value order and the tumor size, while three-point mapping can also be used to demonstrate the correlation between the S11 value order and the tumor size. This proposed multi-detection point technique can be applied to a sensor for other nonlinear property targets. Moreover, a set of antennas with different polarizations, orientations, and arrangements in a network could help to obtain the highest sensitivity and accuracy of the whole system. Full article
(This article belongs to the Special Issue Biomedical Signals, Images and Healthcare Data Analysis)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) The multilayer-head phantom and (<b>b</b>) cross-sectional model in CST.</p>
Full article ">Figure 2
<p>(<b>a</b>) Common locations for meningiomas. Common sites of tumor growth in relationship to adjacent skull, brain, and dural reflections; A,B: Skull base meningiomas and C,D: Falcine meningiomas attached to the dense fibrous tissue of the falx [<a href="#B23-sensors-24-05953" class="html-bibr">23</a>]. Modelling Meningioma tumor locations: (<b>b</b>) exposed and (<b>c</b>) enveloped configurations.</p>
Full article ">Figure 3
<p>Frequency-dependent complex permittivity of meningiomas from 12–18 GHz [<a href="#B13-sensors-24-05953" class="html-bibr">13</a>].</p>
Full article ">Figure 4
<p>(<b>a</b>) Single–band Ku–band patch antenna and (<b>b</b>) its S<sub>11</sub> plot of different substrates.</p>
Full article ">Figure 5
<p>1D polar and 3D far-field plots of (<b>a</b>) FR–4, (<b>b</b>) Rogers 410, and (<b>c</b>) Rogers 430 substrate antenna.</p>
Full article ">Figure 6
<p>Head model and Ku-band single-band antenna setting.</p>
Full article ">Figure 7
<p>(<b>a</b>) Single–band Ku–band disc patch antenna and (<b>b</b>) its S<sub>11</sub> plot.</p>
Full article ">Figure 8
<p>(<b>a</b>) 1D polar plot and (<b>b</b>) 3D far–field plot of a disc antenna at 15 GHz.</p>
Full article ">Figure 9
<p>(<b>a</b>) Dual–band patch, (<b>b</b>) S<sub>11</sub> plot, (<b>c</b>) 1D polar plot, and (<b>d</b>) 3D far–field plot.</p>
Full article ">Figure 10
<p>(<b>a</b>) Nine SRR metamaterial structures. (<b>b</b>) Nine SRR metamaterial S parameters.</p>
Full article ">Figure 10 Cont.
<p>(<b>a</b>) Nine SRR metamaterial structures. (<b>b</b>) Nine SRR metamaterial S parameters.</p>
Full article ">Figure 11
<p>(<b>a</b>) Ku dipole and its (<b>b</b>) S<sub>11,</sub> (<b>c</b>) 1D polar, and (<b>d</b>) 3D far–field plots.</p>
Full article ">Figure 12
<p>(<b>a</b>) Tri–band metamaterial-inspired antenna, (<b>b</b>) its S<sub>11</sub> plot, and 3D far–field plots at (<b>c</b>) 14 GHz, (<b>d</b>) 15 GHz, and (<b>e</b>) 17.5 GHz.</p>
Full article ">Figure 12 Cont.
<p>(<b>a</b>) Tri–band metamaterial-inspired antenna, (<b>b</b>) its S<sub>11</sub> plot, and 3D far–field plots at (<b>c</b>) 14 GHz, (<b>d</b>) 15 GHz, and (<b>e</b>) 17.5 GHz.</p>
Full article ">Figure 13
<p>Testing the head model with a 8 mm diameter tumor using (<b>a</b>) rectangular and (<b>b</b>) disc patch antennas.</p>
Full article ">Figure 14
<p>S<sub>11</sub> plots when testing the rectangular patch using different brain tumor diameters: (<b>a</b>) 2, 4, 16, and 20 mm, and (<b>b</b>) 2, 4, 6, 8, 10, 12, 14, 16, 18, 20 mm.</p>
Full article ">Figure 15
<p>S<sub>11</sub> plots when testing the disc patch using different brain tumor diameters.</p>
Full article ">Figure 16
<p>S<sub>11</sub> plots when testing the dual–band patch: (<b>a</b>) 1st and (<b>b</b>) 2nd band, using brain tumors with different diameters.</p>
Full article ">Figure 17
<p>S<sub>11</sub> plots when testing the tri–band patch: (<b>a</b>) 1st, (<b>b</b>) 2nd, and (<b>c</b>) 3rd band, using brain tumors with different diameters.</p>
Full article ">Figure 17 Cont.
<p>S<sub>11</sub> plots when testing the tri–band patch: (<b>a</b>) 1st, (<b>b</b>) 2nd, and (<b>c</b>) 3rd band, using brain tumors with different diameters.</p>
Full article ">
26 pages, 6242 KiB  
Article
Wireless Sensor Node for Chemical Agent Detection
by Zabdiel Brito-Brito, Jesús Salvador Velázquez-González, Fermín Mira, Antonio Román-Villarroel, Xavier Artiga, Satyendra Kumar Mishra, Francisco Vázquez-Gallego, Jung-Mu Kim, Eduardo Fontana, Marcos Tavares de Melo and Ignacio Llamas-Garro
Chemosensors 2024, 12(9), 185; https://doi.org/10.3390/chemosensors12090185 - 11 Sep 2024
Viewed by 630
Abstract
In this manuscript, we present in detail the design and implementation of the hardware and software to produce a standalone wireless sensor node, called SensorQ system, for the detection of a toxic chemical agent. The proposed wireless sensor node prototype is composed of [...] Read more.
In this manuscript, we present in detail the design and implementation of the hardware and software to produce a standalone wireless sensor node, called SensorQ system, for the detection of a toxic chemical agent. The proposed wireless sensor node prototype is composed of a micro-controller unit (MCU), a radio frequency (RF) transceiver, a dual-band antenna, a rechargeable battery, a voltage regulator, and four integrated sensing devices, all of them integrated in a package with final dimensions and weight of 200 × 80 × 60 mm and 0.422 kg, respectively. The proposed SensorQ prototype operates using the Long-Range (LoRa) wireless communication protocol at 2.4 GHz, with a sensor head implemented on a hetero-core fiber optic structure supporting the surface plasmon resonance (SPR) phenomenon with a sensing section (L = 10 mm) coated with titanium/gold/titanium and a chemically sensitive material (zinc oxide) for the detection of Di-Methyl Methyl Phosphonate (DMMP) vapor in the air, a simulant of the toxic nerve agent Sarin. The transmitted spectra with respect to different concentrations of DMMP vapor in the air were recorded, and then the transmitted power for these concentrations was calculated at a wavelength of 750 nm. The experimental results indicate the feasibility of detecting DMMP vapor in air using the proposed optical sensor head, with DMMP concentrations in the air of 10, 150, and 150 ppm in this proof of concept. We expect that the sensor and wireless sensor node presented herein are promising candidates for integration into a wireless sensor network (WSN) for chemical warfare agent (CWA) detection and contaminated site monitoring without exposure of armed forces. Full article
Show Figures

Figure 1

Figure 1
<p>Hardware architecture of the proposed wireless sensor node.</p>
Full article ">Figure 2
<p>Wireless sensor node: (<b>a</b>) 3D model isometric (top/front/left) view, (<b>b</b>) 3D model lateral view, and (<b>c</b>) integrated and packaged wireless sensor node prototype.</p>
Full article ">Figure 3
<p>Architecture of the SensorQ system. Showing deployed wireless sensor nodes at the bottom of the figure connected to the communications gateway mounted on UAVs. The communications gateway makes data available to the end user through the MQTT protocol and 4G/5G wireless communications links.</p>
Full article ">Figure 4
<p>Wireless sensor node electronics: (<b>a</b>) communications side view and (<b>b</b>) sensors side view. A description of each part according to enclosed numbers is provided in <a href="#chemosensors-12-00185-t003" class="html-table">Table 3</a>.</p>
Full article ">Figure 5
<p>Wireless sensor node antenna: (<b>a</b>) top view, showing the stacked dual band antenna setup and (<b>b</b>) bottom view showing interconnections and power divider network.</p>
Full article ">Figure 6
<p>Gateway electronics. A description of each part according to enclosed numbers is provided in <a href="#chemosensors-12-00185-t005" class="html-table">Table 5</a>.</p>
Full article ">Figure 7
<p>Graphical representation of the proposed sensor probe supporting the SPR effect with stacked material layers deposited on the SMF: longitudinal optical fiber section (left) and optical fiber cross sections (right).</p>
Full article ">Figure 8
<p>Representation of data collection by the WSN composed of one gateway and three sensor nodes operating under low-power listening mode.</p>
Full article ">Figure 9
<p>Representation of the frame-slotted ALOHA’s (FSA) time organization whilst the gateway is collecting data from each sensor node into a defined sequence of frames (top), slot representation (bottom).</p>
Full article ">Figure 10
<p>Wireless sensor node software architecture based on four inter-related layers (L1–L4): L1 is for the Hardware Abstraction Layer, L2 is for the Real-Time Operating System, L3 is for the drivers to access other devices, and L4 is for the Application Layer.</p>
Full article ">Figure 11
<p>Wireless software architecture of the gateway based on four inter-related layers (L1–L4): L1 is for the interface with different peripherals, L2 is for the Raspbian operating system of the Raspberry Pi, L3 is for the MQT client, a GNSS receiver, and a Lora radio transceiver driver, and L4 is for the parallel running tasks.</p>
Full article ">Figure 12
<p>Screenshot of the configuration dashboard, which allows for the manipulation of several parameters regarding the experiment, the MAC layer, the PHY layer, and the commands sections.</p>
Full article ">Figure 13
<p>Deployment of different data collected in the measurements dashboard (screenshot), such as environmental conditions (gas concentration and temperature), the status (RSSI, acceleration, and battery level), and the location (GPS position and altitude) from two sensor node prototypes.</p>
Full article ">Figure 14
<p>Average data collection time depending on (<b>a</b>) the number of slots per frame for a given number of sensor nodes (each node sends 1 data packet of 22 bytes), (<b>b</b>) the number of sensor nodes for a given number of slots per frame (each node sends 1 data packet of 22 bytes), and (<b>c</b>) the number of slots per frame for a given number of sensor nodes (each node sends 10 data packets of 22 bytes or 1 data packet of 220 bytes). All results are presented for SF-6. (<b>a</b>) Data collection time over number of slots (single packet of 22 bytes), (<b>b</b>) data collection time over number of sensor nodes (single packet of 22 bytes), and (<b>c</b>) data collection time over number of slots (10 packets of 22 bytes or 1 packet of 220 bytes).</p>
Full article ">Figure 15
<p>Sensor head experimental setup based on the optical fiber hetero-core structure coated with Ti/Au/Ti/ZnO.</p>
Full article ">Figure 16
<p>Normalized transmitted intensity for different concentrations of DMMP mixed in the air and interaction with our proposed sensing probe (dots: measured data; dashed line: trend).</p>
Full article ">
13 pages, 5820 KiB  
Article
Optic Nerve Sheath Ultrasound Image Segmentation Based on CBC-YOLOv5s
by Yonghua Chu, Jinyang Xu, Chunshuang Wu, Jianping Ye, Jucheng Zhang, Lei Shen, Huaxia Wang and Yudong Yao
Electronics 2024, 13(18), 3595; https://doi.org/10.3390/electronics13183595 - 10 Sep 2024
Viewed by 382
Abstract
The diameter of the optic nerve sheath is an important indicator for assessing the intracranial pressure in critically ill patients. The methods for measuring the optic nerve sheath diameter are generally divided into invasive and non-invasive methods. Compared to the invasive methods, the [...] Read more.
The diameter of the optic nerve sheath is an important indicator for assessing the intracranial pressure in critically ill patients. The methods for measuring the optic nerve sheath diameter are generally divided into invasive and non-invasive methods. Compared to the invasive methods, the non-invasive methods are safer and have thus gained popularity. Among the non-invasive methods, using deep learning to process the ultrasound images of the eyes of critically ill patients and promptly output the diameter of the optic nerve sheath offers significant advantages. This paper proposes a CBC-YOLOv5s optic nerve sheath ultrasound image segmentation method that integrates both local and global features. First, it introduces the CBC-Backbone feature extraction network, which consists of dual-layer C3 Swin-Transformer (C3STR) and dual-layer Bottleneck Transformer (BoT3) modules. The C3STR backbone’s multi-layer convolution and residual connections focus on the local features of the optic nerve sheath, while the Window Transformer Attention (WTA) mechanism in the C3STR module and the Multi-Head Self-Attention (MHSA) in the BoT3 module enhance the model’s understanding of the global features of the optic nerve sheath. The extracted local and global features are fully integrated in the Spatial Pyramid Pooling Fusion (SPPF) module. Additionally, the CBC-Neck feature pyramid is proposed, which includes a single-layer C3STR module and three-layer CReToNeXt (CRTN) module. During upsampling feature fusion, the C3STR module is used to enhance the local and global awareness of the fused features. During downsampling feature fusion, the CRTN module’s multi-level residual design helps the network to better capture the global features of the optic nerve sheath within the fused features. The introduction of these modules achieves the thorough integration of the local and global features, enabling the model to efficiently and accurately identify the optic nerve sheath boundaries, even when the ocular ultrasound images are blurry or the boundaries are unclear. The Z2HOSPITAL-5000 dataset collected from Zhejiang University Second Hospital was used for the experiments. Compared to the widely used YOLOv5s and U-Net algorithms, the proposed method shows improved performance on the blurry test set. Specifically, the proposed method achieves precision, recall, and Intersection over Union (IoU) values that are 4.1%, 2.1%, and 4.5% higher than those of YOLOv5s. When compared to U-Net, the precision, recall, and IoU are improved by 9.2%, 21%, and 19.7%, respectively. Full article
(This article belongs to the Special Issue Deep Learning-Based Object Detection/Classification)
Show Figures

Figure 1

Figure 1
<p>CBC-YOLOv5s optic nerve sheath segmentation algorithm.</p>
Full article ">Figure 2
<p>C3STR module.</p>
Full article ">Figure 3
<p>BoT3 module.</p>
Full article ">Figure 4
<p>CRTN module.</p>
Full article ">Figure 5
<p>Different algorithms for visualization with normal and blurry images.</p>
Full article ">Figure 6
<p>Different algorithms for segmentation examples with normal and blurry images.</p>
Full article ">
11 pages, 3904 KiB  
Article
The Association between Fluoride and Bone Mineral Density in US Children and Adolescents: A Pilot Study
by Haichen Kong, Zihao He, Hui Li, Dan Xing and Jianhao Lin
Nutrients 2024, 16(17), 2948; https://doi.org/10.3390/nu16172948 - 2 Sep 2024
Viewed by 882
Abstract
The aim of this study was to examine the association between fluoride exposure and bone mineral density (BMD) in children and adolescents. We used data from the National Health and Nutrition Examination Survey (NHANES) 2015–2016. The fluoride concentrations in the water samples, plasma [...] Read more.
The aim of this study was to examine the association between fluoride exposure and bone mineral density (BMD) in children and adolescents. We used data from the National Health and Nutrition Examination Survey (NHANES) 2015–2016. The fluoride concentrations in the water samples, plasma samples, and urine samples were measured electrometrically using an ion-specific electrode. Total body less head BMD (TBLH BMD) was measured using dual-energy X-ray absorptiometry (DXA). Weighted generalized linear regression models and restricted cubic splines (RCS) regression models were used to analyze the relationships between the three types of fluoride exposure and TBLH BMD. We performed subgroup analyses stratified by sex. A total of 1413 US children and adolescents were included in our study. In our linear regression models, we found inverse associations between fluoride concentrations in water and plasma and TBLH BMD. Additionally, we discovered a non-linear association between fluoride concentrations in water and plasma and TBLH BMD. No significant association or non-linear relationship was found between urine fluoride levels and TBLH BMD. This nationally representative sample study provides valuable insight into the intricate connection between fluoride exposure and skeletal health in children and adolescents. Full article
(This article belongs to the Section Pediatric Nutrition)
Show Figures

Figure 1

Figure 1
<p>Flow chart of the study participants from NHANES 2015–2016.</p>
Full article ">Figure 2
<p>Dose–response relationships between concentrations of water fluoride and TBHL BMD. (<b>a</b>) Association between concentrations of water fluoride and TBHL BMD in the whole population; the model was adjusted for age, gender, race, Body Mass Index, poverty income ratio, physical activity, and milk product consumption. (<b>b</b>,<b>c</b>) Association between concentrations of water fluoride and TBHL BMD in males and females, respectively; the model was adjusted for age, race, Body Mass Index, poverty income ratio, physical activity, and milk product consumption. The red solid lines with their upper and lower ranges represent the estimated regression coefficient Beta and the 95% confidence interval. The red dotted lines correspond to the fluoride concentration at the curve’s inflection point.</p>
Full article ">Figure 3
<p>Dose–response relationships between concentrations of plasma fluoride and TBHL BMD. (<b>a</b>) Association between concentrations of plasma fluoride and TBHL BMD in the whole population; the model was adjusted for age, gender, race, Body Mass Index, poverty income ratio, physical activity, and milk product consumption. (<b>b</b>,<b>c</b>) Association between concentrations of plasma fluoride and TBHL BMD in males and females, respectively; the model was adjusted for age, race, Body Mass Index, poverty income ratio, physical activity, and milk product consumption. The red solid lines with their upper and lower ranges represent the estimated regression coefficient Beta and the 95% confidence interval. The red dotted lines correspond to the fluoride concentration at the curve’s inflection point.</p>
Full article ">Figure 4
<p>Dose–response relationships between concentrations of urine fluoride and TBHL BMD. (<b>a</b>) Association between concentrations of urine fluoride and TBHL BMD in the whole population; the model was adjusted for age, gender, race, Body Mass Index, poverty income ratio, physical activity, and milk product consumption. (<b>b</b>,<b>c</b>) Association between concentrations of urine fluoride and TBHL BMD in males and females, respectively; the model was adjusted for age, race, Body Mass Index, poverty income ratio, physical activity, and milk product consumption. The red solid lines with their upper and lower ranges represent the estimated regression coefficient Beta and the 95% confidence interval.</p>
Full article ">Figure 5
<p>Correlation between concentrations of different fluoride. (<b>a</b>) Correlation between concentration of water fluoride and concentration of plasma fluoride. (<b>b</b>) Correlation between concentration of water fluoride and concentration of urine fluoride. (<b>c</b>) Correlation between concentration of plasma fluoride and concentration of urine fluoride. The red solid lines represent the fitted curve, while the blue shaded areas surrounding them denote the 95% confidence interval for the regression fit. Density plots along the periphery, either red or blue, represent the distribution of data points.</p>
Full article ">
45 pages, 3449 KiB  
Review
Non-Muscle Myosin II A: Friend or Foe in Cancer?
by Wasim Feroz, Briley SoYoung Park, Meghna Siripurapu, Nicole Ntim, Mary Kate Kilroy, Arwah Mohammad Ali Sheikh, Rosalin Mishra and Joan T. Garrett
Int. J. Mol. Sci. 2024, 25(17), 9435; https://doi.org/10.3390/ijms25179435 - 30 Aug 2024
Viewed by 640
Abstract
Non-muscle myosin IIA (NM IIA) is a motor protein that belongs to the myosin II family. The myosin heavy chain 9 (MYH9) gene encodes the heavy chain of NM IIA. NM IIA is a hexamer and contains three pairs of peptides, [...] Read more.
Non-muscle myosin IIA (NM IIA) is a motor protein that belongs to the myosin II family. The myosin heavy chain 9 (MYH9) gene encodes the heavy chain of NM IIA. NM IIA is a hexamer and contains three pairs of peptides, which include the dimer of heavy chains, essential light chains, and regulatory light chains. NM IIA is a part of the actomyosin complex that generates mechanical force and tension to carry out essential cellular functions, including adhesion, cytokinesis, migration, and the maintenance of cell shape and polarity. These functions are regulated via light and heavy chain phosphorylation at different amino acid residues. Apart from physiological functions, NM IIA is also linked to the development of cancer and genetic and neurological disorders. MYH9 gene mutations result in the development of several autosomal dominant disorders, such as May-Hegglin anomaly (MHA) and Epstein syndrome (EPS). Multiple studies have reported NM IIA as a tumor suppressor in melanoma and head and neck squamous cell carcinoma; however, studies also indicate that NM IIA is a critical player in promoting tumorigenesis, chemoradiotherapy resistance, and stemness. The ROCK-NM IIA pathway regulates cellular movement and shape via the control of cytoskeletal dynamics. In addition, the ROCK-NM IIA pathway is dysregulated in various solid tumors and leukemia. Currently, there are very few compounds targeting NM IIA, and most of these compounds are still being studied in preclinical models. This review provides comprehensive evidence highlighting the dual role of NM IIA in multiple cancer types and summarizes the signaling networks involved in tumorigenesis. Furthermore, we also discuss the role of NM IIA as a potential therapeutic target with a focus on the ROCK-NM IIA pathway. Full article
(This article belongs to the Section Molecular Oncology)
Show Figures

Figure 1

Figure 1
<p>The expression and alteration profile of <span class="html-italic">MYH9</span> at the pancancer level. (<b>A</b>) <span class="html-italic">MYH9</span> mRNA expression from 32 TCGA datasets. (<b>B</b>) The alteration profile of <span class="html-italic">MYH9</span> from the same 32 TCGA datasets. The expression and alteration frequency data were obtained from cBioPortal (<a href="http://www.cbioportal.org" target="_blank">www.cbioportal.org</a>).</p>
Full article ">Figure 2
<p>This figure illustrates the two assembly states of non-muscle myosin II (NMII): the 10S assembly-incompetent state and the 6S assembly-competent state. In the 10S assembly-incompetent state (<b>left</b>), the myosin molecule is folded, with the globular head and heavy chain regions interacting through various tail-binding sites, leading to a compact structure. Many intramolecular interactions keep the 10S state in an inactive stable form (<a href="#ijms-25-09435-t001" class="html-table">Table 1</a>). The interactions involve the Blocked Head (BH), which is the myosin head prevented from binding to actin, and the Free Head (FH), another myosin head that is inhibited but not directly involved in actin binding in the 10S state. The transition to the 6S assembly-competent state (<b>right</b>) occurs upon the phosphorylation of the regulatory light chain (RLC), resulting in an extended, active conformation where the heavy chain regions are aligned, allowing for actin binding and ATPase activity, which are essential for NMII’s role in cell contractility and motility. ELC: Essential Light Chain; RLC: Regulatory Light Chain; FH: Free Head; BH: Blocked Head; TF: Tail–Free Head Interaction; TB: Tail–Blocked Head Interaction; TT: Tail–Tail Interaction.</p>
Full article ">Figure 3
<p>The figure shows the specific kinases involved in the phosphorylation of serine and threonine residues of both RLCs and heavy chains. PKC, protein kinase C; MLCK, myosin light chain kinase; ROCK, Rho-associated protein kinase; TRPM7, transient receptor potential melastatin 7; PKCβ, protein kinase Cβ; CK II, casein kinase II. The figure was adapted from Pecci et al. [<a href="#B28-ijms-25-09435" class="html-bibr">28</a>].</p>
Full article ">Figure 4
<p>This figure illustrates the regulation of myosin II filament formation and activity. Specific serine residues on the myosin heavy chain (S1916, S1927, and S1943) are phosphorylated by various kinases, including MHCKA, MHCKB, MHCKC, TRPM6, TRPM7, PKC, and CK2. The phosphorylation of these sites is depicted as favoring the filamentous state of myosin, which is essential for mechanotransduction and ATP hydrolysis-driven interaction with actin filaments. Phosphatases are the enzymes responsible for dephosphorylation, which may reverse the phosphorylation effect, potentially leading to a shift back to the monomeric state of myosin. Mts1, also known as S100A4, is a calcium-binding protein that regulates myosin II function by modulating filament assembly. It binds to the myosin heavy chain, influencing the balance between monomeric and filamentous forms of myosin II. Mts1 typically inhibits filament formation, thereby controlling myosin’s contractile activity and its ability to interact with actin.</p>
Full article ">Figure 5
<p>The figure shows the structure and orientation of cell migration in the 2D environment. At the front, actin filaments within lamellipodia and filopodia are oriented with their rapidly polymerizing ends in the forward direction. In the main body, actin and myosin filaments form bipolar structures to aid in cell retraction. NM IIA and NM IIB show distinct localizations inside the cell, with NM IIA predominantly being found at the leading edge where actin dynamics are most active. NM IIB is predominant toward the rear end. The region between the leading and trailing edges contains varying concentrations of NM IIA and NM IIB. Additional molecules, such as RhoA, Rac1, Cdc42, Ca<sup>2+</sup> ions, and αPKC, also play significant roles in this cellular organization and migration process.</p>
Full article ">Figure 6
<p>The formation of plasma membrane blebs consists of three phases: initiation, expansion, and retraction. Blebbing stimuli, such as Ca<sup>2+</sup> influx and apoptosis, induce the initiation of membrane protrusion. Actomyosin contractility drives the expansion of blebs, which are devoid of the F-actin cortex. Rho-ROCK signaling then drives bleb retraction via actomyosin contractility. NM IIA contractile forces promote bleb retraction.</p>
Full article ">Figure 7
<p>A schematic representation of the <span class="html-italic">MYH9</span> exons with common mutations found in patients with <span class="html-italic">MYH9</span>-RD. The color coding of exon organization is as follows: black, motor domain; green, neck; orange, coiled coil domain; and brown, non-helical tail.</p>
Full article ">
19 pages, 335 KiB  
Review
A Narrative Review: Repurposing Metformin as a Potential Therapeutic Agent for Oral Cancer
by Jui-Hsiang Li, Pei-Yi Hsin, Yung-Chia Hsiao, Bo-Jun Chen, Zhi-Yun Zhuang, Chiang-Wen Lee, Wei-Ju Lee, Thi Thuy Tien Vo, Chien-Fu Tseng, Shih-Fen Tseng and I-Ta Lee
Cancers 2024, 16(17), 3017; https://doi.org/10.3390/cancers16173017 - 29 Aug 2024
Viewed by 607
Abstract
Oral cancer, particularly oral squamous cell carcinoma (OSCC), is a significant global health challenge because of its high incidence and limited treatment options. Major risk factors, including tobacco use, alcohol consumption, and specific microbiota, contribute to the disease’s prevalence. Recently, a compelling association [...] Read more.
Oral cancer, particularly oral squamous cell carcinoma (OSCC), is a significant global health challenge because of its high incidence and limited treatment options. Major risk factors, including tobacco use, alcohol consumption, and specific microbiota, contribute to the disease’s prevalence. Recently, a compelling association between diabetes mellitus (DM) and oral cancer has been identified, with metformin, a widely used antidiabetic drug, emerging as a potential therapeutic agent across various cancers, including OSCC. This review explores both preclinical and clinical studies to understand the mechanisms by which metformin may exert its anticancer effects, such as inhibiting cancer cell proliferation, inducing apoptosis, and enhancing the efficacy of existing treatments. Preclinical studies demonstrate that metformin modulates crucial metabolic pathways, reduces inflammation, and impacts cellular proliferation, thereby potentially lowering cancer risk and improving patient outcomes. Additionally, metformin’s ability to reverse epithelial-to-mesenchymal transition (EMT), regulate the LIN28/let-7 axis, and its therapeutic role in head and neck squamous cell carcinoma (HNSCC) are examined through experimental models. In clinical contexts, metformin shows promise in enhancing therapeutic outcomes and reducing recurrence rates, although challenges such as drug interactions, complex dosing regimens, and risks such as vitamin B12 deficiency remain. Future research should focus on optimizing metformin’s application, investigating its synergistic effects with other therapies, and conducting rigorous clinical trials to validate its efficacy in OSCC treatment. This dual exploration underscores metformin’s potential to play a transformative role in both diabetes management and cancer care, potentially revolutionizing oral cancer treatment strategies. Full article
(This article belongs to the Special Issue Oral Potentially Malignant Disorders and Oral Cavity Cancer)
14 pages, 35441 KiB  
Article
Smart Ship Draft Reading by Dual-Flow Deep Learning Architecture and Multispectral Information
by Bo Zhang, Jiangyun Li, Haicheng Tang and Xi Liu
Sensors 2024, 24(17), 5580; https://doi.org/10.3390/s24175580 - 28 Aug 2024
Viewed by 443
Abstract
In maritime transportation, a ship’s draft survey serves as a primary method for weighing bulk cargo. The accuracy of the ship’s draft reading determines the fairness of bulk cargo transactions. Human visual-based draft reading methods face issues such as safety concerns, high labor [...] Read more.
In maritime transportation, a ship’s draft survey serves as a primary method for weighing bulk cargo. The accuracy of the ship’s draft reading determines the fairness of bulk cargo transactions. Human visual-based draft reading methods face issues such as safety concerns, high labor costs, and subjective interpretation. Therefore, some image processing methods are utilized to achieve automatic draft reading. However, due to the limitations in the spectral characteristics of RGB images, existing image processing methods are susceptible to water surface environmental interference, such as reflections. To solve this issue, we obtained and annotated 524 multispectral images of a ship’s draft as the research dataset, marking the first application of integrating NIR information and RGB images for automatic draft reading tasks. Additionally, a dual-branch backbone named BIF is proposed to extract and combine spectral information from RGB and NIR images. The backbone network can be combined with the existing segmentation head and detection head to perform waterline segmentation and draft detection. By replacing the original ResNet-50 backbone of YOLOv8, we reached a mAP of 99.2% in the draft detection task. Similarly, combining UPerNet with our dual-branch backbone, the mIoU of the waterline segmentation task was improved from 98.9% to 99.3%. The inaccuracy of the draft reading is less than ±0.01 m, confirming the efficacy of our method for automatic draft reading tasks. Full article
(This article belongs to the Special Issue Sensor-Fusion-Based Deep Interpretable Networks)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>(<b>a</b>) Normal draft image without the issues of surface reflection or character erosion. The reflection of the part selected by the red box in (<b>b</b>) is very obvious, which may confuse the reflection during target detection and affect the accuracy of water segmentation. The characters in (<b>c</b>) have serious corrosion, which affects the recognition accuracy of target detection. The characters in the red box in (<b>d</b>) are submerged, but the water body is relatively clear, so they are still visible in the image and are recognized by the algorithm.</p>
Full article ">Figure 2
<p>MS400 series multi-spectral camera.</p>
Full article ">Figure 3
<p>Five band images (Red, Green, Blue, Red-Edge, NIR) and RGB image obtained by cameras.</p>
Full article ">Figure 4
<p>The overall process of dataset construction.</p>
Full article ">Figure 5
<p>Some examples from our dataset: (<b>a</b>): Original images of the ship. (<b>b</b>): NIR images (840 nm) of the ship. (<b>c</b>): Visualized object detection labels. (<b>d</b>): Visualized segmentation labels.</p>
Full article ">Figure 6
<p>The overall process of draft reading.</p>
Full article ">Figure 7
<p>Overview framework of the image-processing-based automated draft reading methods.</p>
Full article ">Figure 8
<p>Overview of our BIF backbone. The specific architecture of our novel Band Information Fusion backbone, where <math display="inline"><semantics> <msub> <mi>L</mi> <mi>n</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>H</mi> <mi>n</mi> </msub> </semantics></math> represent the number of layers of modules stacked.</p>
Full article ">Figure 9
<p>The architecture of Cross-Modality Process basic module.</p>
Full article ">Figure 10
<p>(<b>a</b>) The illustration of CFM-S in BIF. (<b>b</b>) The illustration of CFM-D in BIF. (<b>c</b>) The detailed structure of MCA module.</p>
Full article ">Figure 11
<p>Illustration of draft calculation. Firstly, the ratio of the vertical distance (<span class="html-italic">r</span>) between each character is computed as the correction factor for the perspective problem. Secondly, we determine the integer digit of the readings based on the classification result (<math display="inline"><semantics> <msub> <mi>v</mi> <mn>1</mn> </msub> </semantics></math>) of the character with “M”. In the following step, we calculate the decimal place of the readings via the vertical position of the waterline, the distance of the character closest to the waterline (<math display="inline"><semantics> <msub> <mi>v</mi> <mn>0</mn> </msub> </semantics></math>), and its numerical category. In the end, the final readings are further obtained by calculating readings with the correction factor.</p>
Full article ">Figure 12
<p>Box plot analysis of different methods in the draft reading task.</p>
Full article ">
20 pages, 27367 KiB  
Article
MCG-RTDETR: Multi-Convolution and Context-Guided Network with Cascaded Group Attention for Object Detection in Unmanned Aerial Vehicle Imagery
by Chushi Yu and Yoan Shin
Remote Sens. 2024, 16(17), 3169; https://doi.org/10.3390/rs16173169 - 27 Aug 2024
Viewed by 599
Abstract
In recent years, object detection in unmanned aerial vehicle (UAV) imagery has been a prominent and crucial task, with advancements in drone and remote sensing technologies. However, detecting targets in UAV images pose challenges such as complex background, severe occlusion, dense small targets, [...] Read more.
In recent years, object detection in unmanned aerial vehicle (UAV) imagery has been a prominent and crucial task, with advancements in drone and remote sensing technologies. However, detecting targets in UAV images pose challenges such as complex background, severe occlusion, dense small targets, and lighting conditions. Despite the notable progress of object detection algorithms based on deep learning, they still struggle with missed detections and false alarms. In this work, we introduce an MCG-RTDETR approach based on the real-time detection transformer (RT-DETR) with dual and deformable convolution modules, a cascaded group attention module, a context-guided feature fusion structure with context-guided downsampling, and a more flexible prediction head for precise object detection in UAV imagery. Experimental outcomes on the VisDrone2019 dataset illustrate that our approach achieves the highest AP of 29.7% and AP50 of 58.2%, surpassing several cutting-edge algorithms. Visual results further validate the model’s robustness and capability in complex environments. Full article
Show Figures

Figure 1

Figure 1
<p>The architecture of the proposed MCG-RTDETR.</p>
Full article ">Figure 2
<p>The structure of the dual convolutional filter. <span class="html-italic">M</span> is the input channel count, <span class="html-italic">N</span> denotes the number of output channels and convolution filters, and <span class="html-italic">G</span> is the group count within dual convolution.</p>
Full article ">Figure 3
<p>The structure of the <math display="inline"><semantics> <mrow> <mn>3</mn> <mo>×</mo> <mn>3</mn> </mrow> </semantics></math> deformable convolution.</p>
Full article ">Figure 4
<p>Diagram of the cascaded group attention module.</p>
Full article ">Figure 5
<p>Diagram of the context-guided downsampling block.</p>
Full article ">Figure 6
<p>The diagram of prediction heads.</p>
Full article ">Figure 7
<p>Illustrative samples of the VisDrone2019 dataset.</p>
Full article ">Figure 8
<p>Visible object detection results of the proposed MCG-RTDETR and some state-of-the-art detectors on complex detection scenes of VisDrone2019 dataset. (<b>a</b>) depicts scenes with occlusion and complex environmental factors, (<b>b</b>) depicts vertical shooting angle during daylight, (<b>c</b>) depicts cloudy with intricate background. (<b>d</b>) depicts very small targets, (<b>e</b>) depicts low-light and night scene, (<b>f</b>) depicts dynamic objects like vehicles at night.</p>
Full article ">Figure 8 Cont.
<p>Visible object detection results of the proposed MCG-RTDETR and some state-of-the-art detectors on complex detection scenes of VisDrone2019 dataset. (<b>a</b>) depicts scenes with occlusion and complex environmental factors, (<b>b</b>) depicts vertical shooting angle during daylight, (<b>c</b>) depicts cloudy with intricate background. (<b>d</b>) depicts very small targets, (<b>e</b>) depicts low-light and night scene, (<b>f</b>) depicts dynamic objects like vehicles at night.</p>
Full article ">Figure 8 Cont.
<p>Visible object detection results of the proposed MCG-RTDETR and some state-of-the-art detectors on complex detection scenes of VisDrone2019 dataset. (<b>a</b>) depicts scenes with occlusion and complex environmental factors, (<b>b</b>) depicts vertical shooting angle during daylight, (<b>c</b>) depicts cloudy with intricate background. (<b>d</b>) depicts very small targets, (<b>e</b>) depicts low-light and night scene, (<b>f</b>) depicts dynamic objects like vehicles at night.</p>
Full article ">Figure 8 Cont.
<p>Visible object detection results of the proposed MCG-RTDETR and some state-of-the-art detectors on complex detection scenes of VisDrone2019 dataset. (<b>a</b>) depicts scenes with occlusion and complex environmental factors, (<b>b</b>) depicts vertical shooting angle during daylight, (<b>c</b>) depicts cloudy with intricate background. (<b>d</b>) depicts very small targets, (<b>e</b>) depicts low-light and night scene, (<b>f</b>) depicts dynamic objects like vehicles at night.</p>
Full article ">
26 pages, 9607 KiB  
Article
A Global Spatial-Spectral Feature Fused Autoencoder for Nonlinear Hyperspectral Unmixing
by Mingle Zhang, Mingyu Yang, Hongyu Xie, Pinliang Yue, Wei Zhang, Qingbin Jiao, Liang Xu and Xin Tan
Remote Sens. 2024, 16(17), 3149; https://doi.org/10.3390/rs16173149 - 26 Aug 2024
Viewed by 434
Abstract
Hyperspectral unmixing (HU) aims to decompose mixed pixels into a set of endmembers and corresponding abundances. Deep learning-based HU methods are currently a hot research topic, but most existing unmixing methods still rely on per-pixel training or employ convolutional neural networks (CNNs), which [...] Read more.
Hyperspectral unmixing (HU) aims to decompose mixed pixels into a set of endmembers and corresponding abundances. Deep learning-based HU methods are currently a hot research topic, but most existing unmixing methods still rely on per-pixel training or employ convolutional neural networks (CNNs), which overlook the non-local correlations of materials and spectral characteristics. Furthermore, current research mainly focuses on linear mixing models, which limits the feature extraction capability of deep encoders and further improvement in unmixing accuracy. In this paper, we propose a nonlinear unmixing network capable of extracting global spatial-spectral features. The network is designed based on an autoencoder architecture, where a dual-stream CNNs is employed in the encoder to separately extract spectral and local spatial information. The extracted features are then fused together to form a more complete representation of the input data. Subsequently, a linear projection-based multi-head self-attention mechanism is applied to capture global contextual information, allowing for comprehensive spatial information extraction while maintaining lightweight computation. To achieve better reconstruction performance, a model-free nonlinear mixing approach is adopted to enhance the model’s universality, with the mixing model learned entirely from the data. Additionally, an initialization method based on endmember bundles is utilized to reduce interference from outliers and noise. Comparative results on real datasets against several state-of-the-art unmixing methods demonstrate the superior of the proposed approach. Full article
(This article belongs to the Topic Computer Vision and Image Processing, 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Schematic diagram of autoencoder architecture: (<b>a</b>) Autoencoder architecture. (<b>b</b>) Several common decoder architectures.</p>
Full article ">Figure 2
<p>The architecture of the proposed AE network for hyperspectral unmixing.</p>
Full article ">Figure 3
<p>The architecture of the Spatial-Spectral Feature Extraction Module.</p>
Full article ">Figure 4
<p>Module of Multi-Head Self-Attention Modules based on Linear Projection.</p>
Full article ">Figure 5
<p>Dataset: (<b>a</b>) Samson dataset. (<b>b</b>) Jasper Ridge dataset. (<b>c</b>) Urban dataset.</p>
Full article ">Figure 6
<p>The flowchart of the proposed endmember initialization method.</p>
Full article ">Figure 7
<p>The results of endmember extraction (Urban dataset): extracted endmembers (blue) and actual endmembers (orange).</p>
Full article ">Figure 8
<p>Visualization results of endmember bundle extraction (Urban dataset).</p>
Full article ">Figure 9
<p>Abundance maps of tree, water, dirt, and road on the Jasper Ridge dataset obtained by different modules.</p>
Full article ">Figure 10
<p>The results of mRMSE and mSAD under different projection dimensions, along with the corresponding computation times (measured in seconds). (<b>a</b>) Samson dataset. (<b>b</b>) Jasper Ridge dataset.</p>
Full article ">Figure 11
<p>Abundance maps of soil, tree, water on the Samson dataset obtained by different methods.</p>
Full article ">Figure 12
<p>Extracted endmember comparison between the different algorithms and the corresponding GTs in the Samson dataset.</p>
Full article ">Figure 13
<p>Abundance maps of tree, water, dirt, road on the Jasper Ridge dataset obtained by different methods.</p>
Full article ">Figure 14
<p>Extracted endmember comparison between the different algorithms and the corresponding GTs in the Jasper Ridge dataset.</p>
Full article ">Figure 15
<p>Abundance maps of asphalt, grass, tree, roof on the Urban dataset obtained by different methods.</p>
Full article ">Figure 16
<p>Extracted endmember comparison between the different algorithms and the corresponding GTs in the Urban dataset.</p>
Full article ">
13 pages, 2403 KiB  
Review
Management of Severe Bilateral Symptomatic Internal Carotid Artery Stenosis: Case Report and Literature Review
by Mircea Robu, Bogdan Radulescu, Irina-Maria Margarint, Anca Dragan, Ovidiu Stiru, Gabriel-Petre Gorecki, Cristian Voica, Vlad Anton Iliescu and Horatiu Moldovan
J. Pers. Med. 2024, 14(9), 893; https://doi.org/10.3390/jpm14090893 - 23 Aug 2024
Viewed by 453
Abstract
Multiple strategies for tandem severe carotid artery stenosis are reported: bilateral carotid artery endarterectomy (CEA), bilateral carotid artery stenting (CAS), and hybrid procedures (CEA and CAS). The management is controversial, considering the reported high risk of periprocedural stroke, hemodynamic distress, and cerebral hyperperfusion [...] Read more.
Multiple strategies for tandem severe carotid artery stenosis are reported: bilateral carotid artery endarterectomy (CEA), bilateral carotid artery stenting (CAS), and hybrid procedures (CEA and CAS). The management is controversial, considering the reported high risk of periprocedural stroke, hemodynamic distress, and cerebral hyperperfusion syndrome. We present the case of a 64-year-old patient with severe symptomatic bilateral internal carotid artery stenosis (95% stenosis on the left internal carotid artery with recent ipsilateral watershed anterior cerebral artery–medial cerebral artery (ACA-MCA) and medial cerebral artery–posterior cerebral artery (MCA-PCA) ischemic strokes and 90% stenosis on the right internal carotid artery with chronic ipsilateral frontal ischemic stroke) managed successfully with staged CEA within a 3-day interval. The patient had a history of coronary angioplasty and stenting. Strategies for brain protection included shunt placement after the evaluation of carotid stump pressure, internal carotid backflow, and near-infrared spectroscopy. A collagen and silver-coated polyester patch was used to complete the endarterectomy using a 6.0 polypropylene continuous suture in both instances. Management also included neurological consults after extubation, dual antiplatelet therapy, head CT between the two surgeries, myocardial ischemia monitoring, and general anesthesia. Staged CEA with a small time interval between surgeries can be an option to treat tandem symptomatic carotid artery stenosis in highly selected patients. The decision should be tailored according to the patient’s characteristics and should also be made by a cardiology specialist, a neurology specialist, and an anesthesia and intensive care physician. Full article
(This article belongs to the Special Issue Review Special Issue: Recent Advances in Personalized Medicine)
Show Figures

Figure 1

Figure 1
<p>Arterial Doppler ultrasound of the carotid system showing bilateral severe stenosis at the level of common carotid artery bifurcation extending at the level of the internal carotid artery. ICA: internal carotid artery; CCA: common carotid artery.</p>
Full article ">Figure 2
<p>Cervical CT angiography showing right severe internal carotid artery stenosis ((<b>A</b>), arrow), left severe internal carotid artery stenosis ((<b>C</b>), arrow), and bilateral severe internal carotid artery stenosis ((<b>B</b>), arrows).</p>
Full article ">Figure 3
<p>Carotid artery angiography showing suboclusive stenosis at the level of the left common and internal carotid artery with patent left vertebral artery. CCA: common carotid artery; ICA: internal carotid artery; ECA: external carotid artery; VA: vertebral artery.</p>
Full article ">Figure 4
<p>(<b>A</b>) A shunt inserted in the left common and internal carotid arteries. (<b>B</b>) Plaque extracted from the left carotid artery. (<b>C</b>) Plaque extracted from the right carotid artery. (<b>D</b>) Final result of left CEA with collagen and silver-coated polyester.</p>
Full article ">
17 pages, 1195 KiB  
Article
Separable CenterNet Detection Network Based on MobileNetV3—An Optimization Approach for Small-Object and Occlusion Issues
by Zhengkuo Jiao, Heng Dong and Naizhe Diao
Mathematics 2024, 12(16), 2524; https://doi.org/10.3390/math12162524 - 15 Aug 2024
Viewed by 421
Abstract
This paper proposes a novel object detection method to address the challenges posed by small objects and occlusion in object detection. This work is performed within the CenterNet framework, leveraging the MobileNetV3 backbone to model the input image’s abstract representation in a lightweight [...] Read more.
This paper proposes a novel object detection method to address the challenges posed by small objects and occlusion in object detection. This work is performed within the CenterNet framework, leveraging the MobileNetV3 backbone to model the input image’s abstract representation in a lightweight manner. A sparse convolutional skip connection is introduced in the bottleneck of MobileNetV3, specifically designed to adaptively suppress redundant and interfering information, thus enhancing feature extraction capabilities. A Dual-Path Bidirectional Feature Pyramid Network (DBi-FPN) is incorporated, allowing for high-level feature fusion through bidirectional flow and significantly improving the detection capabilities for small objects and occlusions. Task heads are applied within the feature space of multi-scale information merged by DBi-FPN, facilitating comprehensive consideration of multi-level representations. A bounding box-area loss function is also introduced, aimed at enhancing the model’s adaptability to object morphologies and geometric distortions. Extensive experiments on the PASCAL VOC 2007 and MS COCO 2017 datasets validate the competitiveness of our proposed method, particularly in real-time applications on resource-constrained devices. Our contributions offer promising avenues for enhancing the accuracy and robustness of object detection systems in complex scenarios. Full article
Show Figures

Figure 1

Figure 1
<p>The pipeline of the proposed method.</p>
Full article ">Figure 2
<p>Visualization comparison with competition methods on VOC 2007 dataset: First column: original image; second column: CenterNet; third column: SSD; fourth column: YOLOv5; fifth column: our method. (<b>a</b>) Comparison of sheep detection results; (<b>b</b>) comparison of horse detection results; (<b>c</b>) comparison of bird detection results; (<b>d</b>) comparison of sheep detection results.</p>
Full article ">
21 pages, 4424 KiB  
Article
CSA-SA-CRTNN: A Dual-Stream Adaptive Convolutional Cyclic Hybrid Network Combining Attention Mechanisms for EEG Emotion Recognition
by Ren Qian, Xin Xiong, Jianhua Zhou, Hongde Yu and Kaiwen Sha
Brain Sci. 2024, 14(8), 817; https://doi.org/10.3390/brainsci14080817 - 15 Aug 2024
Viewed by 520
Abstract
In recent years, EEG-based emotion recognition technology has made progress, but there are still problems of low model efficiency and loss of emotional information, and there is still room for improvement in recognition accuracy. To fully utilize EEG’s emotional information and improve recognition [...] Read more.
In recent years, EEG-based emotion recognition technology has made progress, but there are still problems of low model efficiency and loss of emotional information, and there is still room for improvement in recognition accuracy. To fully utilize EEG’s emotional information and improve recognition accuracy while reducing computational costs, this paper proposes a Convolutional-Recurrent Hybrid Network with a dual-stream adaptive approach and an attention mechanism (CSA-SA-CRTNN). Firstly, the model utilizes a CSAM module to assign corresponding weights to EEG channels. Then, an adaptive dual-stream convolutional-recurrent network (SA-CRNN and MHSA-CRNN) is applied to extract local spatial-temporal features. After that, the extracted local features are concatenated and fed into a temporal convolutional network with a multi-head self-attention mechanism (MHSA-TCN) to capture global information. Finally, the extracted EEG information is used for emotion classification. We conducted binary and ternary classification experiments on the DEAP dataset, achieving 99.26% and 99.15% accuracy for arousal and valence in binary classification and 97.69% and 98.05% in ternary classification, and on the SEED dataset, we achieved an accuracy of 98.63%, surpassing relevant algorithms. Additionally, the model’s efficiency is significantly higher than other models, achieving better accuracy with lower resource consumption. Full article
(This article belongs to the Section Neurotechnology and Neuroimaging)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Emotion model: (<b>a</b>) discrete model, (<b>b</b>) two-dimensional valence–arousal model.</p>
Full article ">Figure 2
<p>Frame diagram of the CSA-SA-CRTNN model. The model consists of four modules, namely the CSAM module, the SA-CRNN module, the MHSA-CRNN module, and the MHSA-TCN module.</p>
Full article ">Figure 3
<p>CSAM Structure Diagram.</p>
Full article ">Figure 4
<p>MHSA Structure Diagram.</p>
Full article ">Figure 5
<p>Structure diagram of MHSA-TCN.</p>
Full article ">Figure 6
<p>Accuracy–epoch relationship diagram. (<b>a</b>) DEAP dataset; (<b>b</b>) SEED dataset.</p>
Full article ">Figure 7
<p>The average accuracy of arousal and valence of DEAP using CSA-SA-CRTNN for each subject. (<b>a</b>) 2-class (<b>b</b>) 3-class.</p>
Full article ">Figure 8
<p>Confusion matrix: (<b>a</b>) 2-class arousal; (<b>b</b>) 2-class valence; (<b>c</b>) 3-class arousal; (<b>d</b>) 3-class valence.</p>
Full article ">Figure 9
<p>Experimental results on the SEED dataset: (<b>a</b>) Average accuracy for each subject; (<b>b</b>) confusion matrix.</p>
Full article ">Figure 10
<p>Comparison of attention mechanisms in different channels.</p>
Full article ">
Back to TopTop