Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (45)

Search Parameters:
Keywords = personalized weight map

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
26 pages, 8647 KiB  
Article
An Explainable CNN and Vision Transformer-Based Approach for Real-Time Food Recognition
by Kintoh Allen Nfor, Tagne Poupi Theodore Armand, Kenesbaeva Periyzat Ismaylovna, Moon-Il Joo and Hee-Cheol Kim
Nutrients 2025, 17(2), 362; https://doi.org/10.3390/nu17020362 - 20 Jan 2025
Cited by 1 | Viewed by 835
Abstract
Background: Food image recognition, a crucial step in computational gastronomy, has diverse applications across nutritional platforms. Convolutional neural networks (CNNs) are widely used for this task due to their ability to capture hierarchical features. However, they struggle with long-range dependencies and global feature [...] Read more.
Background: Food image recognition, a crucial step in computational gastronomy, has diverse applications across nutritional platforms. Convolutional neural networks (CNNs) are widely used for this task due to their ability to capture hierarchical features. However, they struggle with long-range dependencies and global feature extraction, which are vital in distinguishing visually similar foods or images where the context of the whole dish is crucial, thus necessitating transformer architecture. Objectives: This research explores the capabilities of the CNNs and transformers to build a robust classification model that can handle both short- and long-range dependencies with global features to accurately classify food images and enhance food image recognition for better nutritional analysis. Methods: Our approach, which combines CNNs and Vision Transformers (ViTs), begins with the RestNet50 backbone model. This model is responsible for local feature extraction from the input image. The resulting feature map is then passed to the ViT encoder block, which handles further global feature extraction and classification using multi-head attention and fully connected layers with pre-trained weights. Results: Our experiments on five diverse datasets have confirmed a superior performance compared to the current state-of-the-art methods, and our combined dataset leveraging complementary features showed enhanced generalizability and robust performance in addressing global food diversity. We used explainable techniques like grad-CAM and LIME to understand how the models made their decisions, thereby enhancing the user’s trust in the proposed system. This model has been integrated into a mobile application for food recognition and nutrition analysis, offering features like an intelligent diet-tracking system. Conclusion: This research paves the way for practical applications in personalized nutrition and healthcare, showcasing the extensive potential of AI in nutritional sciences across various dietary platforms. Full article
(This article belongs to the Special Issue Digital Transformations in Nutrition)
Show Figures

Figure 1

Figure 1
<p>Global representation of food datasets.</p>
Full article ">Figure 2
<p>Samples from all the various datasets used.</p>
Full article ">Figure 3
<p>Vision Transformer (ViT) Architecture.</p>
Full article ">Figure 4
<p>(<b>a</b>) Scaled dot-product attention and (<b>b</b>) multi-head attention, which consist of serval attention layers running in parallel.</p>
Full article ">Figure 5
<p>R50 + ViT-B_16 hybrid model.</p>
Full article ">Figure 6
<p>An architectural diagram of the pre-trained models used as a schematic representation of transfer learning.</p>
Full article ">Figure 7
<p>Accuracy Comparison across Models on VireoFood172.</p>
Full article ">Figure 8
<p>Top 5 accuracy Comparison across models on VireoFood172.</p>
Full article ">Figure 9
<p>F-1 Score Comparison across Models on VireoFood172.</p>
Full article ">Figure 10
<p>Precision Comparison across models on VireoFood172.</p>
Full article ">Figure 11
<p>Recall Comparison across models on VireoFood172.</p>
Full article ">Figure 12
<p>Precision, Recall, and F1 Score on Combined Dataset.</p>
Full article ">Figure 13
<p>Visualization of Grad-CAM for explainability.</p>
Full article ">Figure 14
<p>Visualization of LIME for explainability.</p>
Full article ">Figure 15
<p>Mobile Application Integration Architecture.</p>
Full article ">Figure 16
<p>Integration into a mobile dietary monitoring, tracking, and nutritional assessment application.</p>
Full article ">
17 pages, 13691 KiB  
Article
MambaPose: A Human Pose Estimation Based on Gated Feedforward Network and Mamba
by Jianqiang Zhang, Jing Hou, Qiusheng He, Zhengwei Yuan and Hao Xue
Sensors 2024, 24(24), 8158; https://doi.org/10.3390/s24248158 - 20 Dec 2024
Viewed by 4622
Abstract
Human pose estimation is an important research direction in the field of computer vision, which aims to accurately identify the position and posture of keypoints of the human body through images or videos. However, multi-person pose estimation yields false detection or missed detection [...] Read more.
Human pose estimation is an important research direction in the field of computer vision, which aims to accurately identify the position and posture of keypoints of the human body through images or videos. However, multi-person pose estimation yields false detection or missed detection in dense crowds, and it is still difficult to detect small targets. In this paper, we propose a Mamba-based human pose estimation. First, we design a GMamba structure to be used as a backbone network to extract human keypoints. A gating mechanism is introduced into the linear layer of Mamba, which allows the model to dynamically adjust the weights according to the different input images to locate the human keypoints more precisely. Secondly, GMamba as the backbone network can effectively solve the long-sequence problem. The direct use of convolutional downsampling reduces selectivity for different stages of information flow. We used slice downsampling (SD) to reduce the resolution of the feature map to half the original size, and then fused local features from four different locations. The fusion of multi-channel information helped the model obtain rich pose information. Finally, we introduced an adaptive threshold focus loss (ATFL) to dynamically adjust the weights of different keypoints. We assigned higher weights to error-prone keypoints to strengthen the model’s attention to these points. Thus, we effectively improved the accuracy of keypoint identification in cases of occlusion, complex background, etc., and significantly improved the overall performance of attitude estimation and anti-interference ability. Experimental results showed that the AP and AP50 of the proposed algorithm on the COCO 2017 validation set were 72.2 and 92.6. Compared with the typical algorithm, it was improved by 1.1% on AP50. The proposed method can effectively detect the keypoints of the human body, and provides stronger robustness and accuracy for the estimation of human posture in complex scenes. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>Detection results of MambaPose. (<b>a</b>) Input image; (<b>b</b>) test result graph; (<b>c</b>) posture image of the human body.</p>
Full article ">Figure 2
<p>The overall structure of MambaPose. Our human pose estimation model consists of three parts: the backbone network, the neck, and the keypoint detection head. We can draw a posture map of the human body based on the keypoints of detection.</p>
Full article ">Figure 3
<p>The overall structure of the neck. The neck uses the PA-FPN structure, and we use GMamba instead of C2F, which can effectively improve the information of keypoints in the human body during fusion. Top-down and bottom-up feature fusion methods are adopted, in which the convolutional layer is downsampled and the UP is upsampled.</p>
Full article ">Figure 4
<p>Overall structure of GMamba.</p>
Full article ">Figure 5
<p>Schematic diagram of the workflow of SS2D. The scan expand operation expands the input image into a series of sub-images. Expand in four symmetrical directions: top-down, bottom-up, left, and right. Then, in the S6 block operation, the subgraph performs in-depth feature extraction. Finally, the extracted sub-images are merged into an output image of the same size as the input image by scanning merging. This process integrates information from all directions, and also ensures the spatial structure of the output feature map, which further enhances the network’s ability to understand and recognize human posture.</p>
Full article ">Figure 6
<p>Overall structure of LFE and GFN.</p>
Full article ">Figure 7
<p>Visualization on the COCO 2017 validation dataset. MambaPose can effectively detect occluded human bodies and small human keypoints. The blue boxes indicate human bodies at different locations and the red dots indicate each joint point, using different-colored lines to connect adjacent joint points.</p>
Full article ">Figure 8
<p>Comparison of the detection capabilities of small targets. Upper, results using YOLO-Pose; lower, results using the method in this paper.</p>
Full article ">Figure 9
<p>Visual rendering of missed detections in dense crowds. Red circles to mark the missing human bodies and keypoints.</p>
Full article ">Figure 10
<p>False detections in images. Upper, detection results using YOLO-Pose; lower, results using the method in this article. YOLO-Pose detects animals and other objects as pedestrians. Red circles indicate false detections.</p>
Full article ">
22 pages, 4829 KiB  
Article
Host Genetics Background Affects Intestinal Cancer Development Associated with High-Fat Diet-Induced Obesity and Type 2 Diabetes
by Aya Ghnaim, Kareem Midlej, Osayd Zohud, Sama Karram, Arne Schaefer, Yael Houri-Haddad, Iqbal M. Lone and Fuad A. Iraqi
Cells 2024, 13(21), 1805; https://doi.org/10.3390/cells13211805 - 31 Oct 2024
Viewed by 928
Abstract
Background: Obesity and type 2 diabetes (T2D) promote inflammation, increasing the risk of colorectal cancer (CRC). High-fat diet (HFD)-induced obesity is key to these diseases through biological mechanisms. This study examined the impact of genetic background on the multimorbidity of intestinal cancer, T2D, [...] Read more.
Background: Obesity and type 2 diabetes (T2D) promote inflammation, increasing the risk of colorectal cancer (CRC). High-fat diet (HFD)-induced obesity is key to these diseases through biological mechanisms. This study examined the impact of genetic background on the multimorbidity of intestinal cancer, T2D, and inflammation due to HFD-induced obesity. Methods: A cohort of 357 Collaborative Cross (CC) mice from 15 lines was fed either a control chow diet (CHD) or HFD for 12 weeks. Body weight was tracked biweekly, and blood glucose was assessed at weeks 6 and 12 via intraperitoneal glucose tolerance tests (IPGTT). At the study’s endpoint, intestinal polyps were counted, and cytokine profiles were analyzed to evaluate the inflammatory response. Results: HFD significantly increased blood glucose levels and body weight, with males showing higher susceptibility to T2D and obesity. Genetic variation across CC lines influenced glucose metabolism, body weight, and polyp development. Mice on HFD developed more intestinal polyps, with males showing higher counts than females. Cytokine analysis revealed diet-induced variations in pro-inflammatory markers like IL-6, IL-17A, and TNF-α, differing by genetic background and sex. Conclusions: Host genetics plays a crucial role in susceptibility to HFD-induced obesity, T2D, CRC, and inflammation. Genetic differences across CC lines contributed to variability in disease outcomes, providing insight into the genetic underpinnings of multimorbidity. This study supports gene-mapping efforts to develop personalized prevention and treatment strategies for these diseases. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>A</b>) Cumulative area under the curve (AUC; min × mg/dL) for the complete population of 15 distinct CC lines following a 6-week dietary challenge comparing HFD to CHD. The X-axis represents the various CC lines, while the Y-axis shows the AUC values. Data were assessed using an ANOVA with a significance level of <span class="html-italic">p</span> &lt; 0.05. (<b>B</b>) The cumulative area under the curve (AUC; minmg/dL) for the total population of 15 different CC lines after a 6-week dietary challenge comparing HFD and CHD. The X-axis displays the different CC lines, and the Y-axis indicates the total AUC values (min × mg/dL). Data were analyzed through ANOVA with <span class="html-italic">p</span> &lt; 0.05. * <span class="html-italic">p</span> &lt; 0.05, ** <span class="html-italic">p</span> &lt; 0.01, *** <span class="html-italic">p</span> &lt; 0.005.</p>
Full article ">Figure 2
<p>(<b>A</b>) shows the overall area under the curve (AUC; min × mg/dL) for 15 distinct CC lines after 6 weeks of dietary challenge comparing HFD to CHD in both male and female populations. The X-axis displays the various CC lines, while the Y-axis indicates the AUC values. Data were evaluated using ANOVA with <span class="html-italic">p</span> &lt; 0.05. (<b>B</b>) shows the overall area under the curve (AUC; min × mg/dL) for 15 different CC lines after four months of dietary challenge comparing HFD to CHD in both male and female populations. The X-axis represents the various CC lines, and the Y-axis shows the AUC values. Data were analyzed using ANOVA with <span class="html-italic">p</span> &lt; 0.05. * <span class="html-italic">p</span> &lt; 0.05, ** <span class="html-italic">p</span> &lt; 0.01, *** <span class="html-italic">p</span> &lt; 0.005.</p>
Full article ">Figure 3
<p>(<b>A</b>) Percentage change in body weight (%∆BW) after 6 weeks of dietary challenge (CHD vs. HFD) for a total of 15 different CC lines (X-axis). The Y-axis represents %∆BW, calculated using the formula (BW6 (g) − BW0 (g))/BW0 × 100. (<b>B</b>) Percentage change in body weight (%∆BW) after 12 weeks of dietary challenge (CHD vs. HFD) across the same 15 CC lines (X-axis). The Y-axis shows %∆BW, calculated as (BW12 (g) − BW0 (g))/BW0 × 100. * <span class="html-italic">p</span> &lt; 0.05, ** <span class="html-italic">p</span> &lt; 0.01, *** <span class="html-italic">p</span> &lt; 0.005.</p>
Full article ">Figure 4
<p>(<b>A</b>) Percent body weight change (%∆BW) after 6 weeks of dietary challenge (CHD vs. HFD) for both populations of 15 distinct CC lines (X-axis). The Y-axis shows %∆BW, calculated as (BW6 (g) − BW0 (g))/BW0 × 100. (<b>B</b>) Percent body weight change (%∆BW) after 12 weeks of dietary challenge (CHD vs. HFD) for both male and female populations of the 15 CC lines (X-axis). The Y-axis represents %∆BW, calculated as (BW12 (g) − BW0 (g))/BW0 × 100. * <span class="html-italic">p</span> &lt; 0.05, ** <span class="html-italic">p</span> &lt; 0.01, *** <span class="html-italic">p</span> &lt; 0.005.</p>
Full article ">Figure 5
<p>(<b>A</b>) The whole intestinal polyps count in the entire population of 15 different CC lines after being maintained for 12 weeks on HFD vs. CHD. The X-axis represents the different CC lines, and the Y-axis represents the polyp number development. Total intestinal polyps = Total small intestine polyp+ colon polyps’ number. Data was analyzed using a one-way analysis of variation (ANOVA). (<b>B</b>) Intestinal polyps count of both populations (female and male) of 15 different CC lines after being maintained for 4 months on HFD vs. CHD. The X-axis represents the different CC lines, and the Y-axis represents the polyp number development. Total intestinal polyps = Total small intestine polyp+ colon polyps’ number. Data was analyzed using ANOVA. * <span class="html-italic">p</span> &lt; 0.05, ** <span class="html-italic">p</span> &lt; 0.01, *** <span class="html-italic">p</span> &lt; 0.005.</p>
Full article ">Figure 6
<p>(<b>A</b>) The small intestine Polyps number among the entire population of 15 different CC lines was maintained for 12 weeks on HFD vs. CHD. The X-axis represents the different CC lines, and the Y-axis represents the polyp number development. Data was analyzed using a one-way analysis of variation (ANOVA). (<b>B</b>) Small intestine polyp number in both populations (female and male) of 15 different CC lines maintained for 12 weeks under dietary challenge (CHD vs. HFD). The X-axis represents the different CC lines, and the Y-axis represents the number of polyps in the small intestine. Data was analyzed using a one-way analysis of variation (ANOVA). * <span class="html-italic">p</span> &lt; 0.05, ** <span class="html-italic">p</span> &lt; 0.01, *** <span class="html-italic">p</span> &lt; 0.005.</p>
Full article ">Figure 7
<p>Total colon polyp number in 15 different CC lines for both population females and males, combined (<b>A</b>) and separated (<b>B</b>) maintained for four months under dietary challenge HFD (42%) vs. CHD (18%). The X-axis presents the different CC lines and the mean of the entire population; the Y-axis presents the number of polyps in the colon for both females and males. Data was analyzed using a one-way analysis of variation (ANOVA). * <span class="html-italic">p</span> &lt; 0.05, ** <span class="html-italic">p</span> &lt; 0.01, *** <span class="html-italic">p</span> &lt; 0.005.</p>
Full article ">Figure 8
<p>Cytokines profiles for ten different pro-inflammatory cytokines in male and female mice of 5 different CC lines at week six during the dietary challenge of either HFD (42%) or CHD (18%). The X-axis presents the different pro-inflammatory cytokines, and the Y-axis presents cytokines levels in blood serum for both female and male mice. Data was analyzed by ANOVA. (<b>A</b>–<b>H</b>) show the results of the following cytokines, IL16, IL4, IFNy, TNFa, RANKL, IL17ACTLA8, respectively.</p>
Full article ">Figure 9
<p>Correlation matrix between different inflammatory cytokines, % ∆BWT6, total AUC 6, for both populations (female and male). A correlation matrix of three different CC lines (IL 557, IL711, IL72) of both populations after 6 weeks of diet challenge CHD vs. HFD intake.</p>
Full article ">
32 pages, 4351 KiB  
Article
Enhancing Brain Tumor Detection Through Custom Convolutional Neural Networks and Interpretability-Driven Analysis
by Kavinda Ashan Kulasinghe Wasalamuni Dewage, Raza Hasan, Bacha Rehman and Salman Mahmood
Information 2024, 15(10), 653; https://doi.org/10.3390/info15100653 - 18 Oct 2024
Cited by 3 | Viewed by 1995
Abstract
Brain tumor detection is crucial for effective treatment planning and improved patient outcomes. However, existing methods often face challenges, such as limited interpretability and class imbalance in medical-imaging data. This study presents a novel, custom Convolutional Neural Network (CNN) architecture, specifically designed to [...] Read more.
Brain tumor detection is crucial for effective treatment planning and improved patient outcomes. However, existing methods often face challenges, such as limited interpretability and class imbalance in medical-imaging data. This study presents a novel, custom Convolutional Neural Network (CNN) architecture, specifically designed to address these issues by incorporating interpretability techniques and strategies to mitigate class imbalance. We trained and evaluated four CNN models (proposed CNN, ResNetV2, DenseNet201, and VGG16) using a brain tumor MRI dataset, with oversampling techniques and class weighting employed during training. Our proposed CNN achieved an accuracy of 94.51%, outperforming other models in regard to precision, recall, and F1-Score. Furthermore, interpretability was enhanced through gradient-based attribution methods and saliency maps, providing valuable insights into the model’s decision-making process and fostering collaboration between AI systems and clinicians. This approach contributes a highly accurate and interpretable framework for brain tumor detection, with the potential to significantly enhance diagnostic accuracy and personalized treatment planning in neuro-oncology. Full article
(This article belongs to the Special Issue Artificial Intelligence and Data Science for Health)
Show Figures

Figure 1

Figure 1
<p>Methodology flowchart.</p>
Full article ">Figure 2
<p>Model performance metrics over epochs (before sampling).</p>
Full article ">Figure 3
<p>Model performance metrics over epochs (after sampling).</p>
Full article ">Figure 4
<p>ROC curve for brain tumor classification.</p>
Full article ">Figure 5
<p>Confusion matrix for brain tumor classification.</p>
Full article ">Figure 6
<p>The saliency map for a glioblastoma case, highlighting the relevant features.</p>
Full article ">Figure 7
<p>The attribution map for a meningioma case, highlighting the relevant features.</p>
Full article ">Figure 8
<p>Selecting MRI image.</p>
Full article ">Figure 9
<p>Results displayed in GUI after a prediction.</p>
Full article ">Figure 10
<p>Results displayed in GUI after the correct prediction.</p>
Full article ">
16 pages, 5621 KiB  
Article
Kinect-Based Gait Analysis System Design and Concurrent Validity in Persons with Anterolateral Shoulder Pain Syndrome, Results from a Pilot Study
by Fredy Bernal, Veronique Feipel and Mauricio Plaza
Sensors 2024, 24(19), 6351; https://doi.org/10.3390/s24196351 - 30 Sep 2024
Viewed by 954
Abstract
As part of an investigation to detect asymmetries in gait patterns in persons with shoulder injuries, the goal of the present study was to design and validate a Kinect-based motion capture system that would enable the extraction of joint kinematics curves during gait [...] Read more.
As part of an investigation to detect asymmetries in gait patterns in persons with shoulder injuries, the goal of the present study was to design and validate a Kinect-based motion capture system that would enable the extraction of joint kinematics curves during gait and to compare them with the data obtained through a commercial motion capture system. The study included eight male and two female participants, all diagnosed with anterolateral shoulder pain syndrome in their right upper extremity with a minimum 18 months of disorder evolution. The participants had an average age of 31.8 ± 9.8 years, a height of 173 ± 18 cm, and a weight of 81 ± 15 kg. The gait kinematics were sampled simultaneously with the new system and the Clinical 3DMA system. Shoulder, elbow, hip, and knee kinematics were compared between systems for the pathological and non-pathological sides using repeated measures ANOVA and 1D statistical parametric mapping. For most variables, no significant difference was found between systems. Evidence of a significant difference between the newly developed system and the commercial system was found for knee flexion–extension (p < 0.004, between 60 and 80% of the gait cycle), and for shoulder abduction–adduction. The good concurrent validity of the new Kinect-based motion analysis system found in this study opens promising perspectives for clinical motion tracking using an affordable and simple system. Full article
Show Figures

Figure 1

Figure 1
<p>Location of reflective markers for Clinical 3DMA system [<a href="#B11-sensors-24-06351" class="html-bibr">11</a>]: 1. Front of the head, 2. Top of the head, 3. Back of the head, 4. Left acromion, 5. Right acromion, 6. Left lateral humeral epicondyle, 7. Right lateral humeral epicondyle, 8. Right trochanter, 9. Left trochanter, 10. Right ulnar styloid process, 11. Left ulnar styloid process, 12. Right lateral femoral epicondyle, 13. Left lateral femoral epicondyle, 14. Right lateral malleolus, 15. Left lateral malleolus, 16. Right 2nd metatarsophalangeal joint, 17. Left 2nd metatarsophalangeal joint, 18. C7 vertebra, 19. Upper part of the sacrum.</p>
Full article ">Figure 2
<p>Kinect-based system—motion capture system.</p>
Full article ">Figure 3
<p>Basic flow diagram of Kinect-based system.</p>
Full article ">Figure 4
<p>Kinect based system—Data Extraction Software 1.0.</p>
Full article ">Figure 5
<p>Kinect based system-Time Extraction Software.</p>
Full article ">Figure 6
<p>Comparison between systems for the non-pathological sides. On the right side of each figure, the blue curve represents the movement captured by Clinical 3DMA, and the red curve data extracted from Kincapsys. Comparison using SPM 1D is shown on the left side of each figure, the shaded area indicating regions of the curves with the significant difference. (<b>a</b>) shoulder flexion-extension, (<b>b</b>) arm elevation, (<b>c</b>) shoulder abd/add, (<b>d</b>) elbow flexion-extension, (<b>e</b>) hip flexion-extension, (<b>f</b>) thigh elevation, (<b>g</b>) hip abd/add, (<b>h</b>) knee flexion–extension.</p>
Full article ">Figure 7
<p>Comparison between systems for the pathological sides. On the right side of each figure, the blue curve represents the movement captured by Clinical 3DMA, and the red curve data extracted from Kincapsys. Comparison using SPM 1D is shown on the left side of each figure, the shaded area indicating regions of the curves with the significant difference. (<b>a</b>) shoulder flexion-extension, (<b>b</b>) arm elevation, (<b>c</b>) shoulder abd/add, (<b>d</b>) elbow flexion-extension, (<b>e</b>) hip flexion-extension, (<b>f</b>) thigh elevation, (<b>g</b>) hip ABD/ADD, (<b>h</b>) knee flexion-extension.</p>
Full article ">
17 pages, 9437 KiB  
Article
Utilizing RT-DETR Model for Fruit Calorie Estimation from Digital Images
by Shaomei Tang and Weiqi Yan
Information 2024, 15(8), 469; https://doi.org/10.3390/info15080469 - 7 Aug 2024
Cited by 3 | Viewed by 2337
Abstract
Estimating the calorie content of fruits is critical for weight management and maintaining overall health as well as aiding individuals in making informed dietary choices. Accurate knowledge of fruit calorie content assists in crafting personalized nutrition plans and preventing obesity and associated health [...] Read more.
Estimating the calorie content of fruits is critical for weight management and maintaining overall health as well as aiding individuals in making informed dietary choices. Accurate knowledge of fruit calorie content assists in crafting personalized nutrition plans and preventing obesity and associated health issues. In this paper, we investigate the application of deep learning models for estimating the calorie content in fruits from digital images, aiming to provide a more efficient and accurate method for nutritional analysis. We create a dataset comprising images of various fruits and employ random data augmentation techniques during training to enhance model robustness. We utilize the RT-DETR model integrated into the ultralytics framework for implementation and conduct comparative experiments with YOLOv10 on the dataset. Our results show that the RT-DETR model achieved a precision rate of 99.01% and mAP50-95 of 94.45% in fruit detection from digital images, outperforming YOLOv10 in terms of F1- Confidence Curves, P-R curves, precision, and mAP. Conclusively, in this paper, we utilize a transformer architecture to detect fruits and estimate their calorie and nutritional content. The results of the experiments provide a technical reference for more accurately monitoring an individual’s dietary intake by estimating the calorie content of fruits. Full article
(This article belongs to the Special Issue Deep Learning for Image, Video and Signal Processing)
Show Figures

Figure 1

Figure 1
<p>The transformer architecture.</p>
Full article ">Figure 2
<p>The images used in data augmentations.</p>
Full article ">Figure 3
<p>Applying HSV augmentation randomly.</p>
Full article ">Figure 4
<p>The effects of four augmentation techniques from the Albumentations Library.</p>
Full article ">Figure 5
<p>The architecture of RT-DETR.</p>
Full article ">Figure 6
<p>F1–confidence curves for YOLOv10 (<b>a</b>) and RT-DETR (<b>b</b>). The graylines are our F1-confidence curves with different parameters and datasets.</p>
Full article ">Figure 7
<p>The P-R curves for YOLOv10 (<b>a</b>) and RT-DETR (<b>b</b>). The graylines indicate the different parameters and datasets for testing our proposed models.</p>
Full article ">Figure 8
<p>(<b>a</b>–<b>e</b>) Prediction of RT-DETR model (<b>left</b>) and YOLOv10 model (<b>right</b>) in various backgrounds.</p>
Full article ">Figure 8 Cont.
<p>(<b>a</b>–<b>e</b>) Prediction of RT-DETR model (<b>left</b>) and YOLOv10 model (<b>right</b>) in various backgrounds.</p>
Full article ">Figure 8 Cont.
<p>(<b>a</b>–<b>e</b>) Prediction of RT-DETR model (<b>left</b>) and YOLOv10 model (<b>right</b>) in various backgrounds.</p>
Full article ">Figure 9
<p>(<b>a</b>,<b>b</b>) Calorie estimation error for the RT-DETR model and the YOLOv10 model in different environments.</p>
Full article ">Figure 10
<p>(<b>a</b>,<b>b</b>) Both RT-DETR and YOLOv10 models incorrectly detected an Ambrosia apple or a Rose apple as a Gala apple. (<b>a</b>) The case in misidentification for an Ambrosia apple (<b>b</b>) The case in misidentification for an Rose apple.</p>
Full article ">
17 pages, 8142 KiB  
Article
DeepSarc-US: A Deep Learning Framework for Assessing Sarcopenia Using Ultrasound Images
by Bahareh Behboodi, Jeremy Obrand, Jonathan Afilalo and Hassan Rivaz
Appl. Sci. 2024, 14(15), 6726; https://doi.org/10.3390/app14156726 - 1 Aug 2024
Cited by 1 | Viewed by 1223
Abstract
Sarcopenia, the age-related loss of skeletal muscle mass, is a core component of frailty that is associated with functional decline and adverse health events in older adults. Unfortunately, the available tools to diagnose sarcopenia are often inaccessible or not user-friendly for clinicians. Point-of-care [...] Read more.
Sarcopenia, the age-related loss of skeletal muscle mass, is a core component of frailty that is associated with functional decline and adverse health events in older adults. Unfortunately, the available tools to diagnose sarcopenia are often inaccessible or not user-friendly for clinicians. Point-of-care ultrasound (US) is a promising tool that has been used to image the quadriceps muscle and measure its thickness (QMT) as a diagnostic criterion for sarcopenia. This measurement can be challenging for clinicians, especially when performed at the bedside using handheld systems or phased-array probes not designed for this use case. In this paper, we sought to automate this measurement using deep learning methods to improve its accuracy, reliability, and speed in the hands of untrained clinicians. In the proposed framework, which aids in better training, particularly when limited data are available, convolutional and transformer-based deep learning models with generic or data-driven pre-trained weights were compared. We evaluated regression (QMT as a continuous output in cm) and classification (QMT as an ordinal output in 0.5 cm bins) approaches, and in the latter, activation maps were generated to interpret the anatomical landmarks driving the model predictions. Finally, we evaluated a segmentation approach to derive QMT. The results showed that both transformer-based models and convolutional neural networks benefit from the proposed framework in estimating QMT. Additionally, the activation maps highlighted the interface between the femur bone and the quadriceps muscle as a key anatomical landmark for accurate predictions. The proposed framework is a pivotal step to enable the application of US-based measurement of QMT in large-scale clinical studies seeking to validate its diagnostic performance for sarcopenia, alone or with ancillary criteria assessing muscle quality or strength. We believe that implementing the proposed framework will empower clinicians to conveniently diagnose sarcopenia in clinical settings and accordingly personalize the care of older patients, leading to improved patient outcomes and a more efficient allocation of healthcare resources. Full article
(This article belongs to the Special Issue Current Updates on Ultrasound for Biomedical Applications)
Show Figures

Figure 1

Figure 1
<p>Examples of the dataset with QMT values of (<b>a</b>) 1.57 cm, (<b>b</b>) 2.17 cm, and (<b>c</b>) 6.75 cm (Note: The pixel spacing varies between images (<b>a</b>–<b>c</b>), so one pixel does not correspond to the same length in cm across these images). The colored dots represent the annotations of the quadriceps muscle and femur bone surfaces (better seen in colored prints). (<b>d</b>) The distribution of QMT across all 486 subjects.</p>
Full article ">Figure 2
<p>Summary of the proposed QMT measurement framework: <span class="html-italic">IW-Regression</span>, <span class="html-italic">CW-Regression</span>, and <span class="html-italic">Seg-Regression</span>. <span class="html-italic">IW-Regression</span> model initialized with ImageNet weights, <span class="html-italic">CW-Regression</span> model initialized with <span class="html-italic">IW-Classifictaion</span> weights, and <span class="html-italic">Seg-Regression</span> model initialized with ImageNet weights.</p>
Full article ">Figure 3
<p>Examples of segmentation masks generated from manual annotations (ground truth) for three patients (<b>a</b>–<b>c</b>) (better seen in colored prints). The predicted masks showcase the segmentation outcomes achieved by the <span class="html-italic">Seg-Regression</span> model (predicted mask). The Dice scores of predicted masks for patients (<b>a</b>), (<b>b</b>), and (<b>c</b>) were 0.63, 0.89, and 0.76, respectively. In this process, the QMT was derived through a post-processing step that involved determining the distance between the horizontal edges of the muscle surface and the femur surface, utilizing Canny edge detection (horizontal edges).</p>
Full article ">Figure 4
<p>Statistical analysis for the (<b>a</b>) <span class="html-italic">IW-Regression</span> and (<b>b</b>) <span class="html-italic">CW-Regression</span> models.</p>
Full article ">Figure 5
<p>Activation maps in classification models: ResNet101 (<b>a</b>,<b>e</b>), DensNet (<b>b</b>,<b>f</b>), ViT-B (<b>c</b>,<b>g</b>), and MAE-B (<b>d</b>,<b>h</b>). The first and second rows represent activation maps for two different subjects. The first row was correctly classified, and the second row was misclassified. (GT: ground truth class label, Pred: predicted class label).</p>
Full article ">Figure 6
<p>Sample activation maps of (<b>a</b>) ViT-B and (<b>b</b>) MAE-B, detecting the body of the muscle.</p>
Full article ">
23 pages, 9356 KiB  
Article
FA-VTON: A Feature Alignment-Based Model for Virtual Try-On
by Yan Wan, Ning Ding and Li Yao
Appl. Sci. 2024, 14(12), 5255; https://doi.org/10.3390/app14125255 - 17 Jun 2024
Viewed by 1106
Abstract
The virtual try-on technology based on 2D images aims to seamlessly transfer provided garments onto target person images. Prior methods mainly concentrated on warping garments and generating images, overlooking the influence of feature alignment on the try-on results. In this study, we initially [...] Read more.
The virtual try-on technology based on 2D images aims to seamlessly transfer provided garments onto target person images. Prior methods mainly concentrated on warping garments and generating images, overlooking the influence of feature alignment on the try-on results. In this study, we initially analyze the distortions present by existing methods and elucidate the critical role of feature alignment in the extraction stage. Building on this, we propose a novel feature alignment-based model (FA-VTON). Specifically, FA-VTON aligns the upsampled higher-level features from both person and garment images to acquire precise boundary information, which serves as guidance for subsequent garment warping. Concurrently, the Efficient Channel Attention mechanism (ECA) is introduced to generate the final result in the try-on generation module. This mechanism enables adaptive adjustment of channel feature weights to extract important features and reduce artifact generation. Furthermore, to make the student network focus on salient regions of each channel, we utilize channel-wise distillation (CWD) to minimize the Kullback–Leibler (KL) divergence between the channel probability maps of the two networks. The experiments show that our model achieves better results in both qualitative and quantitative analyses compared to current methods on the popular virtual try-on datasets. Full article
Show Figures

Figure 1

Figure 1
<p>Comparison of feature maps generated by our method with FS-VTON. We visualize the feature maps by compressing the color channels. (<b>a</b>) Comparison of garment feature maps; (<b>b</b>) comparison of person feature maps.</p>
Full article ">Figure 2
<p>The influence of garment feature extraction on try-on results.</p>
Full article ">Figure 3
<p>The influence of person feature extraction on try-on results.</p>
Full article ">Figure 4
<p>The overall architecture framework of FA-VTON.</p>
Full article ">Figure 5
<p>The structure of the person FEM. The garment FEM has the same structure with different inputs. The dashed box represents the structure of the FASM, consisting of the FAM and FSM.</p>
Full article ">Figure 6
<p>The structure of the FAM and FSM. (<b>a</b>) The workflow of the Feature Alignment Module (FAM). Aligning <math display="inline"><semantics> <mrow> <msubsup> <mi>P</mi> <mi>i</mi> <mrow> <mi>u</mi> <mi>p</mi> </mrow> </msubsup> </mrow> </semantics></math> by learning the spatial position offsets <math display="inline"><semantics> <mrow> <msub> <mi mathvariant="sans-serif">Δ</mi> <mi>i</mi> </msub> </mrow> </semantics></math> of <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>C</mi> <mo stretchy="false">^</mo> </mover> <mrow> <mi>i</mi> <mo>−</mo> <mn>1</mn> </mrow> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msubsup> <mi>P</mi> <mi>i</mi> <mrow> <mi>u</mi> <mi>p</mi> </mrow> </msubsup> </mrow> </semantics></math>; (<b>b</b>) the workflow of the Feature Selection Module (FSM).</p>
Full article ">Figure 7
<p>Overview and details of our Coarse-Fine Warping Module (CFWM). (<b>a</b>) The structure of the CFWM; (<b>b</b>) The specific structure of the Warping Block.</p>
Full article ">Figure 8
<p>Structure of the TGM.</p>
Full article ">Figure 9
<p>The overall architecture of the ECA module.</p>
Full article ">Figure 10
<p>Qualitative results from different models (ACGPN, PF-AFN, SDAFN, DM-VTON, and ours) on the VITON-HD testing dataset.</p>
Full article ">Figure 11
<p>Try-on results of complex clothes.</p>
Full article ">Figure 12
<p>The ablation study of the FASM.</p>
Full article ">Figure 13
<p>The ablation study of ECA.</p>
Full article ">Figure 14
<p>The ablation study of CWD.</p>
Full article ">
24 pages, 4029 KiB  
Article
Personalized Tourist Recommender System: A Data-Driven and Machine-Learning Approach
by Deepanjal Shrestha, Tan Wenan, Deepmala Shrestha, Neesha Rajkarnikar and Seung-Ryul Jeong
Computation 2024, 12(3), 59; https://doi.org/10.3390/computation12030059 - 18 Mar 2024
Cited by 3 | Viewed by 4744
Abstract
This study introduces a data-driven and machine-learning approach to design a personalized tourist recommendation system for Nepal. It examines key tourist attributes, such as demographics, behaviors, preferences, and satisfaction, to develop four sub-models for data collection and machine learning. A structured survey is [...] Read more.
This study introduces a data-driven and machine-learning approach to design a personalized tourist recommendation system for Nepal. It examines key tourist attributes, such as demographics, behaviors, preferences, and satisfaction, to develop four sub-models for data collection and machine learning. A structured survey is conducted with 2400 international and domestic tourists, featuring 28 major questions and 125 variables. The data are preprocessed, and significant features are extracted to enhance the accuracy and efficiency of the machine-learning models. These models are evaluated using metrics such as accuracy, precision, recall, F-score, ROC, and lift curves. A comprehensive database for Pokhara City, Nepal, is developed from various sources that includes attributes such as location, cost, popularity, rating, ranking, and trend. The machine-learning models provide intermediate categorical recommendations, which are further mapped using a personalized recommender algorithm. This algorithm makes decisions based on weights assigned to each decision attribute to make the final recommendations. The system’s performance is compared with other popular recommender systems implemented by TripAdvisor, Google Maps, the Nepal tourism website, and others. It is found that the proposed system surpasses existing ones, offering more accurate and optimized recommendations to visitors in Pokhara. This study is a pioneering one and holds significant implications for the tourism industry and the governing sector of Nepal in enhancing the overall tourism business. Full article
(This article belongs to the Special Issue Intelligent Computing, Modeling and its Applications)
Show Figures

Figure 1

Figure 1
<p>The conceptual framework for a tourist recommender system.</p>
Full article ">Figure 2
<p>Conceptual architecture of a Tourist Recommender System for Nepal.</p>
Full article ">Figure 3
<p>(<b>a</b>) ROC analysis for training dataset for average over the classes. (<b>b</b>) Lift curve of training dataset for average over the classes.</p>
Full article ">Figure 4
<p>Planning model evaluation with average over classes.</p>
Full article ">Figure 5
<p>(<b>a</b>) ROC analysis for training dataset for average over the classes. (<b>b</b>) Lift curve of training dataset for average over the classes.</p>
Full article ">Figure 6
<p>(<b>a</b>) ROC curve analysis for prediction data of access to country class. (<b>b</b>) Decision factor analysis for prediction dataset for average over the classes. (<b>c</b>) Decision factor analysis for prediction dataset for average over the classes.</p>
Full article ">Figure 7
<p>Evaluation of testing data for entertainment class.</p>
Full article ">Figure 8
<p>Evaluation of testing data for popular destination class.</p>
Full article ">Figure 9
<p>(<b>a</b>) ROC analysis for behavioral training dataset for the target class sports and activities. (<b>b</b>) Decision factor analysis for prediction dataset for average over the classes. (<b>c</b>) Decision factor analysis for prediction dataset for average over the classes.</p>
Full article ">
10 pages, 1072 KiB  
Article
Can the ADC Value Be Used as an Imaging “Biopsy” in Endometrial Cancer?
by Octavia Petrila, Ionut Nistor, Narcis Sandy Romedea, Dragos Negru and Viorel Scripcariu
Diagnostics 2024, 14(3), 325; https://doi.org/10.3390/diagnostics14030325 - 2 Feb 2024
Cited by 3 | Viewed by 1683
Abstract
Background: The tumor histological grade is closely related to the prognosis of patients with endometrial cancer (EC). Multiparametric MRI, including diffusion-weighted imaging (DWI), provides information about the cellular density that may be useful to differentiate between benign and malignant uterine lesions. However, correlations [...] Read more.
Background: The tumor histological grade is closely related to the prognosis of patients with endometrial cancer (EC). Multiparametric MRI, including diffusion-weighted imaging (DWI), provides information about the cellular density that may be useful to differentiate between benign and malignant uterine lesions. However, correlations between apparent diffusion coefficient (ADC) values and histopathological grading in endometrial cancer remain controversial. Material and methods: We retrospectively evaluated 92 patients with endometrial cancers, including both endometrioid adenocarcinomas (64) and non-endometrioid adenocarcinomas (28). All patients underwent DWI procedures, and mean ADC values were calculated in a region of interest. These values were then correlated with the tumor grading offered by the histopathological examination, which was considered the gold standard. In this way, the patients were divided into three groups (G1, G2, and G3). The ADC values were then compared to the results offered by the biopsy to see if the DWI sequence and ADC map could replace this procedure. We also compared the mean ADC values to the myometrial invasion (</>50%) and lymphovascular space invasion. Results: We have divided the ADC values into three categories corresponding to three grades: >0.850 × 10−3 mm2/s (ADC1), 0.730–0.849 × 10−3 mm2/s (ADC2) and <0.730 × 10−3 mm2/s (ADC3). The diagnostic accuracy of the ADC value was 85.71% for ADC1, 75.76% for ADC2, and 91.66% for ADC3. In 77 cases out of 92, the category in which they were placed using the ADC value corresponded to the result offered by the histopathological exam with an accuracy of 83.69%. For only 56.52% of patients, the biopsy result included the grading system. For each grading category, the mean ADC value showed better results than the biopsy; for G1 patients, the mean ADC value had an accuracy of 85.71% compared to 66.66% in the biopsy, G2 had 75.76% compared to 68.42%, and G3 had 91.66 compared to 75%. For both deep myometrial invasion and lymphovascular space invasion, there is a close, inversely proportional correlation with the mean ADC value. Conclusions: Mean endometrial tumor ADC on MR-DWI is inversely related to the histological grade, deep myometrial invasion and lymphovascular space invasion. Using this method, the patients could be better divided into risk categories for personalized treatment. Full article
(This article belongs to the Special Issue Diagnosis and Management of Gynecological Cancers: Volume 2)
Show Figures

Figure 1

Figure 1
<p>A 46-year-old woman with histopathologically proven low-grade endometrioid endometrial cancer; sagittal T2-weighted fast spin-echo image (<b>a</b>) and axial view (<b>b</b>) show a hypointense mass in the endometrial cavity, infiltrating more than 50% of the myometrial thickness; axial DWI (<b>c</b>) and ADC map (<b>d</b>) show a hyperintense mass on high b value image (b = 800 s/mm<sup>2</sup>) corresponding to a low-intensity mass on the derived ADC map; the ADC value is 0.948 × 10<sup>−3</sup> mm<sup>2</sup>/s; axial fat-saturated spin-echo T1-weighted image (<b>e</b>) and gadolinium-enhanced fat-saturated spin-echo T1-weighted image (<b>f</b>) show the tumor as a slightly enhanced mass.</p>
Full article ">Figure 2
<p>A 66-year-old woman with histopathologically proven medium-grade endometrioid endometrial cancer; sagittal T2-weighted fast spin-echo image (<b>a</b>) and axial view (<b>b</b>) show a hypointense mass in the endometrial cavity, infiltrating more than 50% of the myometrial thickness and the cervical stroma; also, on the axial view, (<b>b</b>) there is an intramural leiomyoma with no mass effect on the endometrial cavity; axial DWI (<b>c</b>) and ADC map (<b>d</b>) show an inhomogeneous hyperintense mass on high b value image (b = 800 s/mm<sup>2</sup>) corresponding to a low-intensity mass on the derived ADC map; the ADC value is 0.812 × 10<sup>−3</sup> mm<sup>2</sup>/s; axial fat-saturated spin-echo T1-weighted image (<b>e</b>) and gadolinium-enhanced fat-saturated spin-echo T1-weighted image (<b>f</b>) show the tumor as a slightly enhanced mass.</p>
Full article ">Figure 3
<p>A 75-year-old woman with a histopathologically proven high-grade endometrial cancer (mixed type endometrioid and serous); sagittal T2-weighted fast spin-echo image (<b>a</b>) and axial view (<b>b</b>) show a hypointense mass in the endometrial cavity, infiltrating more than 50% of the myometrial thickness and cervical stroma and invading the uterine serosa; the tumor; axial DWI (<b>c</b>) and ADC map (<b>d</b>) show a hyperintense mass on high b value image (b = 800 s/mm<sup>2</sup>) corresponding to a low-intensity mass on the derived ADC map; the ADC value is 0.558 × 10<sup>−3</sup> mm<sup>2</sup>/s; axial fat-saturated spin-echo T1-weighted image (<b>e</b>) and gadolinium-enhanced fat-saturated spin-echo T1-weighted image (<b>f</b>) show the tumor as a slightly enhanced, inhomogeneous mass.</p>
Full article ">
17 pages, 5502 KiB  
Article
Ethnic Disparities in Lipid Metabolism and Clinical Outcomes between Dutch South Asians and Dutch White Caucasians with Type 2 Diabetes Mellitus
by Lushun Yuan, Aswin Verhoeven, Niek Blomberg, Huub J. van Eyk, Maurice B. Bizino, Patrick C. N. Rensen, Ingrid M. Jazet, Hildo J. Lamb, Ton J. Rabelink, Martin Giera and Bernard M. van den Berg
Metabolites 2024, 14(1), 33; https://doi.org/10.3390/metabo14010033 - 3 Jan 2024
Viewed by 2350
Abstract
Type 2 diabetes mellitus (T2DM) poses a higher risk for complications in South Asian individuals compared to other ethnic groups. To shed light on potential mediating factors, we investigated lipidomic changes in plasma of Dutch South Asians (DSA) and Dutch white Caucasians (DwC) [...] Read more.
Type 2 diabetes mellitus (T2DM) poses a higher risk for complications in South Asian individuals compared to other ethnic groups. To shed light on potential mediating factors, we investigated lipidomic changes in plasma of Dutch South Asians (DSA) and Dutch white Caucasians (DwC) with and without T2DM and explore their associations with clinical features. Using a targeted quantitative lipidomics platform, monitoring over 1000 lipids across 17 classes, along with 1H NMR based lipoprotein analysis, we studied 51 healthy participants (21 DSA, 30 DwC) and 92 T2DM patients (47 DSA, 45 DwC) from the MAGNetic resonance Assessment of VICTOza efficacy in the Regression of cardiovascular dysfunction in type 2 dIAbetes mellitus (MAGNA VICTORIA) study. This comprehensive mapping of the circulating lipidome allowed us to identify relevant lipid modules through unbiased weighted correlation network analysis, as well as disease and ethnicity related key mediatory lipids. Significant differences in lipidomic profiles, encompassing various lipid classes and species, were observed between T2DM patients and healthy controls in both the DSA and DwC populations. Our analyses revealed that healthy DSA, but not DwC, controls already exhibited a lipid profile prone to develop T2DM. Particularly, in DSA-T2DM patients, specific lipid changes correlated with clinical features, particularly diacylglycerols (DGs), showing significant associations with glycemic control and renal function. Our findings highlight an ethnic distinction in lipid modules influencing clinical outcomes in renal health. We discover distinctive ethnic disparities of the circulating lipidome and identify ethnicity-specific lipid markers. Jointly, our discoveries show great potential as personalized biomarkers for the assessment of glycemic control and renal function in DSA-T2DM individuals. Full article
(This article belongs to the Section Endocrinology and Clinical Metabolic Research)
Show Figures

Figure 1

Figure 1
<p>Study workflow and design. Abbreviations: T2DM type 2 diabetes mellitus; WGCNA Weighted Correlation Network Analysis.</p>
Full article ">Figure 2
<p>Lipid class abundance between patients with T2DM and healthy controls. (<b>A</b>) Stack plot with the hierarchical cluster in Dutch South Asian. (<b>B</b>) Stack plot with the hierarchical cluster in Dutch white Caucasian. Abbreviations: CE cholesteryl ester; CER ceramide; DCER dihydroceramide; DG diacylglyceride; FA fatty acid; HC healthy control; HexCER hydroxyceramide; LacCER lactosylceramide; LPC lysophosphatidylcholine; LPE lysophosphatidylethanolamine; PA phosphatidic acid; PC phosphatidylcholine; PE phosphatidylethanolamine; PI phosphatidylinositol; PS phosphatidylserine; SM sphingomyelin; T2DM type 2 diabetes mellitus, TG triglyceride.</p>
Full article ">Figure 3
<p>Comparison of differential lipids between patients with T2DM and healthy controls in two ethnicities. (<b>A</b>) Differential lipids per lipid class between patients with T2DM and healthy controls, as well as a comparison across two ethnicities. The colour grey represents lipids with non-significant associations (FDR &gt; 0.05), the colour blue represents lipids lower in T2DM, and the colour red represents lipids higher in T2DM. The dot size represents −log<sub>2</sub>FDR. Venn diagram of (<b>B</b>) lipids lower in T2DM and (<b>C</b>) higher in T2DM than healthy controls (HC) in DSA and DwC. Heatmap of lipids that are commonly/uncommonly (<b>D</b>) lower and (<b>E</b>) higher in DSA and DwC with T2DM. Abbreviations: CE cholesteryl ester; CER ceramide; DCER dihydroceramide; DG diacylglyceride; DwC Dutch white Caucasian; DSA Dutch South Asian; FA fatty acid; FDR false discovery rate; HexCER hydroxyceramide; LacCER lactosylceramide; LPC lysophosphatidylcholine; LPE lysophosphatidylethanolamine; PA phosphatidic acid; PC phosphatidylcholine; PE phosphatidylethanolamine; PI phosphatidylinositol; PS phosphatidylserine; SM sphingomyelin, TG triglyceride.</p>
Full article ">Figure 4
<p>Association of lipid correlation network modules with clinical features in (<b>A</b>) Dutch South Asians with T2DM and (<b>B</b>) Dutch white Caucasians with T2DM. The colour grey denotes a lipid cluster with no significant associations with clinical features, the colour blue denotes a lipid cluster with a negative association with clinical features, and the colour red denotes a lipid cluster with a positive association with clinical features. The correlation coefficients are represented by the size of the dots (Spearman’s rank correlation test). Abbreviations: BP blood pressure; BMI body mass index; HbA1c hemoglobin A1c; HDL high-density lipoprotein; LBM lean body mass; LDL low-density lipoprotein; SAT subcutaneous adipose tissue, VAT visceral adipose tissue.</p>
Full article ">Figure 5
<p>Correlations between key mediatory lipids in diabetic nephropathy-associated module of Dutch South Asians and lipoproteins, kidney function, and glycemic control. (<b>A</b>) Bubble plot depicting the correlations of lipids with lipoproteins, kidney function, and glycemic control in Dutch South Asians with T2DM. (<b>B</b>) Violin plots of lipids between T2DM with and without diabetic nephropathy in Dutch South Asians. (<b>C</b>) Bubble plot depicting the correlations of lipids with lipoproteins, kidney function, and glycemic control in Dutch white Caucasians with T2DM. (<b>D</b>) Violin plots of lipids between T2DM with and without DN in Dutch white Caucasians. Lipids in bold indicated that they were specifically different in Dutch South Asians. The colour grey indicates no significant correlations with clinical features; the colour blue indicates a negative correlation with clinical features, and the colour red indicates a positive correlation with clinical features. The size of the dots represents the correlation coefficients (Pearson’s correlation). The Wilcoxon signed-rank test was performed; * <span class="html-italic">p</span> &lt; 0.05, ** <span class="html-italic">p</span> &lt; 0.01, ns not significant. Abbreviations: DG diacylglyceride; DN diabetic nephropathy; HbA1c hemoglobin A1c; HDL high-density lipoprotein; LDL low-density lipoprotein, T2DM type 2 diabetes mellitus.</p>
Full article ">
29 pages, 5823 KiB  
Article
A Personalized Motion Planning Method with Driver Characteristics in Longitudinal and Lateral Directions
by Di Zeng, Ling Zheng, Yinong Li, Jie Zeng and Kan Wang
Electronics 2023, 12(24), 5021; https://doi.org/10.3390/electronics12245021 - 15 Dec 2023
Cited by 1 | Viewed by 1324
Abstract
Humanlike driving is significant in improving the safety and comfort of automated vehicles. This paper proposes a personalized motion planning method with driver characteristics in longitudinal and lateral directions for highway automated driving. The motion planning is decoupled into path optimization and speed [...] Read more.
Humanlike driving is significant in improving the safety and comfort of automated vehicles. This paper proposes a personalized motion planning method with driver characteristics in longitudinal and lateral directions for highway automated driving. The motion planning is decoupled into path optimization and speed optimization under the framework of the Baidu Apollo EM motion planner. For modeling driver behavior in the longitudinal direction, a car-following model is developed and integrated into the speed optimizer based on a weight ratio hypothesis model of the objective functional, whose parameters are obtained by Bayesian optimization and leave-one-out cross validation using the driving data. For modeling driver behavior in the lateral direction, a Bayesian network (BN), which maps the physical states of the ego vehicle and surrounding vehicles and the lateral intentions of the surrounding vehicles to the driver’s lateral intentions, is built in an efficient and lightweight way using driving data. Further, a personalized reference trajectory decider is developed based on the BN, considering traffic regulations, the driver’s preference, and the costs of the trajectories. According to the actual traffic scenarios in the driving data, a simulation is constructed, and the results validate the human likeness of the proposed motion planning method. Full article
Show Figures

Figure 1

Figure 1
<p>Framework of the motion planning method.</p>
Full article ">Figure 2
<p>Human driving data collection system.</p>
Full article ">Figure 3
<p>An example of SL projection. The red square marks the overlap region where the ego vehicle should not enter.</p>
Full article ">Figure 4
<p>The candidate paths for the three lanes. The curves in olive represent the generated candidate paths for the three lanes. The white dashed lines are the boundaries between lanes. The solid black lines with arrows represent the axes of the coordinate system. The red square marks the overlap region.</p>
Full article ">Figure 5
<p>A QP path example of nudging. The red square marks the overlap region. The orange region depicts the feasible tunnel.</p>
Full article ">Figure 6
<p>Personalized speed decision.</p>
Full article ">Figure 7
<p>QP speed planning result.</p>
Full article ">Figure 8
<p>Comparison of clearance, speed, and acceleration between different ratios.</p>
Full article ">Figure 9
<p>The BN of the personalized reference trajectory decider.</p>
Full article ">Figure 10
<p>An example of traffic scenario.</p>
Full article ">Figure 11
<p>An example of the probability tree.</p>
Full article ">Figure 12
<p>An example of different partitions of a variable.</p>
Full article ">Figure 13
<p>Comparison of trajectories, speed, acceleration, and curvature between different weights of the path optimizer.</p>
Full article ">Figure 14
<p>Bayesian optimization of the parameters of the quadratic ratio model.</p>
Full article ">Figure 15
<p>Comparison of the EM planner, constant model, MLCF model, and hypothesis models in a typical car-following scenario.</p>
Full article ">Figure 16
<p>The simulated trajectories, speed, acceleration, and curvature of the ego vehicle in representative lane-changing and lane-keeping scenarios: (<b>a</b>) the lane-changing scenario; (<b>b</b>) the lane-keeping scenario.</p>
Full article ">
13 pages, 1462 KiB  
Article
Investigating the Feasibility of Predicting KRAS Status, Tumor Staging, and Extramural Venous Invasion in Colorectal Cancer Using Inter-Platform Magnetic Resonance Imaging Radiomic Features
by Mohammed S. Alshuhri, Abdulaziz Alduhyyim, Haitham Al-Mubarak, Ahmad A. Alhulail, Othman I. Alomair, Yahia Madkhali, Rakan A. Alghuraybi, Abdullah M. Alotaibi and Abdullalh G. M. Alqahtani
Diagnostics 2023, 13(23), 3541; https://doi.org/10.3390/diagnostics13233541 - 27 Nov 2023
Cited by 2 | Viewed by 1463
Abstract
(1) Background: Colorectal cancer is the third most common type of cancer with a high mortality rate and poor prognosis. The accurate prediction of key genetic mutations, such as the KRAS status, tumor staging, and extramural venous invasion (EMVI), is crucial for guiding [...] Read more.
(1) Background: Colorectal cancer is the third most common type of cancer with a high mortality rate and poor prognosis. The accurate prediction of key genetic mutations, such as the KRAS status, tumor staging, and extramural venous invasion (EMVI), is crucial for guiding personalized treatment decisions and improving patients’ outcomes. MRI radiomics was assessed to predict the KRAS status and tumor staging in colorectal cancer patients across different imaging platforms to improve the personalized treatment decisions and outcomes. (2) Methods: Sixty colorectal cancer patients (35M/25F; avg. age 56.3 ± 12.9 years) were treated at an oncology unit. The MRI scans included T2-weighted (T2W) and diffusion-weighted imaging (DWI) or the apparent diffusion coefficient (ADC). The manual segmentation of colorectal cancer was conducted on the T2W and DWI/ADC images. The cohort was split into training and validation sets, and machine learning was used to build predictive models. (3) Results: The neural network (NN) model achieved 73% accuracy and an AUC of 0.71 during training for predicting the KRAS mutation status, while during testing, it achieved 62.5% accuracy and an AUC of 0.68. In the case of tumor grading, the support vector machine (SVM) model excelled with a training accuracy of 72.93% and an AUC of 0.7, and during testing, it reached an accuracy of 72% and an AUC of 0.69. (4) Conclusions: ML models using radiomics from ADC maps and T2-weighted images are effective for distinguishing KRAS genes, tumor grading, and EMVI in colorectal cancer. Standardized protocols are essential to improve MRI radiomics’ reliability in clinical practice. Full article
(This article belongs to the Special Issue Application of AI in Diagnosis of Colorectal Cancer)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Typical MR imaging appearance of KARAS + or − tumor (yellow arrow) on (<b>a</b>) ADC and (<b>b</b>) T2-weighted images.</p>
Full article ">Figure 2
<p>Flowchart illustrating the basic steps of radiomics workflow.</p>
Full article ">Figure 3
<p>Shows ICC values of radiomic features between two platforms.</p>
Full article ">Figure 4
<p>(<b>a</b>) ROC curves of train and test to predict DNA. (<b>b</b>) ROC curves of train and test to predict t stage. (<b>c</b>) ROC curves of train and test to predict N stage. (<b>d</b>) ROC curves of train and test to predict MRF. (<b>e</b>) ROC curves of train and test to predict EMVI.</p>
Full article ">
13 pages, 2299 KiB  
Article
Instrumental Gait Analysis and Tibial Plateau Modelling to Support Pre- and Post-Operative Evaluations in Personalized High Tibial Osteotomy
by Claudio Belvedere, Harinderjit Singh Gill, Maurizio Ortolani, Nicoletta Sileoni, Stefano Zaffagnini, Fabio Norvillo, Alisdair MacLeod, Giacomo Dal Fabbro, Alberto Grassi and Alberto Leardini
Appl. Sci. 2023, 13(22), 12425; https://doi.org/10.3390/app132212425 - 17 Nov 2023
Cited by 3 | Viewed by 1495
Abstract
High tibial osteotomy (HTO) is intended to treat medial knee osteoarthritis by realigning the joint such that the loading in the knee during functional activity shifts laterally. The aim of this study was to use a novel methodology combining motion analysis and 3D [...] Read more.
High tibial osteotomy (HTO) is intended to treat medial knee osteoarthritis by realigning the joint such that the loading in the knee during functional activity shifts laterally. The aim of this study was to use a novel methodology combining motion analysis and 3D modelling to assess the efficacy of this surgery in changing the loading location in the knee in a cohort of 25 patients treated with personalized HTO. Pre-operatively and at 6 months post-surgery, weight-bearing CT and gait analysis during level walking were performed on all patients, as well as clinical evaluations using KOOS and VAS scores. CT scans were used to generate a knee bone model and a virtual tibial plateau plane; the intersection pattern between this plane and the ground reaction force (GRF) vector was calculated in the pre- and post-operative gait analyses. Clinical scores improved significantly (p < 0.001) after surgery (pre-/post-operative KOOS and VAS: 56.2 ± 14.0/82.0 ± 8.3 and 6.3 ± 1.7/1.5 ± 1.7). Post-operative GRF-to-tibial plateau intersection patterns were significantly (p < 0.001) more lateral (31.9 ± 19.8% of tibial plateau width) than the pre-operative patterns. Personalized HTO successfully and consistently lateralizes the GRF at the knee, in association with significant improvements in function and pain. The novel combination of 3D bone modelling and motion analysis also has the potential to further aid HTO surgical planning. Full article
(This article belongs to the Special Issue Biomechanics and Human Motion Analysis)
Show Figures

Figure 1

Figure 1
<p>Workflow of the new customized HTO procedure: the original patient’s deformity is assessed on a standing long-leg frontal radiograph (<b>a</b>); virtual surgery is based on a 3D anatomical model of the knee reconstructed from CT images of the patient (<b>b</b>); the HTO fixation plate is then 3D-printed in medical titanium alloy (<b>c</b>) and delivered to the hospital for sterilization and surgical implantation (<b>d</b>); the final alignment of the lower limb is assessed on another standing frontal radiograph (<b>e</b>).</p>
Full article ">Figure 2
<p>Workflow of the recently reported methodology for characterizing GRF data on tibial morphology [<a href="#B25-applsci-13-12425" class="html-bibr">25</a>]: functional data are collected via gait analysis during level walking, using a standard marker set protocol enriched with additional markers around the knee; medical images are acquired with CT scan, and related DICOM files are used to reconstruct 3D models for the bones and the additional markers; a single-value decomposition approach combining the reconstructed and tracked additional markers is then used to register the positions of the GRF vectors on the tibial bone model.</p>
Full article ">Figure 3
<p>Medial–lateral and anterior–posterior patterns of intersections versus % of gait stance. Pre- and post-operative data are reported as mean curves ± standard deviations for all patients, both in millimetres (mm) and in % of the tibial plateau width, along with the associated comparison of pre–post data via SPM (SPM{t}), where t* indicates the critical value of t given the significance level α.</p>
Full article ">Figure 4
<p>Mean pre- and post-operative pattern of intersections (in mm) superimposed on the 3D tibial plateau model from a representative patient (transverse view in the tibial anatomical reference frame) compared with all patients. The standard deviations in the AP and ML direction associated with the first and the second GRF peak in the stance phase (1st and 2nd GRF) to the first and the second sagittal (1st and 2nd flex-ext) and frontal (1st and 2nd abd-add) knee moment peaks and to the transverse knee moment peak (int-ext) are also shown.</p>
Full article ">Figure 5
<p>Hip, knee and ankle joint kinematics on the sagittal, frontal and transverse plane versus % of gait cycle. Pre- and post-operative rotations (in degrees) are reported as mean curves ± standard deviations for all patients, along with the associated comparison of pre–post data via SPM (SPM{t}), where t* indicates the critical value of t given the significance level α.</p>
Full article ">Figure 6
<p>Hip, knee and ankle joint kinetics on the sagittal, frontal and transverse plane versus % of the full gait cycle (stance phase, i.e., with the foot on the forceplate, finishing at about 60% of the cycle). Pre- and post-operative data (in %BWxH) are reported as mean curves ± standard deviations for all patients, along with the associated comparison of pre–post data via SPM (SPM{t}), where t* indicates the critical value of t given the significance level α.</p>
Full article ">
15 pages, 3710 KiB  
Article
Improved YOLOv3 Integrating SENet and Optimized GIoU Loss for Occluded Pedestrian Detection
by Qiangbo Zhang, Yunxiang Liu, Yu Zhang, Ming Zong and Jianlin Zhu
Sensors 2023, 23(22), 9089; https://doi.org/10.3390/s23229089 - 10 Nov 2023
Cited by 8 | Viewed by 1426
Abstract
Occluded pedestrian detection faces huge challenges. False positives and false negatives in crowd occlusion scenes will reduce the accuracy of occluded pedestrian detection. To overcome this problem, we proposed an improved you-only-look-once version 3 (YOLOv3) based on squeeze-and-excitation networks (SENet) and optimized generalized [...] Read more.
Occluded pedestrian detection faces huge challenges. False positives and false negatives in crowd occlusion scenes will reduce the accuracy of occluded pedestrian detection. To overcome this problem, we proposed an improved you-only-look-once version 3 (YOLOv3) based on squeeze-and-excitation networks (SENet) and optimized generalized intersection over union (GIoU) loss for occluded pedestrian detection, namely YOLOv3-Occlusion (YOLOv3-Occ). The proposed network model considered incorporating squeeze-and-excitation networks (SENet) into YOLOv3, which assigned greater weights to the features of unobstructed parts of pedestrians to solve the problem of feature extraction against unsheltered parts. For the loss function, a new generalized intersection over unionintersection over groundtruth (GIoUIoG) loss was developed to ensure the areas of predicted frames of pedestrian invariant based on the GIoU loss, which tackled the problem of inaccurate positioning of pedestrians. The proposed method, YOLOv3-Occ, was validated on the CityPersons and COCO2014 datasets. Experimental results show the proposed method could obtain 1.2% MR−2 gains on the CityPersons dataset and 0.7% mAP@50 improvements on the COCO2014 dataset. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

Figure 1
<p>The architecture of SENet. The input is a feature map. The output is the feature map injected with the attention scores. Firstly, the GAP layer compresses the shape of the input feature map to (1, 1, C). Secondly, two fully connected layers and activated functions obtain the attention scores of all channels, as shown in the feature map marked by diverse colors. Finally, the operation of multiplying channel by channel is implemented between the input feature map and the attention scores to generate the output.</p>
Full article ">Figure 2
<p>The architecture of the proposed YOLOv3-Occ. For an input image, the output is the image with detection boxes. The black arrow represents the data flow. CNN layers consist of a set of CBLs and basic residual blocks, which are used to extract fine-grained features of the input image. The outputs of two basic residual blocks in CNN layers are used as the inputs of two Concat layers in the model.</p>
Full article ">Figure 3
<p>Comparison of mAP@50s among three batch sizes on the CityPersons training set. The curves show how the mAP@50s of these batch sizes change with the number of epochs. The three batch sizes and their corresponding curve colors are represented in the upper left corner of the figure.</p>
Full article ">Figure 4
<p>Comparison of mAPs among three iou-thresholds on the CityPersons training set. The curves show how the mAPs of these iou-thresholds change with the number of epochs. The three iou-thresholds and their corresponding curve colors are represented in the upper left corner of the figure.</p>
Full article ">Figure 5
<p>Visual comparison of the baseline, EMD-RCNN, and our approach. The first row is the results of the baseline. The second row is the results generated by EMD-RCNN. The third row is the results of YOLOv3-Occ. The blue boxes are the detection results, the white boxes are false negatives, and the yellow boxes are false positives.</p>
Full article ">Figure 6
<p>P–R curve on one batch of the COCO validation set. There are eight classes of P–R curves and four of them coincide with the rest of the curves. The eight classes and their corresponding curve colors are represented in the lower right corner of the figure.</p>
Full article ">Figure 7
<p>Comparison of mAP@50s among YOLOv3-Occ, Faster R-CNN with FPN [<a href="#B38-sensors-23-09089" class="html-bibr">38</a>], our baseline, RetinaNet [<a href="#B39-sensors-23-09089" class="html-bibr">39</a>], SSD523 [<a href="#B40-sensors-23-09089" class="html-bibr">40</a>], YOLOv2, YOLOv4 [<a href="#B41-sensors-23-09089" class="html-bibr">41</a>], YOLOv5 [<a href="#B42-sensors-23-09089" class="html-bibr">42</a>] on the COCO validation set. The curves show how the mAP@50s of these methods change with the number of epochs. The eight methods and their corresponding curve colors are represented in the lower right corner of the figure.</p>
Full article ">
Back to TopTop