Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (112)

Search Parameters:
Keywords = MobileFaceNet

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 1518 KiB  
Article
VAS-3D: A Visual-Based Alerting System for Detecting Drowsy Drivers in Intelligent Transportation Systems
by Hadi El Zein, Hassan Harb, François Delmotte, Oussama Zahwe and Samir Haddad
World Electr. Veh. J. 2024, 15(12), 540; https://doi.org/10.3390/wevj15120540 - 21 Nov 2024
Viewed by 474
Abstract
Nowadays, driving accidents are considered one of the most crucial challenges for governments and communities that affect transportation systems and peoples lives. Unfortunately, there are many causes behind the accidents; however, drowsiness is one of the main factors that leads to a significant [...] Read more.
Nowadays, driving accidents are considered one of the most crucial challenges for governments and communities that affect transportation systems and peoples lives. Unfortunately, there are many causes behind the accidents; however, drowsiness is one of the main factors that leads to a significant number of injuries and deaths. In order to reduce its effect, researchers and communities have proposed many techniques for detecting drowsiness situations and alerting the driver before an accident occurs. Mostly, the proposed solutions are visually-based, where a camera is positioned in front of the driver to detect their facial behavior and then determine their situation, e.g., drowsy or awake. However, most of the proposed solutions make a trade-off between detection accuracy and speed. In this paper, we propose a novel Visual-based Alerting System for Detecting Drowsy Drivers (VAS-3D) that ensures an optimal trade-off between the accuracy and speed metrics. Mainly, VAS-3D consists of two stages: detection and classification. In the detection stage, we use pre-trained Haar cascade models to detect the face and eyes of the driver. Once the driver’s eyes are detected, the classification stage uses several pre-trained Convolutional Neural Network (CNN) models to classify the driver’s eyes as either open or closed, and consequently their corresponding situation, either awake or drowsy. Subsequently, we tested and compared the performance of several CNN models, such as InceptionV3, MobileNetV2, NASNetMobile, and ResNet50V2. We demonstrated the performance of VAS-3D through simulations on real drowsiness datasets and experiments on real world scenarios based on real video streaming. The obtained results show that VAS-3D can enhance the accuracy detection of drowsy drivers by at least 7.5% (the best accuracy reached was 95.5%) and the detection speed by up to 57% (average of 0.25 ms per frame) compared to other existing models. Full article
Show Figures

Figure 1

Figure 1
<p>VAS-3D architecture.</p>
Full article ">Figure 2
<p>MRL Eye Dataset screenshot.</p>
Full article ">Figure 3
<p>InceptionV3 architecture adapted in our system.</p>
Full article ">Figure 4
<p>MobileNetV2 architecture adapted in our system.</p>
Full article ">Figure 5
<p>NASNetMobile architecture adapted in our system.</p>
Full article ">Figure 6
<p>ResNet50V2 architecture adapted in VAS-3D.</p>
Full article ">Figure 7
<p>Samples of visual driver behavior detection using HCC.</p>
Full article ">Figure 8
<p>Performance evaluation regarding various scenarios: VAS-3D vs. state-of-the-art. Model1 and Model2 refer to those proposed in [<a href="#B45-wevj-15-00540" class="html-bibr">45</a>] and [<a href="#B46-wevj-15-00540" class="html-bibr">46</a>] respectively.</p>
Full article ">
24 pages, 3029 KiB  
Article
Class and Ethno-Gender Differences in Education and Labour Market Position—An Intersectional Analysis of Ethnic Integration in the UK
by Yaojun Li
Societies 2024, 14(11), 222; https://doi.org/10.3390/soc14110222 - 28 Oct 2024
Viewed by 531
Abstract
This paper analyses the socio-economic disadvantages of women from different ethnic minority heritages in the UK. Using data from the Labour Force Survey (2014–2023), which contains detailed information on parental class and respondents’ socio-economic conditions, we examine four domains of life chances which [...] Read more.
This paper analyses the socio-economic disadvantages of women from different ethnic minority heritages in the UK. Using data from the Labour Force Survey (2014–2023), which contains detailed information on parental class and respondents’ socio-economic conditions, we examine four domains of life chances which are crucial for ethnic integration: educational attainment at the degree level, risks of unemployment, access to professional-managerial (salariat) position and earning power. We proceeded with the gross differences and then examined the differences by ethno-gender status and parental class combinations, controlling for many confounding factors. We also examined the net ethno-gender differences over the life course and the trends of social fluidity over the period covered and across the ethno-gender groups. We found that women from all ethnic origins were doing well in education but faced multiple disadvantages in the labour market, especially in access to the salariat and in earning power. Women of Pakistani/Bangladeshi heritages faced pronounced unemployment risks, especially at the earlier life stages. There is a significant increase in fluidity over the period covered, but this is marked by considerable ethnic and class differences, with Black Caribbean, Black African, Pakistani and Bangladeshi women from more advantaged class origins being unable to secure advantaged class positions and those from working-class families unable to make long-range upward mobility as effectively as White men. Overall, Bangladeshi, Pakistani, Black African and Black Caribbean women are found to be considerably disadvantaged, but there are also signs of social progress. Full article
(This article belongs to the Special Issue Gender and Class: Exploring the Intersections of Power and Inequality)
Show Figures

Figure 1

Figure 1
<p>Intersectional analysis of access to salariat by ethno-gender groups from higher salariat and routine origins. Note: controlling for age, age squared, nativity, education, disability and year of survey.</p>
Full article ">Figure 2
<p>Intersectional analysis of earning power of ethno-gender groups in Class 1 from higher salariat and routine families. Note: controlling for age, age squared, nativity, education, disability, region and year of survey.</p>
Full article ">Figure 3
<p>Degree attainment by ethno-gender groups over the life course. Note: controlling for age, age squared, parental class, nativity, education, disability and year of survey.</p>
Full article ">Figure 4
<p>Risks of unemployment by ethno-gender groups over the life course. Note: controlling for age, age squared, parental class, education, nativity, disability and year of survey.</p>
Full article ">Figure 5
<p>Access to the salariat by ethno-gender groups over the life course. Note: controlling for age, age squared, parental class, nativity, education, disability and year of survey.</p>
Full article ">Figure 6
<p>Annual earning by ethno-gender groups over the life course. Note: controlling for age, age squared, parental class, nativity, education, class, region, caring responsibility, marital status, disability, working hours and year of survey.</p>
Full article ">Figure 7
<p>UNIDIFF parameter estimates and 95% confidence intervals of the origins–destinations associations over time (2014–2023).</p>
Full article ">Figure 8
<p>UNIDIFF parameter estimates and 95% confidence intervals of the origins–destinations associations between ethno-gender groups.</p>
Full article ">
19 pages, 5336 KiB  
Article
Enhancing Situational Awareness with VAS-Compass Net for the Recognition of Directional Vehicle Alert Sounds
by Chiun-Li Chin, Jun-Ren Chen, Wan-Xuan Lin, Hsuan-Chiao Hung, Shang-En Chiang, Chih-Hui Wang, Liang-Ching Lee and Shing-Hong Liu
Sensors 2024, 24(21), 6841; https://doi.org/10.3390/s24216841 - 24 Oct 2024
Viewed by 637
Abstract
People with hearing impairments often face increased risks related to traffic accidents due to their reduced ability to perceive surrounding sounds. Given the cost and usage limitations of traditional hearing aids and cochlear implants, this study aims to develop a sound alert assistance [...] Read more.
People with hearing impairments often face increased risks related to traffic accidents due to their reduced ability to perceive surrounding sounds. Given the cost and usage limitations of traditional hearing aids and cochlear implants, this study aims to develop a sound alert assistance system (SAAS) to enhance situational awareness and improve travel safety for people with hearing impairments. We proposed the VAS-Compass Net (Vehicle Alert Sound–Compass Net), which integrates three lightweight convolutional neural networks: EfficientNet-lite0, MobileNetV3-Small, and GhostNet. Through employing a fuzzy ranking ensemble technique, our proposed model can identify different categories of vehicle alert sounds and directions of sound sources on an edge computing device. The experimental dataset consisted of images derived from the sounds of approaching police cars, ambulances, fire trucks, and car horns from various directions. The audio signals were converted into spectrogram images and Mel-frequency cepstral coefficient images, and they were fused into a complete image using image stitching techniques. We successfully deployed our proposed model on a Raspberry Pi 5 microcomputer, paired with a customized smartwatch to realize an SAAS. Our experimental results demonstrated that VAS-Compass Net achieved an accuracy of 84.38% based on server-based computing and an accuracy of 83.01% based on edge computing. Our proposed SAAS has the potential to significantly enhance the situational awareness, alertness, and safety of people with hearing impairments on the road. Full article
(This article belongs to the Special Issue Wearable Robotics and Assistive Devices)
Show Figures

Figure 1

Figure 1
<p>The architecture and the operational process of the proposed SAAS.</p>
Full article ">Figure 2
<p>Audio signals are converted to image samples in four steps, including audio data collection, preprocessing and augmentation, image formation, and image stitching. The x-axis of the MFCC images represents the time sequence, and the y-axis represents the 13 MFCCs. The x-axis of the spectrogram images represents the time sequence, and the y-axis represents the spectrum. For a detailed explanation, please refer to <a href="#sec3dot4-sensors-24-06841" class="html-sec">Section 3.4</a>.</p>
Full article ">Figure 3
<p>Clarity scores of audio signals at different augmented factors.</p>
Full article ">Figure 4
<p>The stitched images include three spectrogram images and three MFCC images, respectively. In each image, the x-axis represents the time sequence, while the y-axis represents the spectrum or spectrum. For a detailed explanation, please refer to <a href="#sec3dot4-sensors-24-06841" class="html-sec">Section 3.4</a>.</p>
Full article ">Figure 5
<p>The architecture of VAS-Compass Net.</p>
Full article ">Figure 6
<p>The circuit diagram of the alert output device in the SAAS.</p>
Full article ">Figure 7
<p>The main hardware components of the SAAS and their ideal placement on the user’s body.</p>
Full article ">Figure 8
<p>The confusion matrix of VAS-Compass Net in the server-based environment.</p>
Full article ">Figure 9
<p>The confusion matrix of VAS-Compass Net in the edge computing device.</p>
Full article ">Figure 10
<p>Spectrogram images of a police car siren collected by the device (<b>a</b>) on the left side, (<b>b</b>) at the rear, and (<b>c</b>) on the right side. In each spectrogram image, the x-axis represents the time sequence and the y-axis represents the spectrum.</p>
Full article ">Figure 11
<p>MFCC images of a police car siren collected by the device (<b>a</b>) on the left side, (<b>b</b>) at the rear, and (<b>c</b>) on the right side. In each MFCC image, the x-axis represents the time sequence and the y-axis represents the spectrum.</p>
Full article ">Figure 12
<p>Different stitching images for spectrogram images and MFCC images: (<b>a</b>) Spectral_H, (<b>b</b>) MFCC_H, (<b>c</b>) Spectral_V, and (<b>d</b>) MFCC_V.</p>
Full article ">Figure 13
<p>The SAAS indicates a police siren approaching the user from the left side.</p>
Full article ">
17 pages, 5108 KiB  
Article
A Computer Vision Model for Seaweed Foreign Object Detection Using Deep Learning
by Xiang Zhang, Omar Alhendi, Siti Hafizah Ab Hamid, Nurul Japar and Adibi M. Nor
Sustainability 2024, 16(20), 9061; https://doi.org/10.3390/su16209061 - 19 Oct 2024
Viewed by 1093
Abstract
Seaweed foreign object detection has become crucial for food consumption and industrial use. This process not only can prevent potential health issues, but also maintain the overall marketability of seaweed production in the food industry. Traditional methods of inspecting seaweed foreign objects heavily [...] Read more.
Seaweed foreign object detection has become crucial for food consumption and industrial use. This process not only can prevent potential health issues, but also maintain the overall marketability of seaweed production in the food industry. Traditional methods of inspecting seaweed foreign objects heavily rely on human judgment, which deals with large volumes with diverse impurities and can be inconsistent and inefficient. An automation system for real-time seaweed foreign object detection in the inspection process should be adopted. However, automated seaweed foreign object detection has several challenges due to its dependency on visual input inspection, such as an uneven surface and undistinguishable impurities. In fact, limited access to advanced technologies and high-cost equipment would also influence visual input acquisition, thereby hindering the advancement of seaweed foreign object detection in this field. Therefore, we introduce a computer vision model utilizing a deep learning-based algorithm to detect seaweed impurities and classify the samples into ‘clean’ and ‘unclean’ categories. In this study, we managed to identify six types of seaweed impurities including sand sticks, shells, discolored seaweed, grass, worm shells, and mixed impurities. We collected 1204 images and our model’s performance was thoroughly evaluated based on comparisons with three pre-trained models, i.e., Yolov8, ResNet, and MobileNet. Our experiment shows that Yolov8 outperforms the other two models with an accuracy of 98.86%. This study also included the development of an Android application to validate the deep learning engine to ensure its optimal performance. Based on our experiments, the mobile application managed to classify 50 pieces of seaweed samples within 0.2 s each, showcasing its potential use in large-scale production lines and factories. This research demonstrates the impact of Artificial Intelligence on food safety by offering a scalable and efficient solution that can be deployed in other food production processes facing similar challenges. Our approach paves the way for broader industry adoption and advancements in automated foreign object detection systems by optimizing detection accuracy and speed. Full article
Show Figures

Figure 1

Figure 1
<p>Malaysian Red Seaweed.</p>
Full article ">Figure 2
<p>Seaweed Production [<a href="#B13-sustainability-16-09061" class="html-bibr">13</a>].</p>
Full article ">Figure 3
<p>Seaweed foreign objects collected from a local factory. (<b>a</b>) Sand stick attached to the seaweed, (<b>b</b>) shells, (<b>c</b>) discolored seaweed, (<b>d</b>) grass, (<b>e</b>) the worm shells of the Vermetidae family, and (<b>f</b>) mixed impurities.</p>
Full article ">Figure 4
<p>Development of the Seaweed Foreign Object Detection Framework.</p>
Full article ">Figure 5
<p>Data Collection Process.</p>
Full article ">Figure 6
<p>Data Collection Environment. (<b>a</b>) Top-down View, (<b>b</b>) 45-dgree Angle View.</p>
Full article ">Figure 7
<p>Impurity Percentage. (<b>a</b>) clean, (<b>b</b>) impurity amount less than 2%, (<b>c</b>) impurity amount around 3%, (<b>d</b>) impurity amount around 5%, (<b>e</b>) impurity amount more than 10%, (<b>f</b>) impurity amount more than 10%, (<b>g</b>) impurity amount more than 10%.</p>
Full article ">Figure 8
<p>Impurity Amount Percentage Distribution.</p>
Full article ">Figure 9
<p>Model Comparison Result.</p>
Full article ">Figure 10
<p>Model Prediction Result.</p>
Full article ">Figure 11
<p>Yolov8 Model Result.</p>
Full article ">Figure 12
<p>Evolution of Mobile Application Architecture in Factory Implementation.</p>
Full article ">Figure 13
<p>Mobile-based Computer Vison model for Seaweed Foreign Object Detection.</p>
Full article ">Figure 14
<p>UI of C+WeedVision Android Application. (<b>a</b>) unclean result, (<b>b</b>) clean result.</p>
Full article ">
26 pages, 4281 KiB  
Article
Secure and Transparent Lung and Colon Cancer Classification Using Blockchain and Microsoft Azure
by Entesar Hamed I. Eliwa, Amr Mohamed El Koshiry, Tarek Abd El-Hafeez and Ahmed Omar
Adv. Respir. Med. 2024, 92(5), 395-420; https://doi.org/10.3390/arm92050037 - 17 Oct 2024
Viewed by 857
Abstract
Background: The global healthcare system faces challenges in diagnosing and managing lung and colon cancers, which are significant health burdens. Traditional diagnostic methods are inefficient and prone to errors, while data privacy and security concerns persist. Objective: This study aims to develop a [...] Read more.
Background: The global healthcare system faces challenges in diagnosing and managing lung and colon cancers, which are significant health burdens. Traditional diagnostic methods are inefficient and prone to errors, while data privacy and security concerns persist. Objective: This study aims to develop a secure and transparent framework for remote consultation and classification of lung and colon cancer, leveraging blockchain technology and Microsoft Azure cloud services. Dataset and Features: The framework utilizes the LC25000 dataset, containing 25,000 histopathological images, for training and evaluating advanced machine learning models. Key features include secure data upload, anonymization, encryption, and controlled access via blockchain and Azure services. Methods: The proposed framework integrates Microsoft Azure’s cloud services with a permissioned blockchain network. Patients upload CT scans through a mobile app, which are then preprocessed, anonymized, and stored securely in Azure Blob Storage. Blockchain smart contracts manage data access, ensuring only authorized specialists can retrieve and analyze the scans. Azure Machine Learning is used to train and deploy state-of-the-art machine learning models for cancer classification. Evaluation Metrics: The framework’s performance is evaluated using metrics such as accuracy, precision, recall, and F1-score, demonstrating the effectiveness of the integrated approach in enhancing diagnostic accuracy and data security. Results: The proposed framework achieves an impressive accuracy of 100% for lung and colon cancer classification using DenseNet, ResNet50, and MobileNet models with different split ratios (70–30, 80–20, 90–10). The F1-score and k-fold cross-validation accuracy (5-fold and 10-fold) also demonstrate exceptional performance, with values exceeding 99.9%. Real-time notifications and secure remote consultations enhance the efficiency and transparency of the diagnostic process, contributing to better patient outcomes and streamlined cancer care management. Full article
Show Figures

Figure 1

Figure 1
<p>Framework for Remote Lung and Colon CT Scan Consultations.</p>
Full article ">Figure 2
<p>Framework for Classifying Lung and Colon Cancers on the Azure Cloud Platform.</p>
Full article ">Figure 3
<p>A sample from the dataset with the corresponding label.</p>
Full article ">Figure 4
<p>Framework for Lung and Colon Cancer classification using the pre-trained model.</p>
Full article ">Figure 5
<p>Pseudocode Steps for Lung and Colon Cancer classification using the pre-trained model.</p>
Full article ">Figure 6
<p>Accuracy and Loss plot for InceptionV3.</p>
Full article ">Figure 7
<p>The Confusion Matrix for InceptionV3.</p>
Full article ">Figure 8
<p>Accuracy and Loss plot for ResNet50.</p>
Full article ">Figure 9
<p>The Confusion Matrix for ResNet50.</p>
Full article ">Figure 10
<p>Accuracy and Loss plot for DenseNet.</p>
Full article ">Figure 11
<p>The Confusion Matrix for DenseNet.</p>
Full article ">Figure 12
<p>Accuracy and Loss plot for MobileNet.</p>
Full article ">Figure 13
<p>The Confusion Matrix for MobileNet.</p>
Full article ">Figure 14
<p>Training times for the four models for 80–20 split ratio.</p>
Full article ">Figure 15
<p>Testing times for the four models for 80–20 split ratio.</p>
Full article ">Figure 16
<p>Accuracy for the four models for 80-20 split ratio.</p>
Full article ">
15 pages, 879 KiB  
Entry
Synthetic Fuels for Decarbonising UK Rural Transport
by Al-Amin Abba Dabo, Andrew Gough and F. Frank Alparslan
Encyclopedia 2024, 4(4), 1553-1567; https://doi.org/10.3390/encyclopedia4040101 - 15 Oct 2024
Viewed by 717
Definition
Decarbonising transport is a crucial element of the UK’s strategy to achieve net-zero carbon emissions by 2050, as the transport sector is currently the largest contributor to the UK’s greenhouse gas emissions. Rural communities face distinct challenges in this effort due to their [...] Read more.
Decarbonising transport is a crucial element of the UK’s strategy to achieve net-zero carbon emissions by 2050, as the transport sector is currently the largest contributor to the UK’s greenhouse gas emissions. Rural communities face distinct challenges in this effort due to their reliance on internal combustion engines (ICEs) across vehicles and machinery essential for daily life, including farming equipment and private transport. While the upcoming ban on new petrol and diesel vehicles paves the way for the adoption of Electric Vehicles (EVs), this solution may not fully address the unique needs of rural areas where infrastructure limitations and specific mobility requirements pose significant barriers. In this context, synthetic fuels, produced using renewable energy sources, offer a potential alternative. These fuels can be used directly in existing internal combustion engines without requiring major modifications and have the added benefit of reducing overall greenhouse gas emissions by capturing CO2 during production. This entry explores the potential advantages of adopting synthetic fuels, particularly in rural areas, and examines how community-based buying cooperatives could support their wider use through bulk purchasing, cost reduction, and community empowerment. Full article
(This article belongs to the Section Social Sciences)
Show Figures

Figure 1

Figure 1
<p>Net CO<sub>2</sub> impact (gramme CO<sub>2</sub> per gramme of fuel) of synthetic fuel production methods: Sabatier, biomass pyrolysis, and heavy oil upgrading (source: [<a href="#B35-encyclopedia-04-00101" class="html-bibr">35</a>,<a href="#B36-encyclopedia-04-00101" class="html-bibr">36</a>,<a href="#B37-encyclopedia-04-00101" class="html-bibr">37</a>,<a href="#B38-encyclopedia-04-00101" class="html-bibr">38</a>]).</p>
Full article ">Figure 2
<p>Applications of Synthetic Fuels in Rural Transport.</p>
Full article ">
20 pages, 14310 KiB  
Article
Deep Learning Application for Biodiversity Conservation and Educational Tourism in Natural Reserves
by Marco Flórez, Oscar Becerra, Eduardo Carrillo, Manny Villa, Yuli Álvarez, Javier Suárez and Francisco Mendes
ISPRS Int. J. Geo-Inf. 2024, 13(10), 358; https://doi.org/10.3390/ijgi13100358 - 11 Oct 2024
Viewed by 888
Abstract
Natural reserves, such as the Santurbán Moor in Colombia, are ecologically important but face significant threats from activities like mining and agriculture. Preserving biodiversity in these ecosystems is essential for maintaining ecological balance and promoting sustainable tourism practices. Identifying plant species in these [...] Read more.
Natural reserves, such as the Santurbán Moor in Colombia, are ecologically important but face significant threats from activities like mining and agriculture. Preserving biodiversity in these ecosystems is essential for maintaining ecological balance and promoting sustainable tourism practices. Identifying plant species in these reserves accurately is challenging due to environmental variability and species similarities, complicating conservation efforts and educational tourism promotion. This study aims to create and assess a mobile application based on deep learning, called FloraBan, to autonomously identify plant species in natural reserves, enhancing biodiversity conservation and encouraging sustainable and educational tourism practices. The application employs the EfficientNet Lite4 model, trained on a comprehensive dataset of plant images taken in various field conditions. Designed to work offline, the application is particularly useful in remote areas. The model evaluation revealed an accuracy exceeding 90% in classifying plant images. FloraBan was effective under various lighting conditions and complex backgrounds, offering detailed information about each species, including scientific name, family, and conservation status. The ability to function without internet connectivity is a significant benefit, especially in isolated regions like natural reserves. FloraBan represents a notable improvement in the field of automated plant identification, supporting botanical research and efforts to preserve biodiversity in the Santurbán Moor. Additionally, it encourages educational and responsible tourism practices, which align with sustainability goals, providing a useful tool for both tourists and conservationists. Full article
Show Figures

Figure 1

Figure 1
<p>Location of Santurbán moor in Colombia, reference image.</p>
Full article ">Figure 2
<p>Diversity of selected plant species in the Santurbán paramo.</p>
Full article ">Figure 3
<p>Block diagram of the proposed process.</p>
Full article ">Figure 4
<p>Performance curves of the EfficientNet Lite4 model in training and validation.</p>
Full article ">Figure 5
<p>Confusion matrix of the EfficientNet Lite4 model.</p>
Full article ">Figure 6
<p>Plant detection and classification process in the FloraBan application.</p>
Full article ">Figure 7
<p>Screenshots of the FloraBan application interface. (<b>a</b>) Application home screen and image capture screen. (<b>b</b>) Image analysis result screen and detailed information section.</p>
Full article ">Figure 7 Cont.
<p>Screenshots of the FloraBan application interface. (<b>a</b>) Application home screen and image capture screen. (<b>b</b>) Image analysis result screen and detailed information section.</p>
Full article ">
21 pages, 2480 KiB  
Article
Model of Sustainable Household Mobility in Multi-Modal Transportation Networks
by Igor Kabashkin
Sustainability 2024, 16(17), 7802; https://doi.org/10.3390/su16177802 - 7 Sep 2024
Viewed by 942
Abstract
Nowadays, urban and suburban areas face increasing environmental pressures, and encouraging sustainable transportation behaviors at the household level has become crucial. This paper presents a model of a decision support system (DSS) for promoting sustainable household mobility choices in multi-modal transport networks. The [...] Read more.
Nowadays, urban and suburban areas face increasing environmental pressures, and encouraging sustainable transportation behaviors at the household level has become crucial. This paper presents a model of a decision support system (DSS) for promoting sustainable household mobility choices in multi-modal transport networks. The system was modeled using an enhanced Petri Net approach, allowing for the dynamic representation of complex transport networks and multi-modal journey options. The model incorporated various sustainability factors. These were combined into a single environmental impact score, which was considered alongside travel time and cost in the route optimization process. Simulation results demonstrated the DSS’s capability to guide users toward more sustainable mobility choices. The model also showed potential as a tool for policymakers to assess the impact of various sustainable transportation initiatives and infrastructure investments. This paper discussed the versatile applications of the system. It also addressed the limitations of Petri Net models in transportation systems and suggested future research directions. Full article
(This article belongs to the Collection Sustainable Household Behaviors: Consumption and Mobility)
Show Figures

Figure 1

Figure 1
<p>Alternative multi-modal travel routes from origin to destination in household mobility system.</p>
Full article ">Figure 2
<p>Decision-making process for route selection in household mobility system.</p>
Full article ">Figure 3
<p>Set of basic components for the PN model: simple transition (<b>a</b>); multiplier (<b>b</b>); integrator (<b>c</b>); generator (<b>d</b>); absorber (<b>e</b>); and generator of schedules (<b>f</b>).</p>
Full article ">Figure 4
<p>The graph of the transportation network.</p>
Full article ">Figure 5
<p>PN model on the base of the transportation graph.</p>
Full article ">Figure 6
<p>Normalized values of mobility alternatives for different analysis criteria.</p>
Full article ">
17 pages, 4715 KiB  
Article
IoT-MFaceNet: Internet-of-Things-Based Face Recognition Using MobileNetV2 and FaceNet Deep-Learning Implementations on a Raspberry Pi-400
by Ahmad Saeed Mohammad, Thoalfeqar G. Jarullah, Musab T. S. Al-Kaltakchi, Jabir Alshehabi Al-Ani and Somdip Dey
J. Low Power Electron. Appl. 2024, 14(3), 46; https://doi.org/10.3390/jlpea14030046 - 5 Sep 2024
Viewed by 50969
Abstract
IoT applications revolutionize industries by enhancing operations, enabling data-driven decisions, and fostering innovation. This study explores the growing potential of IoT-based facial recognition for mobile devices, a technology rapidly advancing within the interconnected IoT landscape. The investigation proposes a framework called IoT-MFaceNet (Internet-of-Things-based [...] Read more.
IoT applications revolutionize industries by enhancing operations, enabling data-driven decisions, and fostering innovation. This study explores the growing potential of IoT-based facial recognition for mobile devices, a technology rapidly advancing within the interconnected IoT landscape. The investigation proposes a framework called IoT-MFaceNet (Internet-of-Things-based face recognition using MobileNetV2 and FaceNet deep-learning) utilizing pre-existing deep-learning methods, employing the MobileNetV2 and FaceNet algorithms on both ImageNet and FaceNet databases. Additionally, an in-house database is compiled, capturing data from 50 individuals via a web camera and 10 subjects through a smartphone camera. Pre-processing of the in-house database involves face detection using OpenCV’s Haar Cascade, Dlib’s CNN Face Detector, and Mediapipe’s Face. The resulting system demonstrates high accuracy in real-time and operates efficiently on low-powered devices like the Raspberry Pi 400. The evaluation involves the use of the multilayer perceptron (MLP) and support vector machine (SVM) classifiers. The system primarily functions as a closed set identification system within a computer engineering department at the College of Engineering, Mustansiriyah University, Iraq, allowing access exclusively to department staff for the department rapporteur room. The proposed system undergoes successful testing, achieving a maximum accuracy rate of 99.976%. Full article
Show Figures

Figure 1

Figure 1
<p>Proposed Framework (IoT-MFaceNet): where (<b>A</b>) represents the image database, (<b>B</b>) represents the labeled faces in the wild (LFW) database, (<b>C</b>) represents the in-house database with 50 subjects for the web camera, with 10 subjects, (<b>D</b>) represents the in-house database with 10 subjects for a smartphone camera, (<b>E</b>) represents pre-processing, (<b>F</b>) represents pre-trained deep learning, including (<b>F1</b>) the MobileNetV2 technique (<b>F2</b>) and the FaceNet technique, (<b>G</b>) represents the evaluation process represented by the Raspberry Pi type 400, (<b>H</b>) represents classification, and (<b>I</b>) represents the identified person displayed by exploiting the mobile application.</p>
Full article ">Figure 2
<p>The architecture of the MobileNetV2 technique (<b>a</b>) and the architecture of VGG16 and VGG19 of the FaceNet technique (<b>b</b>).</p>
Full article ">Figure 3
<p>Performance of the different models in FPS, where the FPS results are the average FPS of 100 recorded frames. Three different custom-made tops were used to test the proposed performance using the custom-trained CNN with MobileNetV2. The first top has three hidden layers with 384 neurons each, making it the largest custom-made top for MobileNetV2 in this example (384-384-384). The second top, with a configuration of 192-256-128, is the second largest custom-made top. The third top, with a configuration of 128-128-128, is the smallest custom-made top. Additionally, TensorFlow-lite quantization was applied to all three networks.</p>
Full article ">Figure 4
<p>Size of the models created on disk in MB.</p>
Full article ">Figure 5
<p>The hardware design. (<b>A</b>) Example of the hardware connection. The hardware system combines an ultrasonic sensor and webcam to measure and monitor subject distance, capturing images for face recognition. Optimal recognition occurs at distances below 80 cm, with lower confidence beyond that, signaled on the GUI. (<b>B</b>) An instance of the system operating with two subjects is evident. The system accurately identifies the faces of Thoalfeqar and Humam. Both serve as representatives of the fourth class in the Department of Computer Engineering at Mustansiriyah University. Typically, class representatives, including Thoalfeqar and Humam, have access to the department’s rapporteur room along with the department’s staff.</p>
Full article ">Figure 6
<p>The receiver operating characteristic (ROC) without optimization for a custom-trained CNN using MobileNetV2 with an MLP classifier for (<b>a</b>) (128-128-128) custom-made top, (<b>b</b>) (192-256-128) custom-made top, and (<b>c</b>) (384-384-384) custom-made top.</p>
Full article ">Figure 7
<p>The receiver operating characteristic (ROC) with optimization for a custom-trained CNN using MobileNetV2 with an MLP classifier (<b>a</b>) (128-128-128) custom-made top, (<b>b</b>) (192-256-128) custom-made top, and (<b>c</b>) (384-384-384) custom-made top.</p>
Full article ">Figure 8
<p>The receiver operating characteristic (ROC) for the FaceNet with the SVM classifier.</p>
Full article ">
15 pages, 393 KiB  
Article
Facial Age Estimation Using Multi-Stage Deep Neural Networks
by Salah Eddine Bekhouche, Azeddine Benlamoudi, Fadi Dornaika, Hichem Telli and Yazid Bounab
Electronics 2024, 13(16), 3259; https://doi.org/10.3390/electronics13163259 - 16 Aug 2024
Viewed by 1192
Abstract
Over the last decade, the world has witnessed many breakthroughs in artificial intelligence, largely due to advances in deep learning technology. Notably, computer vision solutions have significantly contributed to these achievements. Human face analysis, a core area of computer vision, has gained considerable [...] Read more.
Over the last decade, the world has witnessed many breakthroughs in artificial intelligence, largely due to advances in deep learning technology. Notably, computer vision solutions have significantly contributed to these achievements. Human face analysis, a core area of computer vision, has gained considerable attention due to its wide applicability in fields such as law enforcement, social media, and marketing. However, existing methods for facial age estimation often struggle with accuracy due to limited feature extraction capabilities and inefficiencies in learning hierarchical representations. This paper introduces a novel framework to address these issues by proposing a Multi-Stage Deep Neural Network (MSDNN) architecture. The MSDNN architecture divides each CNN backbone into multiple stages, enabling more comprehensive feature extraction, thereby improving the accuracy of age predictions from facial images. Our framework demonstrates a significant performance improvement over traditional solutions, with its effectiveness validated through comparisons with the EfficientNet and MobileNetV3 architectures. The proposed MSDNN architecture achieves a notable decrease in Mean Absolute Error (MAE) across three widely used public datasets (MORPH2, CACD, and AFAD) while maintaining a virtually identical parameter count compared to the initial backbone architectures. These results underscore the effectiveness and feasibility of our methodology in advancing the field of age estimation, showcasing it as a robust solution for enhancing the accuracy of age prediction algorithms. Full article
Show Figures

Figure 1

Figure 1
<p>General structure of a multistage deep neural network based on a backbone with five stages.</p>
Full article ">Figure 2
<p>Cumulative scores obtained by the proposed approach on the three public datasets.</p>
Full article ">
22 pages, 9193 KiB  
Article
RS-Xception: A Lightweight Network for Facial Expression Recognition
by Liefa Liao, Shouluan Wu, Chao Song and Jianglong Fu
Electronics 2024, 13(16), 3217; https://doi.org/10.3390/electronics13163217 - 14 Aug 2024
Viewed by 925
Abstract
Facial expression recognition (FER) utilizes artificial intelligence for the detection and analysis of human faces, with significant applications across various scenarios. Our objective is to deploy the facial emotion recognition network on mobile devices and extend its application to diverse areas, including classroom [...] Read more.
Facial expression recognition (FER) utilizes artificial intelligence for the detection and analysis of human faces, with significant applications across various scenarios. Our objective is to deploy the facial emotion recognition network on mobile devices and extend its application to diverse areas, including classroom effect monitoring, human–computer interaction, specialized training for athletes (such as in figure skating and rhythmic gymnastics), and actor emotion training. Recent studies have employed advanced deep learning models to address this task, though these models often encounter challenges like subpar performance and an excessive number of parameters that do not align with the requirements of FER for embedded devices. To tackle this issue, we have devised a lightweight network structure named RS-Xception, which is straightforward yet highly effective. Drawing on the strengths of ResNet and SENet, this network integrates elements from the Xception architecture. Our models have been trained on FER2013 datasets and demonstrate superior efficiency compared to conventional network models. Furthermore, we have assessed the model’s performance on the CK+, FER2013, and Bigfer2013 datasets, achieving accuracy rates of 97.13%, 69.02%, and 72.06%, respectively. Evaluation on the complex RAF-DB dataset yielded an accuracy rate of 82.98%. The incorporation of transfer learning notably enhanced the model’s accuracy, with a performance of 75.38% on the Bigfer2013 dataset, underscoring its significance in our research. In conclusion, our proposed model proves to be a viable solution for precise sentiment detection and estimation. In the future, our lightweight model may be deployed on embedded devices for research purposes. Full article
(This article belongs to the Special Issue Advances in Computer Vision and Deep Learning and Its Applications)
Show Figures

Figure 1

Figure 1
<p>Depthwise Separable Convolution.</p>
Full article ">Figure 2
<p>Squeeze and Excitation block.</p>
Full article ">Figure 3
<p>Model structure.</p>
Full article ">Figure 4
<p>CK+ image example.</p>
Full article ">Figure 5
<p>Comparison of Fer 2013 and Bigfer 2013 expression data.</p>
Full article ">Figure 6
<p>(<b>a</b>) Training precision graph of CK+ (blue line), FER2013 (scarlet line), Bigfer2013 (green line), and RAF-DB datasets (brown line). (<b>b</b>) Training loss function values of CK+, FER2013, Bigfer2013, and RAF-DB datasets. (<b>c</b>) Validation accuracy graphs of CK+, FER2013, Bigfer2013, and RAF-DB datasets. (<b>d</b>) Validation loss function values of CK+, FER2013, Bigfer2013, and RAF-DB datasets.</p>
Full article ">Figure 6 Cont.
<p>(<b>a</b>) Training precision graph of CK+ (blue line), FER2013 (scarlet line), Bigfer2013 (green line), and RAF-DB datasets (brown line). (<b>b</b>) Training loss function values of CK+, FER2013, Bigfer2013, and RAF-DB datasets. (<b>c</b>) Validation accuracy graphs of CK+, FER2013, Bigfer2013, and RAF-DB datasets. (<b>d</b>) Validation loss function values of CK+, FER2013, Bigfer2013, and RAF-DB datasets.</p>
Full article ">Figure 7
<p>Confusion matrix (<b>a</b>) and ROC curve (<b>b</b>) of CK+ data (class 0–7 represents neutral, anger, contempt, disgust, fear, happiness, sadness, and surprise).</p>
Full article ">Figure 8
<p>Confusion matrix (<b>a</b>) and ROC curve (<b>b</b>) of FER2013 dataset (class 0–6 represents angry, disgusted, scared, happy, sad, surprised, and neutral).</p>
Full article ">Figure 9
<p>Confusion matrix (<b>a</b>) and ROC curve (<b>b</b>) of Bigfer2013 dataset (class 0–6 represents angry, disgusted, scared, happy, sad, surprised, and neutral).</p>
Full article ">Figure 10
<p>The comparison of accuracy between the proposed model and existing models.</p>
Full article ">Figure 11
<p>The process of model transfer learning (ResNet50, ResNet101, MobileNet, ResNet18, DenSENet 121, SENet 18).</p>
Full article ">Figure 12
<p>(<b>a</b>) The comparison of the training accuracy of RS-Xception on the FER2013 dataset and the transfer learning training accuracy of the other 6 models on the FER2013 dataset. (<b>b</b>) The comparison of the training validation accuracy of RS-Xception on the FER2013 dataset and the transfer learning validation accuracy of the other 6 models on the FER2013 dataset.</p>
Full article ">Figure 13
<p>(<b>a</b>) The confusion matrix of the model after Bigfer2013 transfer learning. (<b>b</b>) The ROC curve of the model after Bigfer2013 transfer learning (class 0-6 represents angry, disgusted, scared, happy, sad, surprised, and neutral).</p>
Full article ">Figure 14
<p>(<b>a</b>) The comparison of the validation accuracy of the model on the BigFer2013 dataset and transfer learning on the Bigfer2013 dataset, and (<b>b</b>) comparison of the loss function value of the model on the BigFer2013 dataset with the value of the loss function of transfer learning on Bigfer2013.</p>
Full article ">Figure 15
<p>Results of expression classification using the proposed model.</p>
Full article ">
23 pages, 6850 KiB  
Article
PlanteSaine: An Artificial Intelligent Empowered Mobile Application for Pests and Disease Management for Maize, Tomato, and Onion Farmers in Burkina Faso
by Obed Appiah, Kwame Oppong Hackman, Belko Abdoul Aziz Diallo, Kehinde O. Ogunjobi, Son Diakalia, Ouedraogo Valentin, Damoue Abdoul-Karim and Gaston Dabire
Agriculture 2024, 14(8), 1252; https://doi.org/10.3390/agriculture14081252 - 30 Jul 2024
Cited by 1 | Viewed by 1059
Abstract
This study presents PlanteSaine, a novel mobile application powered by Artificial Intelligence (AI) models explicitly designed for maize, tomato, and onion farmers in Burkina Faso. Agriculture in Burkina Faso, like many developing nations, faces substantial challenges from plant pests and diseases, posing threats [...] Read more.
This study presents PlanteSaine, a novel mobile application powered by Artificial Intelligence (AI) models explicitly designed for maize, tomato, and onion farmers in Burkina Faso. Agriculture in Burkina Faso, like many developing nations, faces substantial challenges from plant pests and diseases, posing threats to both food security and economic stability. PlanteSaine addresses these challenges by offering a comprehensive solution that provides farmers with real-time identification of pests and diseases. Farmers capture images of affected plants with their smartphones, and PlanteSaine’s AI system analyzes these images to provide accurate diagnoses. The application’s offline functionality ensures accessibility even in remote areas with limited Internet connectivity, while its messaging feature facilitates communication with agricultural authorities for guidance and support. Additionally, PlanteSaine includes an emergency alert mechanism to notify farmers about pest and disease outbreaks, enhancing their preparedness to deal with these threats. An AI-driven framework, featuring an image feature extraction phase with EfficientNetB3 and an artificial neural network (ANN) classifier, was developed and integrated into PlanteSaine. The evaluation of PlanteSaine demonstrates its superior performance compared to baseline models, showcasing its effectiveness in accurately detecting diseases and pests across maize, tomato, and onion crops. Overall, this study highlights the potential of PlanteSaine to revolutionize agricultural technology in Burkina Faso and beyond. Leveraging AI and mobile computing, PlanteSaine provides farmers with accessible and reliable pest and disease management tools, ultimately contributing to sustainable farming practices and enhancing food security. The success of PlanteSaine underscores the importance of interdisciplinary approaches in addressing pressing challenges in global agriculture Full article
(This article belongs to the Section Crop Protection, Diseases, Pests and Weeds)
Show Figures

Figure 1

Figure 1
<p>Sample images from the maize dataset ((<b>a</b>) chenille légionnaire_a, (<b>b</b>) chenille légionnaire_p, (<b>c</b>) curvulariose, (<b>d</b>) feuille saine, (<b>e</b>) helminthosporiose, (<b>f</b>) maladie abiotique, (<b>g</b>) pucerons, (<b>h</b>) rouille, (<b>i</b>) striure).</p>
Full article ">Figure 2
<p>Sample images from the tomato dataset ((<b>a</b>) acarien, (<b>b</b>) alternariose acarien, (<b>c</b>) alternariose, (<b>d</b>) coup du soleil, (<b>e</b>) exces d’azote, (<b>f</b>) feuille saine, (<b>g</b>) fusariose, (<b>h</b>) helicoverpa armigera, (<b>i</b>) flétrissement bactérien, (<b>j</b>) fruit sain, (<b>k</b>) mildiou de la tomate, (<b>l</b>) pourriture apicale, (<b>m</b>) tuta absoluta, (<b>n</b>) virose).</p>
Full article ">Figure 3
<p>Sample images from the Onion dataset ((<b>a</b>) alternariose, (<b>b</b>) chenilles, (<b>c</b>) feuille saine, (<b>d</b>) fusariose, (<b>e</b>) pourriture des bulbes, (<b>f</b>) virose).</p>
Full article ">Figure 4
<p>Baseline Architecture (EfficientNetB3).</p>
Full article ">Figure 5
<p>Proposed Framework.</p>
Full article ">Figure 6
<p>Training and Validation Accuracy for Maize Classifier.</p>
Full article ">Figure 7
<p>Testing accuracies for each class in the Maize dataset (Average accuracy = 0.96).</p>
Full article ">Figure 8
<p>Training and Validation Accuracy for Tomato Classifier.</p>
Full article ">Figure 9
<p>Testing accuracies for each class in the Tomato dataset (Average accuracy = 0.98).</p>
Full article ">Figure 10
<p>Training and Validation Accuracy for Onion Classifier.</p>
Full article ">Figure 11
<p>Testing accuracies for each class in the Onion dataset (Average accuracy = 0.96).</p>
Full article ">Figure 12
<p>Average accuracy, precision, recall, and F1 performance for all models.</p>
Full article ">Figure 13
<p>A demonstration of the two modes of submitting images for PlanteSaine to classify.</p>
Full article ">
17 pages, 4289 KiB  
Article
Image-Acceleration Multimodal Danger Detection Model on Mobile Phone for Phone Addicts
by Han Wang, Xiang Ji, Lei Jin, Yujiao Ji and Guangcheng Wang
Sensors 2024, 24(14), 4654; https://doi.org/10.3390/s24144654 - 18 Jul 2024
Viewed by 761
Abstract
With the popularity of smartphones, a large number of “phubbers” have emerged who are engrossed in their phones regardless of the situation. In response to the potential dangers that phubbers face while traveling, this paper proposes a multimodal danger perception network model and [...] Read more.
With the popularity of smartphones, a large number of “phubbers” have emerged who are engrossed in their phones regardless of the situation. In response to the potential dangers that phubbers face while traveling, this paper proposes a multimodal danger perception network model and early warning system for phubbers, designed for mobile devices. This proposed model consists of surrounding environment feature extraction, user behavior feature extraction, and multimodal feature fusion and recognition modules. The environmental feature module utilizes MobileNet as the backbone network to extract environmental description features from the rear-view image of the mobile phone. The behavior feature module uses acceleration time series as observation data, maps the acceleration observation data to a two-dimensional image space through GADFs (Gramian Angular Difference Fields), and extracts behavior description features through MobileNet, while utilizing statistical feature vectors to enhance the representation capability of behavioral features. Finally, in the recognition module, the environmental and behavioral characteristics are fused to output the type of hazardous state. Experiments indicate that the accuracy of the proposed model surpasses existing methods, and it possesses the advantages of compact model size (28.36 Mb) and fast execution speed (0.08 s), making it more suitable for deployment on mobile devices. Moreover, the developed image-acceleration multimodal phubber hazard recognition network combines the behavior of mobile phone users with surrounding environmental information, effectively identifying potential hazards for phubbers. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>Framework of multimodal phubbing danger detection system.</p>
Full article ">Figure 2
<p>Examples of the nine dangerous states defined in this paper. (<b>a</b>) walking and zebra crossing. (<b>b</b>) climbing stairs. (<b>c</b>) walking and wet surface. (<b>d</b>) walking and darkness. (<b>e</b>) climbing stairs and darkness. (<b>f</b>) static and zebra crossing. (<b>g</b>) static and stairs. (<b>h</b>) static and wet surface. (<b>i</b>) static and darkness.</p>
Full article ">Figure 3
<p>Rules for pairing environment images with sensor data. In the sensor data, the red, green, and blue lines represent the X-axis, Y-axis, and Z-axis, respectively.</p>
Full article ">Figure 4
<p>Data collection sample images. The first row shows the actual scene where the user is located, while the second and third rows depict the corresponding sensor time series and environmental real-life images. The sensor time series for the X, Y, and Z axes are represented by red, green, and blue lines, respectively. (<b>a</b>) climbing stairs. (<b>b</b>) walking and wet surface. (<b>c</b>) walking and zebra crossing. (<b>d</b>) walking and darkness.</p>
Full article ">Figure 5
<p>Multimodal phubbing danger state recognition network.</p>
Full article ">Figure 6
<p>Sensor curves and GADF pseudo-images corresponding to different behaviors. (<b>a</b>) Sample of sensor curve and X-Y-Z pseudo-image when going up and down stairs. (<b>b</b>) Sample of sensor curve and X-Y-Z pseudo-image when walking. (<b>c</b>) Sample of sensor curve and X-Y-Z pseudo-image when stationary.</p>
Full article ">Figure 7
<p>Confusion matrices of test results with GADF input as X-Z and X-Y-Z.</p>
Full article ">Figure 8
<p>Mobile application user interface.</p>
Full article ">Figure 9
<p>Examples of online test results for mobile phone app. The left side of each figure shows the real-time display of the mobile app, while the right side shows the actual photos of the user’s activity state and environment. (<b>a</b>) Walking and wet surfaces. (<b>b</b>) Walking and zebra crossing. (<b>c</b>) Going upstairs. (<b>d</b>) Going downstairs. (<b>e</b>) Static and darkness. (<b>f</b>) Sitting and browsing on the phone.</p>
Full article ">
14 pages, 2793 KiB  
Article
A MobileFaceNet-Based Face Anti-Spoofing Algorithm for Low-Quality Images
by Jianyu Xiao, Wei Wang, Lei Zhang and Huanhua Liu
Electronics 2024, 13(14), 2801; https://doi.org/10.3390/electronics13142801 - 16 Jul 2024
Viewed by 882
Abstract
The Face Anti-Spoofing (FAS) methods plays a very important role in ensuring the security of face recognition systems. The existing FAS methods perform well in short-distance scenarios, e.g., phone unlocking, face payment, etc. However, it is still challenging to improve the generalization of [...] Read more.
The Face Anti-Spoofing (FAS) methods plays a very important role in ensuring the security of face recognition systems. The existing FAS methods perform well in short-distance scenarios, e.g., phone unlocking, face payment, etc. However, it is still challenging to improve the generalization of FAS in long-distance scenarios (e.g., surveillance) due to the varying image quality. In order to address the lack of low-quality images in real scenarios, we build a Low-Quality Face Anti-Spoofing Dataset (LQFA-D) by using Hikvision’s surveillance cameras. In order to deploy the model on an edge device with limited computation, we propose a lightweight FAS network based on MobileFaceNet, in which the Coordinate Attention (CA) attention model is introduced to capture the important spatial information. Then, we propose a multi-scale FAS framework for low-quality images to explore multi-scale features, which includes three multi-scale models. The experimental results of the LQFA-D show that the Average Classification Error Rate (ACER) and detection time of the proposed method are 1.39% and 45 ms per image for the low-quality images, respectively. It demonstrates the effectiveness of the proposed method in this paper. Full article
(This article belongs to the Special Issue Applications of Machine Vision in Robotics)
Show Figures

Figure 1

Figure 1
<p>The overall architecture of LQFA-D collection system.</p>
Full article ">Figure 2
<p>The position between the camera and the subject.</p>
Full article ">Figure 3
<p>Part of the samples in LQFA-D.</p>
Full article ">Figure 4
<p>The data distribution of LQFA-D.</p>
Full article ">Figure 5
<p>The bottleneck of the proposed model.</p>
Full article ">Figure 6
<p>The framework of the proposed method.</p>
Full article ">Figure 7
<p>Detection accuracy of the FAS methods.</p>
Full article ">Figure 8
<p>The computation time of the FAS methods.</p>
Full article ">
14 pages, 1932 KiB  
Article
A Novel Voltage-Abnormal Cell Detection Method for Lithium-Ion Battery Mass Production Based on Data-Driven Model with Multi-Source Time Series Data
by Xiang Wang, Jianjun He, Fuxin Huang, Zhenjie Liu, Aibin Deng and Rihui Long
Energies 2024, 17(14), 3472; https://doi.org/10.3390/en17143472 - 15 Jul 2024
Viewed by 925
Abstract
Before leaving the factory, lithium-ion battery (LIB) cells are screened to exclude voltage-abnormal cells, which can increase the fault rate, troubleshooting difficulty, and degrade pack performance. However, the time interval to obtain the detection results through the existing voltage-abnormal cell method is too [...] Read more.
Before leaving the factory, lithium-ion battery (LIB) cells are screened to exclude voltage-abnormal cells, which can increase the fault rate, troubleshooting difficulty, and degrade pack performance. However, the time interval to obtain the detection results through the existing voltage-abnormal cell method is too long, which can seriously affect production efficiency and delay shipment, especially in the mass production of LIBs when facing a large number of time-critical orders. In this paper, we propose a data-driven voltage-abnormal cell detection method, using a fast model with simple architecture, which can detect voltage-abnormal cells based on the multi-source time series data of the LIB without a time interval. Firstly, our method transforms the different source data of a cell into a multi-source time series data representation and utilizes a recurrent-based data embedding to model the relation within it. Then, a simplified MobileNet is used to extract hidden feature from the embedded data. Finally, we detect the voltage-abnormal cells according to the hidden feature with a cell classification head. The experiment results show that the accuracy and average running time of our model on the voltage-abnormal cell detection task is 95.42% and 0.0509 ms per sample, which is a considerable improvement over existing methods. Full article
Show Figures

Figure 1

Figure 1
<p>The architecture of the proposed method.</p>
Full article ">Figure 2
<p>The details of the multi-source time series data representation.</p>
Full article ">Figure 3
<p>The structure of LSTM cell.</p>
Full article ">Figure 4
<p>The structure of the depthwise separable convolution with bottleneck.</p>
Full article ">Figure 5
<p>The structure of the cell classification head.</p>
Full article ">Figure 6
<p>The influence of the parameters on the voltage-abnormal cell detection task. (<b>a</b>) The influence of the batch size; (<b>b</b>) the influence of the number of LSTM hidden units; (<b>c</b>) the influence of the number of CCH hidden units; (<b>d</b>) the influence of the number of channels in the CCH.</p>
Full article ">
Back to TopTop