1. Introduction
In recent years, there has been an increasing demand for autonomous vehicles in complex indoor environments. Compared with drones, autonomous cars can perform a variety of ground transportation tasks while maintaining good stability. To avoid causing damage to the items being transported, it is crucial for unmanned vehicles to remain stable under all circumstances. Therefore, these vehicles should be equipped with obstacle avoidance algorithms to perform effectively in hazardous situations such as the sudden appearance of moving obstacles, tight corners, and narrow corridors. However, the existing algorithms [
1,
2,
3] face various practical issues. Unmanned vehicles are often limited by heavy sensors (Radar, LiDAR, etc.), which can lead to short battery life, high costs, and large sizes, limiting the use cases and performance of these algorithms. Thus, choosing the most suitable sensors for unmanned vehicles is a critical task. Unmanned vehicles are often limited by heavy sensors (Radar, LiDAR, etc.), which can lead to short battery life, high costs, and large sizes, limiting the use cases and performance of these algorithms. Thus, choosing the most suitable sensors for unmanned vehicles is a critical task.
Common sensors used for these tasks include ultrasonic sensors, LiDAR, and cameras. Ultrasonic sensors excel at measuring short-range obstacles, but their accuracy decreases with distance. LiDAR sensors offer rich information, high precision, long range, and a wide field of view, but are expensive and heavy, reducing unmanned vehicles’ flexibility. In contrast, cameras provide extensive scene information with low power consumption, compact size, and affordability. Therefore, developing high-precision algorithms based on camera sensors is of significant practical importance.
Recent advancements in autonomous navigation have made significant progress, but challenges remain, especially for lightweight vehicles in complex environments. State-of-the-art methods can be broadly categorized into two approaches:
- (1)
Deep learning-based methods: Works such as [
4,
5] utilize neural networks for information extraction from cameras. While highly accurate, they require substantial computational resources, including GPUs, limiting their applicability on lightweight platforms.
- (2)
Traditional feature-based methods: vision-based SLAM using feature extraction can operate on CPU-only platforms but often lack the necessary accuracy for reliable navigation in complex scenarios.
Many works utilize deep learning to extract information from cameras [
4,
5]. These methods require high-performance computing boards with GPUs to run the algorithms, which are not suitable for lightweight small vehicle platforms. Vision-based Simultaneous Localization and Mapping (SLAM) using feature extraction methods can run on CPU-only platforms, obtaining the map information of the scene to make more precise decisions and estimating the necessary pose information for the vehicle. However, additional information is still needed to enhance the SLAM system’s accuracy.
When the environment is entirely constructed by perception technologies in real-time, trajectory planning for the moving vehicle is necessary. Trajectory planning is a challenging module, as there are requirements from both the moving vehicle and the environment. From the vehicle’s perspective, the trajectory must be stable to ensure good stability and comfort. i.e., avoiding damping and sharp changes. From the environment’s perspective, there is a series of obstacles that increase the risk of collision. Therefore, a method that addresses both stability and safety is needed. Control Lyapunov function (CLF) with stability constraints can be combined with the control optimization process to enhance the stability of the control system. Control barrier function (CBF) is used in RGB-D space to improve driving safety [
6]. CLF-CBF-QP is used to ensure the stability and safety of mobile car trajectory planning [
7].
Recent works predominantly utilize depth maps and poses estimated via odometry as inputs for corresponding path-planning tasks [
5,
8,
9,
10,
11]. These approaches often rely on IMUs, GPS, or LiDAR for pose calculation. Meanwhile, pose and depth estimation cannot be effectively globally optimized together, which fails to handle the noise introduced by depth estimation and odometry. This significantly impairs the performance of the planners. Some research employs deep learning to reduce uncertainty in depth estimation and uses reinforcement learning to deal with the noise in inputs [
5,
8], but it places higher demands on the GPUs and CPUs onboard the vehicles. This restricts the application of such works in scenarios with limited computational resources.
To address these limitations and advance the state-of-the-art in autonomous navigation for lightweight vehicles, the proposed method improves enhanced visual SLAM with advanced trajectory planning. The proposed approach offers the following key advancements:
This paper proposes a new autonomous obstacle avoidance algorithm for lightweight self-driving cars. The state-of-the-art visual SLAM algorithm ORBSLAM3 is used to perceive the scene, and its performance is enhanced by eliminating outliers with optical flow epipolar constraints. ORB-SLAM3 is capable of joint optimization of pose and 3D mapping using only the CPU, and it can merge submaps to achieve real-time reconstruction of large-scale scenes whilst storing their 3D maps. Based on the rich pose and 3D information of the scene obtained, a new autonomous obstacle avoidance algorithm for lightweight self-driving cars is proposed. The proposed method employs a trajectory generation method combined with an obstacle shape reconstruction process (SRP) in irregular-shaped environments (CLF-CBF-QP-SRP) for global planning. Global planning trajectories with safety and stability can provide a general reference trajectory for self-driving cars, together with local planning using TEB (Timed Elastic Band) for local planning to ensure avoiding the local collisions. The proposed method can effectively avoid obstacles in the scene whilst minimizing unnecessary movement. The contributions of this work are summarized as follows:
Development of a lightweight, single-camera-based visual SLAM system enhanced with optical flow for outlier culling, capable of perceiving rich environmental information and avoiding obstacles efficiently on CPU-only platforms.
Introduction of a novel path planning algorithm for irregular environments, combining CLF-CBF-QP-SRP for global planning with TEB for local planning, achieving robust collision-free navigation to various target points.
Enhancement of trajectory generation robustness through a unified approach that ensures dynamic stability and safety throughout the entire movement, from the initial point to the target point.
Comprehensive evaluation and comparison with state-of-the-art methods, demonstrating superior performance in generating safe, stable, and efficient trajectories for lightweight autonomous vehicles in complex indoor environments.
Through extensive simulations and comparisons with existing methods, exhaustive experiments demonstrate that the proposed approach outperforms current state-of-the-art techniques in terms of computational efficiency, trajectory stability, and obstacle avoidance capability for lightweight autonomous vehicles.
3. System Overview
As shown in
Figure 1, the proposed system uses visual sensors to map the environment in advance to obtain spatial boundary and obstacle information. During the process of path planning and navigation, the proposed method utilizes images captured by the car’s camera to generate feature points through descriptor matching. The epipolar constraints of the LK optical flow [
29] are used to filter and remove outliers from the scene. Finally, the proposed method uses the relocation algorithm to obtain the current location and attitude information of the car. When the car is moving, the proposed method uses the ORB-SLAM3 algorithm to update the car’s pose information in real time. After the navigation coordinate points are completely set, the system will use the pre-known static cost map for inflation. For global path planning, the proposed method uses CBF to update the relative distance between obstacles and the car in real time to achieve autonomous obstacle avoidance. For complex terrain and new obstacles, the autonomous car uses the Timed Elastic Band (TEB) [
28] local planning algorithm to plan the car’s path locally and use the constraints between the car and surrounding obstacles to speed up iterations to find the optimal path.
3.1. Perception Based on Vision
Compared with LiDAR, visual sensors have the advantages of low cost and small size and are widely used in various autonomous driving platforms. This article uses a visual SLAM system based on a RGB-D camera as shown in
Figure 2, which ensures positioning accuracy and saves costs, facilitating subsequent deployment on low-cost unmanned vehicles and realizing the project as soon as possible.
Point Features Matching and Attributes Updating
A descriptor is a compact representation of a feature point’s local appearance, typically a binary or floating-point vector. In this context, desc(.) represents the function that computes this descriptor for a given point. The proposed method employs a comprehensive point feature-matching methodology for stereo camera configurations. The initial step involves identifying point features from the last frame and current from the RGB-D camera that share the same “grid ID” of these points. Each point feature is assigned a unique “grid ID” based on its spatial locality, facilitating the association of features between frames. To ensure temporal consistency, the algorithm matches features by comparing descriptors and selecting the closest match based on the minimal Euclidean distance:
where
and
are descriptors of points from previous and current frames, respectively. Further validation is performed by examining the cosine similarity of the angle between the direction vectors of matched points, ensuring the matched points align accurately with the expected motion model.
where
and
are the direction vectors of points in consecutive frames. The point feature matching process among consecutive frames is visualized in
Figure 3. In the figure,
represents the relative pose transformation between frames or the last and current image frames. Point features in 3D space, defined by a point
and another point
, result in two points: the point
and another point
when projected onto the image coordinate system
at time
. At the time
t, the same point projected onto the image coordinate system
results in new points: point
and another point
.
Bidirectional cosine similarity is used to match point pairs across frames based on grid IDs, discarding pairs below a similarity threshold. For RGB-D camera frames, feature points’ grayscale centroids and direction vectors will be computed, which are crucial for pose optimization during bundle adjustment.
3.2. Outlier Features Removing
In the proposed SLAM system, the proposed method mitigates the impact of inconsistent feature points on pose estimation by implementing the Lucas–Kanade (LK) method to enforce epipolar constraints, guided by the fundamental matrix F. This ensures geometrically coherent feature point pairs by setting a threshold distance from the epipolar line, allowing us to filter out mismatches. This fusion of techniques bolsters system stability.
To refine feature selection, the proposed algorithm eschews traditional Harris corner [
30] matches in favor of those with lower disparity in central pixel blocks. The proposed method then applies a stringent distance metric to eliminate outliers, which is central to the precision of the proposed algorithm.
For rigor, the RANSAC algorithm [
31] is employed to extract the fundamental matrix
F that maximizes inlier correspondences. With this matrix, features are extracted from one frame to their epipolar counterparts in the subsequent frame.
Consider matched points
and
in consecutive frames with homogeneous coordinates
and
, respectively. The epipolar line
for
is obtained as
. The distance to the epipolar line is defined as
where
F,
, and
represent the fundamental matrix, the first and second rows of the fundamental matrix
F, respectively. Points with
D exceeding a predefined threshold are deemed outliers and discarded.
3.3. Point Features Optimizing Algorithm
In 3D visual SLAM systems, map point features play a critical role in environment modeling and localization. However, factors like sensor noise and dynamic environments can induce errors in the orientation and position of these features. To address this, the proposed method adopts an optimization-based reference keyframe pose correction method to enhance the accuracy and consistency of map point features.
Sim3 (Similarity Transform in 3D) is a seven-dimensional transformation that includes rotation, translation, and scale. It is used here to represent the relationship between different coordinate frames. The process of pose correction is detailed as follows: Let
and
represent the rotation matrix and the translation vector from the reference keyframe to the world coordinate system, respectively. The world coordinates of the map point features to be corrected are assumed to be
, with
denoting the coordinate of the point feature. The Sim3 transformation matrices
and
are used to transform the pose of map point features from the reference keyframe of the current frame to the corrected reference keyframe as follows:
where × denotes matrix multiplication, and
represents the coordinates of the corrected map point feature in the corrected reference keyframe coordinate system. Enhancing the precision of these transformations directly impacts the accuracy of the camera pose and the overall SLAM performance. By refining the keyframe poses through these corrections, the proposed method effectively reduces the impact of positioning errors caused by sensor noise and dynamic environmental factors. This sets a robust foundation for subsequent optimization processes such as motion-only bundle adjustment (BA).
Motion-only BA optimizes the camera orientation
and position
, by minimizing the reprojection error between matched 3D points
in world coordinates and their corresponding image keypoints
, which may be either monocular
or stereo
, with
the set of all matches, as follows:
where
is the robust Huber cost function, and
represents the covariance matrix related to the scale of the keypoint. The optimization problem in Equation (
5) is typically solved using iterative nonlinear least squares methods, such as the Levenberg–Marquardt algorithm. This method is particularly effective for bundle adjustment problems due to its ability to handle the sparsity of the Jacobian matrix efficiently. And
is the projection function for RGB-D cameras defined as
where
represent the coordinates of a point in the world coordinate system.
and
are the normalized image plane coordinates.
and
are the focal lengths of the camera along the
X and
Y axes, respectively.
and
are the coordinates of the principal point, typically at the center of the image.
Z is the depth value directly measured by the RGB-D sensor, providing real-time depth at each image pixel.
Local BA optimizes a subset of covisible keyframes
and all points observed in those frames
. Non-optimized keyframes
, while fixed during optimization, contribute observations of
, adding constraints to enhance map stability without altering their poses. Define
as the set of matches between points in
and keypoints in keyframe
k. The optimization is formulated as follows:
where
represents the 3D position of the
i-th point in
.
and
are the rotation matrix and translation vector for the
l-th keyframe in
. Reprojection error
is quantified as
where
represents the 2D projection coordinates of the j-th feature point on the image plane. The optimization problem involving Equation (
8) is part of the larger Local BA process, which is typically solved using sparse bundle adjustment techniques. These methods, often implemented using libraries such as g2o or Ceres Solver, exploit the problem’s sparsity structure to efficiently optimize over multiple keyframe poses and 3D point positions simultaneously. In contrast to Local BA, Full BA adjusts all keyframes and map points, except the origin keyframe, which remains fixed to resolve scale ambiguity. This extensive optimization ensures the highest accuracy by refining camera poses and landmark positions across the entire map based on all available visual information.
3.4. Global Path Planning by Using CLF-CBF-QP-SRP
In this section, two key concepts are introduced for designing a safe and stable control system: control Lyapunov function (CLF) and control barrier function (CBF). A CLF is a positive definite function that decreases along the trajectories of the system and can be used to ensure asymptotic stability of a desired equilibrium point. A CBF is a function that satisfies some conditions on its Lie derivatives and can be used to enforce state constraints in the operating space. By combining CLF and CBF, a control law can be designed that guarantees both safety and stability of the car.
3.4.1. Vehicle Model
To simplify the computation during the trajectory generation, a kinematic model [
26] is used. The kinematic model is governed by
where
v denotes the velocity of the vehicle,
x and
y denote the longitudinal and lateral coordinates of the vehicle’s midpoint,
is the angular velocity of the vehicle, and
is the course angle of the vehicle.
3.4.2. The Formulation of CLF
To generate a stable trajectory for a moving car, the vehicle model is written as the following nonlinear control system:
where
is the state of the moving car, and
is the control input. Functions
f and
g are smooth vector fields. The control input is subject to the following constraints:
where
is the admissible input set of the control system.
and
are the minimum value and maximum value of inputs. Definition 1 combines an affine constraint with
u to achieve an optimization-based controller.
Definition 1. Assume V is a Lyapunov function when the following condition is satisfied [27]:where and are the Lie-derivatives of and is a class function. The class function is a function with the property of strictly increasing and the initial value . Then s can be stabilized by the following equation: 3.4.3. The Stability Control of Moving Car Using CLF
As the basic formulation CLF is introduced in (
13), the problem is to combine CLF with the dynamics of the moving car. Assume the current position of the moving car is
and the target position is
. The error between the current position and the final position,
, can be used for building the CLF as follows:
where
P is a 3 × 3 symmetric matrix with five parameters used to ensure that the
is positive definite. The format of
P in this paper is formulated as
Therefore, (
10) can be transferred to the format suitable for the target of the moving car.
3.4.4. The Formulation of CBF
CBF is used for the collision-free control, together with CLF for stability. A set
is defined as composing a continuously differentiable function
h:
, yielding
where
is the set that achieves collision-free.
Definition 2. The nonlinear control system is safe if is forward invariant. Assume T is the time interval. The set is forward invariant when each [32]. Definition 3. The function h is a CBF defined on the set if there exists an extended class function α such that the nonlinear control system satisfies [27]:where and are Lie-derivatives of . The set of controls that allow
to be collision-free for all
is formulated as
3.4.5. Safe Control of Moving Car Using CBF
In this paper, all obstacles are considered to be static with square shapes, in contrast to the regular circular shapes used for CBF calculations. While the initial approach simplified obstacles as squares, a more nuanced method is employed to accurately represent and avoid obstacles of various shapes. The obstacle shape reconstruction process begins with point cloud generation, where the RGB-D camera provides a set
of points representing obstacle surfaces in the environment. These points are then clustered to identify distinct obstacles, with each cluster
representing a potential obstacle. For each cluster
, a convex hull
is computed to approximate the obstacle’s shape, defined by a set of vertices
. Based on these convex hulls, barrier functions are constructed for each obstacle. Specifically, for each convex hull
, a barrier function
is formulated as
where
s is the vehicle state,
are the vertices of the convex hull,
are the outward-facing normal vectors of the hull faces, and
is a safety distance. This formulation ensures that
when the vehicle is outside the obstacle (plus safety distance) and
when it’s inside or on the boundary. The overall barrier function for all obstacles is then defined as
This approach allows for more accurate representation and avoidance of obstacles with complex shapes. To compute CBFs, the SRP (Specific Required Parameter) of each obstacle must be determined. As illustrated in
Figure 4, a circle is used to enclose the
ith obstacle, whose radius is calculated as
where
and
are the length and width of the
ith obstacle. To reach the target point in an open space, acquired by the perception stage, there is a moving car together with a set of
obstacles. Each obstacle is represented by
. Assume the position of the obstacle
is denoted by
, thus (
17) is converted to
Thus, for
, the set of controls that ensures the robot car to be safe is expressed as
Assuming that the robot car is defined with a safe radius of
, the safe distance between the car and obstacle
is defined as
. Then the CBF is designed as
3.5. Safe and Stable Control for Self-Driving Cars
Since both the constraints of accurate and safe control have the affine form, real-time solutions can be acquired. Therefore, a QP-based controller that combines CBFs for safety and CLF with (
13) and (
23) for stability is as follows:
CLF-CBF-QP:
where the objective function (
25a) is divided into three parts: the first part minimizes the magnitude of
, the second part adds an extra quadratic cost, and the third part ensures the smoothness of
u.
H and
Q are positive definite matrices,
is the weight coefficient of the relaxation variable
and
is the control value of the last moment.
4. Experimental Evaluation
The simulations were conducted to verify the safety, stability, and efficiency of the proposed route planning algorithm. The experiments were conducted on a Linux machine with the Ubuntu 18.04.6 LTS OS, a 12th generation 16-thread Intel®Core™ i5-12600KF CPU, an NVIDIA GeForce RTX 3070Ti GPU, and 16 GB of RAM. The QP problem is solved using the quadprog solver in MATLAB R2022B.
In order to verify the effectiveness of the proposed system, a closed square experimental scene is built. In order to ensure that the visual algorithm can extract a certain number of feature points in the surrounding environment, TVs, murals, and other items are added to the inside of the room walls. Among them, 9 square hollow tables are used as obstacles. The feature points in the middle of such hollow objects are often not on the obstacles but on the wall behind them. Such a scenario poses certain challenges to the perception algorithm to test. The car is placed on the map as the carrier of the system, rather than the car’s preset starting point. This can effectively test the effectiveness of the proposed system’s relocation system. The car in the experiment is rectangular in shape and has four wheels, equipped with an RGB-D vision sensor on the top of the front end. The initial verification of the obstacle avoidance algorithm (CLF-CBF-QP) will be verified in Matlab, and the experiments of the whole visual navigation system are carried out in simulation environments Gazebo and Rviz.
Figure 5 illustrates one experiment on how to navigate the car from the start point to the destination.
4.1. Simulation Environments Setup
The car is placed at the starting point with coordinates (−4, −4) and the target points with coordinates (0, 1.5). First, the car uses the visual sensor to detect the feature points in the image, as shown in
Figure 5a. After comparing it with the pre-saved atlas, it uses the relocation algorithm to obtain the current location of the car and updates the car’s posture and obstacle distance. Then, the velocity and direction of the car are updated by the route planning algorithm. In order to verify the effectiveness of the proposed algorithm, CBF was first performed on MATLAB for the motion trajectory planning test. The simulation trajectory on MATLAB is shown in the red trajectory in
Figure 5b. The proposed system ensures that it does not encounter obstacles and selects the shortest distance between two points. In order to further verify the effectiveness of the proposed algorithm, the ROS system is used to conduct further simulation tests under Ubuntu. The planned path simulation of the car is displayed as a blue box in Rviz, where the point in the middle of the blue box represents the trajectory of the vehicle. The blue box represents the space the car takes up in the environment. The colored map represents costmap, which is used to represent obstacles and passable areas in the environment. The costmap is represented as a two-dimensional grid, with each grid cell (or pixel) colored according to the “cost” or “safety” it represents. The meaning of the colors is as follows: Blue usually represents low-cost areas, i.e., areas that are relatively safe and free for the robot car. Red usually indicates very high costs and is often directly associated with obstacles. This is an area that robot cars should avoid. Grey represents unknown areas, i.e., those areas that the robot car has not yet explored or whose properties cannot be determined. White represents known free areas, i.e., areas without obstructions.
4.2. Stability and Safety of Trajectory Generation
The stability and safety of trajectory generation for the moving car using CLF-CBF-QP-SRP are validated in this section through different testing target points. As shown in
Figure 6, three different testing scenarios are illustrated. Target points 1, 2, and 3 are located at the lower right, upper right, and upper left of the center point, respectively. These three target points are defined to test the capability of achieving stability and safety in trajectory generation with various obstacles involved. These target points also test the adaptivity of the proposed CLF-CBF-QP-SRP program. In each case, different states of the moving car will be discussed.
4.2.1. Comparison with State-of-the-Art Algorithm
To demonstrate the efficient trajectory generated by the CLF-CBF-QP-SRP, target point 2 is set as a test point to compare with a state-of-the-art algorithm called PU-RRT [
33]. The PU-RRT is capable of generating safe and risk-bounding trajectories and outperforms other adapted RRT algorithms. Therefore, this paper selects PU-RRT for comparison to verify the relatively more efficient trajectory using the CLF-CBF-QP-SRP while maintaining safety. As illustrated in
Figure 6b, a smooth and efficient trajectory without any collisions among a series of candidate trajectories is chosen by the PU-RRT. However, as illustrated in
Figure 6c, the trajectory generated by the proposed CLF-CBF-QP-SRP is shorter than that of PU-RRT. This is because the proposed method aims to find the most efficient trajectory while avoiding collisions. Therefore, the proposed method generates a more efficient trajectory than the PU-RRT, while both can efficiently avoid collisions with objects.
4.2.2. Safe and Efficient Trajectory Generation for Three Target Points
Figure 7 shows the trajectory generation using the LBF-CBF-QP-SRP program from the starting point to three target points, respectively.
Figure 7a illustrates the real environment for trajectory generation.
Figure 7e illustrates the detected environment by visual perception.
Figure 7b illustrates the LBF-CBF-QP-SRP-based global trajectory for target point 1.
Figure 7c illustrates the LBF-CBF-QP-SRP-based global trajectory for target point 2 in MATLAB.
Figure 7d illustrates the LBF-CBF-QP-SRP-based global trajectory for target point 3 in MATLAB.
Figure 7f illustrates the final trajectory with local safe planning for target point 1 in Rviz.
Figure 7g illustrates the final trajectory with local safe planning for target point 2 in Rviz.
Figure 7h illustrates the final trajectory with local safe planning for target point 3 in Rviz. To provide a comprehensive comparison,
Figure 7b–d represent the optimized estimated poses generated by the proposed algorithm, while
Figure 7f–h depict the true poses of the car in the simulated environment. This juxtaposition allows for a direct comparison between the estimated and actual trajectories, demonstrating the accuracy and effectiveness of the proposed approach. From these trajectories, it can be observed that the moving car can avoid collisions and drive close to the boundary of reshaped obstacles, improving the efficiency of reaching the target point. Furthermore, all the trajectories are smooth and stable, as there is no sharp changing. The close correspondence between the estimated and true poses further validates the reliability of the proposed LBF-CBF-QP-SRP program in real-world scenarios. Therefore, the proposed LBF-CBF-QP-SRP program can ensure the safety, smoothness, and stability of driving throughout the whole process.
4.2.3. Safe and Stable Control of Target Point 3
This section illustrates the states, control variables, and total CBF of all obstacles, from the starting point to the target point, among the time axis. As illustrated in
Figure 8a, the total CBF is always larger than 0 throughout the whole time, which suggests it is safe driving.
Figure 8b shows the angular velocity of the moving car. It is obvious that the changing of the angular velocity is smooth and with a small scale, which suggests stability during the driving.
Figure 8c elaborates on the longitudinal position
x, lateral position
y, and course angle
of a moving car. The longitudinal distance has changed with smoothness gradually from −4 to the point near 0, suggesting the smoothness and stability of longitudinal driving. The lateral distance has gradually increased with smoothness from −4 to the point near 2, suggesting the smoothness and stability of lateral driving. The variation of
is also smooth and reaches a stable level at the end of time, revealing the stability of turning among the whole process.
4.2.4. Comparison with Advanced Trajectory Generation Algorithm
To evaluate the performance of the LBF-CBF-QP-SRP program and enhance the validity of the simulation results, it is compared with two benchmark trajectory generation methods: Voronoi diagram [
34] and APF [
35]. As discussed in
Section 2.2, these methods have the advantages of safety and efficiency. Therefore, they are used to be basepoints for comparison.
Figure 9 shows the trajectories generated by the three methods for three different target points, starting from the same initial point
. For target point 1, the LBF-CBF-QP-SRP program produces a smooth and collision-free trajectory. However, the APF method exhibits severe oscillation when it approaches the obstacle in the middle of the third row, which indicates a lack of stability. The Voronoi diagram method collides with the leftmost obstacle in the second row. Moreover, its trajectory is neither smooth nor efficient. For target point 2, the LBF-CBF-QP-SRP program generates a smooth and efficient trajectory, as it drives close to the boundary of the map-centered obstacle. However, the APF method has two noticeable oscillations, with the leftmost obstacle in the third row and the map-centered obstacle, respectively, resulting in unstable driving. The Voronoi diagram method follows a longer trajectory than the other two methods, which implies low efficiency. For target point 3, the LBF-CBF-QP-SRP program ensures a smooth, efficient, and good trajectory. The APF method encounters oscillation with the leftmost obstacle in the second row. The Voronoi diagram method makes frequent turns in some parts of the driving, which reduces the efficiency. For the computational time, each case is tested ten times, and the average time is calculated. The average computational times of LBF-CBF-QP-SRP, APF, and the Voronoi diagram are 2.0176 s, 2.2723 s, and 0.0732 s, respectively. Therefore, the computational time for planning the global trajectory of the proposed method is relatively low.
4.2.5. Comparison with Deep Learning-Based Methods
To further validate the efficiency and effectiveness of the proposed enhanced visual SLAM system, a comparison with state-of-the-art deep learning-based methods was conducted. Specifically, the proposed approach is compared with two recent deep learning-based methods for simultaneous localization, mapping, and navigation, DS-SLAM [
36] and Blitz-SLAM [
37], under the same experimental settings.
Table 1 presents a quantitative comparison of computational resource consumption and performance metrics:
As evident from
Table 1, the proposed method significantly outperforms the deep learning-based approaches in terms of computational resource consumption. The proposed method operates without GPU acceleration, utilizing only 45.3% of CPU resources compared with the 78.6% and 82.1% CPU usage of DS-SLAM and Blitz-SLAM, respectively. Memory usage is also substantially lower, requiring only 2.9 GB compared with 4.8 GB and 5.3 GB for the deep learning methods. In terms of processing time, the proposed method achieves 48 ms per frame, more than twice as fast as the deep learning approaches. This efficiency is crucial for real-time applications in lightweight autonomous vehicles.