Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Dynamic Hand Gesture Recognition Using 3DCNN and LSTM with FSM Context-Aware Model
Next Article in Special Issue
Extrinsic Calibration between Camera and LiDAR Sensors by Matching Multiple 3D Planes
Previous Article in Journal
A Macro Lens-Based Optical System Design for Phototherapeutic Instrumentation
Previous Article in Special Issue
Time-Domain Data Fusion Using Weighted Evidence and Dempster–Shafer Combination Rule: Application in Object Classification
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Artificial Marker and MEMS IMU-Based Pose Estimation Method to Meet Multirotor UAV Landing Requirements

GNSS Research Center, Wuhan University, No. 129 Luoyu Road, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(24), 5428; https://doi.org/10.3390/s19245428
Submission received: 6 November 2019 / Revised: 4 December 2019 / Accepted: 5 December 2019 / Published: 9 December 2019
(This article belongs to the Collection Multi-Sensor Information Fusion)
Figure 1
<p>Examples of the proposed marker and binary encoding matrix.</p> ">
Figure 2
<p>Flow chart of pose estimation by detecting the marker.</p> ">
Figure 3
<p>Image processing results. (<b>a</b>) Original image; (<b>b</b>) binary image after Otsu binarization; (<b>c</b>) polygon contours extracted in (<b>b</b>); (<b>d</b>) quadrilateral contours and the matched marker; (<b>e</b>) the corner points extracted on the marker; (<b>f</b>) the extracted corner points of the central marker when the unmanned air vehicle (UAV) is very close to the ground.</p> ">
Figure 4
<p>Definition of coordinate systems and camera projection model. <math display="inline"><semantics> <mrow> <msub> <mi>O</mi> <mi>c</mi> </msub> </mrow> </semantics></math> is the optic center of the camera and <math display="inline"><semantics> <mrow> <msub> <mi>O</mi> <mi>c</mi> </msub> <mo>−</mo> <msub> <mi>X</mi> <mi>c</mi> </msub> <msub> <mi>Y</mi> <mi>c</mi> </msub> <msub> <mi>Z</mi> <mi>c</mi> </msub> </mrow> </semantics></math> denotes the c-frame; <math display="inline"><semantics> <mrow> <msub> <mi>O</mi> <mi>p</mi> </msub> <mo>−</mo> <mi>u</mi> <mi>v</mi> </mrow> </semantics></math> denotes the p-frame; and <math display="inline"><semantics> <mrow> <msub> <mi>O</mi> <mi>w</mi> </msub> <mo>−</mo> <msub> <mi>X</mi> <mi>w</mi> </msub> <msub> <mi>Y</mi> <mi>w</mi> </msub> <msub> <mi>Z</mi> <mi>w</mi> </msub> </mrow> </semantics></math> denotes the w-frame.</p> ">
Figure 5
<p>System overview of the camera/INS (inertial navigation system) integrated navigation. The vehicle pose is retrieved by the two sensors independently and then fused by error-state extended Kalman filter (EKF), namely loosely coupled.</p> ">
Figure 6
<p>Installation relationship between the camera and inertial measurement unit (IMU).</p> ">
Figure 7
<p>Experimental equipment.</p> ">
Figure 8
<p>Trajectory of the test platform in test 05.</p> ">
Figure 9
<p>Attitude variation of the test platform in test 05.</p> ">
Figure 10
<p>Experimental environment.</p> ">
Figure 11
<p>Position errors of the visual solution in test 03.</p> ">
Figure 12
<p>Attitude errors of the visual solution in test 03.</p> ">
Figure 13
<p>Corner points reprojection errors on the horizontal plane in test 01.</p> ">
Figure 14
<p>Comparison of roll and pitch errors by the visual solution and visual/inertial solution in test 04.</p> ">
Figure 15
<p>Comparison of roll and pitch errors by the attitude and heading reference system (AHRS) solution and visual/inertial solution.</p> ">
Figure 16
<p>Position errors of the visual-inertial solution in the vision outage simulation test.</p> ">
Figure 17
<p>Reprojection error in the vision outage simulation test.</p> ">
Figure 18
<p>Edge extraction and marker detection results in challenging situations. (<b>a</b>) Detected marker at a distance of 8 m with no tilt; (<b>b</b>) detected marker at a distance of 8 m with tilt of 40 degrees.</p> ">
Versions Notes

Abstract

:
The fully autonomous operation of multirotor unmanned air vehicles (UAVs) in many applications requires support of precision landing. Onboard camera and fiducial marker have been widely used for this critical phase due to its low cost and high effectiveness. This paper proposes a six-degrees-of-freedom (DoF) pose estimation solution for UAV landing based on an artificial marker and a micro-electromechanical system (MEMS) inertial measurement unit (IMU). The position and orientation of the landing maker are measured in advance. The absolute position and heading of the UAV are estimated by detecting the marker and extracting corner points with the onboard monocular camera. To achieve continuous and reliable positioning when the marker is occasionally shadowed, IMU data is fused by an extended Kalman filter (EKF). The error terms of the IMU sensor are modeled and estimated. Field experiments show that the positioning accuracy of the proposed system is at centimeter level, and the heading error is less than 0.1 degrees. Comparing to the marker-based approach, the roll and pitch angle errors decreased by 33% and 54% on average. Within five seconds of vision outage, the average drifts of the horizontal and vertical position were 0.41 and 0.09 m, respectively.

1. Introduction

With the rapid development of micro-electromechanical systems (MEMSs) and high-performance flight control processors, applications and research of multirotor unmanned air vehicles (UAVs) have received increasing attention worldwide. Due to the unique structure and stability, multirotor UAVs are widely used in precision agriculture [1,2], rescue [3], surveillance [4,5], etc. These applications also pose new challenges for the precision landing of UAVs [6]. To land safely and precisely, the localization of UAV requires high accuracy, continuity, and reliability.
Global navigation satellite system/inertial navigation system (GNSS/INS)-integrated navigation [7] technologies are the most widely used navigation methods for various UAVs [8,9]. However, the precision of the integrated navigation positioning is completely dependent on the GNSS. Although decimeter-level positioning results can be obtained by a consumer-grade dual-frequency GNSS receiver with sufficient satellites [10,11], in areas with low satellite observation such as city canyons and forests, positioning errors will increase. Moreover, it is totally unavailable for indoor applications. To ease these problems, vision-based pose estimation approaches for UAV landing have attracted extensive research due to their low cost and ability to sense the environment [12,13]. A survey of the vision-based landing techniques for UAVs is presented in [14]. In most applications, e.g., precision agriculture, we can determine the landing point of UAVs where we are allowed to place a customized pad in advance to take advantage of the pertaining features.
Abundant works concerning the design of the fiducial marker system have been proposed in literature, such as ARTag [15], AprilTag [16], and ArUco [17]. The authors in [18,19] use AprilTag markers as their landing pad. In [20], ARTag markers are used for UAV localization. Researchers also have proposed specialized markers to meet precision landing requirements of UAVs.
The authors in [21] use video from the onboard camera and the orientation and altitude of the ground vehicle to calculate the UAV position with respect to (w.r.t.) the target. In [22], an integration algorithm of monocular and stereo vision is exploited to estimate the approach angle and relative height of a UAV by Hough transform and vanishing-line geometry. In [23], the authors introduce a dynamic fiducial marker that can change its appearance according to the requirements of the UAV. A landing marker which consists of the letter “H” inside a black circle is presented in [24]. A landing pad composed of markers of different size is utilized to permit detection from high as well as very low altitude in [25]. In [26], the authors describe a pattern comprising several concentric white rings on a black background. The rings are identified by the unique ratio between their inner and outer border radius. The authors in [27] present a landing marker consisting of three inner circles, each of which is evenly divided into eight fan areas. The area between the inner circles is filled by an even distribution of black and white. They use the template-matching strategy for marker detection in daytime, and a morphology technique named the hit-and-miss algorithm at nighttime. In 2018, the authors extended this work by introducing a method to exploit the trained features extracted from a convolutional neural network (CNN) for marker tracking, which can be performed at very high altitude [28]. The authors in [29] propose a graphics processing unit (GPU)-based pose estimation method based on a marker composed of circles and use experimental covariance measurements to estimate the camera pose during occlusions and vibrations. A comparison of the state of the art on the autonomous landing of UAVs using artificial markers is also shown in this work, regarding the marker shape, positioning error, frame rate, etc. Authors in [27] classify the previous vision-based landing systems for UAVs into two categories (namely passive and active methods) and compare their strength and weakness.
In the above marker-based positioning methods, when the marker is severely occluded, the system will not work. Furthermore, these approaches do not take into consideration the absolute heading angle, although some of them do retrieve the relative heading. To increase the positioning rate as well as cover the short-period marker occlusions, inertial measurements can be introduced to fuse with image data.
The authors in [24] use projective geometry to estimate the five-degrees-of-freedom (DOF) pose from the elliptic projection of the marker. Then the gravity vector estimated by the inertial measurement unit (IMU) is incorporated to resolve the remaining geometric ambiguity. In [25], the authors perform a homography-based method to obtain UAV pose from the marker and integrate it with the IMU data in a loosely coupled way. Their work focus on the landing of UAVs on a moving platform. However, the horizontal angles, which are essential for flight control, are not estimated in the system. Moreover, the IMU sensor errors are not modeled, which can result in significant position error accumulation with time.
In this work, a six-DoF state estimation system based on an artificial marker, a visible light monocular camera, and a MEMS IMU to meet the requirements of multirotor UAV landing is introduced. The IMU used in our system is at dollar-level, which is widely used for commercial smartphones. Referring to the prior works about marker design and detection systems, an improved methodology applicable to UAV landing is proposed. The inner corners of the marker are used for pose estimation when the UAV is too close to the ground and the whole marker is out of the image frame. IMU data is fused by an extended Kalman filter (EKF) to increase the navigation result rate to permit aggressive maneuvers and to cope with short-time marker occlusion. This also results in more smoothing results for both position and attitude. Unlike previous visual/inertial fusion system for UAV landing, models of the error components of IMU are considered and roll and pitch angles are also predicted by INS and updated in the EKF. Field tests illustrate the magnificent performance of our system. It worth mentioning that this work only focuses on the pose estimation for the UAV landing instead of on the whole closed-loop control system.
This paper is organized as follows. In Section 2, the design and detection of the marker as well as the relative state estimation method with the extracted marker is presented. Then the fusion algorithm of pose information from the marker and IMU data is described. Experimental evaluations of the proposed system are provided and discussed in Section 3. Finally, Section 4 draws some conclusions.

2. Methods

2.1. Marker Detection

Similar to previous works, the marker proposed in this work is a big square consisting of small black and white squares. They are distributed on a plane according to a certain rule with the advantages of large information capacity and good expandability.
The marker is constituted by N × N black and white squares of the same size. The number and size of the squares are dependent on the size of the dataset, as well as the amount of encoded information. The border of the marker is black with several black squares properly filled inside. To avoid the central symmetry of the marker, which leads to ambiguous heading definition, it must be ensured that the ultimate marker is not centrosymmetric when determining internal black squares. Our marker is specifically designed for UAV landing, reducing the computational cost while limiting false positives. Unlike the aforementioned markers, the structure inside the border of the proposed marker is much simple. To meet the requirement of estimating pose at different heights along the landing path especially when the UAV is close to the ground, a black square is set at the marker center. At the beginning, the UAV detects the whole marker and adjust its horizontal position and heading while landing. Once it descends to a certain height where the marker is too large to be detected, the center square can be used for guidance. Thus, N should be an odd number.
A marker database is consequently built for different landing points. Every marker has its own specific number and involves a different command for the UAV. We use a binary bit stream to represent the coding matrix of the marker for fast and accurate matching, with values of 0 for black and 1 for white. The ( N 2 ) × ( N 2 ) matrix uniquely represents the encoding information of the marker. Figure 1 shows examples of two different markers ( N = 7 ) and their encoding matrices.
The main process of pose estimation by the marker is shown in Figure 2. First, we perform histogram equalization and morphological processing on the captured image to reduce the noise and then execute binarization. The Otsu [30] algorithm is employed since it provides the optimal image threshold given that the pixel distribution is bimodal, which is true in this scenario. We use the Canny [31] operator to extract the contours in the image and the Douglas–Peucker algorithm to fit the contours to a polygon. After removing the unreasonable contours, all the quadrilateral contours are extracted, and the internal pixels are extracted and clustered to match the encoding matrix in the database. After determining the unique matrix, the vertices of the marker on the image are obtained, and all corners inside the marker are extracted [32]. If the entire marker cannot be extracted, the corners are tracked by calculating the optical flow [33]. When the UAV is too close to the marker, the corners of the central black square are extracted. Since the positions of the corner points are known, the position and orientation of the camera can be estimated by a set of correspondences between real world points and their projection on image. This is introduced in Section 2.2. Figure 3a–f presents the image processing results of the proposed system.

2.2. Pose Estimation by Marker

The relative pose between two coordinate systems can be retrieved by matching the corresponding points. The world coordinate system (w-frame) is established with the center of the marker as the origin. The camera coordinate system (c-frame) is established with the optical center of the camera as the origin, x-axis pointing to the right, y-axis pointing down. The pixel coordinate system (p-frame) is established with the upper left corner of the image as the origin and ( u , v ) parallel to ( x , y ) of the c-frame, respectively. Figure 4 shows the coordinate systems. The camera projection model is shown as Equation (1). The marker can be laid arbitrarily at the landing point, but we have to measure its heading angle in advance for UAV absolute heading estimation. For easy understanding, we assume that the x-axis and y-axis of the w-frame points to the north and east, respectively.
The transformation among coordinate systems is given as
P c = R w c P w + T w c ,
Z c [ u v 1 ] = [ f x 0 c x 0 f y c y 0 0 1 ] [ X c Y c Z c ] = K P c ,
where P c and P w are the coordinates of the corner points in c-frame and w-frame, respectively; ( u , v ) are the coordinates of the corner points in p-frame; R w c and T w c are the rotation matrix and translation vector from w-frame to c-frame, respectively; K is the intrinsic parameter matrix of the camera, which can be calibrated by the chessboard method proposed in [34]; ( c x , c y ) is the camera principal point in p-frame; and f x , f y represent the focal length of the camera in terms of pixel dimensions in the x and y direction, respectively. According to the matching pairs of the corners in w-frame and p-frame, the estimation of the UAV pose ( R w c , T w c ) is a perspective-n-point (PNP) problem that can be solved by [35]. The random sample consensus (RANSAC) method [36] is introduced to exclude corner outliers to ensure the robustness of the visual positioning. Then, we regard it as the initial value and perform the Levenberg–Marquardt [37] approach to optimize it according to the minimum reprojection error criterion, shown as
min R w c , T w c i n ρ i | | P i 1 Z c K ( R w c P w + T w c ) | | ,
where i n denotes the i th point of the all n points; ρ i indicates the weight factor. To reduce the influence of a corner observation with a large error on the result, we use the Cauchy loss function (Equation (4)) to reduce the weight of this observation [38].
ρ ( s ) = log ( 1 + s )
The relative position of the camera w.r.t. the w-frame can be transformed to the absolute pose in the e-frame by the following equations,
δ r e = D δ r n ,
D = [ R M + h 0 0 0 ( R N + h ) cos φ 0 0 0 1 ] ,
where δ r e is the position vector from the origin of w-frame to the camera center projected in the e-frame; δ r n is the camera coordinates in the n-frame; R M is the meridian radius of curvature; R N is the radius of curvature in the prime vertical; φ is the geodetic latitude; and h is the altitude.

2.3. Visual/Inertial Fusion

INS is able to calculate UAV pose from the initial state by IMU measurements alone. By detecting the marker, the relative position and heading of the UAV in the w-frame is retrieved which provides reliable external observations for INS to correct its errors. Since the corner points are on the same plane that is parallel to the horizontal plane, the observability of the horizontal angle (i.e., roll and pitch angle) is weak for PNP. For INS, the external position assist and gravity constraint make the horizontal angle observable. Moreover, the IMU data rate is higher than that of the camera in general. INS mechanization can predict the UAV state during two consecutive image frames. Due to the magnificent complementarity, it is valuable to fuse the two sensors’ data. The algorithm overview is presented in Figure 5.

2.3.1. INS Mechanization and System Model

The main procedure of the INS mechanization can be divided into three steps:
  • Integrate the angular rate to update the vehicle attitude;
  • Transform the acceleration to the n-frame by the updated attitude, then integrate it to obtain the vehicle velocity;
  • Integrate the velocity to calculate the vehicle position.
Refer to [39] for the detailed update equations.
The demands for an accurate estimation of the vehicle pose necessitate the modeling and online estimation of the error components of the sensor. The state vector of the integrated navigation system constructed in this paper can be expressed as
x ( t ) = [ δ r n δ v n ϕ b g b a s g s a ] ,
where δ r n is the position error in the navigation frame (n-frame), which has its origin coinciding with that of the b-frame, with its x-axis pointing north, z-axis pointing down, and y-axis completing a right-hand orthogonal frame, i.e., the north–down–east (NED) system; δ v n is the velocity error in the n-frame; ϕ is the attitude error, composed of roll error, pitch error, and heading error; b g and b a are the residual bias error of the gyroscope and the accelerometer, respectively; s g and s a are the residual scale factor error of the gyroscope and the accelerometer, respectively.
The continuous time system is modeled as,
x ˙ ( t ) = F ( t ) x ( t ) + G ( t ) w ( t ) ,
where, F ( t ) is the system dynamics matrix; G ( t ) is the noise distribution matrix of the system; and w ( t ) is the vector describing the system noise.
Due to the uncertainty in the sensors and the gravity field, the navigation parameters obtained from the INS mechanization equation contain errors. The phi-angle error model [40] is used here. Thus, the differential equation of the inertial navigation error is as follows:
ϕ ˙ = ω i n n × ϕ + δ ω i n n C b n δ ω i b b ,
δ v ˙ n = C b n δ f b + C b n f b × ϕ + v n × ( 2 δ ω i e n + δ ω e n n ) ( 2 ω i e n + ω e n n ) × δ v n + δ g l n ,
δ r ˙ n = ω e n n × δ r n + δ θ × v n + δ v n ,
δ θ = [ δ r E / ( R N + h ) δ r N / ( R M + h ) δ r E tan φ / ( R N + h ) ]
where, b denotes the body frame (b-frame), which has its axes coinciding with the IMU’s body; δ ω i b b and δ f b are the gyroscope measurement error and the accelerometer measurement error, respectively; ω i e n is the rotation rate vector of the e-frame with respect to the inertial frame (i-frame) projected to the n-frame; ω e n n is the rotation rate vector of the n-frame with respect to the e-frame projected to the n-frame; δ g l n is the gravity vector projected to the n-frame; C b n is the rotation matrix from the body frame (b-frame) to the n-frame; δ r N and δ r E are the position error on north and east direction of the n-frame, respectively; The variables b g , b a , s g , and s a are modeled by a first-order Gauss–Markov process.
x ˙ = 1 T x + w x k + 1 = e Δ t k + 1 / T x k + w k ,
where T is the correlation time of the process, and w is the driving white noise process.

2.3.2. Linearized Measurement Model

Since the IMU and the camera cannot be installed at the same place with parallel axes, the pose retrieved from the marker is different from that derived from the INS, which is known as the lever-arm effect and installation angle effect, as described in Equations (14) and (15), respectively. These misalignments have to be calibrated in advance or they will cause significant errors in both position and attitude. For easy understanding, it is assumed that the b-frame coincides with the UAV coordinate system (v-frame).
The installation relationship between the camera and IMU is depicted in Figure 6.
The camera position derived from the IMU and camera is expressed as follows,
r ^ c e = r ^ I M U e + C n e C ^ b n l c b = r I M U e + δ r I M U e + C n e [ I ( ϕ × ) ] C b n l c b = r c e + δ r I M U e + C n e ( C b n l c b × ) ϕ r ˜ c e = r c e + C n e n r c ,
where r ^ c e is the camera position calculated by the IMU-derived position and lever arm; r ˜ c e is the camera position estimated by marker detection; l c b is the lever arm vector from the IMU center to the camera center in the b-frame; n r c is the position error of visual estimation; and [ ] × represents the antisymmetric matrix of a vector.
Since the landing maker can provide a reliable heading reference, it can also be exploited to correct INS errors as an aiding source. The camera attitude derived from the IMU and camera are expressed as follow,
C ^ c n = C ^ b n C c b = [ I ( ϕ × ) ] C b n C c b C ˜ c n = [ I + ( v × ) ] C c n ,
C ^ c n = [ I ( ρ × ) ] C ˜ c n [ I ( ϕ × ) ] = [ I ( ρ × ) ] [ I + ( v × ) ] ,
ρ × = I C ^ c n C ˜ c n T ρ = ϕ + v ,
where C ^ c n and C ˜ c n are the camera attitude matrices derived from the INS and recovered from the visual solution, respectively; C c b is the rotation matrix from the c-frame to the b-frame; v is the heading error of the visual estimation; and ρ is the attitude difference between the vehicle attitude calculated by INS and vision. Hence, the observation vector can be expressed as a combination of the position error observation and heading error observation as follows:
Z = [ z r c n ρ h e a d i n g ] ,
where, ρ h e a d i n g is the third element of ρ and z r c n can be written as
z r c n = C e n ( r ^ c e r ˜ c e ) = C e n ( δ r I M U e + C n e ( C b n l c b × ) ϕ C n e n r c ) = δ r n + ( C b n l c b × ) ϕ n r c .

3. Experimental Results Discussion

3.1. Experiments Description

To demonstrate the performance of the proposed system, we carried out field tests and analyzed the results from two aspects: the retrieved pose errors and reprojection errors of the corner points. The GNSS and high-precision fiber-optic IMU (POS320, MAP Space Time Navigation Technology Co., LTD, Wuhan, China) integrated navigation systems were used as references for the pose. The camera and consumer grade MEMS IMU (ICM20602)-integrated navigation system was the system under test. The camera used in the experiments was a global shutter camera with a 644 × 484 resolution (Mako G-030C, Allied Vision, Stadtroda, Germany). The IMU data rate was 50 fps, and the image frame rate was 20 fps. Figure 7 shows the experimental equipment platform. The spatial displacement of the IMU and the camera was calibrated in advance [41]. Table 1 provides the technical parameters of the two IMUs. The position and heading of the marker in the e-frame were previously obtained by GNSS post processed kinematic (PPK) static measurements. The size of the marker used in the tests was 29 cm × 29 cm.
Without the expensive motion capture system, it is too difficult, if not impossible, to obtain a reliable vehicle pose as ground truth with light load. A machine with four stepper motors, allowing precise movement of the camera in 3D space, was designed and built in [29] to evaluate the estimated position of the camera w.r.t. the maker. Although the high-precision GNSS/INS integrated systems can provide six-DOF pose references, they are usually too heavy (more than 10 kg) to be mounted on the UAV (like the one used in our experiments). Therefore, we hold the experimental device with reference system by hand to mimic the motion and trajectory of the UAV when landing to evaluate the pose accuracy convincingly. The GNSS/INS pose reference system was fixed on a rigid platform with the camera and MEMS IMU. The position and angle misalignment between the two IMUs was calibrated previously to project the ground truth to the MEMS IMU center. The position of the antenna phase center relative to the reference system was also measured in advance. The camera, IMU, and GNSS data were collected for postprocessing. We performed post processed kinematic (PPK) for the GNSS data process, then the smoothing results of the GNSS/INS integrated navigation were calculated as ground truth. The experiments were executed five times on Friendship Square at Wuhan University on 28 February 2019. The trajectory and attitude variation of the test platform in the fifth test is shown in Figure 8 and Figure 9. The experimental environment is shown in Figure 10.

3.2. Results Discussion

The position and attitude errors of the results by the pure visual solution in the third test are shown in Figure 11 and Figure 12, respectively.
The statistical results of the position and attitude errors of the visual solution for five experiments are shown in Table 2 and Table 3, respectively. The maximum and mean values of the error are calculated by the absolute error sequence over the whole article.
The tables show that the positioning error of the visual solution does not exceed 0.01 m in any of the three directions. It can be observed that the positioning results on horizontal direction are less accurate and more unstable than that on vertical direction. Given that the available satellite number was beyond 20 and the carrier-phase ambiguities were resolved to their integer values in our experiments, the accuracy of PPK smoothing positioning could be thought to be at the centimeter level [42]. Meanwhile, it is obvious that the noise influences of visual positioning and PPK positioning are independent of each other, hence the absolute positioning accuracy of the proposed visual solution is equivalent to that of the smoothed PPK.
For the attitude errors, the root mean square error (RMSE) of the roll angle was 0.45° ( 1 σ ), the pitch angle error was 0.43° ( 1 σ ), and the heading angle error was 0.1° ( 1 σ ). Since the corners of the marker are all on the horizontal plane (almost parallel to the p-frame), we can calculate the heading angle very accurately based on the position of the corresponding points on the image, but the roll and pitch estimations are less stable because of the lack of observations in the vertical direction.
The statistical results of the position and attitude errors of the visual/inertial fusion solution in five experiments are shown in Table 4 and Table 5, respectively.
The statistical results show that the positioning errors of the visual/inertial integrated solution are not more than 0.01 m in any of the three directions, which is consistent with the vision solution.
In addition, the reprojection errors was calculated to prove the pose accuracy. The corner points on the image that are not used for estimation (rejected by the RANSAC) were projected into the w-frame by the estimated pose, and the errors comparing to the real 3D corners on the horizontal direction were calculated. The results of the first test are shown in Figure 13, and the statistical results of the five experiments are shown in Table 6.
The statistical results show that the corner reprojection error is 0.022 m ( 1 σ ), which also verifies that the position accuracy of the proposed system is at the centimeter level.
For the attitude errors, it can be observed from the statistical results in Table 5 that the roll angle error is 0.30° ( 1 σ ), the pitch angle error is 0.20° ( 1 σ ), and the heading angle error is 0.07° ( 1 σ ). Comparing to the pure visual solution, the roll and pitch angle accuracies have increased by 33% and 54% on average, respectively. Additionally, Figure 14 shows that the pitch angle and roll angle estimated by the integrated navigation are more stable and smooth than those of the visual solution. Due to the excellent sensibility to vehicle motion of the IMU, the horizontal angle vibration caused by the maneuver is alleviated by the fusion with INS.
The attitude and heading reference systems (AHRS) [43,44] is widely used for continuous attitude estimation for flight control of UAVs since it is independent of GNSS. It exploits the IMU for pitch and roll angle estimation and magnetometer augmentation for improving heading stability. Because there was no magnetometer in our test device, we used the IMU data for AHRS and compared the accuracy and stability of the horizontal angles. The Mahony AHRS algorithm [45] was performed in the test. Figure 15 illustrates the roll and pitch errors of the two systems in the fourth test.
It can be observed that the roll and pitch angle estimation of the proposed system is much more accurate and stable than that of the AHRS. The AHRS provides pitch and roll angles relative to the earth gravity vector, thus, the accuracy of horizontal angles suffers from the vehicle acceleration since it cannot be separated by IMU only. In the tests, we held the platform to move in circles, therefore, the direction-alternating acceleration resulted in the sinusoidal curve of the angles error of the AHRS. As for the horizontal angles estimated by the proposed system, the error of INS-derived position was related to the attitude error, hence the difference between the high-precision position obtained by marker detection and that of INS could correct the error of roll and pitch angles by the filter.
As aforementioned, INS can provide continuous positioning results when the external aiding source is short-term unavailable. To illustrate that our system is robust to failure in marker extraction, we artificially interrupted the visual correction in data processing and calculated the errors of the positioning results. We set five outages for each test, and each outage was set to 5 s. Figure 16 and Figure 17 are the position and reprojection errors of the simulation experiments in the fifth test, respectively.
The position error in the horizontal and vertical direction for the 25 outages in all the tests is 0.41 and 0.09 m ( 1 σ ), respectively, indicating that the proposed system can provide stable and reliable pose estimation when the marker is invalid occasionally. However, the navigation error of INS is accumulated with time as shown in Figure 17, hence long-term failure of marker detection will decay the performance.
To verify the robustness of the proposed system when the UAV is at a relative higher altitude or large-angle maneuver, we held the test platform 8 m away to capture the marker with no slant and with 40 degrees tilting, respectively. As the distance increases, the marker in the image becomes smaller and the field of view becomes larger, hence objects of different shapes around the marker came into view which caused the debris in the figures. Only the polygon containing correct grayscale information would be matched. Figure 18 presents the contours extraction and marker detection results in the two situations. The positioning errors (RMS) on vertical direction for these two tests are 0.12 and 0.23 m, respectively. The errors on horizontal direction are 0.21 and 0.30 m, respectively. It can be observed that the long distance and large tilt angle did not affect the marker detection. However, the positioning error increased since the marker was small on the image and the geometric condition of corner points was poor. Because the precision landing of the UAV depends primarily on the pose accuracy when it is close to the landing pad, this positioning results at such a distance are enough for the UAV to approach the target.
Finally, we record the time spent on each module of the system over 1000 trials on the Tegra K1 developer kit (NVIDIA, Santa Clara, CA, USA), including a 4-Plus-1 ARM Cortex-A15 CPU. The results are shown in Table 7.
It can be observed form the time consumption in Table 7 that the system runs efficiently. The most time-consuming part is marker detection, which requires a large number of pixel operations, while the other parts take relatively shorter times since they only need to process a small amount of low-dimension data. Overall, the average time consumption of the entire system is 32.26 ms, which is able to meet the real-time control requirements of UAVs.

4. Conclusions

In this paper, an artificial marker and MEMS IMU-based six-DoF pose estimation solution to meet the UAV landing requirements is proposed. The marker is specially designed for UAV landing due to its simplicity to generate and effectiveness to detect. An onboard monocular camera is used to recognize the marker and recover the absolute position and attitude of the UAV in the world frame. Then, this information is fused with IMU data by an EKF, thus our system still works in the case of visual failure, which is a fatal problem for previous systems. Improving upon the previous visual/inertial landing systems, the error components of the inertial sensor are modeled, and the horizontal angles are also estimated. Experiments show that the position accuracy is at centimeter level, the heading error is less than 0.1°, and the roll and pitch angle errors are less than 1° with the UAV at a height of about 1.4 m. The visual/inertial fused system improves the roll and pitch accuracy by 33% and 54%, respectively. Moreover, vision outage simulation tests show our system can provide relatively reliable position estimation (0.41 m drift in the horizontal direction and 0.09 m in the vertical direction in 5 s) when the maker occasionally occluded. Tests in challenging situations prove that the marker can be detected at a distance of 8 m and tilt of 40 degrees and the positioning accuracy is at decimeter level. The statistics of time consumption of the proposed system in an onboard computer show that it is able to meet the requirement of real-time control. The proposed system also applies to marker-based state estimation of indoor ground vehicles.
Future works will extend pose estimation to the entire route of UAV flight. Natural features are going to be used to estimate the UAV pose, when there is no marker in the field of vision. The marker can be used to correct the accumulated navigation error as long as it is detected.

Author Contributions

Data curation, J.D. and H.T.; formal analysis, L.C.; investigation, J.D. and H.T.; methodology, Y.W. and L.C.; software, Y.W.; supervision, X.N. and H.Z.; writing—original draft, Y.W.; writing—review and editing, X.N.

Funding

This research was funded by the “National Key Research and Development Program of China” (No. 2018YFB1305001, No. 2016YFB0502202, and No. 2016YFB0501803).

Acknowledgments

We thank Jian Kuang and Qijin Chen for their constructive suggestions for the experiment’s design and data process.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Grenzdörffer, G.; Teichert, B. The photogrammetric potential of low-cost UAVs in forestry and agriculture. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2008, 31, 1207–1214. [Google Scholar]
  2. Herwitz, S.; Johnson, L.F.; Dunagan, S.E.; Higgins, R.G.; Sullivan, D.; Zheng, V.J.; Lobitz, B.M.; Leung, J.G.; Gallmeyer, B.A.; Aoyagi, M.; et al. Imaging from an unmanned aerial vehicle: Agricultural surveillance and decision support. Comput. Electron. Agric. 2004, 44, 49–61. [Google Scholar] [CrossRef]
  3. Waharte, S.; Trigoni, N. Supporting search and rescue operations with UAVs. In Proceedings of the International Conference on Emerging Security Technologies, Canterbury, UK, 6–7 September 2010; pp. 142–147. [Google Scholar]
  4. Barmpounakis, E.N.; Vlahogianni, E.I.; Golias, J.C. Unmanned Aerial Aircraft Systems for transportation engineering: Current practice and future challenges. Int. J. Transp. Sci. Technol. 2016, 5, 111–122. [Google Scholar] [CrossRef]
  5. Liu, Y.; Wang, Q.; Zhuang, Y.; Hu, H. A Novel trail detection and scene understanding framework for a quadrotor UAV with monocular vision. IEEE Sens. J. 2017, 17, 6778–6787. [Google Scholar] [CrossRef]
  6. Merz, T.; Duranti, S.; Conte, G. Autonomous landing of an unmanned helicopter based on vision and inertial sensing. In Experimental Robotics, 9th ed.; Springer: Berlin, Germany, 2006; pp. 343–352. [Google Scholar]
  7. Groves, P.D. Principles of GNSS, Inertial, Multisensor Integrated Navigation Systems; Artech House: Norwood, MA, USA, 2013. [Google Scholar]
  8. Nemra, A.; Aouf, N. Robust INS/GPS sensor fusion for UAV localization using SDRE nonlinear filtering. IEEE Sens. J. 2010, 10, 789–798. [Google Scholar] [CrossRef] [Green Version]
  9. Falco, G.; Gutiérrez, M.C.C.; Serna, E.P.; Zacchello, F.; Bories, S. Low-cost real-time tightly-coupled GNSS/INS navigation system based on carrier-phase double-differences for UAV applications. In Proceedings of the 27th International Technical Meeting of the Satellite Division of The Institute of Navigation (ION GNSS 2014), Tampa, FL, USA, 8–12 September 2014; p. 841857. [Google Scholar]
  10. Dabove, P.; Di Pietra, V. Single-baseline RTK positioning using dual-frequency GNSS receivers inside smartphones. Sensors 2019, 19, 4302. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Robustelli, U.; Baiocchi, V.; Pugliano, G. Assessment of dual frequency GNSS observations from a Xiaomi Mi 8 Android smartphone and positioning performance analysis. Electronics 2019, 8, 91. [Google Scholar] [CrossRef] [Green Version]
  12. Balamurugan, G.; Valarmathi, J.; Naidu, V.P.S. survey on UAV navigation in GPS denied environments. In Proceedings of the International Conference on Signal Processing, Communication, Power and Embedded System (SCOPES), Odisha, India, 3–5 October 2016; pp. 198–204. [Google Scholar]
  13. Shen, S.; Mulgaonkar, Y.; Michael, N.; Kumar, V. Vision-based state estimation for autonomous rotorcraft MAVs in complex environments. In Proceedings of the IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 1758–1764. [Google Scholar]
  14. Gautam, A.; Saripalli, S.P.B.S. A survey of autonomous landing techniques for UAVs. In Proceedings of the International Conference on Unmanned Aircraft Systems (ICUAS), Orlando, FL, USA, 27–30 May 2014; pp. 1210–1218. [Google Scholar]
  15. Fiala, M. ARTag, a fiducial marker system using digital techniques. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 590–596. [Google Scholar]
  16. Olson, E. AprilTag: A robust and flexible visual fiducial system. In Proceedings of the IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 3400–3407. [Google Scholar]
  17. Garrido-Jurado, S.; Muñoz-Salinas, R.; Madrid-Cuevas, F.J.; Marín-Jiménez, M.J. Automatic generation and detection of highly reliable fiducial markers under occlusion. Pattern Recognit. 2014, 47, 2280–2292. [Google Scholar] [CrossRef]
  18. Ling, K. Precision Landing of a Quadrotor UAV on a Moving Target Using Low-Cost Sensors. Master’s Thesis, University of Waterloo, Waterloo, ON, Canada, September 2014. [Google Scholar]
  19. Kyristsis, S.; Antonopoulos, A.; Chanialakis, T.; Stefanakis, E.; Linardos, C.; Tripolitsiotis, A.; Partsinevelos, P. Towards autonomous modular UAV missions: The detection, geo-location and landing paradigm. Sensors 2016, 16, 1844. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Jayatilleke, L.; Zhang, N. Landmark-based localization for Unmanned Aerial Vehicles. In Proceedings of the IEEE International Systems Conference (SysCon), Orlando, FL, USA, 15–18 April 2013; pp. 448–451. [Google Scholar]
  21. Millet, P.; Ready, B.; McLain, T. Vision-based precision landings of a tailsitter UAV. In Proceedings of the AIAA Guidance, Navigation, Control Conference, Chicago, IL, USA, 10–13 August 2009. [Google Scholar]
  22. Pan, X.; Ma, D.; Jin, L.; Jiang, Z. Vision-Based Approach Angle and Height Estimation for UAV Landing. In Proceedings of the Congress on Image and Signal Processing, Hainan, China, 27–30 May 2008; pp. 801–805. [Google Scholar]
  23. Acuna, R.; Willert, V. Dynamic Markers: UAV landing proof of concept. In Proceedings of the Latin American Robotic Symposium, 2018 Brazilian Symposium on Robotics (SBR) and 2018 Workshop on Robotics in Education (WRE), Joao Pessoa, Brazil, 6–10 November 2018; pp. 496–502. [Google Scholar]
  24. Yang, S.; Scherer, S.A.; Zell, A. An onboard monocular vision system for autonomous takeoff, hovering and landing of a micro aerial vehicle. J. Intell. Robot. Syst. 2013, 69, 499–515. [Google Scholar] [CrossRef] [Green Version]
  25. Araar, O.; Aouf, N.; Vitanov, I. Vision based autonomous landing of multirotor UAV on moving platform. J. Intell. Robot. Syst. 2016, 85, 369–384. [Google Scholar] [CrossRef]
  26. Lange, S.; Sunderhauf, N.; Protzel, P. A vision based onboard approach for landing and position control of an autonomous multirotor UAV in GPS-denied environments. In Proceedings of the International Conference on Advanced Robotics, Munich, Germany, 22–26 June 2009; pp. 1–6. [Google Scholar]
  27. Nguyen, P.H.; Kim, K.W.; Lee, Y.W.; Park, K.R. Remote marker-based tracking for UAV landing using visible-light camera sensor. Sensors 2017, 17, 1987. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Nguyen, P.H.; Arsalan, M.; Koo, J.H.; Naqvi, R.A.; Truong, N.Q.; Park, K.R. LightDenseYOLO: A fast and accurate marker tracker for autonomous UAV landing by visible light camera sensor on drone. Sensors 2018, 18, 1703. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Benini, A.; Rutherford, M.J.; Valavanis, K.P. Real-time, GPU-based pose estimation of a UAV for autonomous takeoff and landing. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 3463–3470. [Google Scholar]
  30. Otsu, N. A Threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef] [Green Version]
  31. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, PAMI-8, 679–698. [Google Scholar] [CrossRef]
  32. Shi, J.; Tomasi, C. Good features to track. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 21–23 June 1994; pp. 593–600. [Google Scholar]
  33. Bouguet, J.-Y. Pyramidal implementation of the affine lucas kanade feature tracker description of the algorithm. Intel Corp. 2001, 5, 4. [Google Scholar]
  34. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef] [Green Version]
  35. Lepetit, V.; Moreno-Noguer, F.; Fua, P. EPnP: An accurate O(n) solution to the PnP problem. Int. J. Comput. Vis. 2009, 81, 155. [Google Scholar] [CrossRef] [Green Version]
  36. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  37. Triggs, B.; McLauchlan, P.F.; Hartley, R.I.; Fitzgibbon, A.W. Bundle adjustment—A modern synthesis. In Proceedings of the International Workshop on Vision Algorithms, Corfu, Greece, 21–22 September 1999; pp. 298–372. [Google Scholar]
  38. Agarwal, S.; Mierle, K. Ceres Solver. Available online: http://ceres-solver.org (accessed on 6 November 2019).
  39. Shin, E.H. Estimation Techniques for Low-Cost Inertial Navigation. Ph.D. Thesis, Department of Geomatics Engineering, The University of Calgary, Calgary, AB, Canada, May 2005. [Google Scholar]
  40. Shin, E.-H. Accuarcy Improvement of Low Cost INS/GPS for Land Applications. Master’s Thesis, Department of Geomatics Engineering, University of Calgary, Calgary, AB, Canada, December 2001. [Google Scholar]
  41. Furgale, P.; Rehder, J.; Siegwart, R. Unified temporal and spatial calibration for multi-sensor systems. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 1280–1286. [Google Scholar]
  42. Teunissen, P.; Montenbruck, O. Springer Handbook of Global Navigation Satellite Systems; Springer: Berlin, Germany, 2017. [Google Scholar]
  43. Geiger, W.; Bartholomeyczik, J.; Breng, U.; Gutmann, W.; Hafen, M.; Handrich, E.; Huber, M.; Jackle, A.; Kempfer, U.; Kopmann, H.; et al. MEMS IMU for AHRS applications. In Proceedings of the IEEE/ION Position, Location and Navigation Symposium, Portland, OR, USA, 20–23 April 2008; pp. 225–231. [Google Scholar]
  44. Madgwick, S.O.H.; Harrison, A.J.L.; Vaidyanathan, R. Estimation of IMU and MARG orientation using a gradient descent algorithm. In Proceedings of the IEEE international conference on rehabilitation robotics, Toronto, ON, Canada, 24–28 June 2011; pp. 1–7. [Google Scholar]
  45. Mahony, R.; Hamel, T.; Pflimlin, J. Nonlinear complementary filters on the special orthogonal group. IEEE Trans. Autom. Control 2008, 53, 1203–1218. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Examples of the proposed marker and binary encoding matrix.
Figure 1. Examples of the proposed marker and binary encoding matrix.
Sensors 19 05428 g001
Figure 2. Flow chart of pose estimation by detecting the marker.
Figure 2. Flow chart of pose estimation by detecting the marker.
Sensors 19 05428 g002
Figure 3. Image processing results. (a) Original image; (b) binary image after Otsu binarization; (c) polygon contours extracted in (b); (d) quadrilateral contours and the matched marker; (e) the corner points extracted on the marker; (f) the extracted corner points of the central marker when the unmanned air vehicle (UAV) is very close to the ground.
Figure 3. Image processing results. (a) Original image; (b) binary image after Otsu binarization; (c) polygon contours extracted in (b); (d) quadrilateral contours and the matched marker; (e) the corner points extracted on the marker; (f) the extracted corner points of the central marker when the unmanned air vehicle (UAV) is very close to the ground.
Sensors 19 05428 g003
Figure 4. Definition of coordinate systems and camera projection model. O c is the optic center of the camera and O c X c Y c Z c denotes the c-frame; O p u v denotes the p-frame; and O w X w Y w Z w denotes the w-frame.
Figure 4. Definition of coordinate systems and camera projection model. O c is the optic center of the camera and O c X c Y c Z c denotes the c-frame; O p u v denotes the p-frame; and O w X w Y w Z w denotes the w-frame.
Sensors 19 05428 g004
Figure 5. System overview of the camera/INS (inertial navigation system) integrated navigation. The vehicle pose is retrieved by the two sensors independently and then fused by error-state extended Kalman filter (EKF), namely loosely coupled.
Figure 5. System overview of the camera/INS (inertial navigation system) integrated navigation. The vehicle pose is retrieved by the two sensors independently and then fused by error-state extended Kalman filter (EKF), namely loosely coupled.
Sensors 19 05428 g005
Figure 6. Installation relationship between the camera and inertial measurement unit (IMU).
Figure 6. Installation relationship between the camera and inertial measurement unit (IMU).
Sensors 19 05428 g006
Figure 7. Experimental equipment.
Figure 7. Experimental equipment.
Sensors 19 05428 g007
Figure 8. Trajectory of the test platform in test 05.
Figure 8. Trajectory of the test platform in test 05.
Sensors 19 05428 g008
Figure 9. Attitude variation of the test platform in test 05.
Figure 9. Attitude variation of the test platform in test 05.
Sensors 19 05428 g009
Figure 10. Experimental environment.
Figure 10. Experimental environment.
Sensors 19 05428 g010
Figure 11. Position errors of the visual solution in test 03.
Figure 11. Position errors of the visual solution in test 03.
Sensors 19 05428 g011
Figure 12. Attitude errors of the visual solution in test 03.
Figure 12. Attitude errors of the visual solution in test 03.
Sensors 19 05428 g012
Figure 13. Corner points reprojection errors on the horizontal plane in test 01.
Figure 13. Corner points reprojection errors on the horizontal plane in test 01.
Sensors 19 05428 g013
Figure 14. Comparison of roll and pitch errors by the visual solution and visual/inertial solution in test 04.
Figure 14. Comparison of roll and pitch errors by the visual solution and visual/inertial solution in test 04.
Sensors 19 05428 g014
Figure 15. Comparison of roll and pitch errors by the attitude and heading reference system (AHRS) solution and visual/inertial solution.
Figure 15. Comparison of roll and pitch errors by the attitude and heading reference system (AHRS) solution and visual/inertial solution.
Sensors 19 05428 g015
Figure 16. Position errors of the visual-inertial solution in the vision outage simulation test.
Figure 16. Position errors of the visual-inertial solution in the vision outage simulation test.
Sensors 19 05428 g016
Figure 17. Reprojection error in the vision outage simulation test.
Figure 17. Reprojection error in the vision outage simulation test.
Sensors 19 05428 g017
Figure 18. Edge extraction and marker detection results in challenging situations. (a) Detected marker at a distance of 8 m with no tilt; (b) detected marker at a distance of 8 m with tilt of 40 degrees.
Figure 18. Edge extraction and marker detection results in challenging situations. (a) Detected marker at a distance of 8 m with no tilt; (b) detected marker at a distance of 8 m with tilt of 40 degrees.
Sensors 19 05428 g018
Table 1. Technical parameters of the micro-electromechanical system (MEMS) IMU and POS320.
Table 1. Technical parameters of the micro-electromechanical system (MEMS) IMU and POS320.
ParametersMEMS IMU
(System under Test)
POS320
(Reference System)
Gyro bias stability ( ° / h ) 2000.5
Angle random work ( ° / h ) 0.240.05
Accelerometer bias stability ( m / s 2 ) 0.010.00025
Velocity random work ( m / s / h ) 30.1
Table 2. Statistical results of the position errors of the visual solution.
Table 2. Statistical results of the position errors of the visual solution.
Statistics0102030405
North Error (m)RMSE0.0070.0060.0080.0060.006
MAX0.0240.0290.0250.0210.031
MEAN0.0060.0040.0070.0050.004
East Error (m)RMSE0.0080.0090.0100.0080.007
MAX0.0250.0250.0240.0340.019
MEAN0.0070.0070.0080.0070.006
Down Error (m)RMSE0.0030.0030.0040.0040.003
MAX0.0130.0100.0130.0120.014
MEAN0.0030.0030.0030.0030.002
RMSE: root mean square error; MAX: maximum.
Table 3. Statistical results of the attitude errors of the visual solution.
Table 3. Statistical results of the attitude errors of the visual solution.
Statistics0102030405
Roll Error (°)RMSE0.450.530.540.380.34
MAX1.451.581.751.471.14
MEAN0.370.420.410.310.27
Pitch Error (°)RMSE0.370.490.550.390.35
MAX1.121.331.761.521.05
MEAN0.290.410.450.300.28
Yaw Error (°)RMSE0.050.080.090.050.05
MAX0.150.280.290.130.20
MEAN0.040.050.060.040.04
°: degrees.
Table 4. Statistical results of the position errors by the visual/inertial solution.
Table 4. Statistical results of the position errors by the visual/inertial solution.
Statistics0102030405
North Error (m)RMSE0.0070.0060.0090.0070.006
MAX0.0250.1800.0290.0220.034
MEAN0.0060.0050.0070.0050.005
East Error (m)RMSE0.0090.0090.0100.0080.008
MAX0.0330.0310.0250.0340.023
MEAN0.0070.0070.0080.0070.006
Down Error (m)RMSE0.0040.0040.0050.0040.004
MAX0.0130.0100.0150.0100.012
MEAN0.0030.0030.0030.0040.003
Table 5. Statistical results of the attitude errors by the visual/inertial solution.
Table 5. Statistical results of the attitude errors by the visual/inertial solution.
Statistics0102030405
Roll Error (°)RMSE0.080.440.340.210.43
MAX0.241.661.500.881.86
MEAN0.060.240.200.170.28
Pitch Error (°)RMSE0.170.140.240.270.17
MAX0.830.580.801.030.70
MEAN0.110.080.190.210.12
Yaw Error (°)RMSE0.060.130.070.050.03
MAX0.130.360.320.120.15
MEAN0.050.100.050.040.02
Table 6. Statistical results of the corner reprojection error.
Table 6. Statistical results of the corner reprojection error.
Statistics0102030405
RMSE (m)0.0220.0230.0210.0200.024
MAX (m)0.0420.0490.0490.0330.044
MEAN (m)0.0200.0200.0180.0190.023
Table 7. Running time statistics of different modules of the system.
Table 7. Running time statistics of different modules of the system.
ModuleMAX (ms)MEAN (ms)
Marker detection36.1525.21
Pose estimation by marker9.011.71
Filtering update1.210.25
Total70.3432.26

Share and Cite

MDPI and ACS Style

Wu, Y.; Niu, X.; Du, J.; Chang, L.; Tang, H.; Zhang, H. Artificial Marker and MEMS IMU-Based Pose Estimation Method to Meet Multirotor UAV Landing Requirements. Sensors 2019, 19, 5428. https://doi.org/10.3390/s19245428

AMA Style

Wu Y, Niu X, Du J, Chang L, Tang H, Zhang H. Artificial Marker and MEMS IMU-Based Pose Estimation Method to Meet Multirotor UAV Landing Requirements. Sensors. 2019; 19(24):5428. https://doi.org/10.3390/s19245428

Chicago/Turabian Style

Wu, Yibin, Xiaoji Niu, Junwei Du, Le Chang, Hailiang Tang, and Hongping Zhang. 2019. "Artificial Marker and MEMS IMU-Based Pose Estimation Method to Meet Multirotor UAV Landing Requirements" Sensors 19, no. 24: 5428. https://doi.org/10.3390/s19245428

APA Style

Wu, Y., Niu, X., Du, J., Chang, L., Tang, H., & Zhang, H. (2019). Artificial Marker and MEMS IMU-Based Pose Estimation Method to Meet Multirotor UAV Landing Requirements. Sensors, 19(24), 5428. https://doi.org/10.3390/s19245428

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop