CN107818598B - Three-dimensional point cloud map fusion method based on visual correction - Google Patents
Three-dimensional point cloud map fusion method based on visual correction Download PDFInfo
- Publication number
- CN107818598B CN107818598B CN201710989642.2A CN201710989642A CN107818598B CN 107818598 B CN107818598 B CN 107818598B CN 201710989642 A CN201710989642 A CN 201710989642A CN 107818598 B CN107818598 B CN 107818598B
- Authority
- CN
- China
- Prior art keywords
- point
- point cloud
- points
- cloud map
- neighborhood
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/20—Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/462—Salient features, e.g. scale invariant feature transforms [SIFT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20016—Hierarchical, coarse-to-fine, multiscale or multiresolution image processing; Pyramid transform
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Architecture (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computer Graphics (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
- Processing Or Creating Images (AREA)
Abstract
The invention discloses a three-dimensional point cloud map fusion method based on visual correction, which comprises the following steps: 1) processing two point cloud maps to be fused; 2) extracting 3D-SIFT key points of two three-dimensional point cloud maps; 3) extracting IPFH (IPFH) features on the 3D-SIFT key points; 4) searching for feature matching points by calculating Euclidean distances between feature points; 5) calculating a conversion matrix and rotating a point cloud map; 6) and fusing the two point cloud maps together by adopting an ICP algorithm. According to the method, SIFT features are expanded into three-dimensional point cloud, and 3D-SIFT key points are extracted, so that the robustness of feature to view angle rotation and transformation is guaranteed; by extracting the IPFH characteristics, the problem of wrong weight coefficient of the original FPFH characteristics is solved, and meanwhile, the characteristics of a three-dimensional point are represented by synthesizing the geometric characteristics of the neighborhood points, so that the stability of the algorithm is greatly improved. Through the processing, the method can fuse two three-dimensional point cloud maps with large visual angle difference.
Description
Technical Field
The invention relates to the field of computer vision, mainly relates to three-dimensional point cloud map fusion, and particularly provides a method based on vision correction, which can be used for solving the problem of point cloud map fusion failure caused by overlarge visual angle difference.
Background
With the rapid development of stereo cameras, three-dimensional point cloud data has been widely applied in the field of computer vision, such as robot map construction, navigation, object recognition, and tracking. When the robot is in a wide external environment or a complex internal environment, although SLAM map construction can solve autonomous navigation of the robot, various problems may occur when the map is constructed due to complexity and universality of the environment, for example, navigation fails due to too long time consumed in the map construction process. Therefore, in this case, a plurality of robots are required to construct a map together, and then the respective maps are spliced into a complete map, so that the map sharing is realized, and the autonomous navigation efficiency can be improved.
The point cloud map fusion is as follows: and fusing point clouds from multiple visual angles together by using point cloud maps collected from different visual angles and different positions through rigid body transformation. Besl, P.J. & McKay, N.d. (1992). A method for registration of 3-D maps. IEEE Transactions on Pattern Analysis and Machine understanding, 14(2), 239-. The algorithm can obtain a better fusion result only when an accurate initial value is provided.
In order to calculate an accurate initial value, many researchers calculate the initial value by finding a feature matching. Klasing et al used methods based on optimized averaging to estimate normal vectors, such as the PlanePCA, PlaneSVD, QuadSVD methods. However, the normal vector of the point cloud contains less data information, and the change of the point cloud map view angle can cause the calculation of the normal vector of the point cloud, and finally, the fusion effect of the point cloud map is influenced. Gelfand provides a method for extracting a point cloud volume, however, the method cannot adapt to the point cloud map view transformation. Zou calculates the initial value by extracting the point cloud gaussian curvature, but the gaussian curvature of the same object changes with the change of the point cloud map view angle, which may result in wrong feature matching, and thus the point cloud map fusion fails. Rusu et al propose a method for extracting point cloud FPFH characteristics, but this method has the problems of weight coefficient overflow and characteristic matching error. The RANSAC algorithm is another fusion algorithm, originally proposed by fisherler et al, and adopts random sampling consistency to match a mathematical model, and the algorithm is stable but the complexity of the method can reach O (n3) in the worst case, and the algorithm time consumption is unacceptable when the number of point clouds is large.
Disclosure of Invention
The invention aims to provide a three-dimensional point cloud map fusion method based on visual correction, which is high in fusion precision and high in efficiency and aims to solve the problems in the background technology.
In order to achieve the purpose, the invention provides the following technical scheme:
a three-dimensional point cloud map fusion method based on visual correction comprises the following steps:
1) preprocessing a three-dimensional point cloud map, firstly inputting two three-dimensional point cloud map files in a pcd format so as to obtain two point cloud maps to be fused, wherein the two point cloud maps are respectively target point cloud map pointstgtAnd point cloud map point to be fusedsrcThen, preprocessing the two point cloud maps;
2) respectively extracting point of the target point cloud map for the two point cloud maps in the step 1)tgtAnd point cloud map point to be fusedsrc3D-SIFT key points of (1);
3) respectively extracting IPFH (Internet protocol video frequency) features from the two point cloud map 3D-SIFT key points in the step 2) to obtain point cloud map points to be fusedsrcFeature vector P and target point cloud map pointtgtA feature vector Q, wherein the feature vector P has m feature points, and the feature vector Q has n feature points;
4) calculating Euclidean distances of feature points in the feature vector P and the feature vector Q, and finding out feature matching points according to the minimum distance;
5) calculating a rotation matrix through the feature matching points, and using the matrix to map point of the point cloudtgtRotating to change the visual angle of the point cloud map and reduce the visual angle difference of the two point cloud maps;
6) and fusing the two three-dimensional point cloud maps with reduced view angle difference by adopting an ICP (inductively coupled plasma) accurate fusion algorithm.
As a further scheme of the invention: in step 1), the preprocessing includes denoising and sampling.
As a still further scheme of the invention: in the step 2), the specific steps of extracting the 3D-SIFT key points are as follows:
detecting extreme points in a point cloud scale space, carrying out downsampling on point cloud maps to different degrees to construct a pyramid model, and constructing a Gaussian scale space for each group of point cloud maps, wherein the functions are as follows
Scale space: l (x, y, z, σ) ═ G (x, y, z, k σ) × P (x, y, z);
the parameter k points to the number of groups in the pyramid, and σ is the size of the scale;
constructing a Gaussian difference space: d (x, y, z, k)iσ)=L(x,y,z,ki+1σ)-L(x,y,z,kiσ);
After a DoG space is constructed, searching an extreme point by a comparison method: regarding a certain point, when the value of the certain point is larger than or smaller than all the points in the neighborhood, the certain point is regarded as an extreme point;
II, determining the direction of the key point, and calculating the azimuth angle and the elevation angle of each point in the neighborhood where the extreme point is located, wherein the calculation formula is as follows:
in the formula, d represents the distance between the extreme point and the center point of the neighborhood, theta represents the azimuth angle, phi represents the elevation angle, and x represents0,y0,z0Refers to the center point in the neighborhood of the extreme point;
counting the direction of the azimuth angle theta and the elevation angle phi of each point in the neighborhood: counting the directions of two angles of each point by adopting two histograms, wherein the direction dereferencing range of theta is [0 degrees and 360 degrees ], and counting the direction by adopting an 8-column histogram, wherein each column represents 45 degrees; and phi is taken as [ -90 degrees, 90 degrees ], the direction condition is counted by adopting a 4-column histogram, and each column represents 45 degrees;
after statistics, taking the peak values of the two histograms as the azimuth angle and the elevation angle of the extreme point, and after calculation, each key point comprises x, y and z coordinate information, scale information and direction information, namely (x, y, z, sigma, theta and phi);
III, rotating the coordinate system of a certain key point p to the main direction of the key point, and calculating by adopting the following rotation formula because the main direction is already calculated
Wherein p isiAs a point before rotation, pi' represents a point after rotation;
using rotated point pi'recalculating the several parameters mentioned above, including d, θ, φ, and then recalculating the rotated keypoint p' and the neighborhood point pi'geometric relationship between them, the geometric relationship being referred to as a vector p' pi'an angle with a normal vector n at a key point p before rotation is calculated using the formula where a numerator represents a dot product of two vectors and a denominator represents a length of the two vectors, and a triplet of data information (theta, phi,') is generated for a key point and all neighbor points in the neighborhood
And dividing three angles of the plurality of triples to generate a feature descriptor.
As a still further scheme of the invention: in step 3), the specific steps of extracting the IPFH characteristics in the region around the 3D-SIFT key point are as follows:
determining a k neighborhood by taking a 3D-SIFT key point p (x, y, z, sigma, theta and phi) as a center, constructing point pairs in the neighborhood by pairwise interconnection, and calculating a normal vector of each group of point pairs;
and II, constructing a local u-v-w coordinate system between each group of point pairs, and defining three coordinate axes in the following mode:
four characteristic values are calculated according to the local coordinate system and the normal vector, the calculation method is shown as the following formula, and the four characteristic values are called SPFH
And III, reselecting a k neighborhood from the two point cloud maps, and synthesizing the IPFH by utilizing the SPFH (specific pathogen free) close to each three-dimensional point, wherein the synthesis method is as follows:
ω0refers to the average, ω, of the distance between the query point p and each of the neighborhood pointsiRepresents a query point p and a neighboring point p in the neighborhood spaceiThe distance between them;
and IV, counting the value taking conditions of f1, f3 and f4 of each point, dividing each value into 11 intervals, and describing the IPFH by adopting 33-dimensional data information.
As a still further scheme of the invention: in step 4), IPFH (Internet protocol flash) feature matching is carried out, the corresponding relation is determined, the judgment standard is the distance between feature vectors, and the following matching strategy is adopted, namely, the point cloud map point is takensrcIPFH of a point A, point cloud map pointtgtFinding two points B and C with the nearest distance to the vector A, calculating by the following formula, if the formula condition is satisfied, selecting B as the matching point thereof
As a still further scheme of the invention: in step 5), after the characteristic matching points are found, calculating a conversion matrix by adopting a singular value decomposition method, and carrying out point cloud map point according to the conversion matrixsrcAnd rotating to reduce the visual angle difference between the two point cloud maps and finish the visual angle correction.
Compared with the prior art, the invention has the beneficial effects that:
1. according to the method, the 3D-SIFT is used for extracting key points, the FPFH (field programmable gate noise) features are calculated on the key points, and the initial rough estimation is calculated finally, so that an initial transformation matrix is provided for the traditional point cloud registration method, the iteration times of the registration algorithm can be reduced, the registration success rate and precision of the algorithm are improved, and the calculation cost is reduced.
2. According to the method, SIFT features are expanded into three-dimensional point cloud, and 3D-SIFT key points are extracted, so that the robustness of feature to view angle rotation and transformation is guaranteed; by extracting the IPFH characteristics, the problem of wrong weight coefficient of the original FPFH characteristics is solved, and meanwhile, the characteristics of a three-dimensional point are represented by synthesizing the geometric characteristics of the neighborhood points, so that the stability of the algorithm is greatly improved. Through the processing, the method can fuse two three-dimensional point cloud maps with large visual angle difference.
Drawings
Fig. 1 is a schematic flow chart of a three-dimensional point cloud map fusion method based on visual correction.
FIG. 2 is a schematic diagram of extracting key points by using 3D-SIFT in a three-dimensional point cloud map fusion method based on vision correction.
Fig. 3 is a schematic diagram of feature matching using FPFH in a three-dimensional point cloud map fusion method based on visual rectification.
Fig. 4 is an effect diagram of point cloud fusion performed by the three-dimensional point cloud map fusion method based on visual correction.
Detailed Description
The technical solution of the present invention will be described in further detail with reference to specific embodiments.
Referring to fig. 1, a three-dimensional point cloud map fusion method based on vision correction includes the following steps:
1) preprocessing a three-dimensional point cloud map, firstly inputting two three-dimensional point cloud map files in a pcd format so as to obtain two point cloud maps to be fused, wherein the two point cloud maps are respectively target point cloud map pointstgtAnd point cloud map point to be fusedsrcThen, preprocessing the two point cloud maps, wherein the preprocessing comprises denoising and sampling;
2) respectively extracting point of the target point cloud map for the two point cloud maps in the step 1)tgtAnd point cloud map point to be fusedsrc3D-SIFT key points of (1);
the specific steps of 3D-SIFT key point extraction are as follows:
i, detecting extreme points in a point cloud scale space, carrying out downsampling on point cloud maps to different degrees to construct a pyramid model, constructing a Gaussian scale space for each group of point cloud maps, wherein the used functions are as follows,
scale space: l (x, y, z, σ) ═ G (x, y, z, k σ) × P (x, y, z);
the parameter k points to the number of groups in the pyramid, and σ is the size of the scale;
constructing a Gaussian difference space: d (x, y, z, k)iσ)=L(x,y,z,ki+1σ)-L(x,y,z,kiσ);
After a DoG space is constructed, searching an extreme point by a comparison method: regarding a certain point, when the value of the certain point is larger than or smaller than all the points in the neighborhood, the certain point is regarded as an extreme point;
II, determining the direction of the key point, and calculating the azimuth angle and the elevation angle of each point in the neighborhood where the extreme point is located, wherein the calculation formula is as follows:
in the formula, d represents the distance between the extreme point and the center point of the neighborhood, theta represents the azimuth angle, phi represents the elevation angle, and x represents0,y0,z0Refers to the center point in the neighborhood of the extreme point;
counting the direction of the azimuth angle theta and the elevation angle phi of each point in the neighborhood: counting the directions of two angles of each point by adopting two histograms, wherein the direction dereferencing range of theta is [0 degrees and 360 degrees ], and counting the direction by adopting an 8-column histogram, wherein each column represents 45 degrees; and phi is taken as [ -90 degrees, 90 degrees ], the direction condition is counted by adopting a 4-column histogram, and each column represents 45 degrees;
after statistics, taking the peak values of the two histograms as the azimuth angle and the elevation angle of the extreme point, and after calculation, each key point comprises x, y and z coordinate information, scale information and direction information, namely (x, y, z, sigma, theta and phi);
III, rotating the coordinate system of a certain key point p to the main direction of the key point, and calculating by adopting the following rotation formula because the main direction is already calculated
Wherein p isiAs a point before rotation, pi' represents a point after rotation;
using rotated point pi'recalculating the several parameters mentioned above, including d, θ, φ, and then recalculating the rotated keypoint p' and the neighborhood point pi'geometric relationship between them, the geometric relationship being referred to as a vector p' pi'an angle with a normal vector n at a key point p before rotation is calculated using the formula where a numerator represents a dot product of two vectors and a denominator represents a length of the two vectors, and a triplet of data information (theta, phi,') is generated for a key point and all neighbor points in the neighborhood
Dividing three angles of a plurality of triples to generate a feature descriptor;
3) respectively extracting IPFH (Internet protocol video frequency) features from the two point cloud map 3D-SIFT key points in the step 2) to obtain point cloud map points to be fusedsrcFeature vector P and target point cloud map pointtgtA feature vector Q, wherein the feature vector P has m feature points, and the feature vector Q has n feature points;
the specific steps of extracting the IPFH characteristics in the region around the 3D-SIFT key point are as follows:
determining a k neighborhood by taking a 3D-SIFT key point p (x, y, z, sigma, theta and phi) as a center, constructing point pairs in the neighborhood by pairwise interconnection, and calculating a normal vector of each group of point pairs;
and II, constructing a local u-v-w coordinate system between each group of point pairs, and defining three coordinate axes in the following mode:
four characteristic values are calculated according to the local coordinate system and the normal vector, the calculation method is shown as the following formula, and the four characteristic values are called SPFH
And III, reselecting a k neighborhood from the two point cloud maps, and synthesizing the IPFH by utilizing the SPFH (specific pathogen free) close to each three-dimensional point, wherein the synthesis method is as follows:
ω0refers to the average, ω, of the distance between the query point p and each of the neighborhood pointsiRepresents a query point p and a neighboring point p in the neighborhood spaceiThe distance between them;
IV, counting the value conditions of f1, f3 and f4 of each point, dividing each value into 11 intervals, and describing the IPFH by adopting 33-dimensional data information;
4) calculating Euclidean distances of feature points in the feature vector P and the feature vector Q, and finding out feature matching points according to the minimum distance;
IPFH characteristic matching, determining corresponding relation, judging the distance between characteristic vectors as a standard, and adopting a matching strategy of taking point cloud map pointsrcIPFH of a point A, point cloud map pointtgtFinding two points B and C with the nearest distance to the A vector, passing throughCalculating by a surface formula, if formula conditions are met, selecting B as a matching point thereof
5) Calculating a rotation matrix through the feature matching points, and using the matrix to map point of the point cloudtgtRotating to change the visual angle of the point cloud map and reduce the visual angle difference of the two point cloud maps;
after finding the characteristic matching points, calculating a conversion matrix by adopting a singular value decomposition method, and carrying out point cloud map point according to the conversion matrixsrcRotating to reduce the visual angle difference between the two point cloud maps and finish the visual angle correction;
6) and fusing the two three-dimensional point cloud maps with reduced view angle difference by adopting an ICP (inductively coupled plasma) accurate fusion algorithm.
In fig. 2, an effect diagram of extracting key points by using 3D-SIFT in a three-dimensional point cloud map fusion method based on visual rectification is shown. The SIFT feature is an image local feature which keeps certain stability to rotation, scale scaling, brightness change invariance, view angle change, affine transformation and noise.
In fig. 3, a schematic diagram of further extracting FPFH features on key points extracted by 3D-SIFT and performing feature matching in a three-dimensional point cloud map fusion method based on visual correction is shown. FPFH is an improvement to PFH characteristics, is a point characteristic representation method for describing local geometric characteristic information around a sample point in the form of a statistical histogram, calculates Euclidean distances of characteristic points in source and target point clouds, finds characteristic matching points according to the minimum distance, and calculates a rotation matrix through the characteristic matching points.
In fig. 4, the test and experimental results in the three-dimensional point cloud map fusion method based on visual correction are shown, the experiment employs three sets of data sets, wherein the first two sets of data are from indoor point cloud data acquired by using an x tion Pro stereo sensor; another set is from the outdoor partial city point cloud provided by the PCL collected by the 3D laser scanner. The first column and the second column of the picture respectively represent source point clouds and target point clouds, the third column and the fourth column are point clouds after fusion, and the result after the two point clouds are fused is represented by different colors in the fourth column.
According to the method, the 3D-SIFT is used for extracting key points, the FPFH (field programmable gate noise) features are calculated on the key points, and the initial rough estimation is calculated finally, so that an initial transformation matrix is provided for the traditional point cloud registration method, the iteration times of the registration algorithm can be reduced, the registration success rate and precision of the algorithm are improved, and the calculation cost is reduced.
According to the method, SIFT features are expanded into three-dimensional point cloud, and 3D-SIFT key points are extracted, so that the robustness of feature to view angle rotation and transformation is guaranteed; by extracting the IPFH characteristics, the problem of wrong weight coefficient of the original FPFH characteristics is solved, and meanwhile, the characteristics of a three-dimensional point are represented by synthesizing the geometric characteristics of the neighborhood points, so that the stability of the algorithm is greatly improved. Through the processing, the method can fuse two three-dimensional point cloud maps with large visual angle difference.
While the preferred embodiments of the present invention have been described in detail, the present invention is not limited to the above embodiments, and various changes can be made without departing from the spirit of the present invention within the knowledge of those skilled in the art.
Claims (5)
1. A three-dimensional point cloud map fusion method based on visual correction is characterized by comprising the following steps:
1) preprocessing a three-dimensional point cloud map, firstly inputting two three-dimensional point cloud map files in a pcd format so as to obtain two point cloud maps to be fused, wherein the two point cloud maps are respectively target point cloud map pointstgtAnd point cloud map point to be fusedsrcThen, preprocessing the two point cloud maps;
2) respectively extracting point of the target point cloud map for the two point cloud maps in the step 1)tgtAnd point cloud map point to be fusedsrc3D-SIFT key points of (1);
3) for two points in step 2)Respectively extracting IPFH (internet protocol video frequency) features from the cloud map 3D-SIFT key points to obtain point cloud map points to be fusedsrcFeature vector P and target point cloud map pointtgtA feature vector Q, wherein the feature vector P has m feature points, and the feature vector Q has n feature points;
4) calculating Euclidean distances of feature points in the feature vector P and the feature vector Q, and finding out feature matching points according to the minimum distance;
5) calculating a rotation matrix through the feature matching points, and using the matrix to map point of the point cloudtgtRotating to change the visual angle of the point cloud map and reduce the visual angle difference of the two point cloud maps;
6) fusing the two three-dimensional point cloud maps with reduced view angle difference by adopting an ICP (inductively coupled plasma) accurate fusion algorithm;
in step 3), the specific steps of extracting the IPFH characteristics in the region around the 3D-SIFT key point are as follows:
determining a k neighborhood by taking a 3D-SIFT key point p (x, y, z, sigma, theta and phi) as a center, constructing point pairs in the neighborhood by pairwise interconnection, and calculating a normal vector of each group of point pairs;
and II, constructing a local u-v-w coordinate system between each group of point pairs, and defining three coordinate axes in the following mode:
four characteristic values are calculated according to the local coordinate system and the normal vector, the calculation method is shown as the following formula, and the four characteristic values are called SPFH
And III, reselecting a k neighborhood from the two point cloud maps, and synthesizing the IPFH by utilizing the SPFH (specific pathogen free) close to each three-dimensional point, wherein the synthesis method is as follows:
ω0refers to the average, ω, of the distance between the query point p and each of the neighborhood pointsiRepresents a query point p and a neighboring point p in the neighborhood spaceiThe distance between them;
and IV, counting the value taking conditions of f1, f3 and f4 of each point, dividing each value into 11 intervals, and describing the IPFH by adopting 33-dimensional data information.
2. The three-dimensional point cloud map fusion method based on visual correction as claimed in claim 1, wherein in step 1), the preprocessing includes denoising and sampling.
3. The three-dimensional point cloud map fusion method based on visual correction as claimed in claim 1, wherein in the step 2), the specific steps of 3D-SIFT key point extraction are as follows:
detecting extreme points in a point cloud scale space, carrying out downsampling on point cloud maps to different degrees to construct a pyramid model, and constructing a Gaussian scale space for each group of point cloud maps, wherein the functions are as follows
Scale space: l (x, y, z, σ) ═ G (x, y, z, k σ) × P (x, y, z);
the parameter k points to the number of groups in the pyramid, and σ is the size of the scale;
constructing a Gaussian difference space: d (x, y, z, k)iσ)=L(x,y,z,ki+1σ)-L(x,y,z,kiσ);
After a DoG space is constructed, searching an extreme point by a comparison method: regarding a certain point, when the value of the certain point is larger than or smaller than all the points in the neighborhood, the certain point is regarded as an extreme point;
II, determining the direction of the key point, and calculating the azimuth angle and the elevation angle of each point in the neighborhood where the extreme point is located, wherein the calculation formula is as follows:
in the formula, d represents the distance between the extreme point and the center point of the neighborhood, theta represents the azimuth angle, phi represents the elevation angle, and x represents0,y0,z0Refers to the center point in the neighborhood of the extreme point;
counting the direction of the azimuth angle theta and the elevation angle phi of each point in the neighborhood: counting the directions of two angles of each point by adopting two histograms, wherein the direction dereferencing range of theta is [0 degrees and 360 degrees ], and counting the direction by adopting an 8-column histogram, wherein each column represents 45 degrees; and phi is taken as [ -90 degrees, 90 degrees ], the direction condition is counted by adopting a 4-column histogram, and each column represents 45 degrees;
after statistics, taking the peak values of the two histograms as the azimuth angle and the elevation angle of the extreme point, and after calculation, each key point comprises x, y and z coordinate information, scale information and direction information, namely (x, y, z, sigma, theta and phi);
III, rotating the coordinate system of a certain key point p to the main direction of the key point, and calculating by adopting the following rotation formula because the main direction is already calculated
Wherein p isiIs a point before rotation, p'iRepresents the rotated point;
using rotated point p'iRecalculating the aforementioned parameters, including d, θ, φ, and then calculating the rotated keypoint p and neighborhood point p'iThe geometric relationship between them is referred to as a vector p' pi'an angle with a normal vector n at a key point p before rotation is calculated using the formula where a numerator represents a dot product of two vectors and a denominator represents a length of the two vectors, and a triplet of data information (theta, phi,') is generated for a key point and all neighbor points in the neighborhood
And dividing three angles of the plurality of triples to generate a feature descriptor.
4. The three-dimensional point cloud map fusion method based on vision correction as claimed in claim 1, wherein in step 4), IPFH feature matching is performed to determine the corresponding relationship, the judgment criterion is the distance between feature vectors, and the following matching strategy is adopted, namely, point cloud map point is takensrcIPFH of a point A, point cloud map pointtgtFinding two points B and C with the nearest distance to the vector A, calculating by the following formula, if the formula condition is satisfied, selecting B as the matching point thereof
5. The three-dimensional point cloud map fusion method based on visual correction as claimed in claim 4, wherein in step 5), after the feature matching points are found, the transformation matrix is calculated by using a singular value decomposition method, and the point cloud map points are mapped according to the transformation matrixsrcAnd rotating to reduce the visual angle difference between the two point cloud maps and finish the visual angle correction.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710989642.2A CN107818598B (en) | 2017-10-20 | 2017-10-20 | Three-dimensional point cloud map fusion method based on visual correction |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710989642.2A CN107818598B (en) | 2017-10-20 | 2017-10-20 | Three-dimensional point cloud map fusion method based on visual correction |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107818598A CN107818598A (en) | 2018-03-20 |
CN107818598B true CN107818598B (en) | 2020-12-25 |
Family
ID=61606939
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710989642.2A Active CN107818598B (en) | 2017-10-20 | 2017-10-20 | Three-dimensional point cloud map fusion method based on visual correction |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107818598B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11830207B2 (en) * | 2020-05-18 | 2023-11-28 | Beijing Baidu Netcom Science Technology Co., Ltd. | Method, apparatus, electronic device and readable storage medium for point cloud data processing |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110211167A (en) * | 2019-06-14 | 2019-09-06 | 北京百度网讯科技有限公司 | Method and apparatus for handling point cloud data |
CN110930495A (en) * | 2019-11-22 | 2020-03-27 | 哈尔滨工业大学(深圳) | Multi-unmanned aerial vehicle cooperation-based ICP point cloud map fusion method, system, device and storage medium |
CN111310818B (en) * | 2020-02-10 | 2021-05-18 | 贝壳找房(北京)科技有限公司 | Feature descriptor determining method and device and computer-readable storage medium |
CN111429574B (en) * | 2020-03-06 | 2022-07-15 | 上海交通大学 | Mobile robot positioning method and system based on three-dimensional point cloud and vision fusion |
CN112414417B (en) * | 2020-11-17 | 2021-11-26 | 智邮开源通信研究院(北京)有限公司 | Automatic driving map generation method and device, electronic equipment and readable storage medium |
CN112902905A (en) * | 2021-01-20 | 2021-06-04 | 西安电子科技大学 | High-definition 3D scanning-based ground object spectrum testing method and system |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105118059A (en) * | 2015-08-19 | 2015-12-02 | 哈尔滨工程大学 | Multi-scale coordinate axis angle feature point cloud fast registration method |
CN106296693A (en) * | 2016-08-12 | 2017-01-04 | 浙江工业大学 | Based on 3D point cloud FPFH feature real-time three-dimensional space-location method |
CN106780459A (en) * | 2016-12-12 | 2017-05-31 | 华中科技大学 | A kind of three dimensional point cloud autoegistration method |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10360469B2 (en) * | 2015-01-15 | 2019-07-23 | Samsung Electronics Co., Ltd. | Registration method and apparatus for 3D image data |
-
2017
- 2017-10-20 CN CN201710989642.2A patent/CN107818598B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105118059A (en) * | 2015-08-19 | 2015-12-02 | 哈尔滨工程大学 | Multi-scale coordinate axis angle feature point cloud fast registration method |
CN106296693A (en) * | 2016-08-12 | 2017-01-04 | 浙江工业大学 | Based on 3D point cloud FPFH feature real-time three-dimensional space-location method |
CN106780459A (en) * | 2016-12-12 | 2017-05-31 | 华中科技大学 | A kind of three dimensional point cloud autoegistration method |
Non-Patent Citations (1)
Title |
---|
Aligning Point Cloud Views using Persistent Feature Histograms;Radu Bogdan Rusu 等;《2008 IEEE/RSJ International Conference on Intelligent Robots and Systems》;20081014;第3384-3391页 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11830207B2 (en) * | 2020-05-18 | 2023-11-28 | Beijing Baidu Netcom Science Technology Co., Ltd. | Method, apparatus, electronic device and readable storage medium for point cloud data processing |
Also Published As
Publication number | Publication date |
---|---|
CN107818598A (en) | 2018-03-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107818598B (en) | Three-dimensional point cloud map fusion method based on visual correction | |
CN107301654B (en) | Multi-sensor high-precision instant positioning and mapping method | |
US8830229B2 (en) | Recognition and pose determination of 3D objects in 3D scenes | |
Micusik et al. | Descriptor free visual indoor localization with line segments | |
CN112634451A (en) | Outdoor large-scene three-dimensional mapping method integrating multiple sensors | |
Yu et al. | Robust robot pose estimation for challenging scenes with an RGB-D camera | |
US20140105506A1 (en) | Recognition and pose determination of 3d objects in multimodal scenes | |
CN114972459B (en) | Point cloud registration method based on low-dimensional point cloud local feature descriptor | |
CN111145232A (en) | Three-dimensional point cloud automatic registration method based on characteristic information change degree | |
CN104040590A (en) | Method for estimating pose of object | |
Xu et al. | Ring++: Roto-translation-invariant gram for global localization on a sparse scan map | |
CN107025449B (en) | Oblique image straight line feature matching method constrained by local area with unchanged visual angle | |
CN114088081B (en) | Map construction method for accurate positioning based on multistage joint optimization | |
CN111768447A (en) | Monocular camera object pose estimation method and system based on template matching | |
CN109766903B (en) | Point cloud model curved surface matching method based on curved surface features | |
Zhu et al. | A review of 6d object pose estimation | |
CN110111375A (en) | A kind of Image Matching elimination of rough difference method and device under Delaunay triangulation network constraint | |
CN104463953A (en) | Three-dimensional reconstruction method based on inertial measurement unit and RGB-D sensor | |
CN111998862A (en) | Dense binocular SLAM method based on BNN | |
CN117132630A (en) | Point cloud registration method based on second-order spatial compatibility measurement | |
WO2023131203A1 (en) | Semantic map updating method, path planning method, and related apparatuses | |
CN114463396B (en) | Point cloud registration method utilizing plane shape and topological graph voting | |
Song et al. | Bundledslam: An accurate visual slam system using multiple cameras | |
CN117351078A (en) | Target size and 6D gesture estimation method based on shape priori | |
Liu et al. | Robust 3-d object recognition via view-specific constraint |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |