CN112164100B - Image registration method based on graph convolution neural network - Google Patents
Image registration method based on graph convolution neural network Download PDFInfo
- Publication number
- CN112164100B CN112164100B CN202011025452.7A CN202011025452A CN112164100B CN 112164100 B CN112164100 B CN 112164100B CN 202011025452 A CN202011025452 A CN 202011025452A CN 112164100 B CN112164100 B CN 112164100B
- Authority
- CN
- China
- Prior art keywords
- matching
- point
- graph
- layer
- pairs
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 27
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 21
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 26
- 238000011176 pooling Methods 0.000 claims description 15
- 239000011159 matrix material Substances 0.000 claims description 13
- 230000006870 function Effects 0.000 claims description 11
- 238000005096 rolling process Methods 0.000 claims description 8
- 238000013507 mapping Methods 0.000 claims description 7
- 238000010606 normalization Methods 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims description 4
- 238000000605 extraction Methods 0.000 claims description 3
- 238000010586 diagram Methods 0.000 description 8
- 238000004590 computer program Methods 0.000 description 7
- 238000012545 processing Methods 0.000 description 4
- 238000003860 storage Methods 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 2
- 238000003475 lamination Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 244000291564 Allium cepa Species 0.000 description 1
- UDHXJZHVNHGCEC-UHFFFAOYSA-N Chlorophacinone Chemical compound C1=CC(Cl)=CC=C1C(C=1C=CC=CC=1)C(=O)C1C(=O)C2=CC=CC=C2C1=O UDHXJZHVNHGCEC-UHFFFAOYSA-N 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000001427 coherent effect Effects 0.000 description 1
- 238000013527 convolutional neural network Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
- G06T7/33—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
The application relates to an image registration method based on a graph convolution neural network, which comprises the following steps: obtaining key points of the image pairs and initializing key point matching; inputting the initial matching pairs into a multi-layer perceptron to obtain characteristic point information of each matching pair; inputting the initial matching pairs into a graph convolution neural network to obtain local graph space characteristic information of each matching point pair; the characteristic point information of the matched point pairs and the spatial characteristic information of the partial graph are combined and then input into a multi-layer perceptron to learn the combined characteristic, and the final characteristic is output; and calculating a loss value by utilizing the output final characteristics, and adopting a back propagation algorithm to adjust network parameters. The application effectively improves the registration precision.
Description
Technical Field
The application relates to the technical field of computer vision, in particular to an image registration method based on a graph convolution neural network.
Background
More and more computer vision products are incorporated into our daily lives, and complex data in real life are increasingly demanding on computer vision algorithms. Estimating the geometric relationship between the two images is a fundamental problem for computer vision neighborhood, playing an important role in Structure from Motion and Simultaneous Localization and Mapping. In image registration, conventional outlier removal algorithms such as RANSAC (Fischler MA, bolles RC. Random sample coherent: a paradigm for model fitting with applications to image analysis and automated cartographic. Communications of the ACM.1981Jun 1;24 (6): 381-95) are standard algorithms and are also the most popular outlier removal algorithms. GMS (Bian J, lin WY, matsushita Y, yeung SK, nguyen TD, cheng MM.Gms: grid-based motion statistics for fast, ultra-robustfeature core approach.InProceding of the IEEE Conference on Computer Vision and Pattern Recognition 2017 (pp.4181-4190)). As the performance of deep learning has been higher in recent years, algorithms for image matching using deep learning have been increasing. However, there is no solution for image registration using a graph convolutional neural network (GCN), firstly because no effective graph convolution can be used for image registration, and secondly because no effective neighbor-taking strategy can be adopted due to non-rigid transformation.
Disclosure of Invention
Therefore, the application aims to provide an image registration method based on a graph convolution neural network, which effectively improves registration accuracy.
The application is realized by adopting the following scheme: an image registration method based on a graph convolution neural network specifically comprises the following steps:
obtaining key points of the image pairs and initializing key point matching;
inputting the initial matching pairs into a multi-layer perceptron to obtain characteristic point information of each matching pair;
inputting the initial matching pairs into a graph convolution neural network to obtain local graph space characteristic information of each matching point pair;
the characteristic point information of the matched point pairs and the spatial characteristic information of the partial graph are combined and then input into a multi-layer perceptron to learn the combined characteristic, and the final characteristic is output;
and calculating a loss value by utilizing the output final characteristics, and adopting a back propagation algorithm to adjust network parameters.
Further, the obtaining the keypoints of the image pair and initializing the keypoint matching pair and the data set are specifically as follows:
extracting hypothesis matching pairs of key point coordinates in the 2D registration image by adopting a feature extraction algorithm to obtainWherein N is the number of matching point pairs, p i Matching pairs for one of the hypotheses, +.>And->Is the coordinates of a matched pair.
Further, the inputting the initial matching pairs into the multi-layer perceptron to obtain the characteristic point information of each matching pair specifically comprises the following steps:
step S21: adopting a layer of shared perceptron to make the set P= [ P ] of initial matching pair 1 ;p 2 ;...;p i ;...;p N ]Mapping toWherein N is the number of matching point pairs, and M is the number of characteristic channels;
step S22: inputting the N matching pairs into a basic residual error network structure to obtain mapping characteristic output of each matching pairWherein C is 1 Is the dimension of the feature channel.
Further, the basic residual network structure includes a layer of shared perceptron MLP, an instance normalization layer IN, a batch normalization layer BN, and a corrective linear unit ReLU.
Further, the graph rolling neural network comprises more than one graph rolling module, more than one pooling layer and a shared perceptron network MLP; each graph convolution module comprises a graph convolution layer and a ReLU layer;
extracting graph convolution characteristics of input data through a graph convolution layer, carrying out pooling operation on the extracted graph convolution characteristics through a pooling layer, constructing nearest neighbors of each matching pair again after pooling each time to input the nearest neighbors to the next graph convolution layer, merging the graph convolution characteristics under different spatial relations in a channel dimension, and finally learning the combined characteristics by adopting a shared perceptron network (MLP) to obtain local graph spatial characteristic information of each matching point pairWhere N is the number of pairs of matching points of the input, C 2 Is the dimension of the feature channel.
Further, after combining the feature point information of the matching point pairs with the spatial feature information of the local map, inputting the feature point information into a multi-layer perceptron to learn the combined feature, and outputting final features specifically as follows:
characteristic point information out of each matching point pair 1 And local map spatial feature information out 2 Through feature channel dimension combination, obtain
Will beSequentially inputting into more than one basic residual error network and a layer of perceptron layer to obtain final logic output value +.>
Further, the calculating the loss value by using the output final feature and adjusting the network parameter by adopting a back propagation algorithm specifically comprises:
calculation of final output by weighted eight-point algorithmIs>
The loss value is calculated using the loss function and the network parameters are adjusted by a back propagation algorithm.
Further, the loss value loss is calculated as:
wherein, I ess To be weighted by eightPoint algorithm prediction final outputIs>And an essential matrix error between the true essential matrix E; l (L) cls For the binary cross entropy loss function, z represents the truth label, s represents the logic value of the network output, i.e.>
Compared with the prior art, the application has the following beneficial effects: the method introduces the graph convolution based on space into an image registration method, combines the advantages of the graph convolution with a neural network module commonly used in image registration at the present stage, and extracts the characteristics. Therefore, the application can improve the matching precision finally. Experimental results show that the application achieves the most advanced performance on the reference data set.
Drawings
FIG. 1 is a flow chart of a method according to an embodiment of the application.
Fig. 2 is a schematic diagram of a graph rolling network according to an embodiment of the application.
Fig. 3 is a schematic diagram of an overall network structure according to an embodiment of the present application.
Detailed Description
The application will be further described with reference to the accompanying drawings and examples.
It should be noted that the following detailed description is exemplary and is intended to provide further explanation of the application. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of exemplary embodiments according to the present application. As used herein, the singular is also intended to include the plural unless the context clearly indicates otherwise, and furthermore, it is to be understood that the terms "comprises" and/or "comprising" when used in this specification are taken to specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof.
As shown in fig. 1 and 3, the present embodiment provides an image registration method based on a graph convolution neural network, which specifically includes the following steps:
obtaining key points of the image pairs and initializing key point matching;
inputting the initial matching pairs into a multi-layer perceptron to obtain characteristic point information of each matching pair;
inputting the initial matching pairs into a graph convolution neural network to obtain local graph space characteristic information of each matching point pair;
the characteristic point information of the matched point pairs and the spatial characteristic information of the partial graph are combined and then input into a multi-layer perceptron to learn the combined characteristic, and the final characteristic is output;
and calculating a loss value by utilizing the output final characteristics, and adopting a back propagation algorithm to adjust network parameters.
In this embodiment, the obtaining the keypoints of the image pair and initializing the keypoint matching pair and the data set are specifically:
extracting hypothesis matching pairs of key point coordinates in the 2D registration image by adopting a feature extraction algorithm to obtainWherein N is the number of matching point pairs, p i Matching pairs for one of the hypotheses, +.>And->Is the coordinates of a matched pair.
In this embodiment, the inputting the initial matching pair into the multi-layer perceptron to obtain the feature point information of each matching pair specifically includes the following steps:
step S21: employing one layer sharingThe perceptron sets the initial matching pair p= [ P ] 1 ;p 2 ;...;p i ;...;p N ]Mapping toWherein N is the number of matching point pairs, and M is the number of characteristic channels;
step S22: inputting the N matching point pairs into a basic residual error network structure, and obtaining the mapping characteristic output of each matching point pairWherein C is 1 Is the dimension of the feature channel.
IN this embodiment, the basic residual network structure includes a layer of shared perceptron MLP, an instance normalization layer IN, a batch normalization layer BN, and a corrective linear unit ReLU.
As shown in fig. 2, in the present embodiment, the graph rolling neural network includes more than one graph rolling module (GCN), more than one pooling layer, and a shared perceptron network MLP; each graph convolution module comprises a graph convolution layer and a ReLU layer;
wherein, the graph roll lamination conv calculates a convolution kernel K S Receptive fieldCosine similarity of (c). The method comprises the following steps:
in the receptive fieldIs of a graph structure, p n In order to center the receptive field,is p n M euclidean distance based nearest neighbors given by a pre-calculation;<·>representation ofDot product; w (·) represents the weight parameter given by the network learning; here, f (x) =w·x+b, w is a weight matrix, and b is a deviation; graph convolution kernel K S ={k C ,k 1 ,k 2 ,...,k S The } represents +.1 convolution kernel node by S +.>Composition, k C =(0,0,0,0);And calculating the cosine similarity of the characteristics of each neighbor and the weight of each convolution kernel. Wherein d m,n =p m -p n Representing the direction of the receptive field center to the neighbors.
Extracting graph convolution characteristics of input data through a graph convolution layer, carrying out pooling operation on the extracted graph convolution characteristics through a pooling layer, and inputtingPooling into->r is the pooling rate; it should be noted that, because the original matching pair after pooling has changes to the number, scale and spatial relationship, after each pooling, the nearest neighbor of each matching pair is built again to be input to the next graph volume lamination, then the graph volume features under different spatial relationships are merged in a channel dimension to obtain multi-scale features, finally a small shared perceptron network MLP is adopted to learn the combined features (multi-scale features) to obtain the local graph space feature information of each matching point pair>Where N is the number of pairs of matching points of the input, C 2 Is the dimension of the feature channel.
In this embodiment, after combining the feature point information of the matching point pair with the spatial feature information of the partial graph, the feature point information is input to a multi-layer perceptron to learn a combined feature, and the final feature is output specifically as follows:
characteristic point information out of each matching point pair 1 And local map spatial feature information out 2 Through feature channel dimension combination, obtain
Will beSequentially inputting into more than one basic residual error network and a layer of perceptron layer to obtain logic value +.>N is the number of input matching pairs.
In this embodiment, the calculating the loss value by using the output final feature and adjusting the network parameter by using the back propagation algorithm specifically includes:
calculation of final output by weighted eight-point algorithmIs>
The loss value is calculated using the loss function and the network parameters are adjusted by a back propagation algorithm.
In this embodiment, the loss value loss is calculated as:
wherein, I ess To predict logical values by a weighted eight-point algorithmIs>And the true value essential matrix E, specifically adopting geometric loss: />Here P 1 ,P 2 Is a corresponding matching point, t [i] Is the i-th element of vector t, +.>Representing the true value essence matrix E and the point P 1 Distance of->Representing the true value essence matrix E and the point P 2 Distance of->Essence matrix representing network predictions->To P 1 Is a distance of (3). l (L) cls For the binary cross entropy loss function, z represents the truth label, s represents the logic value of the network output, i.e.>
The present example performed quantitative and qualitative experiments of the proposed method with the current most advanced matching method on a common data set (yfcs 100M), and the results showed that the method of the present example was significantly superior to other algorithms. The following table is a quantitative comparison of the accuracy and recall of the F-measurements of this example with several other matching algorithms. The comparison method is Ranac, LPM, pointCN. From the table, it can be seen that the application significantly improves the detection accuracy, and the best effect is obtained in the 4 methods.
F-score | Precision | Recall | |
Ransac | 0.1914 | 0.2222 | 0.1879 |
LPM | 0.2213 | 0.2415 | 0.2579 |
PointCN | 0.3319 | 0.2745 | 0.5588 |
Method of the present embodiment | 0.3975 | 0.3237 | 0.6234 |
It will be appreciated by those skilled in the art that embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
The above description is only a preferred embodiment of the present application, and is not intended to limit the application in any way, and any person skilled in the art may make modifications or alterations to the disclosed technical content to the equivalent embodiments. However, any simple modification, equivalent variation and variation of the above embodiments according to the technical substance of the present application still fall within the protection scope of the technical solution of the present application.
Claims (6)
1. An image registration method based on a graph convolution neural network is characterized by comprising the following steps:
obtaining key points of the image pairs and initializing key point matching;
inputting the initial matching pairs into a multi-layer perceptron to obtain characteristic point information of each matching pair;
inputting the initial matching pairs into a graph convolution neural network to obtain local graph space characteristic information of each matching point pair;
the characteristic point information of the matched point pairs and the spatial characteristic information of the partial graph are combined and then input into a multi-layer perceptron to learn the combined characteristic, and the final characteristic is output;
calculating a loss value by utilizing the output final characteristics, and adjusting network parameters by adopting a back propagation algorithm;
the method for calculating the loss value by utilizing the output final characteristics and adopting a back propagation algorithm to adjust network parameters comprises the following steps:
calculation of final output by weighted eight-point algorithmIs>
Calculating a loss value by using a loss function, and adjusting network parameters by using a back propagation algorithm;
the loss value loss is calculated as follows:
wherein, I ess To predict final output by weighted eight-point algorithmIs of the matrix of the nature of (2)/>And an essential matrix error between the true essential matrix E; l (L) cls For a binary cross entropy loss function, z represents a truth value label, and s represents the weight of network output;
the geometric loss is adopted specifically:here P 1 ,P 2 Is a corresponding matching point, t [i] Is the i-th element of vector t, +.>Representing the true value essence matrix E and the point P 1 Distance of->Representing the true value essence matrix E and the point P 2 Distance of->Essence matrix representing network predictions->To P 1 Is a distance of (2); l (L) cls For the binary cross entropy loss function, z represents the truth label, s represents the logic value of the network output, i.e.>
2. The image registration method based on a graph convolution neural network according to claim 1, wherein the obtaining the keypoints of the image pair and initializing the keypoint matching pair and the data set are specifically as follows:
adopting a feature extraction algorithm to extract hypothesis matching pairs of key point coordinates in the 2D registration image to obtain P=[p 1 ;p 2 ;...;p i ;...;p N ]∈R N×4 ,Wherein N is the number of matching point pairs, p i Matching pairs for one of the hypotheses, +.>And->Is the coordinates of a matched pair.
3. The image registration method based on a graph convolution neural network according to claim 1, wherein the step of inputting the initial matching pairs into the multi-layer perceptron to obtain the feature point information of each matching pair specifically comprises the following steps:
step S21: adopting a layer of shared perceptron to make the set P= [ P ] of initial matching pair 1 ;p 2 ;...;p i ;...;p N ]Mapping toWherein N is the number of matching point pairs, and M is the number of characteristic channels;
step S22: inputting the N matching point pairs into a basic residual error network structure, and obtaining the mapping characteristic output of each matching point pairWherein C is 1 Is the dimension of the feature channel.
4. A method of image registration based on a graph roll-up neural network according to claim 3, wherein the basic residual network structure comprises a layer of shared perceptron MLP, an instance normalization layer IN, a batch normalization layer BN, and a corrective linear unit ReLU.
5. The image registration method based on a graph rolling neural network according to claim 1, wherein the graph rolling neural network comprises more than one graph rolling module, more than one pooling layer and a shared perceptron network (MLP); each graph convolution module comprises a graph convolution layer and a ReLU layer;
extracting graph convolution characteristics of input data through a graph convolution layer, carrying out pooling operation on the extracted graph convolution characteristics through a pooling layer, constructing nearest neighbors of each matching pair again after pooling each time to input the nearest neighbors to the next graph convolution layer, merging the graph convolution characteristics under different spatial relations in a channel dimension, and finally learning the combined characteristics by adopting a shared perceptron network (MLP) to obtain local graph spatial characteristic information of each matching point pairWhere N is the number of pairs of matching points of the input, C 2 Is the dimension of the feature channel.
6. The image registration method based on the graph convolution neural network according to claim 1, wherein the feature point information of the matching point pair and the local graph space feature information are input into a multi-layer perceptron learning combination feature after being combined, and the final feature is output specifically as follows:
characteristic point information out of each matching point pair 1 And local map spatial feature information out 2 Through feature channel dimension combination, obtain
Will beSequentially inputting into more than one basic residual error network and a layer of perceptron layer to obtain logic value +.>
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011025452.7A CN112164100B (en) | 2020-09-25 | 2020-09-25 | Image registration method based on graph convolution neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011025452.7A CN112164100B (en) | 2020-09-25 | 2020-09-25 | Image registration method based on graph convolution neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112164100A CN112164100A (en) | 2021-01-01 |
CN112164100B true CN112164100B (en) | 2023-12-12 |
Family
ID=73864043
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011025452.7A Active CN112164100B (en) | 2020-09-25 | 2020-09-25 | Image registration method based on graph convolution neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112164100B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112801206B (en) * | 2021-02-23 | 2022-10-14 | 中国科学院自动化研究所 | Image key point matching method based on depth map embedded network and structure self-learning |
CN114677502B (en) * | 2022-05-30 | 2022-08-12 | 松立控股集团股份有限公司 | License plate detection method with any inclination angle |
CN117253060B (en) * | 2023-09-04 | 2024-06-21 | 江苏势通生物科技有限公司 | Image matching method, image matching device, storage medium and computer equipment |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109064502A (en) * | 2018-07-11 | 2018-12-21 | 西北工业大学 | The multi-source image method for registering combined based on deep learning and artificial design features |
CN110335337A (en) * | 2019-04-28 | 2019-10-15 | 厦门大学 | A method of based on the end-to-end semi-supervised visual odometry for generating confrontation network |
CN111462867A (en) * | 2020-04-05 | 2020-07-28 | 武汉诶唉智能科技有限公司 | Intelligent mobile medical method and system based on 5G network and block chain |
CN111488937A (en) * | 2020-04-15 | 2020-08-04 | 闽江学院 | Image matching method based on multi-scale neighbor deep neural network |
CN111488938A (en) * | 2020-04-15 | 2020-08-04 | 闽江学院 | Image matching method based on two-step switchable normalized depth neural network |
-
2020
- 2020-09-25 CN CN202011025452.7A patent/CN112164100B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109064502A (en) * | 2018-07-11 | 2018-12-21 | 西北工业大学 | The multi-source image method for registering combined based on deep learning and artificial design features |
CN110335337A (en) * | 2019-04-28 | 2019-10-15 | 厦门大学 | A method of based on the end-to-end semi-supervised visual odometry for generating confrontation network |
CN111462867A (en) * | 2020-04-05 | 2020-07-28 | 武汉诶唉智能科技有限公司 | Intelligent mobile medical method and system based on 5G network and block chain |
CN111488937A (en) * | 2020-04-15 | 2020-08-04 | 闽江学院 | Image matching method based on multi-scale neighbor deep neural network |
CN111488938A (en) * | 2020-04-15 | 2020-08-04 | 闽江学院 | Image matching method based on two-step switchable normalized depth neural network |
Non-Patent Citations (1)
Title |
---|
基于偏好统计数据表征的鲁棒几何模型拟合方法;郭翰林 等;《计算机学报》;第43卷(第7期);第1199-1214页 * |
Also Published As
Publication number | Publication date |
---|---|
CN112164100A (en) | 2021-01-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112164100B (en) | Image registration method based on graph convolution neural network | |
CN111192292B (en) | Target tracking method and related equipment based on attention mechanism and twin network | |
US20220027603A1 (en) | Fast, embedded, hybrid video face recognition system | |
CN111695415B (en) | Image recognition method and related equipment | |
Dai et al. | MS2DG-Net: Progressive correspondence learning via multiple sparse semantics dynamic graph | |
CN111968150B (en) | Weak surveillance video target segmentation method based on full convolution neural network | |
Guo et al. | JointPruning: Pruning networks along multiple dimensions for efficient point cloud processing | |
CN107871103B (en) | Face authentication method and device | |
CN112070058A (en) | Face and face composite emotional expression recognition method and system | |
CN116580257A (en) | Feature fusion model training and sample retrieval method and device and computer equipment | |
CN112232134A (en) | Human body posture estimation method based on hourglass network and attention mechanism | |
CN115908984B (en) | Training method and device for image clustering model | |
CN112308128B (en) | Image matching method based on attention mechanism neural network | |
CN112084895A (en) | Pedestrian re-identification method based on deep learning | |
CN110826534B (en) | Face key point detection method and system based on local principal component analysis | |
CN110083734B (en) | Semi-supervised image retrieval method based on self-coding network and robust kernel hash | |
CN109993070B (en) | Pedestrian re-identification method based on global distance scale loss function | |
CN114463340A (en) | Edge information guided agile remote sensing image semantic segmentation method | |
CN112990356A (en) | Video instance segmentation system and method | |
CN115619822A (en) | Tracking method based on object-level transformation neural network | |
CN115512693B (en) | Audio recognition method, acoustic model training method, device and storage medium | |
CN115205554A (en) | Retrieval method based on semantic concept extraction | |
CN117437557A (en) | Hyperspectral image classification method based on double-channel feature enhancement | |
CN115100694A (en) | Fingerprint quick retrieval method based on self-supervision neural network | |
CN111506691B (en) | Track matching method and system based on depth matching model |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |