Nothing Special   »   [go: up one dir, main page]

CN115130663B - Heterogeneous network attribute completion method based on graph neural network and attention mechanism - Google Patents

Heterogeneous network attribute completion method based on graph neural network and attention mechanism Download PDF

Info

Publication number
CN115130663B
CN115130663B CN202211043710.3A CN202211043710A CN115130663B CN 115130663 B CN115130663 B CN 115130663B CN 202211043710 A CN202211043710 A CN 202211043710A CN 115130663 B CN115130663 B CN 115130663B
Authority
CN
China
Prior art keywords
attribute
node
target node
nodes
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202211043710.3A
Other languages
Chinese (zh)
Other versions
CN115130663A (en
Inventor
于彦伟
王凯
董军宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ocean University of China
Original Assignee
Ocean University of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ocean University of China filed Critical Ocean University of China
Priority to CN202211043710.3A priority Critical patent/CN115130663B/en
Publication of CN115130663A publication Critical patent/CN115130663A/en
Application granted granted Critical
Publication of CN115130663B publication Critical patent/CN115130663B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Devices For Executing Special Programs (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a heterogeneous network attribute completion method based on a graph neural network and an attention mechanism, and belongs to the technical field of data processing. Firstly, capturing nodes similar to target nodes (nodes with missing attributes) from a network by combining the cosine similarity between K-Nearest Neighbor and the attributes, and adaptively converting the attributes of the nodes into network characterization of the feature domain of the target nodes; based on the attention mechanism of the graph neural network and the transducer, carrying out hierarchical analysis on the topological structure in the network and the attribute information of the nodes, and further obtaining the network representation of the target node in the spatial domain; finally, combining network characterization of the spatial domain and the characteristic domain, and carrying out model parameter learning based on a loss function of Euclidean distance so as to complement the missing attribute. Through practical verification, the attribute completion method provided by the invention has the characteristics of high efficiency and high accuracy.

Description

Heterogeneous network attribute completion method based on graph neural network and attention mechanism
Technical Field
The invention relates to a heterogeneous network attribute completion method based on a graph neural network and an attention mechanism, and belongs to the technical field of data processing.
Background
Networks are ubiquitous in our real life, and most of the connections between objects in the real world can be represented as networks, for example, friend relationships between users can be regarded as social networks, reference relationships between papers can be regarded as reference networks, and connection relationships between road segments can be regarded as a traffic network. The above networks are all made up of nodes of the same type, so they are also called homogeneous networks. More widely present in the real world is a heterogeneous network, the nodes constituting the heterogeneous network being of different types, for example: shopping network is composed of users and commodities, and academic network is composed of authors, the units where authors are located and papers. Although these networks contain huge amounts of data, the lack of attributes in the networks (e.g., in shopping networks, some users are reluctant to upload their age information, in academic networks, keywords of papers are not filled in completely by authors) also presents a great challenge to the exploitation of the potential value contained in the networks. The data mining method for the network can effectively improve the efficiency of data mining of the network by complementing the missing attribute in the network, but the method is complex, and the considered factors are many, such as the connection relationship between network nodes, and the relationship between the existing attribute and the missing attribute of the network nodes. How to efficiently and accurately complement missing data in a network is also becoming increasingly important to academia and industry.
In the field of attribute completion, a conventional attribute completion method generally performs analysis completion from connection relationships of nodes in a network and semantic text information. The above method does not take into account the topology of the entire network. In recent years, the graph neural network method shows higher efficiency and accuracy in mining network information, such as graph convolution neural network and graph annotation force network show excellent performance in capturing network topology structure and node attribute information. The development of the graph neural network also brings new possibilities to the field of attribute completion, such as a method for carrying out joint learning based on the graph convolution neural network and combining attribute completion and commodity recommendation, and attribute completion is carried out by using the representation learned by the graph neural network. These methods have achieved significant results in attribute completion, but there is still room for improvement.
By analyzing and summarizing the existing attribute completion method, the existing method has the following defects: 1) The network characterization in the heterogeneous network is different from that of all nodes in the homogeneous network, and the types of all nodes in the heterogeneous network are the same, so that the missing attribute of the nodes in the heterogeneous network is complemented, and the homogeneous node information and the heterogeneous node information need to be comprehensively considered. This problem has plagued many network characterization models. 2) The graph neural network cannot efficiently capture high-order node information in the network. GCN essentially generates a characterization by aggregating the surrounding node information of the target node under a semi-supervised framework. In the attribute completion problem, the attribute of the target node may be related to not only the surrounding nodes but also the higher-order nodes. If the higher-order information of the target node can be captured through stacking the layer number of GCNs, the node-to-node characterization in the network is more and more similar with the increase of the layer number of GCNs, the specificity is lost, and the accuracy of attribute complementation is further affected.
Disclosure of Invention
In order to solve the heterogeneous network attribute completion problem more effectively, the invention aims to provide an attribute completion method based on a graph neural network and an attention mechanism so as to further improve the efficiency and accuracy of attribute completion and provide a method and technical support for the heterogeneous network attribute completion problem.
In order to achieve the aim of the invention, the invention adopts the following specific technical scheme:
an attribute completion method based on a graph neural network and an attention mechanism comprises the following steps:
s1: acquiring a heterogeneous attribute network with missing attributes, wherein a node with missing attributes is called a target node, and a node with complete attributes is called a source node;
s2: k source nodes most similar to the existing attribute of the target node are selected by adopting a K nearest neighbor (K-nearest neighbor) algorithm to complete the attribute in the feature space;
s3: after K source nodes which are most similar to the target node are screened out by utilizing cosine similarity, a learnable parameter is endowed to each source node to dynamically adjust the weight of the attribute of each source node for complementing the attribute of the target node, and attribute characterization endowed with weight of the feature space is obtained;
s4: the nodes directly connected with the target node in the heterogeneous attribute network are aggregated with the target node to obtain a low-order representation of the structural space; the specific aggregation mode is realized by simplifying a graph neural network (Simplifying graph convolutional networks);
s5: firstly, obtaining a high-order node of a target node by using a random walk mode;
s6: different weights are given to nodes in the high-order node sequence based on a transducer, and high-order characterization of the given weights of the structural space is obtained;
s7: splicing the attribute representation of the characteristic space, the low-order representation of the structural space and the high-order representation of the structural space, and then sending the spliced representation into a multi-layer perceptron to convert the representation into the attribute of the target node;
s8: the method comprises the steps of manually removing part of attributes of part of nodes by adopting a supervised learning mode, then complementing the attributes by reconstruction prediction, continuously training through the difference between the complemented attributes and the true attributes, and finally complementing the attribute missing values of other nodes by utilizing a trained model.
Further, in the step S1, the heterogeneous network with the missing attribute is defined asRepresenting the set of vertices in the graph, +.>Representing the set of edges in the diagram, +.>Attribute matrix representing vertices in the graph, +.>For the number of vertices in the graph, < > and->For each dimension of vertex feature, +.>Is a marking matrix, when->In the time-course of which the first and second contact surfaces,the corresponding attribute in (a) is missing; on the contrary, let(s)>The corresponding attributes of (a) are complete.
Further, in the S2, a large number of nodes are typically included in the heterogeneous network, and it is inefficient and impractical to complement the missing attribute of the target node with the attributes of all the source nodes; therefore, K nearest neighbors are adopted to select K source nodes which are most similar to the existing attribute of the target node to complete the attribute in the feature space, and cosine similarity (formula (1)) is used to measure the similarity of two attribute vectors;
(1)
wherein ,representing the similarity of two nodes, +.>Representing the existing properties of the target node +.>Representing an attribute in the source node corresponding to the existing attribute of the target node,/->The larger the value, the higher the similarity between the target node and the source node.
Further, in the S3, there may be a plurality of factors affecting the attribute complement relationship of the source node to the target node in the heterogeneous network, for example: whether nodes are directly connected with each other, weights on the connected node edges, and the like. The dynamic adjustment is performed according to formula (2):
(2)
wherein ,for the characterization of the feature space of the target node, +.>Is->Is most similar->Set of individual source nodes->A leachable adjustment weight for each source node, +.>Is a feature vector of the source node.
Further, in the step S4, after the feature space finishes the feature learning of the target node, the feature needs to be learned on the target node in the structural space; the specific formula (3) is as follows:
(3)
wherein ,to simplify the drawing neural network +.>Weight matrix of layer,/>Is->Layer output, in the first layer of the reduced graph neural network +.>Is an adjacency matrix.
The graph neural network essentially captures topological structure information and attribute information of nodes in the network by aggregating neighbor attributes of the target nodes, but as the layers of the graph neural network are stacked, the characteristic aggregated by the target nodes is over-smoothed, that is to say, the characteristics obtained by aggregating the characteristic of the target nodes lose specificity and distinguishing. Therefore, a simplified graph neural network is used in the structure space for aggregating neighbor nodes of the target node, and only one layer of network is used, and the specificity of the characterization of the target node is reserved while the topological structure and the attribute characteristics are captured.
Further, the step S5 is specifically: in order to capture high-order information of a target node in a heterogeneous network, the high-order node of the target node is obtained by using a random walk mode, wherein the random walk is a mode of traversing the node in the graph combining the advantages of depth-first traversal and breadth-first traversal in the graph, and the specific traversing mode is shown in a formula (4):
(4)
wherein Representing node->To node->Is the weight of the edge of (2);
for each target node, taking the target node as a root node, and carrying out random walk based on the root node to acquire a node sequence. Because directly connected source nodes have been used in the reduced graph neural network, these nodes are deleted in the random walk node sequence.
Further, in S6, the influence of the node sequence generated by the random walk on the target node is different, so the nodes in the node sequence are given different weights based on the transformers. The method comprises the following steps:
s6-1: firstly, carrying out linear change on the characteristics of a target node and a source node sequence, then calculating the weights of the target node and the nodes in each sequence based on the linear change, and carrying out softmax normalization operation on the obtained weights for the stability of calculation;
s6-2: then, the node features in the source node sequence are again subjected to a new linear transformation independent of the weight calculation.
S6-3: and finally, giving weight to the node characteristics of the source node sequence after each linear transformation, and accumulating the node characteristics to obtain the node characteristics of the target node higher order, wherein the calculation method is shown in a formula (5):
(5)
wherein Representing node->Characterization aggregated by a simplified graph neural network, +.>Representing node->Is characterized by (1)>,/> and />Projection parameter matrix representing a parameter which can be learned, +.>Representing node->Higher order network characterization of (2).
S6-4: on this basis, the attention mechanism is extended to a multi-headed attention mechanism to capture multiple dependency relationships between the target node and the Gao Jieyuan node. Then, the multiple dependency relationships are sent to an average pooling layer to obtain final high-order network characterization, and the calculation method is shown in a formula (6):
(6)
wherein ,representing node->In->Obtained higher order network characterization in the sub-attentive mechanism,/->Representing the total number of times that the attention mechanism calculation needs to be made.
Further, in S7, a specific calculation manner of the splicing is shown in formula (7):
(7)
wherein Representing vector concatenation operations,/->For the predicted target node->Is a property value of (a).
Further, the step S8 specifically includes:
s8-1: after the predicted attribute of the target node is obtained, the original existing attribute is reserved, the predicted attribute is filled in the missing attribute to complete the task of attribute completion, and the filling method is shown in a formula (8):
(8)
wherein ,representing predicted attributes of all nodes after attribute completion, +.>Representing Hadamard product, ->A matrix with all 1 elements:
s8-2: the difference between the predicted attribute and the real attribute of the loss function measurement is set based on the Euclidean distance formula, and the specific calculation mode is as shown in formula (9):
(9)
wherein Representing the set of target nodes>For node->True missing attribute value, +_>Is a nodePredicted missing attribute values.
The invention considers the relevance between the target node and the source node in the attribute space and the relevance in the structural space simultaneously when the heterogeneous network attribute is fully filled, and is specifically expressed as follows: in attribute space useThe neighbor algorithm finds the most similar +.>The source node then adaptively gives it the corresponding +.>And adding the weight values to obtain the attribute space characterization of the target node. In the structural space, a simplified graph neural network is used for aggregating first-order neighbor information of the target node to obtain low-order representation of the target node; a multi-headed attention mechanism based on a transducer and random walk is then used to capture high-order neighbor information of the target node in the structural space to obtain a high-order representation of the target node. And finally, fusing the three characterizations, updating parameters of the whole model under the guidance of a loss function based on Euclidean distance, and finally complementing the missing attribute of the target node.
The invention has the advantages and beneficial effects that:
compared with the traditional attribute completion method, the method introduces a framework for network characterization learning to perform attribute completion, and can better capture the structural characteristics of the target node in the network; the first-order neighbors of the target node are aggregated by using the simplified graph neural network, so that the accuracy of the obtained network representation can be ensured on the premise of improving the attribute completion time efficiency; the method uses a random walk and transformer-based attention mechanism to capture the high-order neighbor information of the target node, and solves the problem of node characteristic smoothing caused by simple stacking of graph neural network modules to a certain extent.
Through practical verification, the attribute completion method provided by the invention has the characteristics of high efficiency and high accuracy.
Drawings
Fig. 1 is an overall flow chart of the present invention.
Fig. 2 is a frame diagram of the present invention.
Fig. 3 is a first topological structure diagram of weights between nodes obtained in the present invention.
Fig. 4 is a second topology diagram of the weights between nodes obtained in the present invention.
FIG. 5 is a flow chart of the present invention for obtaining higher order node characterizations based on an attention mechanism.
Detailed Description
The invention will be further described with reference to fig. 1-5 and the accompanying examples.
Example 1:
an attribute completion method based on a graph neural network and an attention mechanism is shown in a whole flow chart in figure 1. The method comprises the following steps:
s1: acquiring a heterogeneous attribute network with missing attributes, wherein a node with missing attributes is called a target node, and a node with complete attributes is called a source node; defining heterogeneous property networks with missing properties asRepresenting the set of vertices in the graph +.> Representing the set of edges in the diagram, +.>Attribute matrix representing vertexes in the graph, n is the number of vertexes in the graph, +.>For each dimension of vertex feature, +.>Is a marking matrix if +.>Then indicate +.>The corresponding attribute in (a) is missing; otherwise, it means ++>The corresponding attributes in (a) are complete;
s2: k source nodes most similar to the existing attribute of the target node are selected by adopting a K nearest neighbor (K-nearest neighbor) algorithm to complete the attribute in the feature space; the typically inclusion of a large number of nodes in a heterogeneous network, it is inefficient and impractical to complement the missing attributes of the target node with the attributes of all source nodes; therefore, K nearest neighbors are adopted to select K source nodes which are most similar to the existing attribute of the target node to complete the attribute in the feature space, and cosine similarity (formula (1)) is used to measure the similarity of two attribute vectors;
(1)
wherein ,representing the similarity of two nodes, +.>Representing the existing properties of the target node +.>Representing an attribute in the source node corresponding to the existing attribute of the target node,/->The larger the value is, the higher the similarity between the target node and the source node is;
s3: after K source nodes which are most similar to the target node are screened out by utilizing cosine similarity, a learnable parameter is endowed to each source node to dynamically adjust the weight of the attribute of each source node for complementing the attribute of the target node, and attribute characterization endowed with weight of the feature space is obtained; there may be a number of factors in the heterogeneous network that affect the attribute completion relationship of the source node to the target node, such as: whether nodes are directly connected with each other, weights on the connected node edges, and the like. The dynamic adjustment is performed according to formula (2):
(2)
wherein ,for the characterization of the feature space of the target node, +.>Is->Is most similar->Set of individual source nodes->A leachable adjustment weight for each source node, +.>Is a feature vector of the source node;
s4: the nodes directly connected with the target node in the heterogeneous attribute network are aggregated with the target node to obtain a low-order representation of the structural space; the specific aggregation mode is realized by simplifying a graph neural network (Simplifying graph convolutional networks); after the feature space finishes the representation learning of the target node, the target node needs to be learned and characterized in the structural space; the specific formula (3) is as follows:
(3)
wherein ,to simplify the drawing neural network +.>Weight matrix of layer,/>Is->Layer output, in the first layer of the reduced graph neural network +.>Is an adjacency matrix;
s5: firstly, obtaining a high-order node of a target node by using a random walk mode; in order to capture high-order information of a target node in a heterogeneous network, the high-order node of the target node is obtained by using a random walk mode, wherein the random walk is a mode of traversing the node in the graph combining the advantages of depth-first traversal and breadth-first traversal in the graph, and the specific traversing mode is shown in a formula (4):
(4)
wherein Representing node->To node->Is the weight of the edge of (2);
for each target node, taking the target node as a root node, and carrying out random walk based on the root node to acquire a node sequence. Because directly connected source nodes have been used in the reduced graph neural network, these nodes are deleted in the random walk node sequence;
s6: different weights are given to nodes in the high-order node sequence based on a transducer, and high-order characterization of the given weights of the structural space is obtained; the method comprises the following steps:
s6-1: firstly, carrying out linear change on the characteristics of a target node and a source node sequence, then calculating the weights of the target node and the nodes in each sequence based on the linear change, and carrying out softmax normalization operation on the obtained weights for the stability of calculation;
s6-2: then, the node features in the source node sequence are again subjected to a new linear transformation independent of the weight calculation.
S6-3: and finally, giving weight to the node characteristics of the source node sequence after each linear transformation, and accumulating the node characteristics to obtain the node characteristics of the target node higher order, wherein the calculation method is shown in a formula (5):
(5)
wherein Representing node->Characterization aggregated by a simplified graph neural network, +.>Representing node->Is characterized by (1)>,/> and />Projection parameter matrix representing a parameter which can be learned, +.>Representing node->Higher order network characterization of (2).
S6-4: on this basis, the attention mechanism is extended to a multi-headed attention mechanism to capture multiple dependency relationships between the target node and the Gao Jieyuan node. Then, the multiple dependency relationships are sent to an average pooling layer to obtain final high-order network characterization, and the calculation method is shown in a formula (6):
(6)
wherein ,representing node->In->Obtained higher order network characterization in the sub-attentive mechanism,/->Representing the number of times a total need for attention mechanism calculations;
s7: splicing the attribute representation of the characteristic space, the low-order representation of the structural space and the high-order representation of the structural space, and then sending the spliced representation into a multi-layer perceptron to convert the representation into the attribute of the target node; the specific calculation mode of the splicing is shown in a formula (7):
(7)
wherein Representing vector concatenation operations,/->For the predicted target node->Is a property value of (a).
S8: the method comprises the steps of adopting a supervised learning mode, firstly manually removing part of attributes of part of nodes, then complementing the attributes through reconstruction prediction, continuously training through the difference between the complemented attributes and the true attributes, and finally complementing attribute missing values of other nodes by using a model which is completed through training; the method comprises the following steps:
s8-1: after the predicted attribute of the target node is obtained, the original existing attribute is reserved, the predicted attribute is filled in the missing attribute to complete the task of attribute completion, and the filling method is shown in a formula (8):
(8)
wherein ,representing predicted attributes of all nodes after attribute completion, +.>Representing Hadamard product, ->A matrix with all 1 elements;
s8-2: the difference between the predicted attribute and the real attribute of the loss function measurement is set based on the Euclidean distance formula, and the specific calculation mode is as shown in formula (9):
(9)
wherein Representing the set of target nodes>For node->True missing attribute value, +_>Is a nodePredicted missing attribute values.
Example 2: this example uses example 1 as a basic method, and a module design is performed.
The attribute completion system based on the graph neural network and the attention mechanism comprises a data preprocessing module, a marking matrix construction module, a feature space characterization learning module, a low-order neighbor characterization learning module, a high-order neighbor characterization learning module, a characterization fusion module and an attribute reasoning module, wherein the details of the parts are as follows as shown in fig. 2:
the data preprocessing module is used for: firstly normalizing attribute characteristics in an original data set, dividing the data into a training set, a test set and a verification set, randomly removing attribute information of nodes in the training set, and recording the removed attribute information as a true value to guide a model to learn.
The marking matrix construction module: traversing the node attributes in the dataset, marking the missing attribute values of the nodes, and further forming an attribute marking matrix
The feature space representation learning module: and selecting nodes similar to the target node in the structural space, then giving weight to the nodes, and summing the characteristics of the nodes to obtain the characterization of the attribute space of the target node.
The low-order neighbor characterization learning module: and (3) aggregating the attributes of the first-order neighbor nodes of the target node by using the simplified graph neural network, and acquiring the low-order neighbor characterization of the target node as shown in fig. 3.
The high-order neighbor characterization learning module: the attribute of the high-order neighbor node of the target node is aggregated through random walk and a attention mechanism based on a transducer, and as shown in fig. 4 and 5, the high-order neighbor characterization of the target node is obtained.
The characterization fusion module: and fusing the feature space characterization, the low-order neighbor node characterization and the high-order neighbor node characterization of the target node.
The attribute reasoning module: predicting the fusion characterization of the target node through a multi-layer perceptron to obtain the predicted attribute of the target node, and filling the predicted attribute into the corresponding missing attribute through a marking matrix.
Example 3: the embodiment performs instance verification based on the method and the system
In order to verify the accuracy of model attribute completion, the invention provides the following three data sets: experiments were performed on database systems and program logic networks (DataBase systems and Logic Programming, DBLP), international computer society networks (Association for Computing Machinery, ACM) and internet movie databases (Internet Movie Database, IMDb), using Heat Kernel and corestation as evaluation metrics, and compared to seven existing models.
The seven existing models are respectively: matrix complement (Matrix Completion, MC), maximum likelihood estimation (Expectation Maximization, EM), multi-layer perceptron (Multilayer Perceptron, MLP)
) Support vector regression (Support Vector Regression, SVR), heterogeneous graph annotation force network (heterogeneous graph attention networks, HGAT), adaptive graph neural network (Adaptive Graph Convolutional Network, AGCN), and heterogeneous graph neural network by attribute-complementing (Heterogeneous Graph Neural Network via Attribute Completion, HGNN-AC).
Table 1 results of comparative experiments
The final experimental results are shown in Table 1, where AC-HEN is the method provided by the present invention. It can be seen that on three real data sets, both the Heat Kernel and the corestation of the attribute completion method provided by the invention are significantly higher than those of other methods, which means that the model constructed by the invention is superior to other existing models, and the accuracy of attribute completion is higher.
The above-mentioned plan is merely an implementation method in the present invention, but the scope of the present invention is not limited thereto, and all those skilled in the art should understand that the conceivable substitutions or alterations are included in the scope of the present invention, so the scope of the present invention shall be defined by the scope of the claims.

Claims (9)

1. The attribute completion system based on the graph neural network and the attention mechanism is characterized by comprising a data preprocessing module, a marking matrix construction module, a feature space characterization learning module, a low-order neighbor characterization learning module, a high-order neighbor characterization learning module, a characterization fusion module and an attribute reasoning module: the data preprocessing module is used for: attribute features in an original dataset are first normalized, the dataset comprising: database system and program logic network, internet computer science network and Internet film database; then dividing the data into a training set, a testing set and a verification set, randomly removing attribute information of nodes in the training set, and recording the removed attribute information as a true value to guide a model to learn; the marking matrix construction module: traversing the node attributes in the data set, marking the missing attribute values of the nodes, and further forming an attribute marking matrix M; the feature space representation learning module: selecting nodes similar to the target node in the structural space, then giving weight to the nodes, and summing the characteristics of the nodes to obtain the representation of the attribute space of the target node; the low-order neighbor characterization learning module: aggregating the attribute of the first-order neighbor node of the target node by using the simplified graph neural network to obtain the low-order neighbor characterization of the target node; the high-order neighbor characterization learning module: the attribute of a high-order neighbor node of the target node is aggregated through random walk and a transducer-based attention mechanism, and the high-order neighbor characterization of the target node is obtained; the characterization fusion module: fusing the feature space representation, the low-order neighbor node representation and the high-order neighbor node representation of the target node; the attribute reasoning module: predicting the fusion characterization of the target node through a multi-layer perceptron to obtain a predicted attribute of the target node, and filling the predicted attribute into a corresponding missing attribute through a marking matrix; the completion method of the attribute completion system based on the graph neural network and the attention mechanism comprises the following steps:
s1: acquiring a heterogeneous attribute network with missing attributes, wherein a node with missing attributes is called a target node, and a node with complete attributes is called a source node;
s2: k source nodes which are most similar to the existing attribute of the target node are selected by adopting a K neighbor algorithm to complete the attribute in the feature space;
s3: after K source nodes which are most similar to the target node are screened out by utilizing cosine similarity, a learned parameter is endowed to each source node to dynamically adjust the weight of the attribute of each source node for complementing the attribute of the target node, and attribute characterization of the characteristic space endowed with the weight is obtained;
s4: the nodes directly connected with the target node in the heterogeneous attribute network are aggregated with the target node to obtain a low-order representation of the structural space; the specific aggregation mode is realized by simplifying a graph neural network;
s5: firstly, obtaining a high-order node of a target node by using a random walk mode;
s6: different weights are given to nodes in the high-order node sequence based on a transducer, and high-order characterization of the given weights of the structural space is obtained;
s7: splicing the attribute representation of the characteristic space, the low-order representation of the structural space and the high-order representation of the structural space, and then sending the spliced representation into a multi-layer perceptron to convert the representation into the attribute of the target node;
s8: the method comprises the steps of manually removing part of attributes of part of nodes by adopting a supervised learning mode, then complementing the attributes by reconstruction prediction, continuously training through the difference between the complemented attributes and the true attributes, and finally complementing the attribute missing of other nodes by utilizing a trained model.
2. The attribute completion system according to claim 1, wherein in S1, a heterogeneous network with missing attributes is defined as Represents the set of vertices in the graph, ε represents the set of edges in the graph, +.>Attribute matrix representing vertexes in the graph, n is the number of vertexes in the graph, m is the dimension of each vertex feature,/for each vertex feature>Is a marking matrix, when->When (I)>The corresponding attribute in (a) is missing; on the contrary, let(s)>The corresponding attributes of (a) are complete.
3. The attribute completion system according to claim 1, wherein in S2, K source nodes most similar to the existing attribute of the target node are selected by K neighbors to perform attribute completion in the feature space, and cosine similarity is used to measure similarity of two attribute vectors;
wherein ,Su,v Representing the similarity of the two nodes,representing the existing properties of the target node +.>Representing the attribute corresponding to the existing attribute of the target node in the source node, S u,v The larger the value, the higher the similarity between the target node and the source node.
4. The property completion system of claim 1, wherein in S3, the dynamic adjustment is performed according to formula (2):
wherein ,for characterization of the feature space of the target node, N k (V u ) Is in contact with the target node V u Is the most similar set of k source nodes,/-)>For each source node corresponding to a leachable adjustment weight, X v Is a feature vector of the source node.
5. The attribute completion system according to claim 1, wherein in S4, after the feature space finishes the feature learning of the target node, the feature needs to be learned for the target node in the structural space; the specific formula (3) is as follows:
H (l) =A*H (l-1) *W (l) (3)
wherein ,W(l) To simplify the weight matrix of the first layer in the graph neural network, H (l) Is the output of the first layer, at the first layer of the reduced graph neural networkA is an adjacency matrix.
6. The attribute completion system of claim 1, wherein S5 is specifically: in order to capture high-order information of a target node in a heterogeneous network, the high-order node of the target node is obtained by using a random walk mode, wherein the random walk is a mode of a node in a traversal diagram combining the advantages of depth-first traversal and breadth-first traversal, and the specific traversal mode is shown in a formula (4):
wherein wvt Representing node v v To node v t Is a weight of an edge of (c).
7. The attribute completion system of claim 1, wherein S6 is specifically:
s6-1: firstly, carrying out linear change on the characteristics of a target node and a source node sequence, then calculating the weights of the target node and the nodes in each sequence based on the linear change, and carrying out softmax normalization operation on the obtained weights for the stability of calculation;
s6-2: then, performing new linear transformation independent of weight calculation on the node characteristics in the source node sequence again;
s6-3: finally, giving weight to the node characteristics of the source node sequence after each linear transformation, and accumulating the node characteristics to obtain the node characteristics of the target node higher order; the calculation method is shown in the formula (5):
wherein Representing node v u Characterization aggregated by a simplified graph neural network, +.>Representing node v u Is characterized by (1)> and />Projection parameter matrix representing a parameter which can be learned, +.>Representing node v u Higher order network characterization of (2);
s6-4: expanding the attention mechanism to a multi-head attention mechanism to capture multiple dependency relations between a target node and a Gao Jieyuan node; then, the multiple dependency relationships are sent to an average pooling layer to obtain final high-order network characterization, and the calculation method is shown in a formula (6):
wherein ,representing node v u The higher order network characterization obtained in the ith attention mechanism, # head represents the total number of times that the attention mechanism calculation needs to be performed.
8. The attribute completion system according to claim 1, wherein in S7, a specific calculation manner of the stitching is as shown in formula (7):
where Concat (,) represents a vector concatenation operation,for the predicted target node v u Is a property value of (a).
9. The attribute completion system of claim 1, wherein S8 is specifically:
s8-1: after the predicted attribute of the target node is obtained, the original existing attribute is reserved, the predicted attribute is filled in the missing attribute to complete the task of attribute completion, and the filling method is shown in a formula (8):
wherein ,indicating the predicted attribute of all nodes after attribute complementation, wherein, as follows, the Hadamard product is indicated, E is the matrix with the element of 1;
s8-2: the difference between the predicted attribute and the real attribute of the loss function measurement is set based on the Euclidean distance formula, and the specific calculation mode is as shown in formula (9):
wherein vtarget A set of target nodes is represented and,for node v u True missing attribute value, +_>For node v u Predicted missing attribute values.
CN202211043710.3A 2022-08-30 2022-08-30 Heterogeneous network attribute completion method based on graph neural network and attention mechanism Active CN115130663B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211043710.3A CN115130663B (en) 2022-08-30 2022-08-30 Heterogeneous network attribute completion method based on graph neural network and attention mechanism

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211043710.3A CN115130663B (en) 2022-08-30 2022-08-30 Heterogeneous network attribute completion method based on graph neural network and attention mechanism

Publications (2)

Publication Number Publication Date
CN115130663A CN115130663A (en) 2022-09-30
CN115130663B true CN115130663B (en) 2023-10-13

Family

ID=83388013

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211043710.3A Active CN115130663B (en) 2022-08-30 2022-08-30 Heterogeneous network attribute completion method based on graph neural network and attention mechanism

Country Status (1)

Country Link
CN (1) CN115130663B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115759199B (en) * 2022-11-21 2023-09-26 山东大学 Multi-robot environment exploration method and system based on hierarchical graph neural network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113094484A (en) * 2021-04-07 2021-07-09 西北工业大学 Text visual question-answering implementation method based on heterogeneous graph neural network
WO2021179640A1 (en) * 2020-03-10 2021-09-16 深圳大学 Graph model-based short video recommendation method, intelligent terminal and storage medium
WO2021179838A1 (en) * 2020-03-10 2021-09-16 支付宝(杭州)信息技术有限公司 Prediction method and system based on heterogeneous graph neural network model
CN114692867A (en) * 2022-03-24 2022-07-01 大连理工大学 Network representation learning algorithm combining high-order structure and attention mechanism
CN114723037A (en) * 2022-02-25 2022-07-08 上海理工大学 Heterogeneous graph neural network computing method for aggregating high-order neighbor nodes

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021179640A1 (en) * 2020-03-10 2021-09-16 深圳大学 Graph model-based short video recommendation method, intelligent terminal and storage medium
WO2021179838A1 (en) * 2020-03-10 2021-09-16 支付宝(杭州)信息技术有限公司 Prediction method and system based on heterogeneous graph neural network model
CN113094484A (en) * 2021-04-07 2021-07-09 西北工业大学 Text visual question-answering implementation method based on heterogeneous graph neural network
CN114723037A (en) * 2022-02-25 2022-07-08 上海理工大学 Heterogeneous graph neural network computing method for aggregating high-order neighbor nodes
CN114692867A (en) * 2022-03-24 2022-07-01 大连理工大学 Network representation learning algorithm combining high-order structure and attention mechanism

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
"Self-training method based on GCN for semi-supervised short text classification";Hongyan Cui等;《Information Sciences》;全文 *
基于注意力机制增强图卷积神经网络的个性化新闻推荐;杨宝生;;兰州文理学院学报(自然科学版)(第05期);全文 *
网络表示学习算法综述;丁钰;魏浩;潘志松;刘鑫;;计算机科学(第09期);全文 *

Also Published As

Publication number Publication date
CN115130663A (en) 2022-09-30

Similar Documents

Publication Publication Date Title
CN113961759B (en) Abnormality detection method based on attribute map representation learning
CN109389151B (en) Knowledge graph processing method and device based on semi-supervised embedded representation model
Liu et al. Fast attributed multiplex heterogeneous network embedding
CN110442802B (en) Multi-behavior preference prediction method for social users
CN113255895A (en) Graph neural network representation learning-based structure graph alignment method and multi-graph joint data mining method
CN115130663B (en) Heterogeneous network attribute completion method based on graph neural network and attention mechanism
CN116628597A (en) Heterogeneous graph node classification method based on relationship path attention
Chen et al. Heterogeneous graph convolutional network with local influence
CN115699058A (en) Feature interaction through edge search
CN117349743A (en) Data classification method and system of hypergraph neural network based on multi-mode data
CN117556148B (en) Personalized cross-domain recommendation method based on network data driving
Zhang et al. Graph alternate learning for robust graph neural networks in node classification
Wang et al. Community discovery algorithm of complex network attention model
Pamir et al. Electricity theft detection for energy optimization using deep learning models
Sun et al. LSEnet: Lorentz Structural Entropy Neural Network for Deep Graph Clustering
CN117408336A (en) Entity alignment method for structure and attribute attention mechanism
CN117236374A (en) Layering interpretation method based on fully developed material graph neural network
CN115174421B (en) Network fault prediction method and device based on self-supervision unwrapping hypergraph attention
CN114298854B (en) Weak supervision user identity linking method combining learning representation and alignment
CN116467466A (en) Knowledge graph-based code recommendation method, device, equipment and medium
CN116976402A (en) Training method, device, equipment and storage medium of hypergraph convolutional neural network
CN113159976B (en) Identification method for important users of microblog network
Zhou et al. A structure distinguishable graph attention network for knowledge base completion
CN114611668A (en) Vector representation learning method and system based on heterogeneous information network random walk
Ghadirian et al. Hybrid adaptive modularized tri-factor non-negative matrix factorization for community detection in complex networks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant