Nothing Special   »   [go: up one dir, main page]

CN111831783B - Method for extracting chapter-level relation - Google Patents

Method for extracting chapter-level relation Download PDF

Info

Publication number
CN111831783B
CN111831783B CN202010644404.XA CN202010644404A CN111831783B CN 111831783 B CN111831783 B CN 111831783B CN 202010644404 A CN202010644404 A CN 202010644404A CN 111831783 B CN111831783 B CN 111831783B
Authority
CN
China
Prior art keywords
entity
representation
document
sentence
abstract semantic
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010644404.XA
Other languages
Chinese (zh)
Other versions
CN111831783A (en
Inventor
张世琨
叶蔚
李博
胡文蕙
张君福
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Peking University Software Engineering Co ltd
Original Assignee
Beijing Peking University Software Engineering Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Peking University Software Engineering Co ltd filed Critical Beijing Peking University Software Engineering Co ltd
Priority to CN202010644404.XA priority Critical patent/CN111831783B/en
Publication of CN111831783A publication Critical patent/CN111831783A/en
Application granted granted Critical
Publication of CN111831783B publication Critical patent/CN111831783B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/31Indexing; Data structures therefor; Storage structures
    • G06F16/313Selection or weighting of terms for indexing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • G06F16/353Clustering; Classification into predefined classes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Databases & Information Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Molecular Biology (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Animal Behavior & Ethology (AREA)
  • Machine Translation (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention provides a chapter-level relation extraction method, relates to the technical field of natural language processing, and mainly solves the technical problems of calculation resource consumption and logic reasoning between a target entity and a non-target entity aiming at chapter-level documents. The invention comprises the following steps: inputting a document to be processed, wherein the document is a chapter-level document; processing the document based on bidirectional attention constraint to obtain abstract semantic representations of entities and sentences, wherein the abstract semantic representations have global information and logical reasoning information; and judging the relation type of the target entity pair in the document based on the abstract semantic representation. The developer can efficiently and accurately extract the chapter-level relation by using the method disclosed by the invention, and simultaneously solves two main problems of chapter-level relation extraction, namely the problem of calculation cost caused by traversing all entities to generate alternative samples and the problem of logic reasoning between target entities and non-target entities.

Description

Method for extracting chapter-level relation
Technical Field
The invention relates to the technical field of natural language processing, in particular to a chapter-level relation extraction method in the technical field of bidirectional attention constraint and graph convolution neural network intersection. The invention obtains abstract representation of entity and document by constraint of bidirectional attention mechanism, and uses chapter-level relation extraction technique of graph convolution neural network for logical reasoning.
Background
The relation extraction is based on given context and two target entities, and judgesTwo target entitiesWhat the relationship between them is. Relationship extraction is one of the very important techniques for constructing large-scale knowledge-graphs, and can also be used to assist some downstream tasks, such as question-answering systems, etc. The research on relation extraction is mainly divided into two directions, one direction is based on a traditional machine learning thought for constructing a large number of artificial feature engineering, and the other direction is a deep learning method for building a neural network.
In traditional machine learning solutions, some researchers construct different features for an entire sentence and for a given target entity pair. The solution to traditional machine learning is very dependent on feature engineering and expertise, so that a model needs a large amount of manpower to construct features, meanwhile, the generalization capability of the model can greatly fluctuate along with the change of a data set, and in recent years, a large number of researchers use a deep learning method to solve the problem of relation extraction. In terms of deep learning, researchers have focused their research directions on the application of Convolutional Neural Networks (CNNs), recurrent Neural Networks (RNNs), and attention mechanisms, as well as on improvements to the above models, thereby also yielding a number of relevant solutions based on deep learning approaches.
The existing task definition and method for extracting the relation are all tasks based on sentence level, namely, aiming at a sentence as a context, aiming at two target entities appearing in the current sentence, the relation of the target entities is judged. However, in a practical application scenario, the text processed in most cases is not just as simple at the sentence level, relatively speaking, a text with one chapter is more common, a large number of entities appear in each chapter, and at the same time, each target entity appears multiple times in one document, so some classical methods in relation extraction at the sentence level are not available in this case, such as entity location vectors, PCNN, and so on. Meanwhile, the length of the text at the article level is often hundreds or even thousands, which is far greater than that of the text at the sentence level, the text at the overlength causes difficult extraction of the information of the whole document, the information is depended on for a long distance, and good representation and interaction of the entity are difficult, so that the effect of extracting the model by the chapter-level relation is greatly reduced. Meanwhile, because a large number of entities exist, each combination of the entities should become a potential target entity pair, and by traversing all the entity combinations in each article, a sentence and target entity sample is constructed, and a sentence-level relation extraction model is applied, so that the calculation data consumption is greatly increased. More importantly, due to the increase of the text length, a large number of entities exist in the input document, and the relationship between two target entities is obtained through not only direct interaction but also logic reasoning among non-target entities, and can be obtained through one or more times of reasoning among the entities, so that the accurate classification result cannot be obtained for the relationship needing reasoning.
Disclosure of Invention
One of the purposes of the invention is to provide a chapter-level relation extraction method, which solves the technical problems of calculation resource consumption and logic reasoning between target entities and non-target entities aiming at chapter-level problems in the prior art. Numerous advantageous effects can be achieved in the preferred embodiments of the present invention, as described in detail below.
In order to achieve the above purpose, the present invention provides the following technical solutions:
the invention relates to a chapter-level relation extraction method, which comprises the following steps:
inputting a document to be processed, wherein the document is a chapter-level document;
processing the document based on bidirectional attention constraint to obtain abstract semantic representations of entities and sentences, wherein the abstract semantic representations have global information and logical reasoning information;
and judging the relation type of the target entity pair in the document based on the abstract semantic representation.
Further, the processing the document based on the bidirectional attention constraint to obtain an abstract semantic representation of an entity and a sentence includes:
obtaining a representation of each entity and a representation of each sentence in the document;
acquiring a global entity interaction matrix based on the representation of each entity;
calculating the global entity interaction matrix and the representation of each sentence based on a first attention mechanism to obtain the representation of each sentence based on entity attention;
calculating the representation of each sentence based on the entity attention based on the graph convolution network to obtain the representation of each sentence with logic reasoning information;
based on a second attention mechanism, calculating the global entity interaction matrix and the representation of each sentence with logical reasoning information to obtain a new global entity interaction matrix with global information and logical reasoning information, and taking the new global entity interaction matrix as the abstract semantic representation of the entity and the sentence.
Further, the obtaining a representation of each entity and a representation of each sentence in the document includes:
acquiring an abstract semantic representation of the document;
a representation of each entity and a representation of each sentence in the document is obtained based on the abstract semantic representation of the document.
Further, the obtaining a representation of each sentence in the document based on the abstract semantic representation of the document includes:
and extracting the composition of each sentence from the abstract semantic representation of the document according to the starting position and the ending position of each sentence, carrying out maximum pooling processing on the composition of each sentence, and taking a sentence vector obtained after the maximum pooling processing as the representation of each sentence.
Further, the obtaining a representation of each entity in the document based on the abstract semantic representation of the document includes:
and determining words forming the entities according to the positions of each entity, calculating an average word vector of the words, taking the average word vector as an entity vector, and taking the average vector of the entity vectors of the same entity as the representation of each entity.
Further, the obtaining the abstract semantic representation of the document includes:
performing word vector conversion on the document to obtain a word vector matrix of the document;
and carrying out bidirectional LSTM operation on the word vector matrix to obtain the abstract semantic representation of the document.
Further, the obtaining a global entity interaction matrix based on the representation of each entity includes:
information interaction is carried out on every two entities, and two entity vectors are converted into one entity interaction vector;
and forming a global entity interaction matrix by all entity interaction vectors.
Further, the matrix weights adopted by the first attention mechanism and the second attention mechanism are the same.
Further, the determining the relationship type of the target entity pair in the document based on the abstract semantic representation includes:
performing column-wise weighting operation on the new global entity interaction matrix to obtain a global entity interaction vector;
and inputting the global entity interaction vector into a preset classification function for distinguishing the relationship types of the entity pairs to obtain a classification result, and taking the classification result as the relationship type of the target entity pair in the document.
Further, the classification function employs a softmax function.
The chapter-level relation extraction method provided by the invention has at least the following beneficial technical effects:
the invention can efficiently and simultaneously acquire entity representation no matter how many different entities are in the current document by representing each entity and encoding the relative positions among the entities, and further inputs the relationship results of all potential entity pairs in the document once. In addition, the present invention employs a bi-directional attention mechanism for the problem of logical reasoning between non-target entities, i.e. first obtaining each sentence and the coded representation of each entity, then for each entity, traversing the obtaining so whether the entity has a relationship to the combination and, if so, which relationship in a given set of relationships.
The developer can efficiently and accurately extract the chapter-level relation by using the method disclosed by the invention, and simultaneously solves two main problems of chapter-level relation extraction, namely the problem of calculation cost caused by traversing all entities to generate alternative samples and the problem of logic reasoning between target entities and non-target entities.
Drawings
In order to more clearly illustrate the embodiments of the invention or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described, it being obvious that the drawings in the following description are only some embodiments of the invention, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a flow chart of a chapter level relationship extraction method of the present invention;
FIG. 2 is a schematic diagram of a chapter level relationship extraction method of the present invention;
FIG. 3 is a schematic flow chart of the present invention for processing the document based on bi-directional attention constraints;
fig. 4 is a schematic representation of an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the technical solutions of the present invention will be described in detail below. It will be apparent that the described embodiments are only some, but not all, embodiments of the invention. All other embodiments, based on the examples herein, which are within the scope of the invention as defined by the claims, will be within the scope of the invention as defined by the claims.
Referring to fig. 1 and 2, the present invention is a chapter level relation extraction method, which includes the steps of:
s1: inputting a document to be processed, wherein the document is a chapter-level document;
s2: processing the document based on bidirectional attention constraint to obtain abstract semantic representations of entities and sentences, wherein the abstract semantic representations have global information and logical reasoning information;
s3: and judging the relation type of the target entity pair in the document based on the abstract semantic representation.
It should be noted that the bidirectional attention mechanism is based on the following intuitive observations: if a given sentence contains a large amount of information that determines what the target entity pair is, then the target entity pair, in turn, will find a closer relationship with the given sentence when calculating the relationship with each sentence in the document.
The invention mainly aims at the documents of the chapter level, namely the texts which are often hundreds or thousands in length and far larger than the sentence level, and a large number of target entities exist in the documents of the chapter level. According to the relation extraction method for the target entity in the chapter-level document based on the bidirectional attention constraint, the input chapter-level document is subjected to a series of transformation and information extraction, whether the target entity pair combination contains a relation or not is finally obtained, and if the relation is contained, the relation is which in a given relation set. The developer can efficiently and accurately extract the chapter-level relation, and simultaneously solve two main problems of chapter-level relation extraction, namely the calculation cost problem caused by traversing all entities to generating the alternative sample and the logic reasoning problem between the target entity and the non-target entity.
Referring to fig. 3, the processing the document based on the bidirectional attention constraint to obtain an abstract semantic representation of entities and sentences includes:
s21: obtaining a representation of each entity and a representation of each sentence in the document;
s22: acquiring a global entity interaction matrix based on the representation of each entity;
s23: calculating the global entity interaction matrix and the representation of each sentence based on a first attention mechanism to obtain the representation of each sentence based on entity attention;
s24: calculating the representation of each sentence based on the entity attention based on the graph convolution network to obtain the representation of each sentence with logic reasoning information;
s25: based on a second attention mechanism, calculating the global entity interaction matrix and the representation of each sentence with logical reasoning information to obtain a new global entity interaction matrix with global information and logical reasoning information, and taking the new global entity interaction matrix as the abstract semantic representation of the entity and the sentence.
In step S21, the obtaining a representation of each entity and a representation of each sentence in the document includes:
acquiring an abstract semantic representation of the document;
a representation of each entity and a representation of each sentence in the document is obtained based on the abstract semantic representation of the document.
Wherein the obtaining the abstract semantic representation of the document comprises:
performing word vector conversion on the document to obtain a word vector matrix of the document;
and carrying out bidirectional LSTM operation on the word vector matrix to obtain the abstract semantic representation of the document.
It should be noted that, the input document to be processed is encoded by using the pre-training word vector matrix, and is converted into word vectors, so as to obtain word vector representation of the document, and abstract semantic representation of the document is further obtained through the bidirectional LSTM.
Suppose there are k different entities in the document than there are n words, m sentences. Let the i-th word be denoted as w i There is a pre-trained word vector matrix of size N x d, where N is the number of words in the word vector matrix and d is the dimension of each word in the word vector matrix. Converting each word into a word vector using a pre-trained word vector matrix pair, i.e., for w i And finding out the corresponding word vectors, and splicing word vectors of all words in the document to obtain word vector representation of the whole document, wherein the size of the word vector representation is n multiplied by d.
Assuming that the number of hidden units in the bidirectional LSTM is H, after passing through the bidirectional LSTM, the abstract semantics of the document are represented as H, and the size is n×2h.
The obtaining a representation of each sentence in the document based on the abstract semantic representation of the document includes:
and extracting the composition of each sentence from the abstract semantic representation of the document according to the starting position and the ending position of each sentence, carrying out maximum pooling processing on the composition of each sentence, and taking a sentence vector obtained after the maximum pooling processing as the representation of each sentence.
It should be noted that, when the starting position and the ending position of each sentence are preset in the article to be processed, each sentence in the document is easily extracted, and the vector representation of each sentence is obtained by using the maximum pooling operation.
Assuming that a certain sentence in a document starts and ends with the a-th word and the b-th word, respectively, the sentence is expressed as:
wherein,the size of (2) is 1×2h.
The obtaining a representation of each entity in the document based on the abstract semantic representation of the document includes:
and determining words forming the entities according to the positions of each entity, calculating an average word vector of the words, taking the average word vector as an entity vector, and taking the average vector of the entity vectors of the same entity as the representation of each entity.
It should be noted that, assuming that a given target entity appears 3 times, the entity expressions are e respectively 1 1 ,e 2 1 And e 3 1 Then the representation of the current entity The size of (2) is 1×2h.
In step S22, the obtaining a global entity interaction matrix based on the representation of each entity includes:
information interaction is carried out on every two entities, and two entity vectors are converted into one entity interaction vector;
and forming a global entity interaction matrix by all entity interaction vectors.
It should be noted that, traversing all entity pairs in the current documentSimultaneously considering the sequence of the two entities (namely the directions of the entities), and then carrying out information interaction on the two entities, wherein the interaction method is to transform two entity vectors into an advanced entity interaction vector v through a bilinear layer ij
Assuming that the two entity vectors areThe operation of the bilinear layer is v ij =σ(e i W v e j +b v ) Wherein W is v ,b v Is a trainable parameter.
All entity interaction vectors are constructed as a global entity interaction matrix. Assuming that there are k entities in the current document, there are k (k-1) interaction vectors in the constructed global entity interaction matrix.
Step S23: calculating the global entity interaction matrix and the representation of each sentence based on a first attention mechanism to obtain the representation of each sentence based on entity attention; the method comprises the following specific steps:
the attention mechanism operation is performed on each sentence representation and the global entity interaction matrix, so that the invention obtains a certain logic reasoning capability, and the expression of each sentence based on the entity attention is obtained.
Specifically, from the foregoing, it can be known that a sentence is expressed asIf the representation of a certain entity interaction vector is v, the attention weight of the current sentence and the entity interaction vector is calculated as +.>Calculating the current sentence and each entity interaction vector to obtain k (k-1) attention weights, and then weighting all entity interaction vectors to obtain the representation of each sentence based on the attention of the entity interaction vector>I.e. < ->All sentences are combined into a feature matrix X by columns.
Step S24: calculating the representation of each sentence based on the entity attention based on the graph convolution network to obtain the representation of each sentence with logic reasoning information; the method comprises the following specific steps:
and carrying out graph convolution operation on the feature matrix X of the sentence through the following transformation: l=ρ (AXW) 2 ) Wherein ρ is a nonlinear activation function, e.g. sigmoid function, W 2 Is a trainable parameter. The L thus obtained is the same size as X, each column of which is a representation L of a sentence.
It should be noted that, through a graph convolution neural network layer, the sequential connection information and the logical reasoning information between each sentence are obtained. After passing this layer, each sentence representation with sequential contact information and logical inference information will be obtained.
Step S25: based on a second attention mechanism, calculating the global entity interaction matrix and the representation of each sentence with logical reasoning information to obtain a new global entity interaction matrix with global information and logical reasoning information, and taking the new global entity interaction matrix as the abstract semantic representation of the entity and the sentence. The method specifically comprises the following steps:
specifically, the step S24 can learn that a sentence is expressed asIf the representation of a certain entity interaction vector is v, the attention weight of the current entity interaction vector and sentence is calculated as +.>Calculating the current entity interaction vector and each sentence to obtain m attention weights, and then weighting all sentence vectors to obtain each entity interaction vector based on the attention of the sentenceCharacterization->I.e. < ->
It should be noted that, the matrix weights adopted by the first attention mechanism and the second attention mechanism are the same. This is because, given two target entities and contexts, the degree of association between the target entity and context should be consistent with the degree of association between the context and target entity, so that the resulting matrix for both attention mechanisms is constrained to be as close as possible. Through bi-directional attention constraints, a representation of each entity for each sentence is derived, and then a new global entity interaction matrix with global information and logical reasoning information is obtained, which is used for classification of relation extraction.
In step S3, determining a relationship type of the target entity pair in the document based on the abstract semantic representation includes:
performing column-wise weighting operation on the new global entity interaction matrix to obtain a global entity interaction vector;
and inputting the global entity interaction vector into a preset classification function for distinguishing the relationship types of the entity pairs to obtain a classification result, and taking the classification result as the relationship type of the target entity pair in the document.
Preferably, the classification function employs a softmax function.
For each entity interaction vector generated from step S25Determining which relationship it belongs to, using the softmax function:
taking the obtained roughThe relationship type corresponding to y with the maximum value is used as the entity interaction vectorIs a relationship of (3).
It should be noted that, assuming that there are predefined 3 relationships, after a global entity interaction vector passes through softmax, 3 probability values are obtained, and respectively corresponding to the probability values of the 3 relationships, which probability is the largest is predicted as which relationship. For example, the target entities of the document are: the method comprises the steps of carrying out classification function operation on Beijing capital airport, beijing and Beijing subway airport lines, wherein the obtained relationship is belongings, capital and roads after the target entity pair is subjected to the classification function operation, and the probability that the relationship corresponds to 0.7, 0.2 and 0.1 respectively is that the relationship of the target entity pair is belonged to.
The implementation is illustrated with respect to fig. 4 as follows:
1. firstly, acquiring representations of all entities (such as Beijing capital airport, beijing subway airport line and the like) and sentences in an input document;
2. acquiring an attention weight matrix based on given target entities and sentences, and new entity and sentence representations;
3. after the new sentence representation passes through the graph convolution neural network, the sentence representation with logic reasoning information is obtained, and then the second attention mechanism calculation is carried out with the entity representation;
4. constraint is carried out on the obtained calculation results of the two attentiveness mechanisms, and interaction information of each entity pair is calculated at the same time;
5. the interaction information for each entity pair is used to determine whether or not a given target entity pair has a relationship, and if so, what.
The foregoing is merely illustrative of the present invention, and the present invention is not limited thereto, and any person skilled in the art will readily recognize that variations or substitutions are within the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.

Claims (9)

1. A chapter-level relationship extraction method, comprising:
inputting a document to be processed, wherein the document is a chapter-level document;
processing the document based on bidirectional attention constraint to obtain abstract semantic representations of entities and sentences, wherein the abstract semantic representations have global information and logical reasoning information;
judging the relation type of a target entity pair in the document based on the abstract semantic representation;
the processing the document based on the bidirectional attention constraint to obtain abstract semantic representation of entities and sentences comprises the following steps:
obtaining a representation of each entity and a representation of each sentence in the document;
acquiring a global entity interaction matrix based on the representation of each entity;
calculating the global entity interaction matrix and the representation of each sentence based on a first attention mechanism to obtain the representation of each sentence based on entity attention;
calculating the representation of each sentence based on the entity attention based on the graph convolution network to obtain the representation of each sentence with logic reasoning information;
based on a second attention mechanism, calculating the global entity interaction matrix and the representation of each sentence with logical reasoning information to obtain a new global entity interaction matrix with global information and logical reasoning information, and taking the new global entity interaction matrix as the abstract semantic representation of the entity and the sentence.
2. The extraction method according to claim 1, wherein the obtaining the representation of each entity and the representation of each sentence in the document comprises:
acquiring an abstract semantic representation of the document;
a representation of each entity and a representation of each sentence in the document is obtained based on the abstract semantic representation of the document.
3. The extraction method according to claim 2, wherein the obtaining a representation of each sentence in the document based on the abstract semantic representation of the document comprises:
and extracting the composition of each sentence from the abstract semantic representation of the document according to the starting position and the ending position of each sentence, carrying out maximum pooling processing on the composition of each sentence, and taking a sentence vector obtained after the maximum pooling processing as the representation of each sentence.
4. The extraction method according to claim 2, wherein the obtaining a representation of each entity in the document based on the abstract semantic representation of the document comprises:
and determining words forming the entities according to the positions of each entity, calculating an average word vector of the words, taking the average word vector as an entity vector, and taking the average vector of the entity vectors of the same entity as the representation of each entity.
5. The extraction method according to claim 2, wherein said obtaining an abstract semantic representation of said document comprises:
performing word vector conversion on the document to obtain a word vector matrix of the document;
and carrying out bidirectional LSTM operation on the word vector matrix to obtain the abstract semantic representation of the document.
6. The extraction method according to claim 4, wherein the obtaining a global entity interaction matrix based on the representation of each entity comprises:
information interaction is carried out on every two entities, and two entity vectors are converted into one entity interaction vector;
and forming a global entity interaction matrix by all entity interaction vectors.
7. The extraction method according to claim 1, wherein the matrix weights used by the first and second attention mechanisms are the same.
8. The extraction method according to claim 1, wherein said determining a relationship type of a target entity pair in the document based on the abstract semantic representation comprises:
performing column-wise weighting operation on the new global entity interaction matrix to obtain a global entity interaction vector;
and inputting the global entity interaction vector into a preset classification function for distinguishing the relationship types of the entity pairs to obtain a classification result, and taking the classification result as the relationship type of the target entity pair in the document.
9. The extraction method according to claim 8, wherein the classification function employs a softmax function.
CN202010644404.XA 2020-07-07 2020-07-07 Method for extracting chapter-level relation Active CN111831783B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010644404.XA CN111831783B (en) 2020-07-07 2020-07-07 Method for extracting chapter-level relation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010644404.XA CN111831783B (en) 2020-07-07 2020-07-07 Method for extracting chapter-level relation

Publications (2)

Publication Number Publication Date
CN111831783A CN111831783A (en) 2020-10-27
CN111831783B true CN111831783B (en) 2023-12-08

Family

ID=72900251

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010644404.XA Active CN111831783B (en) 2020-07-07 2020-07-07 Method for extracting chapter-level relation

Country Status (1)

Country Link
CN (1) CN111831783B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113435190B (en) * 2021-05-18 2022-08-05 北京理工大学 Chapter relation extraction method integrating multilevel information extraction and noise reduction
CN113297838A (en) * 2021-05-21 2021-08-24 华中科技大学鄂州工业技术研究院 Relationship extraction method based on graph neural network
CN113468325A (en) * 2021-06-09 2021-10-01 广西电网有限责任公司 Document level relation extraction method based on associated sentence selection and relation graph reasoning
CN113822019B (en) * 2021-09-22 2024-07-12 科大讯飞股份有限公司 Text normalization method, related device and readable storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106354710A (en) * 2016-08-18 2017-01-25 清华大学 Neural network relation extracting method
CN108280062A (en) * 2018-01-19 2018-07-13 北京邮电大学 Entity based on deep learning and entity-relationship recognition method and device
CN111008529A (en) * 2019-07-24 2020-04-14 贵州大学 Chinese relation extraction method based on neural network
CN111274800A (en) * 2020-01-19 2020-06-12 浙江大学 Inference type reading understanding method based on relational graph convolution network

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108280112B (en) * 2017-06-22 2021-05-28 腾讯科技(深圳)有限公司 Abstract generation method and device and computer equipment

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106354710A (en) * 2016-08-18 2017-01-25 清华大学 Neural network relation extracting method
CN108280062A (en) * 2018-01-19 2018-07-13 北京邮电大学 Entity based on deep learning and entity-relationship recognition method and device
CN111008529A (en) * 2019-07-24 2020-04-14 贵州大学 Chinese relation extraction method based on neural network
CN111274800A (en) * 2020-01-19 2020-06-12 浙江大学 Inference type reading understanding method based on relational graph convolution network

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
Exploiting Entity BIO Tag Embeddings and Multi-task Learning for Relation Extraction with Imbalanced Data;Ye, Wei* 等;computer Science;1-10 *
Improve relation extraction with dual attention-guided graph convolutional networks;Zhixin Li 等;Neural Computing and Applications;1773–1784 *
图文知识图谱中的关系抽取算法研究;李智超;中国优秀硕士学位论文全文数据库 信息科技辑(第11期);I138-572 *
基于BiLSTM并结合自注意力机制和句法信息的隐式篇章关系分类;凡子威 等;计算机科学(第05期);214-220 *
基于双向LSTM和自注意力机制的中文关系抽取研究;刘鉴 等;山西大学学报(自然科学版)(第01期);8-13 *
基于联合注意力机制的篇章级机器翻译;李京谕 等;中文信息学报(第12期);45-53 *
基于语境交互感知和模式筛选的隐式篇章关系识别;郭凤羽 等;计算机学报(第05期);901-915 *
结合实体共现信息与句子语义特征的关系抽取方法;马语丹 等;中国科学:信息科学(第11期);1533-1545 *

Also Published As

Publication number Publication date
CN111831783A (en) 2020-10-27

Similar Documents

Publication Publication Date Title
CN109284506B (en) User comment emotion analysis system and method based on attention convolution neural network
CN111831783B (en) Method for extracting chapter-level relation
CN109033068B (en) Method and device for reading and understanding based on attention mechanism and electronic equipment
Mo et al. Large language model (llm) ai text generation detection based on transformer deep learning algorithm
CN111783394B (en) Training method of event extraction model, event extraction method, system and equipment
CN106502985B (en) neural network modeling method and device for generating titles
CN110287323B (en) Target-oriented emotion classification method
CN113987187B (en) Public opinion text classification method, system, terminal and medium based on multi-label embedding
CN113553412B (en) Question-answering processing method, question-answering processing device, electronic equipment and storage medium
Mukherjee et al. Utilization of oversampling for multiclass sentiment analysis on amazon review dataset
WO2023159767A1 (en) Target word detection method and apparatus, electronic device and storage medium
CN111145914B (en) Method and device for determining text entity of lung cancer clinical disease seed bank
Zhang et al. A BERT fine-tuning model for targeted sentiment analysis of Chinese online course reviews
CN111597815A (en) Multi-embedded named entity identification method, device, equipment and storage medium
CN111859979A (en) Ironic text collaborative recognition method, ironic text collaborative recognition device, ironic text collaborative recognition equipment and computer readable medium
CN111400494A (en) Sentiment analysis method based on GCN-Attention
CN111597816A (en) Self-attention named entity recognition method, device, equipment and storage medium
CN114064852A (en) Method and device for extracting relation of natural language, electronic equipment and storage medium
CN116644760A (en) Dialogue text emotion analysis method based on Bert model and double-channel model
Wang et al. Deep learning sentiment classification based on weak tagging information
CN114492661B (en) Text data classification method and device, computer equipment and storage medium
CN116595023A (en) Address information updating method and device, electronic equipment and storage medium
Touati-Hamad et al. Arabic quran verses authentication using deep learning and word embeddings
CN114356990A (en) Base named entity recognition system and method based on transfer learning
CN113139057A (en) Domain-adaptive chemical potential safety hazard short text classification method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant