CN115618022B - Low-resource relation extraction method based on data synthesis and two-stage self-training - Google Patents
Low-resource relation extraction method based on data synthesis and two-stage self-training Download PDFInfo
- Publication number
- CN115618022B CN115618022B CN202211630125.3A CN202211630125A CN115618022B CN 115618022 B CN115618022 B CN 115618022B CN 202211630125 A CN202211630125 A CN 202211630125A CN 115618022 B CN115618022 B CN 115618022B
- Authority
- CN
- China
- Prior art keywords
- training
- data
- model
- stage
- self
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000012549 training Methods 0.000 title claims abstract description 114
- 238000000605 extraction Methods 0.000 title claims abstract description 43
- 230000015572 biosynthetic process Effects 0.000 title claims abstract description 38
- 238000003786 synthesis reaction Methods 0.000 title claims abstract description 38
- 238000000034 method Methods 0.000 claims abstract description 22
- 230000006870 function Effects 0.000 claims description 10
- 230000008569 process Effects 0.000 claims description 8
- 238000004821 distillation Methods 0.000 claims description 6
- 238000005070 sampling Methods 0.000 claims description 6
- 241000764238 Isis Species 0.000 claims description 3
- 238000009966 trimming Methods 0.000 claims description 2
- 238000002372 labelling Methods 0.000 abstract description 6
- 238000009826 distribution Methods 0.000 description 4
- 102100033814 Alanine aminotransferase 2 Human genes 0.000 description 1
- 101710096000 Alanine aminotransferase 2 Proteins 0.000 description 1
- 230000004913 activation Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000013140 knowledge distillation Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 238000003062 neural network model Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 102000004169 proteins and genes Human genes 0.000 description 1
- 108090000623 proteins and genes Proteins 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000001308 synthesis method Methods 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/36—Creation of semantic tools, e.g. ontology or thesauri
- G06F16/367—Ontology
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Animal Behavior & Ethology (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Machine Translation (AREA)
Abstract
The invention relates to the field of data synthesis, and discloses a low-resource relation extraction method based on data synthesis and two-stage self-training. The two-stage self-training framework sequentially uses the non-labeling generated data and the labeled data to train the model in each iteration, so that on one hand, the model can be promoted to cooperatively learn from the non-labeling generated data and the labeled data, and on the other hand, the influence of noise of the generated data is effectively reduced. The method is fit for the low resource condition in the real scene, and can more effectively utilize the existing pre-training language model.
Description
Technical Field
The invention relates to the field of data synthesis, in particular to a low-resource relation extraction method based on data synthesis and two-stage self-training.
Background
The relation extraction system is used as an important technical support for knowledge extraction and map construction, aims to mine the relation among entities in unstructured texts, and becomes a research hotspot in the field of natural language processing in recent years. While neural network models, and in particular pre-trained language models, have made significant breakthroughs in the task of relationship extraction, training these models requires a large amount of labeling data. However, in many real-world scenarios, it is often time-consuming and labor-consuming to obtain high-quality annotation data, and thus how to build a relationship extraction system with good performance under limited resources and data becomes a significant challenge.
Remote supervision is widely studied as an effective method of constructing large-scale relational extraction data sets by aligning entities in text with an existing knowledge base, automatically annotating the relational extraction data. However, due to the difference of the relation mode and the text corpus, the data of the remote supervision annotation can be greatly different from the downstream task, and further optimization of the model performance is inhibited. For example, with the reliance on existing knowledge bases, current remote supervision mostly uses Wikidata as the source of relational triples, and wikipedia as the remote supervision corpus. This limits the pattern and text of the annotation data to daily knowledge, but the downstream tasks may involve other specialized knowledge in different fields, such as semantic relationships between names, interactions of chemical proteins.
Considering that the large-scale language model has demonstrated strong language generating capability in a plurality of fields such as news manuscripts, commodity categories, daily conversations and the like, the invention uses the large-scale language model for synthesizing data of a relation extraction task to solve the problem of scarcity of training data in a low-resource scene and the problem of field difference of remote supervision. On the other hand, for better use of the generated data, the present invention proposes a two-stage self-learning framework.
Disclosure of Invention
In order to solve the technical problems, the invention provides a low-resource relation extraction method based on data synthesis and two-stage self-training, and a self-learning framework generally iterates to label and learn pseudo labels without labeled data so as to guide continuous improvement of model performance. However, in each iteration of the two-stage self-learning framework of the present invention, training is performed using the generated data in the first stage and training is performed using the annotation data in the second stage. Because the annotation data is introduced at a later stage of the training process, the training process in the form of the sequence improves the attention of the model to the annotation data. Furthermore, the present invention formulates the training of the generated data as a knowledge distillation process using soft pseudo tags, rather than assigning hard tags to them. In general, the two-stage self-training framework utilizes the generated data, solves the problem of less labeled data under low resources, and further improves the performance of the knowledge extraction system.
In order to solve the technical problems, the invention adopts the following technical scheme:
a low-resource relation extraction method based on data synthesis and two-stage self-training comprises the following steps:
step one, data synthesis based on marked training data: converting the training data into a linear natural language sequence by adding a position symbol into the training data; constructing a data synthesis model based on a large-scale language model, and performing fine adjustment on the data synthesis model through training data; repeatedly performing the data synthesis process using polynomial sampling until an unlabeled generated dataset meeting the preset conditions is obtainedGenerating data with a place holder;in order to generate a sequence of words in the data,,respectively isIs a subject and an object of (a);to generate an amount of data;
step two, two-stage self-learning: in training data setTraining self-coding language model eta, and then generating data set with added position signUsing self-encoding languageModel eta classification, soft pseudo label marking:;
order theSoft pseudo tag set of (a),Is a soft pseudo tag; training multiple self-coding language models by using K different random seeds, marking as teacher model eta, and marking a soft pseudo-label set marked by a kth teacher model asThe method comprises the steps of carrying out a first treatment on the surface of the Initializing a new self-coding language model, which is recorded as student modelFor student modelTwo-stage training strategies are applied: in the first stage training, distillation training is performed using the generated data with soft pseudo tags:the method comprises the steps of carrying out a first treatment on the surface of the Model studentsOptimized as a student modelCalculate distillation loss:
Represents KL divergence; in the second stage training, the student is modeledIn training data setTraining is carried out on:;as a standard cross-entropy loss function,the student model is obtained after the second stage training iteration; the next time the two-stage training strategy is executed, the student is modeledAs a teacher model η; repeating the two-stage training strategy until a data set is generatedEach generated data is marked with a soft pseudo tag;
step three, relation extraction: constructing a relation extraction model based on a self-coding language model; the generated data and the training data are collectively called as training examples, and the relationship labels of the training examples are called as real labelsInputting the training examples into a relation extraction model, and calculating the relation predicted by the relation extraction modelTie labelAnd training instance's true tagsTo train a relationship extraction model.
Further, in the first step, the fine tuning loss of the data synthesis model is allThe method comprises the steps of carrying out a first treatment on the surface of the The fine tuning mode of the data synthesis model and the pre-training mode of the data synthesis model are the same, whereinRepresenting probability functions, LLM represents a large-scale language model,is a word sequence in training data added with position symbolThe word(s) in (a) is (are),for special start symbol, after finishing trimmingAnd generating a pre-addition mark prompt.
Further, in the second step, the two-stage training strategy is repeatedly executed for T times; from generating data sets in each iterationSampling 1/T of generated data, and generating a data set after T iterationsAll the generated data in the table are marked with softAnd (5) pseudo tags.
Further, in the first step, when a placer is added to the training data:
wherein ,representing the word sequence after adding the placeholders in the training data,andto be used for marking training data word sequence bodyThe location identifier of the location(s),andfor marking training data word sequence objectsA location identifier of the location.
Further, when the relation label predicted by the relation extraction model in the third step is obtained, vector representations h of corresponding positions of word sequences in the training examples are spliced to be classified, and the relation label predicted by the relation extraction model is obtained:
wherein In order to activate the function,in order to be a fully connected network,is a vector representation of the subject of the object,for vector representation of objects, [;]representing a stitching operation.
Compared with the prior art, the invention has the beneficial technical effects that:
the invention provides a low-resource relation extraction method based on data synthesis and two-stage self-training, which comprises a data synthesis method and a two-stage self-training framework; the data synthesis effectively relieves the problems of less marked data and high marked cost in the current relation extraction task. The two-stage self-training framework sequentially uses the non-label generated data and the labeled data to train the self-coding language model in each round of iteration, so that on one hand, the self-coding language model can be promoted to cooperatively learn from the non-label generated data and the labeled data, and on the other hand, the influence of noise of the generated data is effectively reduced. The relation extraction system fits the low resource condition in the real scene, can more effectively utilize the existing pre-training language model, explores the potential thereof, and has wide application prospect.
Drawings
Figure 1 is a diagram of the overall model architecture of the present invention.
Detailed Description
A preferred embodiment of the present invention will be described in detail with reference to the accompanying drawings.
The invention is provided with a training data setTag set corresponding to training data set,For training dataCorresponding relation label, and,to train all annotated sets of relationship labels in the dataset,the number of training data in the training data set. Here, theIs the word sequence in the ith training data of the training data set,Representing word sequencesThe first of (3)Individual words, L being word sequencesThe length of the tube is equal to the length,andrespectively word sequencesIs provided with a main body and an object,representing subject in word sequenceThe starting position of (a) isThe end position is,Representing an object in a word sequenceThe starting position of (a) isThe end position is。
Generating a datasetCorresponding soft pseudo tag set,In the form of a soft pseudo tag,in order to generate a sequence of words in the data,,respectively isIs a subject and an object of (a);the amount of data is generated for the generated dataset.
Generating dataAnd training dataCollectively referred to as training examplesThe relationship label of the training example is called a real labelI.e. soft pseudo tagsSum relationship labelCollectively referred to as genuine labels。
The goal of the relation extraction is to learn a functionThe function isBy passing throughPredicting real tags。
The relation extraction system provided by the invention is shown in fig. 1, and comprises the following three parts: 1. extracting a model of the relation; 2. a data synthesis model oriented to a low-resource scene; 3. two-stage self-learning framework.
1. Relation extraction model
The subject of the relation extraction model employs a BERT-like autorecoding language model rather than an autoregressive language model, as the autorecoding language model generally performs better on language understanding class tasks. In view of previous studies on extraction models, the present invention has added special placeholders to word sequences of training examples of input relationship extraction models to emphasize the locations of subjects and objects in the word sequences.
wherein Representing the training example word sequence after the addition of a particular placer,andis used for marking the main bodyThe specific placer of the location(s),andis used for labeling objectsA special place identifier of the location.
After the coding of the relation extraction model, vector representations h of corresponding positions of the word sequences are spliced to classify:
wherein Is the function of the activation and,is a fully-connected network, and the network is a fully-connected network,is a vector representation of the subject,is a vector representation of the object, [;]representing the operation of the splicing operation,the relation label distribution predicted by the relation extraction model is calculatedAndtraining a relational extraction model.
The method of generating the training examples is described subsequently.
2. Data synthesis model oriented to low-resource scene
The present embodiment employs a data synthesis model based on a large-scale language model (LLM). In consideration of the strong language generating capability of the large-scale language model, the method and the device solve the problem of scarcity of training data and the problem of far supervision field difference in a low-resource scene by extracting data required by tasks through the large-scale language model synthesis relation based on the labeling data. The training instance of the relational extraction model has a specific structure:i.e. a fact of relationship is formed by a piece of text(i.e., word sequence)Comprises a main bodyObject and its manufacturing methodAnd (5) determining. Training data is first converted into a linear natural language sequence using a manner similar to equation one:
the data synthesis model may be based on any large-scale language model, such as GPT-2. Then, fine tuning is carried out on the large-scale language model based on the marked training data, the fine tuning mode is the same as the pre-training mode of the data synthesis model, and the fine tuning loss of the data synthesis modelThe method comprises the following steps:
wherein Representing probability functions, LLM represents a large-scale language model,is a word sequenceIn a word, and special start symbol<bos>(i.e. beginning of sentence) asIs added before the sequence. Note that here the relationship labels in the training data are ignoredThis is therefore an unconditional generation process. This can eliminate noise due to tag-semantic inconsistencies and enable the data synthesis model to model itself, helping the data synthesis model learn from unlabeled generated data. After finishing the fine tuning, only the special start symbol is needed<bos>A marker is added before to prompt generation and the generation process is repeatedly performed using polynomial sampling until a generation dataset is obtained that meets expectations。
3. Two-stage self-learning framework
Self-learning is a learning algorithm widely used in semi-supervised learning. Typically, to jointly learn from an unlabeled dataset and a labeled dataset, it is necessary to iteratively sample from the unlabeled dataset and assign pseudo tags to the sampled data, and then combine them with the labeled dataset to retrain the model. However, this naive merged design builds on a strong assumption that the unlabeled dataset must have exactly the same distribution as the labeled dataset, and the resulting data does not strictly meet this assumption.
To this end, the invention proposes a different two-stage self-learning framework: the model is trained separately using the unlabeled generated dataset and the labeled training dataset in turn, rather than combining them together for training. First, in the marked training data setTraining a self-coding language model (such as BERT model) eta, and then generating data with unlabeled position symbolsUsing the self-coding language model eta classification to realize soft pseudo tag marking:
order theSoft pseudo tag set of (a)Note that the superscript "≡" here indicates "soft pseudo tags", i.e. only the distribution of tag categories is maintained rather than further taking their argmax value. To further eliminate fluctuations in pseudo tags, K different random seeds are used to train multiple self-encoding language models, denoted as teacher model η, and the kth teacher model labeled soft pseudo tag set is denoted as。
Soft pseudo tags are soft pseudo tags, which are concepts corresponding to hard tags. The hard label can directly label samples with 0 and 1 discrete labels to represent positive and negative samples, and the soft label keeps the distribution of sample label types and marks the samples according to 0-1. In contrast, hard tags are simple to label, but are not microscopically optimized; soft tags are smoother and more expressive. The soft pseudo tag in the invention can play a role in enhancing the expression capability of the generated data tag.
Then reinitialize a new student modelAnd applies a two-stage training strategy. In the first stage training, training is performed using the generated data with soft pseudo tags:
this can be seen as a distillation process, in which a data set is generatedWith the help of (a), knowledge is transferred from teacher model eta to student modelOptimizing it as a student modelCalculate distillation loss:
Representing KL divergence. Then in the second stage training, the student is modeledIn a labeled training data setTraining is carried out on:
here, theIs a standard cross entropy loss function,is a student model obtained after the second stage training iteration. Then in the next iteration, the student is modeledEta re-labeling as teacher model。
The whole process is repeated for T times. According to standard self-learning settings, from within each iterationSampling 1/T generated data, and after T iterations, allThe generated data in the system is marked with a soft pseudo tag.
Through the two-stage self-learning mode, the semantic gap between the existing universal knowledge base and downstream task data can be well spanned, and meanwhile, the interference caused by data noise generation is effectively reduced.
It will be evident to those skilled in the art that the invention is not limited to the details of the foregoing illustrative embodiments, and that the present invention may be embodied in other specific forms without departing from the spirit or essential characteristics thereof. The present embodiments are therefore to be considered in all respects as illustrative and not restrictive, the scope of the invention being indicated by the appended claims rather than by the foregoing description, and all changes which come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein.
Furthermore, it should be understood that although the present disclosure describes embodiments, not every embodiment is provided with a single embodiment, and that this description is provided for clarity only, and that the disclosure is not limited to specific embodiments, and that the embodiments may be combined appropriately to form other embodiments that will be understood by those skilled in the art.
Claims (5)
1. A low-resource relation extraction method based on data synthesis and two-stage self-training comprises the following steps:
step one, data synthesis based on marked training data: converting the training data into a linear natural language sequence by adding a position symbol into the training data; constructing a data synthesis model based on a large-scale language model, and performing fine adjustment on the data synthesis model through training data; repeatedly performing the data synthesis process using polynomial sampling until an unlabeled generated dataset meeting the preset conditions is obtainedGenerating data with a place holder;in order to generate a sequence of words in the data,,respectively isIs a subject and an object of (a);to generate an amount of data;
step two, two-stage self-learning: in training data setTraining self-coding language model eta, and then generating data set with added position signUsing the self-coding language model eta classification to realize soft pseudo tag marking:;
order theSoft pseudo tag set of (a),Is a soft pseudo tag; training multiple self-coding language models by using K different random seeds, marking as teacher model eta, and marking a soft pseudo-label set marked by a kth teacher model asThe method comprises the steps of carrying out a first treatment on the surface of the Initializing a new self-coding language model, which is recorded as student modelFor student modelTwo-stage training strategies are applied: in the first stage training, distillation training is performed using the generated data with soft pseudo tags:the method comprises the steps of carrying out a first treatment on the surface of the Model studentsOptimized as a student modelCalculate distillation loss:
Represents KL divergence; in the second stage training, the student is modeledIn training data setTraining is carried out on:;as a standard cross-entropy loss function,is thatThe set of labels to be used in the corresponding set of labels,the student model is obtained after the second stage training iteration; the next time the two-stage training strategy is executed, the student is modeledAs a teacher model η; repeating the two-stage training strategy until a data set is generatedEach generated data is marked with a soft pseudo tag;
step three, relation extraction: constructing a relation extraction model based on a self-coding language model; the generated data and the training data are collectively called as training examples, and the relationship labels of the training examples are called as real labelsInputting the training examples into a relation extraction model, and calculating relation labels predicted by the relation extraction modelAnd training instance's true tagsTo train a relationship extraction model.
2. The method for extracting low-resource relationship based on data synthesis and two-stage self-training according to claim 1, wherein in the first step, fine tuning loss of the data synthesis model isThe method comprises the steps of carrying out a first treatment on the surface of the The fine tuning mode of the data synthesis model and the pre-training mode of the data synthesis model are the same, whereinRepresenting probability functions, LLM represents a large-scale language model,is a word sequence in training data added with position symbolThe word(s) in (a) is (are),for special start symbol, after finishing trimmingAnd generating a pre-addition mark prompt.
3. The method for extracting low-resource relation based on data synthesis and two-stage self-training according to claim 1, wherein in the second step, the two-stage training strategy is repeatedly executed T times; from generating data sets in each iterationSampling 1/T of generated data, and generating a data set after T iterationsAll the generated data in (a) are marked with soft pseudo tags.
4. The method for extracting low-resource relation based on data synthesis and two-stage self-training according to claim 1, wherein when a placer is added to the training data in the first step:
5. The low-resource relationship extraction method based on data synthesis and two-stage self-training according to claim 1, wherein: when the relation label predicted by the relation extraction model in the third step is obtained, vector representations h of corresponding positions of word sequences in the training examples are spliced to be classified, and the relation label predicted by the relation extraction model is obtained:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211630125.3A CN115618022B (en) | 2022-12-19 | 2022-12-19 | Low-resource relation extraction method based on data synthesis and two-stage self-training |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211630125.3A CN115618022B (en) | 2022-12-19 | 2022-12-19 | Low-resource relation extraction method based on data synthesis and two-stage self-training |
Publications (2)
Publication Number | Publication Date |
---|---|
CN115618022A CN115618022A (en) | 2023-01-17 |
CN115618022B true CN115618022B (en) | 2023-04-28 |
Family
ID=84879772
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211630125.3A Active CN115618022B (en) | 2022-12-19 | 2022-12-19 | Low-resource relation extraction method based on data synthesis and two-stage self-training |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115618022B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116415005B (en) * | 2023-06-12 | 2023-08-18 | 中南大学 | Relationship extraction method for academic network construction of scholars |
CN117174240B (en) * | 2023-10-26 | 2024-02-09 | 中国科学技术大学 | Medical image report generation method based on large model field migration |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113420548A (en) * | 2021-06-24 | 2021-09-21 | 杭州电子科技大学 | Entity extraction sampling method based on knowledge distillation and PU learning |
WO2021184311A1 (en) * | 2020-03-19 | 2021-09-23 | 中山大学 | Method and apparatus for automatically generating inference questions and answers |
CN114386371A (en) * | 2022-03-25 | 2022-04-22 | 中国科学技术大学 | Method, system, equipment and storage medium for correcting Chinese spelling error |
CN114528835A (en) * | 2022-02-17 | 2022-05-24 | 杭州量知数据科技有限公司 | Semi-supervised specialized term extraction method, medium and equipment based on interval discrimination |
CN114912456A (en) * | 2022-07-19 | 2022-08-16 | 北京惠每云科技有限公司 | Medical entity relationship identification method and device and storage medium |
CN115270797A (en) * | 2022-09-23 | 2022-11-01 | 山东省计算中心(国家超级计算济南中心) | Text entity extraction method and system based on self-training semi-supervised learning |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11783008B2 (en) * | 2020-11-06 | 2023-10-10 | Adobe Inc. | Machine-learning tool for generating segmentation and topic metadata for documents |
-
2022
- 2022-12-19 CN CN202211630125.3A patent/CN115618022B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021184311A1 (en) * | 2020-03-19 | 2021-09-23 | 中山大学 | Method and apparatus for automatically generating inference questions and answers |
CN113420548A (en) * | 2021-06-24 | 2021-09-21 | 杭州电子科技大学 | Entity extraction sampling method based on knowledge distillation and PU learning |
CN114528835A (en) * | 2022-02-17 | 2022-05-24 | 杭州量知数据科技有限公司 | Semi-supervised specialized term extraction method, medium and equipment based on interval discrimination |
CN114386371A (en) * | 2022-03-25 | 2022-04-22 | 中国科学技术大学 | Method, system, equipment and storage medium for correcting Chinese spelling error |
CN114912456A (en) * | 2022-07-19 | 2022-08-16 | 北京惠每云科技有限公司 | Medical entity relationship identification method and device and storage medium |
CN115270797A (en) * | 2022-09-23 | 2022-11-01 | 山东省计算中心(国家超级计算济南中心) | Text entity extraction method and system based on self-training semi-supervised learning |
Also Published As
Publication number | Publication date |
---|---|
CN115618022A (en) | 2023-01-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110597997B (en) | Military scenario text event extraction corpus iterative construction method and device | |
CN107273355B (en) | Chinese word vector generation method based on word and phrase joint training | |
CN115618022B (en) | Low-resource relation extraction method based on data synthesis and two-stage self-training | |
CN110232192A (en) | Electric power term names entity recognition method and device | |
CN110209832B (en) | Method, system and computer equipment for judging upper and lower relationship | |
CN114065738B (en) | Chinese spelling error correction method based on multitask learning | |
CN115858758A (en) | Intelligent customer service knowledge graph system with multiple unstructured data identification | |
US11869484B2 (en) | Apparatus and method for automatic generation and update of knowledge graph from multi-modal sources | |
CN111222318A (en) | Trigger word recognition method based on two-channel bidirectional LSTM-CRF network | |
CN114676255A (en) | Text processing method, device, equipment, storage medium and computer program product | |
CN110516240B (en) | Semantic similarity calculation model DSSM (direct sequence spread spectrum) technology based on Transformer | |
CN115270797A (en) | Text entity extraction method and system based on self-training semi-supervised learning | |
CN113326367B (en) | Task type dialogue method and system based on end-to-end text generation | |
CN114564563A (en) | End-to-end entity relationship joint extraction method and system based on relationship decomposition | |
CN118093834A (en) | AIGC large model-based language processing question-answering system and method | |
CN112905750B (en) | Method and equipment for generating optimization model | |
CN112015921B (en) | Natural language processing method based on learning auxiliary knowledge graph | |
CN112861827B (en) | Sign language translation method and system using single language material translation | |
CN116306653A (en) | Regularized domain knowledge-aided named entity recognition method | |
CN117035077A (en) | Difficulty controllable problem generation method based on soft template and inverse fact reasoning | |
CN106815211B (en) | Method for document theme modeling based on cyclic focusing mechanism | |
CN116484858A (en) | Text abstract generation method based on diffusion model | |
CN116595169A (en) | Question-answer intention classification method for coal mine production field based on prompt learning | |
CN115906854A (en) | Multi-level confrontation-based cross-language named entity recognition model training method | |
CN115730599A (en) | Chinese patent key information identification method based on structBERT, computer equipment, storage medium and program product |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |