CN113516198B - Cultural resource text classification method based on memory network and graphic neural network - Google Patents
Cultural resource text classification method based on memory network and graphic neural network Download PDFInfo
- Publication number
- CN113516198B CN113516198B CN202110864647.9A CN202110864647A CN113516198B CN 113516198 B CN113516198 B CN 113516198B CN 202110864647 A CN202110864647 A CN 202110864647A CN 113516198 B CN113516198 B CN 113516198B
- Authority
- CN
- China
- Prior art keywords
- text
- graph
- word
- words
- model
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/205—Parsing
- G06F40/211—Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/284—Lexical analysis, e.g. tokenisation or collocates
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Evolutionary Computation (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Evolutionary Biology (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Machine Translation (AREA)
Abstract
A cultural resource text classification method based on a memory network and a graph neural network mainly comprises three modules, a pre-trained two-way long-short term memory network (BiLSTM) module, a text graph construction module and a graph neural network (GCN) module, wherein firstly, a text data set is pre-trained by utilizing a two-way long-short term memory network model and text and word characteristics containing time sequence information are obtained, secondly, a global text graph consisting of the text and the words is constructed according to the co-occurrence relation among the words and the importance of the words in the text, the node characteristics of the text graph are initialized by adopting the characteristics extracted by the pre-trained two-way long-short term memory network module, then, further characterization learning is carried out on the node characteristics of the global text graph through a two-layer graph convolution neural network, and a final text classification result is obtained, so that the method can be used for improving the classification accuracy of cultural resource texts.
Description
Technical Field
The invention belongs to the technical field of natural language processing, and particularly relates to a cultural resource text classification method based on a two-way long-short-term memory network and a graphic neural network (BiLSTM-GCN).
Background
Cultural resources are natural or social resources which can meet the requirements of human culture and provide basis for cultural industry, and can be divided into various types. The culture resources are effectively classified, and the storage, the mining and the reuse of the culture resources can be promoted.
Text classification is a technology for automatically classifying and labeling texts according to certain classification rules and classification standards, and is also a basic task of natural language processing, and is used in many applications such as emotion analysis, data mining, news filtering and the like. Early text classification, which relies on manual labeling and some set rules to classify text data, was extremely inefficient, which was not possible to process hundreds of millions of levels of text data at any time in the current big data age. With the progress of the age and the development of computer science and technology, text classification realizes automatic classification by training and learning text data through a computer. The manual classification workload is greatly reduced, the working efficiency is improved, and the classification effect is even better than that of manual classification.
The text classification method mainly comprises a traditional machine learning method and a deep learning method, the traditional machine learning-based text classification method is used for realizing text classification by manually extracting features and utilizing a shallow classifier, the text classification task is very mature, however, as the effect requirements of people on text classification are higher and higher, the current text classification method mainly uses the deep learning-based method, most of deep learning models usually represent unstructured text data into data which can be understood by a computer, and then training is carried out through a large number of marked data sets, so that important features in the text are extracted, and a final classification result is obtained.
The graph neural network is a deep learning network based on graph data, and compared with the traditional network, the graph neural network can better perform characterization learning on the graph structure data, and is widely applied to tasks such as social networks, recommendation systems, molecular activity prediction and the like. The text in the natural language processing task also contains rich graph structure information, including word co-occurrence, syntactic semantics and text context, and the graph neural network can fully utilize the graph structure information of the text data. Current text classification methods with respect to graphic neural networks rarely take into account time series problems in the text, which greatly limit the effectiveness of text classification.
Disclosure of Invention
In order to overcome the problem that the text classification method based on the graph convolution neural network does not fully consider the time sequence problem contained in the text, the invention aims to provide the cultural resource text classification method based on the memory network and the graph neural network, and the method can introduce the time sequence information in the text in advance so as to achieve a better text classification effect.
In order to achieve the above purpose, the technical scheme adopted by the invention is as follows:
a text classification method based on BiLSTM-GCN comprises a BiLSTM pre-training module, a text diagram construction module and a GCN module, and is characterized by comprising the following steps:
1) BiLSTM pre-training module: pre-training the text by using a bidirectional LSTM and acquiring the text and word characteristics after pre-training;
LSTM is a deep learning network based on modeling of time step sequence, its structural features determine that hidden layer at each moment is related to previous input, compared with RNN, LSTM can avoid wellAvoiding the problem of gradient vanishing due to long-term dependence, assuming a text consisting of n words { w ] 1 ,w 2 ,...,w n Each word is represented by a d-dimensional vector, and is initialized by using a glove word vector trained by Stanford, wherein the initialized word vector is { v } 1 ,v 2 ,...,v n In order to get better word vectors for the model, the word vectors are updated continuously with iteration during training. Each text is modeled using bi-directional LSTM, one LSTM is semantic representation of the text from beginning to end, and the other LSTM is semantic representation of the text from end to end, the input at the current time being related not only to the previous state but also to the following state.
Each word v t Both forward and backward hidden statesAnd->
h t The two hidden states are concatenated together, which fuses the word context information. The specific formula is as follows:
through the previous operation, text and word characteristics containing time series information can be extracted, then the full connection layer is needed to classify samples, and the model is used for hiding the first state h 1 And last hidden state h n As the characteristics of the text, the characteristics are input into a full connection layer, and then an activation function is connected to realize text classification, wherein the specific formula is as follows:
y=softmax(MLP(h 1 ,h 2 ));
2) The text diagram construction module: constructing a large global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words;
constructing a global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words, the edge weights between words are expressed based on co-occurrence relations between words, PMI values or cosine similarity between words are adopted, the edge weights between the texts and the words are expressed based on importance of the words in the text, and the edge weights are obtained by adopting TF-IDF values or a keyword extraction algorithm;
3) GCN module: the method comprises the steps of realizing characterization learning of text graph nodes through a two-layer graph convolution neural network, and further realizing final text classification;
the word characteristics v and the text characteristics h which contain time sequence information can be obtained through the step 1), the characteristics are initialization characteristics of the global text graph nodes, and the characteristics are put into the GCN, so that the characteristic information of the text graph can be further extracted. Assuming that the constructed undirected text graph is g, the node number is n, and the word node number is n word Wen Benjie the number of points is n text A represents the adjacency matrix of the text graph g, I N Is an identity matrix, D is a degree matrix of the text graph g, W 0 ,W 1 All are weight matrixes, and X is a node characteristic matrix of the text graph g.
The GCN model adopts a graph convolution neural network of semi-supervised learning, the whole model can train part of node data with a small number of labels, then the rest of nodes without labels are classified, and the specific implementation formula is as follows:
the loss function is the cross entropy loss of all text, defined as follows:
Y D is a text with labels, i.e. training set and validation set in the model, F is the dimension of the final output feature of the model, equal to the number of classifications, y pred Is a label for model prediction.
The beneficial effects of the invention are as follows:
the text and word characteristics are initialized through the pre-trained BiLSTM module, time sequence information contained in the text can be introduced in advance before graph convolution is carried out, the defect that the global text graph cannot contain the time sequence information is overcome, and then the text graph is characterized and learned through a two-layer graph convolution neural network, so that the effect of text classification can be greatly improved.
Drawings
FIG. 1 is a schematic diagram of a method of the present invention.
Detailed Description
The present invention will be further described with reference to the accompanying drawings and examples, but the present invention is not limited to the following examples.
As shown in FIG. 1, the text classification method based on BiLSTM-GCN comprises a BiLSTM pre-training module, a text diagram construction module and a GCN module, and is characterized by comprising the following steps:
1) BiLSTM pre-training module: pre-training the text by using a bidirectional LSTM and acquiring the text and word characteristics after pre-training;
LSTM is a deep learning network based on modeling of time step sequence, its structural features determine that hidden layer at each moment is related to previous input, compared with RNN, LSTM can well avoid gradient vanishing problem caused by long-time dependence, and a text is assumed to be composed of n words { w 1 ,w 2 ,...,w n Each word is represented by a d-dimensional vector, and is initialized by using a glove word vector trained by Stanford, wherein the initialized word vector is { v } 1 ,v 2 ,...,v n In order to get better word vectors for the model, the word vectors are updated continuously with iteration during training. Each text is modeled using bi-directional LSTM, one LSTM is semantic representation of the text from beginning to end, and the other LSTM is semantic representation of the text from end to end, the input at the current time being related not only to the previous state but also to the following state.
Each word v t Both forward and backward hidden statesAnd->
h t The two hidden states are concatenated together, which fuses the word context information. The specific formula is as follows:
through the previous operation, text and word characteristics containing time series information can be extracted, then the full connection layer is needed to classify samples, and the model is used for hiding the first state h 1 And last hidden state h n As the characteristics of the text, the characteristics are input into a full connection layer, and then an activation function is connected to realize text classification, wherein the specific formula is as follows:
y=softmax(MLP(h 1 ,h 2 ));
2) The text diagram construction module: constructing a large global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words;
constructing a global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words, the edge weights between words are expressed based on co-occurrence relations between words, PMI values or cosine similarity between words are adopted, the edge weights between the texts and the words are expressed based on importance of the words in the text, and the edge weights are obtained by adopting TF-IDF values or a keyword extraction algorithm;
3) GCN module: the method comprises the steps of realizing characterization learning of text graph nodes through a two-layer graph convolution neural network, and further realizing final text classification;
word features v and text features h containing time sequence information can be obtained through the step 1), the features are initialization features of the global text graph nodes, and the features are put into GCN and can be furtherFeature information of the text graph is extracted. Assuming that the constructed undirected text graph is g, the node number is n, and the word node number is n word Wen Benjie the number of points is n text A represents the adjacency matrix of the text graph g, I N Is an identity matrix, D is a degree matrix of the text graph g, W 0 ,W 1 All are weight matrixes, and X is a node characteristic matrix of the text graph g.
The GCN model adopts a graph convolution neural network of semi-supervised learning, the whole model can train part of node data with a small number of labels, then the rest of nodes without labels are classified, and the specific implementation formula is as follows:
the loss function is the cross entropy loss of all text, defined as follows:
Y D is a text with labels, i.e. training set and validation set in the model, F is the dimension of the final output feature of the model, equal to the number of classifications, y pred Is a label for model prediction.
Experimental comparative analysis process
The experiment of the invention is based on Python3.7 programming language and Pytorch frame, the word vector uses 100-dimensional, 200-dimensional and 300-dimensional Glove word vector of Stanford pre-training, and the initial learning rate of the whole model is 10 -3 L2 weight decay is 10 -4 Dropou after full connection layert is set to 0.5 and the batch size is set to 64. If the validation set does not drop in loss after consecutive 5 generations of training, training is stopped.
This patent was tested on the R8 and R52 datasets to verify the validity of the model and its performance.
R8, R52 are each news classification data sets, wherein each sample contains a piece of news text and a corresponding hot topic label. Wherein, R8 contains 8 news categories, the training set and the test set respectively have 5485 samples and 2189 samples, R52 contains 52 news categories, and the training set and the test set respectively have 6532 samples and 2568 samples. In the R8, R52 data sets, 10% of the training set data were randomly selected as the validation set. The comparative experiments are mainly divided into algorithms based on a traditional deep learning algorithm comprising CNN, LSTM, biLSTM and Fastext and algorithms based on a graph neural network comprising Text-GCN.
The accuracy and the F1 value can be used as evaluation indexes of the performance of the classification model, as shown in table 1, the text classification model based on the GCN is stronger than the traditional deep learning text classification model on several classification data sets, and the text classification model based on the GCN can learn rich graph information in the text due to the fact that graph convolution is adopted for learning text characteristics, so that better classification effect can be achieved. According to the BiLSTM-GCN model, time sequence information of a text is introduced in advance on the basis of GCN, characteristic information of the text is enriched, experiments show that the accuracy of the BiLSTM-GCN model on an R8 data set is improved by 0.51%, and the accuracy of the BiLSTM-GCN model on an R52 data set is improved by 0.7%, and the fact that a better classification effect is achieved by directly using one-hot characteristics after words brought by the BiLSTM model are introduced and the text characteristic information is achieved.
TABLE 1 BiLSTM-GCN vs. different model accuracy
TABLE 2 comparison of BiLSTM-GCN with different model F1 values
The accuracy is only used, so that the effect of the model is not sufficiently evaluated, the recall rate and the accuracy are considered by the F1 value, and the effect of the model provided by the invention can be more sufficiently judged. As can be seen from table 2, the model proposed by the present invention has an improvement of 1.32% over the R8 dataset and 1.84% over the R52 dataset over the baseline GCN model F1, which further indicates that the model proposed by us is more effective in text classification tasks.
Claims (1)
1. A cultural resource text classification method based on a memory network and a graphic neural network is characterized by comprising the following steps:
1) BiLSTM pre-training module: pre-training the text by using a bidirectional LSTM and acquiring the text and word characteristics after pre-training;
assume that a text consists of n words { w } 1 ,w 2 ,…,w n Each word is represented by a d-dimensional vector, and is initialized by using a glove word vector trained by Stanford, wherein the initialized word vector is { v } 1 ,v 2 ,…v n In order to make the model obtain better word vectors, the word vectors are updated continuously along with iteration in the training process; modeling each text by adopting a bidirectional LSTM, wherein one LSTM is used for carrying out semantic representation of the text from the beginning of the sentence to the end of the sentence, and the other LSTM is used for carrying out semantic representation of the text from the end of the sentence to the beginning of the sentence, and the input at the current moment is not only related to the previous state but also related to the following state;
each word v t Both forward and backward hidden statesAnd->h t The two hidden states are spliced together, and the two hidden states fuse the context information of the word, and the specific formula is as follows:
through the previous operation, text and word characteristics containing time series information can be extracted, then the full connection layer is needed to classify samples, and the model is used for hiding the first state h 1 And last hidden state h n As the characteristics of the text, the characteristics are input into a full connection layer, and then an activation function is connected to realize text classification, wherein the specific formula is as follows:
y=soft max(MLP(h 1 ,h n ));
2) The text diagram construction module: constructing a large global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words;
constructing a global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words, the edge weights between words are expressed based on co-occurrence relations between words, PMI values or cosine similarity between words are adopted, the edge weights between the texts and the words are expressed based on importance of the words in the text, and the edge weights are obtained by adopting TF-IDF values or a keyword extraction algorithm;
3) GCN module: the method comprises the steps of realizing characterization learning of text graph nodes through a two-layer graph convolution neural network, and further realizing final text classification;
the word characteristics v and text characteristics h containing time sequence information can be obtained through the step 1), the characteristics are initialization characteristics of the global text graph nodes, the characteristics are put into GCN, the characteristic information of the text graph can be further extracted, the constructed undirected text graph is assumed to be g, the node number is N, and the word node number is N word Wen Benjie the number of points is N text A represents the adjacency matrix of the text graph g, I N Is an identity matrix, W 0 ,W 1 The node feature matrices are weight matrices, and X is a node feature matrix of the text graph g;
the GCN model adopts a graph convolution neural network of semi-supervised learning, the whole model can train part of node data with a small number of labels, then classify the rest of nodes without labels, and the specific implementation formula is as follows:
the loss function is the cross entropy loss of all text, defined as follows:
Y D is a text with labels, i.e. training set and validation set in the model, F is the dimension of the final output feature of the model, equal to the number of classifications, y pred Is a label for model prediction.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110864647.9A CN113516198B (en) | 2021-07-29 | 2021-07-29 | Cultural resource text classification method based on memory network and graphic neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110864647.9A CN113516198B (en) | 2021-07-29 | 2021-07-29 | Cultural resource text classification method based on memory network and graphic neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113516198A CN113516198A (en) | 2021-10-19 |
CN113516198B true CN113516198B (en) | 2023-09-22 |
Family
ID=78068765
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110864647.9A Active CN113516198B (en) | 2021-07-29 | 2021-07-29 | Cultural resource text classification method based on memory network and graphic neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113516198B (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114048305B (en) * | 2021-11-05 | 2024-06-07 | 山东大学 | Class case recommendation method of administrative punishment document based on graph convolution neural network |
CN114154022A (en) * | 2021-11-12 | 2022-03-08 | 东北大学 | Scheme-source cable classification processing method based on hierarchical graph convolution neural network model |
CN114398479A (en) * | 2021-11-13 | 2022-04-26 | 重庆邮电大学 | Text classification method, device and medium based on time sequence interaction graph neural network |
CN117475199A (en) * | 2023-10-16 | 2024-01-30 | 深圳市泰洲科技有限公司 | Intelligent classification method for customs declaration commodity |
CN117455421B (en) * | 2023-12-25 | 2024-04-16 | 杭州青塔科技有限公司 | Subject classification method and device for scientific research projects, computer equipment and storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106021461A (en) * | 2016-05-17 | 2016-10-12 | 深圳市中润四方信息技术有限公司 | Text classification method and text classification system |
CN109597891A (en) * | 2018-11-26 | 2019-04-09 | 重庆邮电大学 | Text emotion analysis method based on two-way length Memory Neural Networks in short-term |
US10339423B1 (en) * | 2017-06-13 | 2019-07-02 | Symantec Corporation | Systems and methods for generating training documents used by classification algorithms |
WO2021008037A1 (en) * | 2019-07-15 | 2021-01-21 | 平安科技(深圳)有限公司 | A-bilstm neural network-based text classification method, storage medium, and computer device |
CN112966503A (en) * | 2021-03-22 | 2021-06-15 | 山东建筑大学 | Aspect level emotion analysis method |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11481418B2 (en) * | 2020-01-02 | 2022-10-25 | International Business Machines Corporation | Natural question generation via reinforcement learning based graph-to-sequence model |
-
2021
- 2021-07-29 CN CN202110864647.9A patent/CN113516198B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106021461A (en) * | 2016-05-17 | 2016-10-12 | 深圳市中润四方信息技术有限公司 | Text classification method and text classification system |
US10339423B1 (en) * | 2017-06-13 | 2019-07-02 | Symantec Corporation | Systems and methods for generating training documents used by classification algorithms |
CN109597891A (en) * | 2018-11-26 | 2019-04-09 | 重庆邮电大学 | Text emotion analysis method based on two-way length Memory Neural Networks in short-term |
WO2021008037A1 (en) * | 2019-07-15 | 2021-01-21 | 平安科技(深圳)有限公司 | A-bilstm neural network-based text classification method, storage medium, and computer device |
CN112966503A (en) * | 2021-03-22 | 2021-06-15 | 山东建筑大学 | Aspect level emotion analysis method |
Non-Patent Citations (3)
Title |
---|
Graph Convolutional Networks for Text Classification;Liang Yao 等;《arXiv》;20181113;第1-9页 * |
基于BERT的文本情感分析;刘思琴等;《信息安全研究》;20200305(第03期);第30-37页 * |
基于预训练模型的文本分类网络TextCGA;杨玮祺等;《现代计算机》;20200425(第12期);第53-38页 * |
Also Published As
Publication number | Publication date |
---|---|
CN113516198A (en) | 2021-10-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113516198B (en) | Cultural resource text classification method based on memory network and graphic neural network | |
CN110866117B (en) | Short text classification method based on semantic enhancement and multi-level label embedding | |
CN110413986B (en) | Text clustering multi-document automatic summarization method and system for improving word vector model | |
CN111274394B (en) | Method, device and equipment for extracting entity relationship and storage medium | |
CN112699247B (en) | Knowledge representation learning method based on multi-class cross entropy contrast complement coding | |
CN110321563B (en) | Text emotion analysis method based on hybrid supervision model | |
CN111274790B (en) | Chapter-level event embedding method and device based on syntactic dependency graph | |
CN111078833B (en) | Text classification method based on neural network | |
CN114896388A (en) | Hierarchical multi-label text classification method based on mixed attention | |
CN105975478A (en) | Word vector analysis-based online article belonging event detection method and device | |
CN112925904B (en) | Lightweight text classification method based on Tucker decomposition | |
CN107688576B (en) | Construction and tendency classification method of CNN-SVM model | |
CN113806547B (en) | Deep learning multi-label text classification method based on graph model | |
CN112749274A (en) | Chinese text classification method based on attention mechanism and interference word deletion | |
CN114461804B (en) | Text classification method, classifier and system based on key information and dynamic routing | |
CN111506732A (en) | Text multi-level label classification method | |
CN113239143B (en) | Power transmission and transformation equipment fault processing method and system fusing power grid fault case base | |
CN113254655B (en) | Text classification method, electronic device and computer storage medium | |
CN111985243A (en) | Emotion model training method, emotion analysis device and storage medium | |
CN113343690A (en) | Text readability automatic evaluation method and device | |
CN113051887A (en) | Method, system and device for extracting announcement information elements | |
CN116049387A (en) | Short text classification method, device and medium based on graph convolution | |
CN114925205B (en) | GCN-GRU text classification method based on contrast learning | |
CN111709225A (en) | Event cause and effect relationship judging method and device and computer readable storage medium | |
CN115659947A (en) | Multi-item selection answering method and system based on machine reading understanding and text summarization |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |