CN117422065A - Natural language data processing system based on reinforcement learning algorithm - Google Patents
Natural language data processing system based on reinforcement learning algorithm Download PDFInfo
- Publication number
- CN117422065A CN117422065A CN202311323531.XA CN202311323531A CN117422065A CN 117422065 A CN117422065 A CN 117422065A CN 202311323531 A CN202311323531 A CN 202311323531A CN 117422065 A CN117422065 A CN 117422065A
- Authority
- CN
- China
- Prior art keywords
- feature
- classification
- news
- training
- vector
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
- 238000012545 processing Methods 0.000 title claims abstract description 73
- 230000002787 reinforcement Effects 0.000 title claims abstract description 62
- 238000004422 calculation algorithm Methods 0.000 title claims abstract description 61
- 238000012549 training Methods 0.000 claims abstract description 105
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 33
- 239000013598 vector Substances 0.000 claims description 294
- 238000006073 displacement reaction Methods 0.000 claims description 31
- 238000000605 extraction Methods 0.000 claims description 29
- 230000006870 function Effects 0.000 claims description 29
- 230000011218 segmentation Effects 0.000 claims description 23
- 238000005457 optimization Methods 0.000 claims description 22
- 238000010586 diagram Methods 0.000 claims description 17
- 239000011159 matrix material Substances 0.000 claims description 13
- 230000004913 activation Effects 0.000 claims description 10
- 238000003672 processing method Methods 0.000 claims description 10
- 230000004927 fusion Effects 0.000 claims description 9
- 238000004364 calculation method Methods 0.000 claims description 6
- 238000013528 artificial neural network Methods 0.000 claims description 4
- 238000011176 pooling Methods 0.000 claims description 3
- 238000005070 sampling Methods 0.000 claims description 3
- 230000009467 reduction Effects 0.000 claims description 2
- 238000012546 transfer Methods 0.000 claims description 2
- 238000000034 method Methods 0.000 abstract description 10
- 230000008569 process Effects 0.000 abstract description 8
- 230000008901 benefit Effects 0.000 description 5
- 238000004590 computer program Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000003058 natural language processing Methods 0.000 description 2
- 238000013139 quantization Methods 0.000 description 2
- 238000006467 substitution reaction Methods 0.000 description 2
- 239000003795 chemical substances by application Substances 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000000280 densification Methods 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000005065 mining Methods 0.000 description 1
- 238000003062 neural network model Methods 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/205—Parsing
- G06F40/216—Parsing using statistical methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/09—Supervised learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/092—Reinforcement learning
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Biophysics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Evolutionary Computation (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Probability & Statistics with Applications (AREA)
- Databases & Information Systems (AREA)
- Image Analysis (AREA)
Abstract
The application relates to the field of natural language data processing, and particularly discloses a natural language data processing system based on a reinforcement learning algorithm, which firstly collects a news data set to be classified, wherein the news data set to be classified comprises: and respectively extracting and analyzing the characteristics of the news text data and the news picture data through a trained convolutional neural network model. In this way, the model is trained and learned by using a reinforcement learning algorithm, and in the training process, the system selects classification actions according to the current state by interacting with the environment, and updates parameters of the strategy network according to the reward signals so as to optimize classification performance. In this way, a faster and more accurate classification of news data can be achieved.
Description
Technical Field
The present application relates to the field of natural language data processing, and more particularly, to a natural language data processing system based on reinforcement learning algorithms.
Background
Reinforcement learning is a learning paradigm in which one agent interacts with the environment to achieve a learning goal. Reinforcement learning has achieved remarkable achievements in recent years, and its application in the field of natural language processing is increasing. For example, they have applications in text classification, text generation, machine translation, and the like. Among them, the text classification problem has been a classical problem in the field of natural language processing, and the research of classification processing of texts such as news has been relatively thorough, but there are still a great deal of details and many additional difficulties encountered in commercialization implementation process to be considered. Because news is classified into many sub-scenes, the methods used in the different classification scenes may be completely different, and the news data includes text data, picture data, and the like. Whereas conventional news classification systems typically rely on marked training data, this means that a significant amount of manual labeling work is required, which is costly and time consuming.
Accordingly, a natural language data processing system based on reinforcement learning algorithms is desired to achieve faster and more accurate classification of news datasets.
Disclosure of Invention
The present application has been made in order to solve the above technical problems. Embodiments of the present application provide a natural language data processing system based on a reinforcement learning algorithm, which first collects a news data set to be classified, wherein the news data set to be classified includes: and respectively extracting and analyzing the characteristics of the news text data and the news picture data through a trained convolutional neural network model. In this way, the model is trained and learned by using a reinforcement learning algorithm, and in the training process, the system selects classification actions according to the current state by interacting with the environment, and updates parameters of the strategy network according to the reward signals so as to optimize classification performance. In this way, a faster and more accurate classification of news data can be achieved.
According to a first aspect of the present application, there is provided a natural language data processing system based on a reinforcement learning algorithm, comprising:
the data acquisition unit is used for acquiring a news data set to be classified, wherein the news data set to be classified comprises: news text data and news picture data;
The context coding unit is used for enabling the news text data to pass through a trained context coder comprising an embedded layer so as to obtain semantic feature vectors;
the multi-scale convolution coding unit is used for obtaining a multi-scale text understanding feature vector through the trained multi-scale neighborhood feature extraction module of the semantic feature vector;
the picture convolutional coding unit is used for enabling the news picture data to pass through a convolutional neural network which is completed through training and serves as a filter so as to obtain picture understanding feature vectors;
the feature fusion unit is used for fusing the multi-scale text understanding feature vector and the picture understanding feature vector to obtain a classification feature vector;
the feature optimization unit is used for carrying out displacement order based on interpolation vectors on the classification feature vectors so as to obtain optimized classification feature vectors;
and the result generation unit is used for enabling the optimized classification feature vector to pass through a multi-label classifier to obtain a classification result, wherein the classification result is used for representing classification labels corresponding to the news data set to be classified.
With reference to the first aspect of the present application, in a natural language data processing system based on a reinforcement learning algorithm of the first aspect of the present application, the context encoding unit includes: the segmentation processing subunit is used for carrying out segmentation processing on the news text data to obtain segment sequences corresponding to each news; the word segmentation processing subunit is used for carrying out word segmentation processing on the segment sequence to obtain a word sequence; an embedding vectorization subunit, configured to map each word in the word sequence into a word embedding vector by using an embedding layer of a sequence encoder of the Clip model to obtain a sequence of word embedding vectors; a context coding subunit, configured to perform global context semantic coding on the sequence of word embedding vectors using a Bert model based on a converter of a sequence encoder of the Clip model to obtain a plurality of feature vectors; and the cascading subunit is used for cascading the plurality of feature vectors to obtain the semantic feature vectors.
With reference to the first aspect of the present application, in a natural language data processing system based on a reinforcement learning algorithm of the first aspect of the present application, the multi-scale convolution encoding unit includes: the first scale convolution coding unit is used for respectively carrying out one-dimensional convolution coding on the semantic feature vectors by using a first convolution layer of the multi-scale neighborhood feature extraction module according to the following first convolution formula so as to obtain the first scale feature vectors; wherein the first convolution formula is:
wherein a is the width of the first convolution kernel in the X direction, F (a) is a first convolution kernel parameter vector, G (X-a) is a local vector matrix operated with a first convolution kernel function, w is the size of the first convolution kernel, X represents the semantic feature vector, and Cov (X) represents one-dimensional convolution encoding of the semantic feature vector respectively; the second scale convolution encoding unit is used for carrying out one-dimensional convolution encoding on the semantic feature vector by using a second convolution layer of the multi-scale neighborhood feature extraction module according to the following second convolution formula so as to obtain a second scale feature vector; wherein the second convolution formula is:
wherein b is the width of the second convolution kernel in the X direction, F (b) is a second convolution kernel parameter vector, G (X-b) is a local vector matrix operated with a second convolution kernel function, m is the size of the second convolution kernel, X represents the semantic feature vector, and Cov (X) represents one-dimensional convolution encoding of the semantic feature vector respectively; and a concatenation unit, configured to concatenate the first scale feature vector and the second scale feature vector to obtain the multi-scale text understanding feature vector.
In the natural language data processing system based on the reinforcement learning algorithm, the natural language data processing system based on the reinforcement learning algorithm further comprises a training module for training the context encoder including the embedded layer, the multi-scale neighborhood feature extraction module and the convolutional neural network as a filter;
wherein, training module includes:
the training data acquisition unit is used for acquiring training data, wherein the training data comprises the news text data, the news picture data and classification tag values corresponding to the news data set to be classified;
the training context coding unit is used for enabling the news text data to pass through a context coder comprising an embedded layer to obtain training semantic feature vectors;
the training multi-scale convolution coding unit is used for enabling the training semantic feature vector to pass through a multi-scale neighborhood feature extraction module to obtain a training multi-scale text understanding feature vector;
the training picture convolutional coding unit is used for enabling the news picture data to pass through a convolutional neural network serving as a filter so as to obtain training picture understanding feature vectors;
the training feature fusion unit is used for fusing the training multi-scale text understanding feature vector and the training picture understanding feature vector to obtain a training classification feature vector;
The training feature optimization unit is used for carrying out displacement order based on interpolation vectors on the training classification feature vectors so as to obtain training optimization classification feature vectors;
the classification loss unit is used for enabling the training optimization classification feature vector to pass through a classifier to obtain a classification loss function value;
and the training unit is used for training the context encoder containing the embedded layer, the multi-scale neighborhood feature extraction module and the convolution neural network serving as the filter by using the classification loss function value.
With reference to the first aspect of the present application, in a natural language data processing system based on a reinforcement learning algorithm of the first aspect of the present application, the classification loss unit is configured to: processing the training optimized classification feature vector by using the classifier according to the following calculation formula to obtain the classification result; wherein, the calculation formula is: o=softmax { (W) n ,B n ):...:(W 1 ,B 1 ) X, where W 1 To W n Is a weight matrix, B 1 To B n X is a training optimization classification feature vector for the bias vector; and calculating a cross entropy value between the classification result and a true value as the classification loss function value.
According to a second aspect of the present application, there is provided a natural language data processing method based on a reinforcement learning algorithm, including:
Collecting a news data set to be classified, wherein the news data set to be classified comprises: news text data and news picture data;
passing the news text data through a trained context encoder comprising an embedded layer to obtain semantic feature vectors;
the semantic feature vector passes through a trained multi-scale neighborhood feature extraction module to obtain a multi-scale text understanding feature vector;
the news picture data is passed through a convolutional neural network which is completed through training and serves as a filter so as to obtain picture understanding feature vectors;
fusing the multi-scale text understanding feature vector and the picture understanding feature vector to obtain a classification feature vector;
performing displacement order based on interpolation vectors on the classification feature vectors to obtain optimized classification feature vectors;
and the optimized classification feature vector passes through a multi-label classifier to obtain a classification result, wherein the classification result is used for representing classification labels corresponding to the news data set to be classified.
In the natural language data processing method based on the reinforcement learning algorithm, the natural language data processing method based on the reinforcement learning algorithm further comprises a training module for training the context encoder including the embedded layer, the multi-scale neighborhood feature extraction module and the convolutional neural network as a filter;
Wherein, training module includes:
acquiring training data, wherein the training data comprises the news text data, the news picture data and classification tag values corresponding to the news data set to be classified;
passing the news text data through a context encoder comprising an embedded layer to obtain training semantic feature vectors;
the training semantic feature vector passes through a multi-scale neighborhood feature extraction module to obtain a training multi-scale text understanding feature vector;
the news picture data is passed through a convolutional neural network serving as a filter to obtain training picture understanding feature vectors;
fusing the training multi-scale text understanding feature vector and the training picture understanding feature vector to obtain a training classification feature vector;
performing displacement order based on interpolation vectors on the training classification feature vectors to obtain training optimization classification feature vectors;
the training optimization classification feature vector passes through a classifier to obtain a classification loss function value;
training the context encoder including the embedded layer, the multi-scale neighborhood feature extraction module, and the convolutional neural network as a filter with the class loss function value.
Compared with the prior art, the natural language data processing system based on the reinforcement learning algorithm firstly collects a news data set to be classified, wherein the news data set to be classified comprises: and respectively extracting and analyzing the characteristics of the news text data and the news picture data through a trained convolutional neural network model. In this way, the model is trained and learned by using a reinforcement learning algorithm, and in the training process, the system selects classification actions according to the current state by interacting with the environment, and updates parameters of the strategy network according to the reward signals so as to optimize classification performance. In this way, a faster and more accurate classification of news data can be achieved.
Drawings
The foregoing and other objects, features and advantages of the present application will become more apparent from the following more particular description of embodiments of the present application, as illustrated in the accompanying drawings. The accompanying drawings are included to provide a further understanding of embodiments of the application and are incorporated in and constitute a part of this specification, illustrate the application and not constitute a limitation to the application. In the drawings, like reference numerals generally refer to like parts or steps.
FIG. 1 illustrates a schematic block diagram of a natural language data processing system based on a reinforcement learning algorithm in accordance with an embodiment of the present application.
FIG. 2 illustrates a schematic block diagram of a context encoding unit in a natural language data processing system based on a reinforcement learning algorithm according to an embodiment of the present application.
FIG. 3 illustrates a schematic block diagram of a multi-scale convolutional encoding unit in a natural language data processing system based on a reinforcement learning algorithm in accordance with an embodiment of the present application.
Fig. 4 illustrates a schematic block diagram of a picture convolutional encoding unit in a natural language data processing system based on a reinforcement learning algorithm according to an embodiment of the present application.
FIG. 5 illustrates a schematic block diagram of a feature optimization unit in a natural language data processing system based on a reinforcement learning algorithm in accordance with an embodiment of the present application.
FIG. 6 illustrates a schematic block diagram of training modules in a natural language data processing system based on a reinforcement learning algorithm in accordance with an embodiment of the present application.
FIG. 7 illustrates a flow chart of a natural language data processing method based on a reinforcement learning algorithm according to an embodiment of the present application.
Fig. 8 illustrates a schematic diagram of a system architecture of a natural language data processing method based on a reinforcement learning algorithm according to an embodiment of the present application.
Fig. 9 illustrates a block diagram of an electronic device according to an embodiment of the present application.
Detailed Description
Hereinafter, example embodiments according to the present application will be described in detail with reference to the accompanying drawings. It should be apparent that the described embodiments are only some of the embodiments of the present application and not all of the embodiments of the present application, and it should be understood that the present application is not limited by the example embodiments described herein.
Exemplary System
FIG. 1 illustrates a schematic block diagram of a natural language data processing system based on a reinforcement learning algorithm in accordance with an embodiment of the present application. As shown in fig. 1, the natural language data processing system 100 based on the reinforcement learning algorithm according to the embodiment of the present application includes: a data acquisition unit 110, configured to acquire a news data set to be classified, where the news data set to be classified includes: news text data and news picture data; a context encoding unit 120, configured to pass the news text data through a trained context encoder including an embedded layer to obtain a semantic feature vector; the multi-scale convolution encoding unit 130 is configured to pass the semantic feature vector through a trained multi-scale neighborhood feature extraction module to obtain a multi-scale text understanding feature vector; a picture convolutional encoding unit 140, configured to pass the news picture data through a convolutional neural network that is completed through training and serves as a filter to obtain a picture understanding feature vector; a feature fusion unit 150, configured to fuse the multi-scale text understanding feature vector and the picture understanding feature vector to obtain a classification feature vector; a feature optimization unit 160, configured to perform displacement order based on interpolation vectors on the classification feature vectors to obtain optimized classification feature vectors; the result generating unit 170 is configured to pass the optimized classification feature vector through a multi-tag classifier to obtain a classification result, where the classification result is used to represent a classification tag corresponding to the news data set to be classified.
In this embodiment of the present application, the data obtaining unit 110 is configured to collect a news data set to be categorized, where the news data set to be categorized includes: news text data and news picture data. It should be understood that news classification refers to a task of classifying input news data according to a category to which it belongs. Moreover, news data may have not only text data but also image data. Therefore, when classifying the news data set, it can be classified more accurately by news text data and news picture data. Specifically, in the technical scheme of the application, the text data and the news picture data to be news in the news data set to be classified are collected.
In this embodiment of the present application, the context encoding unit 120 is configured to pass the news text data through a trained context encoder including an embedded layer to obtain a semantic feature vector. It should be understood that, the news text data is usually long text, and it is difficult to efficiently and accurately extract the semantic features of the news text data, and the segment sequence corresponding to each news can be obtained by performing segmentation processing on the news text data, so that the subsequent mining of the semantic features of the news is facilitated, and the accuracy of semantic understanding of the news text data is further improved. In addition, considering that certain correlation exists between the segment sequences, in order to accurately acquire the global semantic understanding feature distribution information of the segment sequences in the high-dimensional feature space, word segmentation processing is further carried out on the segment sequences to obtain word sequences. Further, in order to prevent the word sequence confusion phenomenon after the word segmentation processing, the word sequence can be encoded by a trained context encoder containing an embedded layer to obtain global semantic distribution characteristics in the news text data. Specifically, in the technical scheme of the application, the semantic feature vector is passed through a trained multi-scale neighborhood feature extraction module to obtain a multi-scale text understanding feature vector.
FIG. 2 illustrates a schematic block diagram of a context encoding unit in a natural language data processing system based on a reinforcement learning algorithm according to an embodiment of the present application. As shown in fig. 2, the context encoding unit 120 includes: a segmentation processing subunit 121, configured to perform segmentation processing on the news text data to obtain a segment sequence corresponding to each news; a word segmentation processing subunit 122, configured to perform word segmentation processing on the segment sequence to obtain a word sequence; an embedding vectorization subunit 123, configured to map each word in the word sequence into a word embedding vector by using an embedding layer of a sequence encoder of the Clip model to obtain a sequence of word embedding vectors; a context encoding subunit 124, configured to perform global-based context semantic encoding on the sequence of word embedded vectors using a Bert model based on a converter of the sequence encoder of the Clip model to obtain a plurality of feature vectors; a concatenation subunit 125, configured to concatenate the plurality of feature vectors to obtain the semantic feature vector.
In this embodiment of the present application, the multi-scale convolutional encoding unit 130 is configured to pass the semantic feature vector through a trained multi-scale neighborhood feature extraction module to obtain a multi-scale text understanding feature vector. Considering that not only a certain association exists between adjacent words or sentences in the news text data, a certain association may also exist between non-adjacent words and sentences and words and sentences with different spans. That is, when the high-dimensional semantic distribution features of the news text data are extracted, not only the global semantic distribution features of the news text data are extracted, but also the multi-scale neighborhood high-dimensional semantic understanding associated features of the news text data are extracted according to different scales. Specifically, in the technical scheme of the application, the semantic feature vector is passed through a trained multi-scale neighborhood feature extraction module to obtain a multi-scale text understanding feature vector. The multi-scale neighborhood feature extraction module can use two one-dimensional convolution cores with different scales to carry out convolution coding on the semantic feature vectors so as to obtain two feature vectors with different scales, and then the two feature vectors with different scales are cascaded to obtain multi-scale neighborhood high-dimensional semantic understanding associated features of news text data.
FIG. 3 illustrates a schematic block diagram of a multi-scale convolutional encoding unit in a natural language data processing system based on a reinforcement learning algorithm in accordance with an embodiment of the present application. As shown in fig. 3, the multi-scale convolutional encoding unit 130 includes: a first scale convolution encoding unit 131, configured to perform one-dimensional convolution encoding on the semantic feature vectors by using a first convolution layer of the multi-scale neighborhood feature extraction module according to the following first convolution formula, so as to obtain the first scale feature vectors;
wherein the first convolution formula is:
wherein a is the width of the first convolution kernel in the X direction, F (a) is a first convolution kernel parameter vector, G (X-a) is a local vector matrix operated with a first convolution kernel function, w is the size of the first convolution kernel, X represents the semantic feature vector, and Cov (X) represents one-dimensional convolution encoding of the semantic feature vector respectively; a second scale convolution encoding unit 132, configured to perform one-dimensional convolution encoding on the semantic feature vector by using a second convolution layer of the multi-scale neighborhood feature extraction module according to the following second convolution formula to obtain the second scale feature vector;
wherein the second convolution formula is:
Wherein b is the width of the second convolution kernel in the X direction, F (b) is a second convolution kernel parameter vector, G (X-b) is a local vector matrix operated with a second convolution kernel function, m is the size of the second convolution kernel, X represents the semantic feature vector, and Cov (X) represents one-dimensional convolution encoding of the semantic feature vector respectively; and a concatenation unit 133, configured to concatenate the first scale feature vector and the second scale feature vector to obtain the multi-scale text understanding feature vector.
In this embodiment of the present application, the picture convolutional encoding unit 140 is configured to pass the news picture data through a trained convolutional neural network serving as a filter to obtain a picture understanding feature vector. It should be understood that the news picture data in the news data set is image data, which can be encoded by a convolutional neural network having excellent performance in the image field. In addition, considering that each position in the feature map contains local feature information of the position, when the high-dimensional semantic understanding distribution feature of the news picture data is extracted, we pay more attention to the feature representation of the whole picture, and the feature map dimension-reducing feature vector obtained after the news picture data is encoded by the convolutional neural network can compress the local feature information into the whole feature representation. Therefore, the distribution characteristics of the news picture data based on global high-dimensional semantic understanding can be obtained, and the accuracy of the global semantic understanding of the news picture data is improved. Specifically, in the technical scheme of the application, the news picture data is passed through a convolutional neural network which is completed through training and serves as a filter so as to obtain a picture understanding feature vector.
Fig. 4 illustrates a schematic block diagram of a picture convolutional encoding unit in a natural language data processing system based on a reinforcement learning algorithm according to an embodiment of the present application. As shown in fig. 4, the picture convolutional encoding unit 140 includes: an encoding subunit 141, configured to convolutionally encode the news picture data using the convolutional neural network as a filter to obtain a picture understanding feature map; the dimension reduction subunit 142 is configured to globally average and pool each feature matrix along the channel dimension of the picture understanding feature map to obtain the picture understanding feature vector. Wherein the encoding subunit 141 is configured to: input data are respectively carried out in forward transfer of layers by using the convolutional neural network model: performing three-dimensional convolution processing on the input data based on the convolution neural network serving as a filter to obtain a convolution feature map; carrying out mean pooling treatment based on a local feature matrix on the convolution feature map to obtain a pooled feature map; non-linear activation is carried out on the pooled feature map so as to obtain an activated feature map; the output of the last layer of the convolutional neural network serving as the filter is the picture understanding feature map, and the input of the first layer of the convolutional neural network serving as the filter is the news picture data.
In this embodiment of the present application, the feature fusion unit 150 is configured to fuse the multi-scale text understanding feature vector and the picture understanding feature vector to obtain a classification feature vector. It should be appreciated that text and pictures are two different information expressions that can complement and enrich each other, and fusing the multi-scale text understanding feature vector and the picture understanding feature vector can provide a more comprehensive and diversified news semantic understanding feature information representation, which helps to better understand and describe the high-dimensional semantic understanding distribution features of the news data to be classified. Specifically, in the technical scheme of the application, the multi-scale text understanding feature vector and the picture understanding feature vector are fused to obtain a classification feature vector.
In a specific embodiment of the present application, the feature fusion unit 150 includes: fusing the multi-scale text understanding feature vector and the picture understanding feature vector using a cascading function to obtain the classification feature vector, wherein the cascading function is formulated as:
f(X i ,X j )=Relu( W f[θ(X i ),φ(X j )])
wherein W is f ,θ(X i ) And phi (X) j ) All representing the point convolution of the input, relu as the activation function, [ ]Representing the splicing operation, X i Feature values, X, representing respective positions in the multi-scale text understanding feature vector j And representing the characteristic value of each position in the picture understanding characteristic vector.
In this embodiment, the feature optimization unit 160 is configured to perform displacement order based on interpolation vectors on the classification feature vectors to obtain optimized classification feature vectors. Considering that feature redundancy exists in the classification feature vector and the feature values of all positions in the classification feature vector have differences in spatial position significance, namely the feature values of all positions in the classification feature vector have different contribution degrees to final classification judgment, the classification result obtained by the classification feature vector through the classifier is poor in accuracy and difficult to guarantee in robustness.
In view of the above technical problems, in the technical solution of the present application, vector granularity segmentation is first performed on the classification feature vector, that is, a feature value set of the classification feature vector is subjected to subgroup processing to obtain a plurality of classification local features. In a specific example of the present application, vector segmentation is performed on the classification feature vector to obtain a plurality of classification local feature vectors. Considering that after vector segmentation, each classified local feature vector loses information due to sparsification of feature dimensions, after vector segmentation, the classified local feature vectors are respectively passed through an upsampling module based on linear difference values to obtain a plurality of dense classified local feature vectors, that is, the characteristic densification of the classified local features is performed through linear interpolation. Further, a displacement offset value of each of the densely classified local feature vectors in a feature vector set formed by the plurality of densely classified local feature vectors is calculated, the displacement offset value representing a quantization outlier of each of the densely classified local feature vectors based on a feature distribution pattern. It will be appreciated that the higher the quantization outliers, the poorer the contribution of each densely classified local feature vector to the final classification decision. Then, the displacement characteristic values of the dense classification local characteristic vectors are arranged into displacement order input vectors, and then the displacement order characteristic vectors are obtained through a displacement order characteristic extractor comprising a one-dimensional convolution layer and a Sigmoid activation layer, namely, high-dimensional implicit mode characteristics among quantized outliers are captured through one-dimensional convolution coding and Sigmoid coding, and class probability domain mapping is carried out through a Sigmoid activation function. And finally, taking the characteristic value of each position in the displacement order characteristic vector as a weight, respectively weighting each classified local characteristic vector to obtain a plurality of weighted classified local characteristic vectors, and carrying out vector splicing on the plurality of weighted classified local characteristic vectors to obtain the optimized classified characteristic vector.
In this way, the displacement order based on the interpolation vector is utilized to perform feature selection, specifically, local features with distinguishing ability in the classification feature vector are enhanced according to the difference of the quantized outliers, and local features with weak distinguishing ability in the classification feature vector are restrained.
FIG. 5 illustrates a schematic block diagram of a feature optimization unit in a natural language data processing system based on a reinforcement learning algorithm in accordance with an embodiment of the present application. As shown in fig. 5, the feature optimization unit 160 includes: a vector segmentation subunit 161, configured to perform vector segmentation on the classification feature vector to obtain a plurality of classification local feature vectors; a sampling subunit 162, configured to pass the plurality of classified local feature vectors through an upsampling module based on a linear difference value to obtain a plurality of densely classified local feature vectors, respectively; a euclidean distance calculating subunit 163 configured to calculate, for each of the plurality of dense classification local feature vectors, euclidean distances between the each dense classification local feature vector and all other dense classification local feature vectors to obtain a plurality of euclidean distance values and calculate a sum of the plurality of euclidean distance values as a displacement feature value of the each dense classification local feature vector; a feature activation subunit 164, configured to arrange the displacement feature values of the densely classified local feature vectors into displacement order input vectors, and obtain displacement order feature vectors through a displacement order feature extractor including a one-dimensional convolution layer and a Sigmoid activation layer; a vector weighting subunit 165, configured to respectively weight the classified local feature vectors with respect to the feature values of the positions in the displacement order feature vector as weights, so as to obtain a plurality of weighted classified local feature vectors; the vector stitching subunit 166 is configured to perform vector stitching on the plurality of weighted classified local feature vectors to obtain the optimized classified feature vector.
In this embodiment of the present application, the result generating unit 170 is configured to pass the optimized classification feature vector through a multi-tag classifier to obtain a classification result, where the classification result is used to represent a classification tag corresponding to the news data set to be classified.
It should be appreciated that training of the context encoder including the embedded layer, the multi-scale neighborhood feature extraction module, and the convolutional neural network as a filter is required prior to news classification using the above-described data-based neural network model. That is, in the natural language data processing system based on the reinforcement learning algorithm according to the embodiment of the present application, the training module is further configured to train the context encoder including the embedded layer, the multi-scale neighborhood feature extraction module, and the convolutional neural network as the filter.
FIG. 6 illustrates a schematic block diagram of training modules in a natural language data processing system based on a reinforcement learning algorithm in accordance with an embodiment of the present application. As shown in fig. 6, the training module 200 includes: a training data obtaining unit 210, configured to obtain training data, where the training data includes the news text data, the news picture data, and a classification tag value corresponding to the news data set to be classified; a training context encoding unit 220, configured to pass the news text data through a context encoder including an embedded layer to obtain training semantic feature vectors; a training multi-scale convolution encoding unit 230, configured to pass the training semantic feature vector through a multi-scale neighborhood feature extraction module to obtain a training multi-scale text understanding feature vector; a training picture convolutional encoding unit 240 for passing the news picture data through a convolutional neural network as a filter to obtain training picture understanding feature vectors; the training feature fusion unit 250 is configured to fuse the training multi-scale text understanding feature vector and the training picture understanding feature vector to obtain a training classification feature vector; a training feature optimization unit 260, configured to perform displacement order based on interpolation vectors on the training classification feature vectors to obtain training optimization classification feature vectors; a classification loss unit 270, configured to pass the training optimized classification feature vector through a classifier to obtain a classification loss function value; a training unit 280, configured to train the context encoder including the embedded layer, the multi-scale neighborhood feature extraction module, and the convolutional neural network as a filter with the classification loss function value.
In this embodiment, the classification loss unit 270 is configured to pass the training optimized classification feature vector through a classifier to obtain a classification loss function value.
In a specific embodiment of the present application, the classification loss unit 270 is configured to: processing the training optimized classification feature vector by using the classifier according to the following calculation formula to obtain the classification result; wherein, the calculation formula is: o=softmax { (W) n ,B n ):...:(W 1 ,B 1 ) X, where W 1 To W n Is a weight matrix, B 1 To B n X is a training optimization classification feature vector for the bias vector; and calculating a cross entropy value between the classification result and a true value as the classification loss function value.
In summary, the natural language data processing system 100 based on the reinforcement learning algorithm according to the embodiment of the present application is illustrated, which first collects a news data set to be classified, wherein the news data set to be classified includes: and respectively extracting and analyzing the characteristics of the news text data and the news picture data through a trained convolutional neural network model. In this way, the model is trained and learned by using a reinforcement learning algorithm, and in the training process, the system selects classification actions according to the current state by interacting with the environment, and updates parameters of the strategy network according to the reward signals so as to optimize classification performance. In this way, a faster and more accurate classification of news data can be achieved.
As described above, the natural language data processing system 100 based on the reinforcement learning algorithm according to the embodiment of the present application may be implemented in various terminal devices, for example, a server or the like in which the natural language data processing algorithm based on the reinforcement learning algorithm is deployed. In one example, the natural language data processing system 100 according to the reinforcement learning algorithm may be integrated into the terminal device as a software module and/or hardware module. For example, the reinforcement learning algorithm-based natural language data processing system 100 may be a software module in the operating system of the terminal device, or may be an application developed for the terminal device; of course, the reinforcement learning algorithm-based natural language data processing system 100 can also be one of a plurality of hardware modules of the terminal device.
Alternatively, in another example, the reinforcement learning algorithm-based natural language data processing system 100 and the terminal device may be separate devices, and the reinforcement learning algorithm-based natural language data processing system 100 may be connected to the terminal device through a wired and/or wireless network and transmit interactive information in a contracted data format.
Exemplary method
FIG. 7 illustrates a flow chart of a natural language data processing method based on a reinforcement learning algorithm according to an embodiment of the present application. Fig. 8 illustrates a schematic diagram of a system architecture of a natural language data processing method based on a reinforcement learning algorithm according to an embodiment of the present application. As shown in fig. 7 and 8, a natural language data processing method based on a reinforcement learning algorithm according to an embodiment of the present application includes: s110, collecting a news data set to be classified, wherein the news data set to be classified comprises: news text data and news picture data; s120, passing the news text data through a trained context encoder comprising an embedded layer to obtain semantic feature vectors; s130, passing the semantic feature vector through a trained multi-scale neighborhood feature extraction module to obtain a multi-scale text understanding feature vector; s140, the news picture data is passed through a convolutional neural network which is completed through training and serves as a filter so as to obtain picture understanding feature vectors; s150, fusing the multi-scale text understanding feature vector and the picture understanding feature vector to obtain a classification feature vector; s160, a feature optimization unit, which is used for carrying out displacement order based on interpolation vectors on the classification feature vectors so as to obtain optimized classification feature vectors; and S170, the optimized classification feature vector passes through a multi-label classifier to obtain a classification result, wherein the classification result is used for representing classification labels corresponding to the news data set to be classified.
Here, it will be understood by those skilled in the art that the specific functions and operations of the respective steps in the above-described reinforcement learning algorithm-based natural language data processing method have been described in detail in the above description of the reinforcement learning algorithm-based natural language data processing system with reference to fig. 1, and thus, repetitive descriptions thereof will be omitted.
Exemplary electronic device
Next, an electronic device according to an embodiment of the present application is described with reference to fig. 9.
Fig. 9 illustrates a block diagram of an electronic device according to an embodiment of the present application.
As shown in fig. 9, the electronic device 10 includes one or more processors 11 and a memory 12.
The processor 11 may be a Central Processing Unit (CPU) or other form of processing unit having data processing and/or instruction execution capabilities, and may control other components in the electronic device 10 to perform desired functions.
Memory 12 may include one or more computer program products that may include various forms of computer-readable storage media, such as volatile memory and/or non-volatile memory. The volatile memory may include, for example, random Access Memory (RAM) and/or cache memory (cache), and the like. The non-volatile memory may include, for example, read Only Memory (ROM), hard disk, flash memory, and the like. One or more computer program instructions may be stored on the computer readable storage medium that can be executed by the processor 11 to implement the reinforcement learning algorithm based natural language data processing system and/or other desired functions of the various embodiments of the present application described above.
In one example, the electronic device 10 may further include: an input device 13 and an output device 14, which are interconnected by a bus system and/or other forms of connection mechanisms (not shown).
The input means 13 may comprise, for example, a keyboard, a mouse, etc.
The output device 14 can output various information including a decoded value and the like to the outside. The output means 14 may include, for example, a display, speakers, a printer, and a communication network and remote output devices connected thereto, etc.
The basic principles of the present application have been described above in connection with specific embodiments, however, it should be noted that the advantages, benefits, effects, etc. mentioned in the present application are merely examples and not limiting, and these advantages, benefits, effects, etc. are not to be considered as necessarily possessed by the various embodiments of the present application. Furthermore, the specific details disclosed herein are for purposes of illustration and understanding only, and are not intended to be limiting, as the application is not intended to be limited to the details disclosed herein as such.
It should be understood that the specific examples herein are intended only to facilitate a better understanding of the embodiments of the present application by those skilled in the art and are not intended to limit the scope of the embodiments of the present application. In addition, in various embodiments of the present application, the sequence number of each process does not mean the sequence of execution, and the execution sequence of each process should be determined by its functions and internal logic, and should not constitute any limitation on the implementation process of the embodiments of the present application.
The foregoing is merely specific embodiments of the present application, but the scope of the present application is not limited thereto, and any person skilled in the art can easily think about changes or substitutions within the technical scope of the present application, and the changes and substitutions are intended to be covered by the scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.
Claims (10)
1. A natural language data processing system based on a reinforcement learning algorithm, comprising:
the data acquisition unit is used for acquiring a news data set to be classified, wherein the news data set to be classified comprises: news text data and news picture data;
the context coding unit is used for enabling the news text data to pass through a trained context coder comprising an embedded layer so as to obtain semantic feature vectors;
the multi-scale convolution coding unit is used for obtaining a multi-scale text understanding feature vector through the trained multi-scale neighborhood feature extraction module of the semantic feature vector;
the picture convolutional coding unit is used for enabling the news picture data to pass through a convolutional neural network which is completed through training and serves as a filter so as to obtain picture understanding feature vectors;
The feature fusion unit is used for fusing the multi-scale text understanding feature vector and the picture understanding feature vector to obtain a classification feature vector;
the feature optimization unit is used for carrying out displacement order based on interpolation vectors on the classification feature vectors so as to obtain optimized classification feature vectors;
and the result generation unit is used for enabling the optimized classification feature vector to pass through a multi-label classifier to obtain a classification result, wherein the classification result is used for representing classification labels corresponding to the news data set to be classified.
2. The reinforcement learning algorithm-based natural language data processing system of claim 1, wherein the context encoding unit comprises:
the segmentation processing subunit is used for carrying out segmentation processing on the news text data to obtain segment sequences corresponding to each news;
the word segmentation processing subunit is used for carrying out word segmentation processing on the segment sequence to obtain a word sequence;
an embedding vectorization subunit, configured to map each word in the word sequence into a word embedding vector by using an embedding layer of a sequence encoder of the Clip model to obtain a sequence of word embedding vectors;
a context coding subunit, configured to perform global context semantic coding on the sequence of word embedding vectors using a Bert model based on a converter of a sequence encoder of the Clip model to obtain a plurality of feature vectors;
And the cascading subunit is used for cascading the plurality of feature vectors to obtain the semantic feature vectors.
3. The reinforcement learning algorithm-based natural language data processing system of claim 2, wherein the multi-scale convolution encoding unit comprises:
the first scale convolution coding unit is used for respectively carrying out one-dimensional convolution coding on the semantic feature vectors by using a first convolution layer of the multi-scale neighborhood feature extraction module according to the following first convolution formula so as to obtain the first scale feature vectors;
wherein the first convolution formula is:
wherein a is the width of the first convolution kernel in the X direction, F (a) is a first convolution kernel parameter vector, G (X-a) is a local vector matrix operated with a first convolution kernel function, w is the size of the first convolution kernel, X represents the semantic feature vector, and Cov (X) represents one-dimensional convolution encoding of the semantic feature vector respectively;
the second scale convolution encoding unit is used for carrying out one-dimensional convolution encoding on the semantic feature vector by using a second convolution layer of the multi-scale neighborhood feature extraction module according to the following second convolution formula so as to obtain a second scale feature vector;
Wherein the second convolution formula is:
wherein b is the width of the second convolution kernel in the X direction, F (b) is a second convolution kernel parameter vector, G (X-b) is a local vector matrix operated with a second convolution kernel function, m is the size of the second convolution kernel, X represents the semantic feature vector, and Cov (X) represents one-dimensional convolution encoding of the semantic feature vector respectively;
and the cascading unit is used for cascading the first scale feature vector and the second scale feature vector to obtain the multi-scale text understanding feature vector.
4. A natural language data processing system based on a reinforcement learning algorithm according to claim 3, wherein the picture convolution encoding unit comprises:
the coding subunit is used for performing convolutional coding on the news picture data by using the convolutional neural network serving as a filter so as to obtain a picture understanding characteristic diagram;
and the dimension reduction subunit is used for carrying out global mean value pooling on each feature matrix of the picture understanding feature image along the channel dimension so as to obtain the picture understanding feature vector.
5. The reinforcement learning algorithm-based natural language data processing system of claim 4, wherein the encoding subunit is configured to: input data are respectively carried out in forward transfer of layers by using the convolutional neural network model:
Performing three-dimensional convolution processing on the input data based on the convolution neural network serving as a filter to obtain a convolution feature map;
carrying out mean pooling treatment based on a local feature matrix on the convolution feature map to obtain a pooled feature map; and
non-linear activation is carried out on the pooled feature map so as to obtain an activated feature map;
the output of the last layer of the convolutional neural network serving as the filter is the picture understanding feature map, and the input of the first layer of the convolutional neural network serving as the filter is the news picture data.
6. The reinforcement learning algorithm-based natural language data processing system of claim 5, wherein the feature fusion unit comprises:
fusing the multi-scale text understanding feature vector and the picture understanding feature vector using a cascading function to obtain the classification feature vector, wherein the cascading function is formulated as:
f(X i ,X j )=Relu(W f [θ(X i ),φ(X j )])
wherein W is f ,θ(X i ) And phi (X) j ) All representing the point convolution of the input, relu as the activation function, []Representing the splicing operation, X i Feature values, X, representing respective positions in the multi-scale text understanding feature vector j And representing the characteristic value of each position in the picture understanding characteristic vector.
7. The reinforcement learning algorithm-based natural language data processing system of claim 6, wherein the feature optimization unit comprises:
the vector segmentation subunit is used for carrying out vector segmentation on the classified feature vectors to obtain a plurality of classified local feature vectors;
the sampling subunit is used for respectively passing the plurality of classified local feature vectors through an up-sampling module based on linear difference values to obtain a plurality of dense classified local feature vectors;
a euclidean distance calculating subunit configured to calculate, for each of the plurality of dense classification local feature vectors, euclidean distances between the each dense classification local feature vector and all other dense classification local feature vectors to obtain a plurality of euclidean distance values and calculate a sum of the plurality of euclidean distance values as a displacement feature value of the each dense classification local feature vector;
a feature activation subunit, configured to arrange the displacement feature values of the dense classification local feature vectors into displacement order input vectors, and obtain displacement order feature vectors through a displacement order feature extractor that includes a one-dimensional convolution layer and a Sigmoid activation layer;
The vector weighting subunit is used for taking the characteristic value of each position in the displacement order characteristic vector as a weight and respectively weighting each classified local characteristic vector to obtain a plurality of weighted classified local characteristic vectors;
and the vector splicing subunit is used for carrying out vector splicing on the plurality of weighted classified local feature vectors to obtain the optimized classified feature vector.
8. The reinforcement learning algorithm-based natural language data processing system of claim 7, further comprising a training module for training the context encoder including an embedded layer, the multi-scale neighborhood feature extraction module, and the convolutional neural network as a filter;
wherein, training module includes:
the training data acquisition unit is used for acquiring training data, wherein the training data comprises the news text data, the news picture data and classification tag values corresponding to the news data set to be classified;
the training context coding unit is used for enabling the news text data to pass through a context coder comprising an embedded layer to obtain training semantic feature vectors;
The training multi-scale convolution coding unit is used for enabling the training semantic feature vector to pass through a multi-scale neighborhood feature extraction module to obtain a training multi-scale text understanding feature vector;
the training picture convolutional coding unit is used for enabling the news picture data to pass through a convolutional neural network serving as a filter so as to obtain training picture understanding feature vectors;
the training feature fusion unit is used for fusing the training multi-scale text understanding feature vector and the training picture understanding feature vector to obtain a training classification feature vector;
the training feature optimization unit is used for carrying out displacement order based on interpolation vectors on the training classification feature vectors so as to obtain training optimization classification feature vectors;
the classification loss unit is used for enabling the training optimization classification feature vector to pass through a classifier to obtain a classification loss function value;
and the training unit is used for training the context encoder containing the embedded layer, the multi-scale neighborhood feature extraction module and the convolution neural network serving as the filter by using the classification loss function value.
9. The reinforcement learning algorithm-based natural language data processing system of claim 8, wherein the classification loss unit is configured to: processing the training optimized classification feature vector by using the classifier according to the following calculation formula to obtain the classification result;
Wherein, the calculation formula is: o=softmax { (W) n ,B n ):…:(W 1 ,B 1 ) X, where W 1 To W n Is a weight matrix, B 1 To B n X is a training optimization classification feature vector for the bias vector; and
and calculating a cross entropy value between the classification result and a true value as the classification loss function value.
10. A natural language data processing method based on reinforcement learning algorithm is characterized by comprising the following steps:
collecting a news data set to be classified, wherein the news data set to be classified comprises: news text data and news picture data;
passing the news text data through a trained context encoder comprising an embedded layer to obtain semantic feature vectors;
the semantic feature vector passes through a trained multi-scale neighborhood feature extraction module to obtain a multi-scale text understanding feature vector;
the news picture data is passed through a convolutional neural network which is completed through training and serves as a filter so as to obtain picture understanding feature vectors;
fusing the multi-scale text understanding feature vector and the picture understanding feature vector to obtain a classification feature vector;
performing displacement order based on interpolation vectors on the classification feature vectors to obtain optimized classification feature vectors;
And the optimized classification feature vector passes through a multi-label classifier to obtain a classification result, wherein the classification result is used for representing classification labels corresponding to the news data set to be classified.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311323531.XA CN117422065A (en) | 2023-10-13 | 2023-10-13 | Natural language data processing system based on reinforcement learning algorithm |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311323531.XA CN117422065A (en) | 2023-10-13 | 2023-10-13 | Natural language data processing system based on reinforcement learning algorithm |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117422065A true CN117422065A (en) | 2024-01-19 |
Family
ID=89522117
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311323531.XA Withdrawn CN117422065A (en) | 2023-10-13 | 2023-10-13 | Natural language data processing system based on reinforcement learning algorithm |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117422065A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117952650A (en) * | 2024-01-30 | 2024-04-30 | 和源顺(湖州)工艺品有限公司 | Handicraft e-commerce sales management system based on big data |
-
2023
- 2023-10-13 CN CN202311323531.XA patent/CN117422065A/en not_active Withdrawn
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117952650A (en) * | 2024-01-30 | 2024-04-30 | 和源顺(湖州)工艺品有限公司 | Handicraft e-commerce sales management system based on big data |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN115203380B (en) | Text processing system and method based on multi-mode data fusion | |
CN108615036B (en) | Natural scene text recognition method based on convolution attention network | |
CN108959246B (en) | Answer selection method and device based on improved attention mechanism and electronic equipment | |
CN109062901B (en) | Neural network training method and device and name entity recognition method and device | |
CN112733866A (en) | Network construction method for improving text description correctness of controllable image | |
CN111160348A (en) | Text recognition method for natural scene, storage device and computer equipment | |
CN113449528B (en) | Address element extraction method and device, computer equipment and storage medium | |
CN114254071B (en) | Querying semantic data from unstructured documents | |
CN115951883B (en) | Service component management system of distributed micro-service architecture and method thereof | |
CN116245513A (en) | Automatic operation and maintenance system and method based on rule base | |
CN117893859A (en) | Multi-mode text image classification method and device, electronic equipment and storage medium | |
CN115311598A (en) | Video description generation system based on relation perception | |
CN111145914A (en) | Method and device for determining lung cancer clinical disease library text entity | |
CN117422065A (en) | Natural language data processing system based on reinforcement learning algorithm | |
CN117194605A (en) | Hash encoding method, terminal and medium for multi-mode medical data deletion | |
Belharbi et al. | Deep neural networks regularization for structured output prediction | |
CN115565177A (en) | Character recognition model training method, character recognition device, character recognition equipment and medium | |
CN114692624A (en) | Information extraction method and device based on multitask migration and electronic equipment | |
CN113704466B (en) | Text multi-label classification method and device based on iterative network and electronic equipment | |
CN117079292A (en) | Method and device for extracting identity card information, computer equipment and storage medium | |
CN112487231B (en) | Automatic image labeling method based on double-image regularization constraint and dictionary learning | |
CN115270792A (en) | Medical entity identification method and device | |
CN116089605A (en) | Text emotion analysis method based on transfer learning and improved word bag model | |
CN115731600A (en) | Small target detection method based on semi-supervision and feature fusion | |
CN114140858A (en) | Image processing method, device, equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
WW01 | Invention patent application withdrawn after publication |
Application publication date: 20240119 |
|
WW01 | Invention patent application withdrawn after publication |