CN109087670B - Emotion analysis method, system, server and storage medium - Google Patents
Emotion analysis method, system, server and storage medium Download PDFInfo
- Publication number
- CN109087670B CN109087670B CN201811005214.2A CN201811005214A CN109087670B CN 109087670 B CN109087670 B CN 109087670B CN 201811005214 A CN201811005214 A CN 201811005214A CN 109087670 B CN109087670 B CN 109087670B
- Authority
- CN
- China
- Prior art keywords
- target
- emotion
- data
- feature points
- intonation
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000008451 emotion Effects 0.000 title claims abstract description 202
- 238000004458 analytical method Methods 0.000 title claims abstract description 37
- 238000000034 method Methods 0.000 claims abstract description 27
- 230000006399 behavior Effects 0.000 claims description 20
- 238000012545 processing Methods 0.000 claims description 11
- 238000004364 calculation method Methods 0.000 claims description 7
- 238000007405 data analysis Methods 0.000 claims description 7
- 238000004590 computer program Methods 0.000 claims description 5
- 238000000605 extraction Methods 0.000 claims description 2
- 238000010586 diagram Methods 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 6
- 238000012216 screening Methods 0.000 description 5
- 230000002996 emotional effect Effects 0.000 description 4
- 230000006870 function Effects 0.000 description 3
- 230000014509 gene expression Effects 0.000 description 3
- 230000008921 facial expression Effects 0.000 description 2
- 239000013307 optical fiber Substances 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 208000027534 Emotional disease Diseases 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000035479 physiological effects, processes and functions Effects 0.000 description 1
- 230000008707 rearrangement Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000005477 standard model Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/63—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/02—Feature extraction for speech recognition; Selection of recognition unit
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Child & Adolescent Psychology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Hospice & Palliative Care (AREA)
- Psychiatry (AREA)
- Signal Processing (AREA)
- Information Transfer Between Computers (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The embodiment of the invention discloses an emotion analysis method, an emotion analysis system, a server and a storage medium, wherein the method comprises the following steps: identifying language keywords and target tones included in the acquired voice data; analyzing and determining voice characteristic points according to the acquired language keywords and the target intonation; generating a target emotion model based on the voice feature points, and calibrating the target voice feature points on the target emotion model; matching the target emotion model with standard emotion models in a standard emotion model library to adjust the calibrated target voice feature points on the target emotion model and recording the change data of the target voice feature points; and matching the change data with the tone characteristic data and the psychological behavior characteristic data in the database, and outputting the emotion or emotion change data of the user according to the matching result. Therefore, the purposes of objectively and accurately analyzing the emotion change of the user and helping the user to manage the emotion can be achieved.
Description
Technical Field
The invention relates to the technical field of data analysis, in particular to an emotion analysis method, system, server and storage medium.
Background
Human emotional expression is of exceptional importance to individuals, as changes in mind and physiology can occur over time if not released at the right time. Today, with the rapid development of economy, people are increasingly inattentive to the care of special groups. Therefore, there is an urgent need for a method for managing emotions by helping a special group having emotion problems objectively analyze emotions through an intelligent method.
At present, the most common method is to establish a harmonious man-machine environment by endowing a computer system with the emotional ability to recognize, understand, express and adapt to people through an emotion computer technology so as to help users objectively analyze emotions and manage emotions. Common methods for analyzing the emotion of the user include analyzing the emotion of the user through a voice recognition technology and a face recognition technology. The user emotion is analyzed only by recognizing sensitive words included in the voice using the voice recognition technology, and the accuracy is low. When the face recognition technology is used, the emotion of a user is analyzed by collecting the expression characteristics of the face and analyzing the expression characteristics. However, the facial expressions usually last for a relatively short time and even disappear once, which is inconvenient to collect, and sometimes the facial expressions express opposite emotions, so that sometimes the analysis of the emotions of the user through face recognition is not accurate. Therefore, it is difficult to achieve the purpose of helping the user manage the emotion by objectively and accurately analyzing the emotion change of the user.
Disclosure of Invention
The embodiment of the invention provides an emotion analysis method, an emotion analysis system, a server and a storage medium, which aim to objectively and accurately analyze emotion changes of a user and help the user to manage emotion.
In a first aspect, an embodiment of the present invention provides an emotion analysis method, including:
identifying language keywords and target tones included in the acquired voice data;
analyzing and determining voice feature points according to the acquired language keywords and the target intonation, wherein the voice feature points are keywords and intonations which mark user emotion in the language keywords and the target intonation;
generating a target emotion model based on the voice feature points, and calibrating the target voice feature points on the target emotion model;
matching the target emotion model with standard emotion models in a standard emotion model library to adjust the calibrated target voice feature points on the target emotion model and recording the change data of the target voice feature points;
and matching the change data with the tone characteristic data and the psychological behavior characteristic data in the database, and outputting the emotion or emotion change data of the user according to the matching result.
In a second aspect, an embodiment of the present invention provides an emotion analysis system, including:
the recognition module is used for recognizing the language keywords and the target intonation included in the acquired voice data;
the feature point analysis module is used for analyzing and determining voice feature points according to the acquired language key words and the target intonation, wherein the voice feature points are the key words and the intonations which mark the emotion of the user in the language key words and the target intonation;
the modeling module is used for generating a target emotion model based on the analyzed and determined voice feature points and calibrating the target voice feature points on the target emotion model;
the adjustment recording module is used for matching the target emotion model with a standard emotion model in a standard emotion model library so as to adjust the calibrated target voice feature points on the target emotion model and record the change data of the target voice feature points;
and the emotion analysis module is used for matching the change data with the tone characteristic data and the psychological behavior characteristic data in the database and outputting the emotion or emotion change data of the user according to the matching result.
In a third aspect, an embodiment of the present invention further provides a server, including:
one or more processors;
a memory for storing one or more programs,
the one or more programs, when executed by the one or more processors, cause the one or more processors to implement a method of emotion analysis as in any of the embodiments of the invention.
In a fourth aspect, the embodiments of the present invention further provide a computer-readable storage medium, on which a computer program is stored, which when executed by a processor, implements the emotion method according to any of the embodiments of the present invention.
According to the emotion analysis method, the system, the server and the storage medium provided by the embodiment of the invention, the target emotion model is established according to the extracted voice feature points, the target voice feature points are calibrated on the target emotion model so as to achieve the purpose of screening out the voice feature points capable of accurately showing the emotion of the user, then the target emotion model is matched with the standard emotion model in the standard emotion model library, the calibrated target voice feature points on the target emotion model are adjusted to enable the target emotion model to be closer to the standard emotion model, and then the target emotion model is matched with the tone feature data and the psychological behavior feature data in the database according to the change data of the target voice feature points, so that the emotion change of the user is quickly and accurately analyzed.
Drawings
Fig. 1 is a schematic flow chart of a method for emotion analysis according to an embodiment of the present invention;
fig. 2 is a schematic flow chart of a method for analyzing emotion according to a second embodiment of the present invention;
fig. 3 is a schematic structural diagram of an emotion analyzing system provided in the third embodiment of the present invention;
fig. 4 is a structural diagram of a server according to a fourth embodiment of the present invention.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the invention and are not limiting of the invention. It should be further noted that, for the convenience of description, only some of the structures related to the present invention are shown in the drawings, not all of the structures.
Example one
Fig. 1 is a flowchart of an emotion analyzing method provided in an embodiment of the present invention, which is applicable to objectively analyzing and managing emotions for a specific group having an emotional problem, and which may be executed by an emotion analyzing system, which may be configured in a server, for example. The method specifically comprises the following steps:
and S110, identifying the language key words and the target intonation included in the acquired voice data.
The voice data of the user can be collected through a preset rule, illustratively, the voice data of the user within a period of time or the voice data of the user within different periods of time in one day are collected according to a preset time rule, and specifically, the voice data can be collected through corresponding voice collection software.
For the collected voice data, the semantic content and intonation included in the voice data can be recognized through a voice recognition technology, wherein the intonation includes volume, speed, tone and the like of the voice data. Illustratively, semantic content, i.e., textual content, in the speech data may be recognized through acoustic model and language model analysis.
And extracting language keywords and target intonations from the identified semantic content and intonation according to the voice recognition result, wherein the target intonation comprises at least one of volume, speed, tone and respective change trend of voice data. Illustratively, a participle thesaurus can be used to remove meaningless words in semantic content, and simultaneously, language keywords which can indicate the emotion of a user are extracted; for the identified intonation, the target intonation meeting the preset condition is selected, illustratively, the target intonation is selected from the intonation whose volume exceeds the maximum preset threshold and is lower than the minimum preset threshold, or the target intonation is selected from the intonation whose speed exceeds a certain preset threshold.
And S120, analyzing and determining voice feature points according to the acquired language keywords and the target intonation, wherein the voice feature points are the keywords and the intonation which mark the emotion of the user in the language keywords and the target intonation.
And further analyzing and screening the acquired language keywords and target tones, and determining the keywords and the tones which can obviously indicate the emotion of the user as voice feature points, wherein the voice feature points comprise the keyword feature points and the tone feature points. Illustratively, the language keywords can be screened through an emotion sensitive word bank established in advance, and the screened keywords are determined as keyword feature points, wherein the emotion sensitive word bank comprises words frequently spoken under various different emotions of the user. Since the target intonation is usually displayed in the form of a waveform map, a point with a relatively obvious variation trend can be used as an intonation feature point, for example, a point with a suddenly increased speech speed.
And S130, generating a target emotion model based on the voice feature points, and calibrating the target voice feature points on the target emotion model.
And generating a target emotion model according to the determined voice feature points so as to analyze the emotion of the user according to the target emotion model. And calibrating target voice feature points on the target emotion model, wherein the target voice feature points can be a part with more prominent features in the voice feature points determined in the step S120, so that further screening of the emotional features of the user is realized, and the emotional features of the user are more obvious.
And S140, matching the target emotion model with a standard emotion model in a standard emotion model library to adjust the calibrated target voice feature points on the target emotion model and record the change data of the target voice feature points.
Matching the target emotion model with standard emotion models in a standard emotion model library, namely matching the voice feature points calibrated on the target emotion model with the feature points on the standard emotion model, including matching of keywords and tones, finely adjusting the calibrated target voice feature points on the target emotion model according to the matching result and the data trend of each target voice feature point, and recording the change data of the target voice feature points. Illustratively, the target voice feature point is adjusted from a position A to a position B, and the change of each feature of the target voice feature point between the positions A and B, such as the change of the speed of speech, the change of the tone, the change of the volume and the like, is recorded.
S150, matching the change data with the tone characteristic data and the psychological behavior characteristic data in the database, and outputting the emotion or emotion change data of the user according to the matching result.
And outputting emotion or emotion change data of the user according to the matching result of the change data of the target voice feature point and the tone feature data and the psychological behavior feature data in the database. Illustratively, if the speech rate frequency is high, the volume is large, and the tone change is large, the emotion of the user is considered to be urgent; if the speech rate frequency is slow, the volume is moderate, and the tone is gentle, the emotion of the user is considered to be stable.
In this embodiment, a target emotion model is established according to the extracted voice feature points, the target voice feature points are calibrated on the target emotion model to achieve the purpose of screening out the voice feature points capable of accurately showing the emotion of the user, then the target emotion model is matched with emotion models in a standard emotion model library, the calibrated target voice feature points on the target emotion model are adjusted to enable the target emotion model to be closer to the standard model, and then the target emotion model is matched with tone feature data and psychological behavior feature data in the database according to change data of the target voice feature points, so that the emotion change of the user is quickly and accurately analyzed.
Example two
Fig. 2 is a schematic flow chart of an emotion analysis method provided in the second embodiment of the present invention, where the present embodiment performs optimization based on the above embodiment, and the method includes:
s210, identifying the language keywords and the target intonation included in the acquired voice data.
S220, analyzing and determining voice feature points according to the obtained language keywords and the target intonation, wherein the voice feature points are the keywords and the intonation which mark the emotion of the user in the language keywords and the target intonation.
S230, carrying out convolution operation on the voice feature points to obtain a convolution operation result, generating a target emotion model based on the convolution operation result, and calibrating extreme points in the convolution operation result as target voice feature points.
In this embodiment, the voice feature points obtained in S220 are subjected to convolution operation to obtain a convolution operation result, and a target emotion model is generated according to the data trend of each feature point in the convolution operation result, where the generated target emotion model may be a waveform diagram. Further, in order to screen feature points capable of obviously indicating the emotion of the user from the obtained voice feature points, the voice feature points corresponding to extreme points, such as peak and trough positions, in the model map are marked as target voice feature points.
S240, matching the target emotion model with a standard emotion model in a standard emotion model library to adjust the calibrated target voice feature points on the target emotion model and recording the change data of the target voice feature points.
And S250, matching the change data with the tone characteristic data and the psychological behavior characteristic data in the database, and outputting the emotion or emotion change data of the user according to the matching result.
Further, in this embodiment, in order to improve the accuracy of emotion analysis of the user, the intonation feature data and the psychological behavior feature data in the database are updated through big data analysis calculation. Illustratively, the internet data is periodically collected, and the collected internet data is analyzed and calculated, and the intonation characteristic data and the psychological behavior characteristic data in the database are continuously updated, wherein the database is a local database or an internet database.
In this embodiment, a target emotion model is established by performing convolution operation on the acquired voice feature points, and the target voice feature points are calibrated according to extreme values in the calculation results, so that the purpose of screening the feature points which obviously indicate the emotion of the user is achieved, the accuracy of emotion analysis of the user is improved, and the tone feature data and the psychological behavior feature data in the database are periodically updated, so that the emotion analysis is more accurate, and the emotion management of the user is facilitated.
EXAMPLE III
Fig. 3 is a schematic structural diagram of an emotion analyzing system provided in the third embodiment of the present invention, and as shown in fig. 3, the system includes:
the recognition module 310 is configured to recognize a language keyword and a target intonation included in the acquired voice data;
the feature point analysis module 320 is configured to analyze and determine a voice feature point according to the obtained language keyword and the target intonation, where the voice feature point is a keyword and an intonation that identify a user emotion in the language keyword and the target intonation;
the modeling module 330 is configured to generate a target emotion model based on the analyzed and determined voice feature points, and calibrate the target voice feature points on the target emotion model;
the adjustment recording module 340 is configured to match the target emotion model with a standard emotion model in a standard emotion model library, so as to adjust a calibrated target voice feature point on the target emotion model and record change data of the target voice feature point;
and the emotion analysis module 350 is configured to match the change data with the intonation feature data and the psychological behavior feature data in the database, and output emotion or emotion change data of the user according to a matching result.
In this embodiment, the feature point analysis module determines a voice feature point according to a recognition result of the recognition module, the modeling module establishes a target emotion model according to the voice feature point, the adjustment recording module performs fine adjustment on the target emotion model and records a change of the target voice feature point, and the emotion analysis module analyzes emotion and emotion change of a user according to change data of the target voice feature point. Therefore, the purpose of helping the user to manage the emotion by objectively and accurately analyzing the emotion change of the user can be achieved.
On the basis of the above embodiment, the identification module includes:
the acquisition unit is used for acquiring voice data of a user according to a preset rule;
the recognition unit is used for recognizing semantic content and intonation included in the voice data;
and the extraction unit is used for extracting the language key words and the target intonation from the identified semantic content and the intonation, wherein the target intonation comprises the volume, the speed and the tone of the voice data and the respective change trend.
On the basis of the above embodiment, the modeling module includes:
and the modeling unit is used for performing convolution operation on the voice feature points to obtain a convolution operation result and generating a target emotion model based on the convolution operation result.
On the basis of the above embodiment, the modeling module includes:
and the calibration unit is used for calibrating the extreme value points in the convolution operation result as target voice characteristic points.
On the basis of the above embodiment, the system further includes:
and the big data processing and updating module is used for updating the intonation characteristic data and the psychological behavior characteristic data in the database through big data analysis and calculation, wherein the database is a local database or an Internet database.
The emotion analysis system provided by the embodiment of the invention can execute the emotion analysis method provided by any embodiment of the invention, and has corresponding functional modules and beneficial effects of the execution method.
Example four
Fig. 4 is a schematic structural diagram of a server according to a fourth embodiment of the present invention. FIG. 4 illustrates a block diagram of an exemplary server 12 suitable for use in implementing embodiments of the present invention. The server 12 shown in fig. 4 is only an example, and should not bring any limitation to the function and the scope of use of the embodiment of the present invention.
As shown in FIG. 4, the server 12 is in the form of a general purpose computing device. The components of the server 12 may include, but are not limited to: one or more processors or processing units 16, a system memory 28, and a bus 18 that couples various system components including the system memory 28 and the processing unit 16.
The server 12 typically includes a variety of computer system readable media. Such media may be any available media that is accessible by server 12 and includes both volatile and nonvolatile media, removable and non-removable media.
The system memory 28 may include computer system readable media in the form of volatile memory, such as Random Access Memory (RAM)30 and/or cache memory 32. The server 12 may further include other removable/non-removable, volatile/nonvolatile computer system storage media. By way of example only, storage system 34 may be used to read from and write to non-removable, nonvolatile magnetic media (not shown in FIG. 4, and commonly referred to as a "hard drive"). Although not shown in FIG. 4, a magnetic disk drive for reading from and writing to a removable, nonvolatile magnetic disk (e.g., a "floppy disk") and an optical disk drive for reading from or writing to a removable, nonvolatile optical disk (e.g., a CD-ROM, DVD-ROM, or other optical media) may be provided. In these cases, each drive may be connected to bus 18 by one or more data media interfaces. Memory 28 may include at least one program product having a set (e.g., at least one) of program modules that are configured to carry out the functions of embodiments of the invention.
A program/utility 40 having a set (at least one) of program modules 42 may be stored, for example, in memory 28, such program modules 42 including, but not limited to, an operating system, one or more application programs, other program modules, and program data, each of which examples or some combination thereof may comprise an implementation of a network environment. Program modules 42 generally carry out the functions and/or methodologies of the described embodiments of the invention.
The server 12 may also communicate with one or more external devices 14 (e.g., keyboard, pointing device, display 24, etc.), with one or more devices that enable a user to interact with the server 12, and/or with any devices (e.g., network card, modem, etc.) that enable the server 12 to communicate with one or more other computing devices. Such communication may be through an input/output (I/O) interface 22. Also, the server 12 may communicate with one or more networks (e.g., a Local Area Network (LAN), a Wide Area Network (WAN), and/or a public network, such as the Internet) via the network adapter 20. As shown, the network adapter 20 communicates with the other modules of the server 12 via the bus 18. It should be understood that although not shown in the figures, other hardware and/or software modules may be used in conjunction with the server 12, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID systems, tape drives, and data backup storage systems, among others.
The processing unit 16 executes various functional applications and data processing by running a program stored in the system memory 28, for example, to implement the emotion analysis classification method provided by the embodiment of the present invention, including:
identifying language keywords and target tones included in the acquired voice data;
analyzing and determining voice feature points according to the acquired language keywords and the target intonation, wherein the voice feature points are keywords and intonations which mark user emotion in the language keywords and the target intonation;
generating a target emotion model based on the voice feature points, and calibrating the target voice feature points on the target emotion model;
matching the target emotion model with standard emotion models in a standard emotion model library to adjust the calibrated target voice feature points on the target emotion model and recording the change data of the target voice feature points;
and matching the change data with the intonation characteristics and the psychological behavior characteristics in the database, and outputting the emotion or emotion change data of the user according to the matching result.
In one embodiment, processing unit 16 implements a method of emotion analysis by executing a program stored in system memory 28, further comprising:
acquiring voice data of a user according to a preset rule;
recognizing semantic content and intonation included in the voice data;
extracting language keywords and target intonations from the identified semantic content and intonations, wherein the target intonation comprises at least one of volume, speed, tone and respective trend of change of the voice data.
In one embodiment, processing unit 16 implements a method of emotion analysis by executing a program stored in system memory 28, further comprising:
and carrying out convolution operation on the voice feature points to obtain a convolution operation result, and generating a target emotion model based on the convolution operation result.
In one embodiment, processing unit 16 implements a method of emotion analysis by executing a program stored in system memory 28, further comprising:
and marking the extreme value points in the convolution operation result as target voice characteristic points.
In one embodiment, processing unit 16 implements a method of emotion analysis by executing a program stored in system memory 28, further comprising:
and updating the intonation characteristic data and the psychological behavior characteristic data in the database through big data analysis and calculation, wherein the database is a local database or an Internet database.
EXAMPLE five
An embodiment of the present invention further provides a computer-readable storage medium, on which a computer program is stored, where the computer program, when executed by a processor, implements the emotion analysis method provided in the embodiment of the present invention, and the method includes:
identifying language keywords and target tones included in the acquired voice data;
analyzing and determining voice feature points according to the acquired language keywords and the target intonation, wherein the voice feature points are keywords and intonations which mark user emotion in the language keywords and the target intonation;
generating a target emotion model based on the voice feature points, and calibrating the target voice feature points on the target emotion model;
matching the target emotion model with standard emotion models in a standard emotion model library to adjust the calibrated target voice feature points on the target emotion model and recording the change data of the target voice feature points;
and matching the change data with the intonation characteristics and the psychological behavior characteristics in the database, and outputting the emotion or emotion change data of the user according to the matching result.
In one embodiment, the program when executed by the processor may further implement:
acquiring voice data of a user according to a preset rule;
recognizing semantic content and intonation included in the voice data;
extracting language keywords and target intonations from the identified semantic content and intonations, wherein the target intonation comprises at least one of volume, speed, tone and respective trend of change of the voice data.
In one embodiment, the program when executed by the processor may further implement:
and carrying out convolution operation on the voice feature points to obtain a convolution operation result, and generating a target emotion model based on the convolution operation result.
In one embodiment, the program when executed by the processor may further implement:
and marking the extreme value points in the convolution operation result as target voice characteristic points.
In one embodiment, the program when executed by the processor may further implement:
and updating the intonation characteristic data and the psychological behavior characteristic data in the database through big data analysis and calculation, wherein the database is a local database or an Internet database.
Computer storage media for embodiments of the invention may employ any combination of one or more computer-readable media. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any type of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet service provider).
It is to be noted that the foregoing is only illustrative of the preferred embodiments of the present invention and the technical principles employed. It will be understood by those skilled in the art that the present invention is not limited to the particular embodiments described herein, but is capable of various obvious changes, rearrangements and substitutions as will now become apparent to those skilled in the art without departing from the scope of the invention. Therefore, although the present invention has been described in greater detail by the above embodiments, the present invention is not limited to the above embodiments, and may include other equivalent embodiments without departing from the spirit of the present invention, and the scope of the present invention is determined by the scope of the appended claims.
Claims (10)
1. A method of sentiment analysis, the method comprising:
identifying language keywords and target tones included in the acquired voice data;
analyzing and determining voice feature points according to the acquired language keywords and the target intonation, wherein the voice feature points are keywords and intonations which mark user emotion in the language keywords and the target intonation;
generating a target emotion model based on the voice feature points, and calibrating the target voice feature points on the target emotion model;
matching the target emotion model with standard emotion models in a standard emotion model library to adjust the calibrated target voice feature points on the target emotion model and recording the change data of the target voice feature points;
and matching the change data with the tone characteristic data and the psychological behavior characteristic data in the database, and outputting the emotion or emotion change data of the user according to the matching result.
2. The method according to claim 1, wherein the recognizing the language keyword and the target intonation included in the acquired voice data comprises:
acquiring voice data of a user according to a preset rule;
recognizing semantic content and intonation included in the voice data;
extracting language keywords and target intonations from the identified semantic content and intonations, wherein the target intonation comprises at least one of volume, speed, tone and respective trend of change of the voice data.
3. The method of claim 1, wherein generating a target emotion model based on the speech feature points comprises:
and carrying out convolution operation on the voice feature points to obtain a convolution operation result, and generating a target emotion model based on the convolution operation result.
4. The method of claim 3, wherein the step of calibrating the target speech feature points on the target emotion model comprises:
and marking the extreme value points in the convolution operation result as target voice characteristic points.
5. The method of claim 1, further comprising:
and updating the intonation characteristic data and the psychological behavior characteristic data in the database through big data analysis and calculation, wherein the database is a local database or an Internet database.
6. An emotion analysis system, characterized in that the system comprises:
the recognition module is used for recognizing the language keywords and the target intonation included in the acquired voice data;
the feature point analysis module is used for analyzing and determining voice feature points according to the acquired language key words and the target intonation, wherein the voice feature points are the key words and the intonations which mark the emotion of the user in the language key words and the target intonation;
the modeling module is used for generating a target emotion model based on the analyzed and determined voice feature points and calibrating the target voice feature points on the target emotion model;
the adjustment recording module is used for matching the target emotion model with a standard emotion model in a standard emotion model library so as to adjust the calibrated target voice feature points on the target emotion model and record the change data of the target voice feature points;
and the emotion analysis module is used for matching the change data with the tone characteristic data and the psychological behavior characteristic data in the database and outputting the emotion or emotion change data of the user according to the matching result.
7. The system of claim 6, wherein the identification module comprises:
the acquisition unit is used for acquiring voice data of a user according to a preset rule;
the recognition unit is used for recognizing semantic content and intonation included in the voice data;
and the extraction unit is used for extracting the language key words and the target intonation from the identified semantic content and the intonation, wherein the target intonation comprises at least one of the volume, the speed, the tone and the change trend of the voice data.
8. The system of claim 6, further comprising:
and the big data processing and updating module is used for updating the intonation characteristic data and the psychological behavior characteristic data in the database through big data analysis and calculation, wherein the database is a local database or an Internet database.
9. A server, characterized in that the server comprises:
one or more processors;
a memory for storing one or more programs,
the one or more programs, when executed by the one or more processors, cause the one or more processors to implement the emotion analysis method as recited in any of claims 1-5.
10. A computer-readable storage medium, on which a computer program is stored, which program, when being executed by a processor, is adapted to carry out the emotion analyzing method as set forth in any one of claims 1 to 5.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811005214.2A CN109087670B (en) | 2018-08-30 | 2018-08-30 | Emotion analysis method, system, server and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811005214.2A CN109087670B (en) | 2018-08-30 | 2018-08-30 | Emotion analysis method, system, server and storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109087670A CN109087670A (en) | 2018-12-25 |
CN109087670B true CN109087670B (en) | 2021-04-20 |
Family
ID=64840278
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811005214.2A Active CN109087670B (en) | 2018-08-30 | 2018-08-30 | Emotion analysis method, system, server and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109087670B (en) |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109801105A (en) * | 2019-01-17 | 2019-05-24 | 深圳壹账通智能科技有限公司 | Service methods of marking, device, equipment and storage medium based on artificial intelligence |
CN110047508A (en) * | 2019-03-12 | 2019-07-23 | 平安科技(深圳)有限公司 | Voice recording method, device, computer equipment and storage medium |
CN110110135A (en) * | 2019-04-17 | 2019-08-09 | 西安极蜂天下信息科技有限公司 | Voice characteristics data library update method and device |
CN112037820B (en) * | 2019-05-16 | 2023-09-05 | 杭州海康威视数字技术股份有限公司 | Security alarm method, device, system and equipment |
CN112037821A (en) * | 2019-06-03 | 2020-12-04 | 阿里巴巴集团控股有限公司 | Visual representation method and device of voice emotion and computer storage medium |
CN111354377B (en) * | 2019-06-27 | 2022-11-18 | 深圳市鸿合创新信息技术有限责任公司 | Method and device for recognizing emotion through voice and electronic equipment |
CN110399837B (en) * | 2019-07-25 | 2024-01-05 | 深圳智慧林网络科技有限公司 | User emotion recognition method, device and computer readable storage medium |
CN112307816B (en) * | 2019-07-29 | 2024-08-20 | 北京地平线机器人技术研发有限公司 | In-vehicle image acquisition method and device, electronic equipment and storage medium |
CN110534135A (en) * | 2019-10-18 | 2019-12-03 | 四川大学华西医院 | A method of emotional characteristics are assessed with heart rate response based on language guidance |
CN112631137A (en) * | 2020-04-02 | 2021-04-09 | 张瑞华 | Intelligent household control method and intelligent control equipment applied to biological feature recognition |
CN112071304B (en) * | 2020-09-08 | 2024-03-15 | 深圳市天维大数据技术有限公司 | Semantic analysis method and device |
CN113284494B (en) * | 2021-05-25 | 2023-12-01 | 北京基智科技有限公司 | Voice assistant recognition method, device, equipment and computer readable storage medium |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104050965A (en) * | 2013-09-02 | 2014-09-17 | 广东外语外贸大学 | English phonetic pronunciation quality evaluation system with emotion recognition function and method thereof |
WO2018061839A1 (en) * | 2016-09-29 | 2018-04-05 | 株式会社村田製作所 | Transmission device, transmission method, and transmission program |
CN108010516A (en) * | 2017-12-04 | 2018-05-08 | 广州势必可赢网络科技有限公司 | Semantic independent speech emotion feature recognition method and device |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010105396A1 (en) * | 2009-03-16 | 2010-09-23 | Fujitsu Limited | Apparatus and method for recognizing speech emotion change |
CN101789990A (en) * | 2009-12-23 | 2010-07-28 | 宇龙计算机通信科技(深圳)有限公司 | Method and mobile terminal for judging emotion of opposite party in conservation process |
CN103593054B (en) * | 2013-11-25 | 2018-04-20 | 北京光年无限科技有限公司 | A kind of combination Emotion identification and the question answering system of output |
CN104102627B (en) * | 2014-07-11 | 2016-10-26 | 合肥工业大学 | A kind of multi-modal noncontact sentiment analysis record system |
US9576190B2 (en) * | 2015-03-18 | 2017-02-21 | Snap Inc. | Emotion recognition in video conferencing |
CN105159979A (en) * | 2015-08-27 | 2015-12-16 | 广东小天才科技有限公司 | friend recommendation method and device |
CN106548788B (en) * | 2015-09-23 | 2020-01-07 | 中国移动通信集团山东有限公司 | Intelligent emotion determining method and system |
CN107948417A (en) * | 2017-11-22 | 2018-04-20 | 周燕红 | A kind of method, apparatus, terminal and the storage medium of voice data monitoring |
CN108010512B (en) * | 2017-12-05 | 2021-04-30 | 广东小天才科技有限公司 | Sound effect acquisition method and recording terminal |
-
2018
- 2018-08-30 CN CN201811005214.2A patent/CN109087670B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104050965A (en) * | 2013-09-02 | 2014-09-17 | 广东外语外贸大学 | English phonetic pronunciation quality evaluation system with emotion recognition function and method thereof |
WO2018061839A1 (en) * | 2016-09-29 | 2018-04-05 | 株式会社村田製作所 | Transmission device, transmission method, and transmission program |
CN108010516A (en) * | 2017-12-04 | 2018-05-08 | 广州势必可赢网络科技有限公司 | Semantic independent speech emotion feature recognition method and device |
Also Published As
Publication number | Publication date |
---|---|
CN109087670A (en) | 2018-12-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109087670B (en) | Emotion analysis method, system, server and storage medium | |
US10522136B2 (en) | Method and device for training acoustic model, computer device and storage medium | |
WO2021000408A1 (en) | Interview scoring method and apparatus, and device and storage medium | |
CN110069608B (en) | Voice interaction method, device, equipment and computer storage medium | |
WO2021000497A1 (en) | Retrieval method and apparatus, and computer device and storage medium | |
CN106897439B (en) | Text emotion recognition method, device, server and storage medium | |
US10529340B2 (en) | Voiceprint registration method, server and storage medium | |
WO2019148586A1 (en) | Method and device for speaker recognition during multi-person speech | |
CN110675862A (en) | Corpus acquisition method, electronic device and storage medium | |
CN110880329A (en) | Audio identification method and equipment and storage medium | |
US20200219487A1 (en) | Information processing apparatus and information processing method | |
CN109299227B (en) | Information query method and device based on voice recognition | |
CN113707173B (en) | Voice separation method, device, equipment and storage medium based on audio segmentation | |
CN113628627A (en) | Electric power industry customer service quality inspection system based on structured voice analysis | |
CN111180025A (en) | Method and device for representing medical record text vector and inquiry system | |
WO2020233381A1 (en) | Speech recognition-based service request method and apparatus, and computer device | |
CN112397073A (en) | Audio data processing method and device | |
CN108847251B (en) | Voice duplicate removal method, device, server and storage medium | |
CN113096687A (en) | Audio and video processing method and device, computer equipment and storage medium | |
CN113593523A (en) | Speech detection method and device based on artificial intelligence and electronic equipment | |
CN113241060B (en) | Security early warning method and system | |
JP2019124897A (en) | Speech analysis apparatus, speech analysis system, and program | |
CN110276001B (en) | Checking page identification method and device, computing equipment and medium | |
CN110442714B (en) | POI name normative evaluation method, device, equipment and storage medium | |
CN113761206A (en) | Intelligent information query method, device, equipment and medium based on intention recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |