Nothing Special   »   [go: up one dir, main page]

CN106936991A - The method and terminal of a kind of automatic regulating volume - Google Patents

The method and terminal of a kind of automatic regulating volume Download PDF

Info

Publication number
CN106936991A
CN106936991A CN201710124718.5A CN201710124718A CN106936991A CN 106936991 A CN106936991 A CN 106936991A CN 201710124718 A CN201710124718 A CN 201710124718A CN 106936991 A CN106936991 A CN 106936991A
Authority
CN
China
Prior art keywords
parameter
feature parameter
noise
environment
speech model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN201710124718.5A
Other languages
Chinese (zh)
Inventor
朱益
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Jinli Communication Equipment Co Ltd
Original Assignee
Shenzhen Jinli Communication Equipment Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Jinli Communication Equipment Co Ltd filed Critical Shenzhen Jinli Communication Equipment Co Ltd
Priority to CN201710124718.5A priority Critical patent/CN106936991A/en
Publication of CN106936991A publication Critical patent/CN106936991A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72448User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
    • H04M1/72454User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions according to context-related or environment-related conditions
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0272Voice signal separating
    • G10L21/028Voice signal separating using properties of sound source
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/60Substation equipment, e.g. for use by subscribers including speech amplifiers
    • H04M1/6033Substation equipment, e.g. for use by subscribers including speech amplifiers for providing handsfree use or a loudspeaker mode in telephone sets
    • H04M1/6041Portable telephones adapted for handsfree use
    • H04M1/605Portable telephones adapted for handsfree use involving control of the receiver volume to provide a dual operational mode at close or far distance from the user
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
    • G10L15/063Training
    • G10L2015/0635Training updating or merging of old and new templates; Mean values; Weighting
    • G10L2015/0636Threshold criteria for the updating
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L2015/088Word spotting
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L2021/02087Noise filtering the noise being separate speech, e.g. cocktail party

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Quality & Reliability (AREA)
  • Artificial Intelligence (AREA)
  • Environmental & Geological Engineering (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Telephone Function (AREA)

Abstract

The embodiment of the invention discloses the method and terminal of a kind of automatic regulating volume, methods described includes:The current environment voice signal of detection;The characteristic parameter of environment voice signal is obtained, whether characteristic parameter is non-noise signal for environment-identification voice signal;Whether it is non-noise signal according to characteristic parameter and default speech model environment-identification voice signal, speech model includes the scope of characteristic parameter in non-noise region;If non-noise signal, current volume is reduced to preset value.The method of automatic regulating volume provided in an embodiment of the present invention can effectively prevent user from causing missing or omitting for information because video or audio broadcast sound volume are too high with automatic regulating volume.

Description

The method and terminal of a kind of automatic regulating volume
Technical field
The present invention relates to electronic technology field, more particularly to a kind of automatic regulating volume method and terminal.
Background technology
With the development of intelligent electronic device, intelligent electronic device brings great convenience to the life of user, especially It is that smart mobile phone has become one of daily necessities of user.It should be appreciated that user often listens music or viewing to regard using mobile phone Volume is larger and cause user usually not hear or do not hear other people dialogues with itself when frequently, due to playing music or video Content and cause missing or omitting for information, and the volume adjusting mode of existing mobile phone is often the control of user's manually operated volume Panel or operating physical button realize regulation, therefore cannot in time avoid above-mentioned situation by automatic regulating volume in time Occur.
The content of the invention
The embodiment of the present invention provides a kind of method and terminal of automatic regulating volume, can be reduced and used with automatic regulating volume Family because the broadcast sound volume of video or audio is too high cause to be linked up face to face with other people when there is the situation missed or omit of information.
In a first aspect, the embodiment of the invention provides a kind of method of automatic regulating volume, the above method includes:
The current environment voice signal of detection;The characteristic parameter of environment voice signal is obtained, characteristic parameter is used for identification ring Whether border voice signal is non-noise signal;According to characteristic parameter and default speech model environment-identification voice signal whether It is non-noise signal, speech model includes the scope of characteristic parameter in non-noise region;If non-noise signal, by current sound Amount is reduced to preset value.
On the other hand, a kind of terminal is the embodiment of the invention provides, the terminal includes:Detection unit, collecting unit, identification Unit and control unit,
Detection unit is used to detect current environment voice signal;Collecting unit is used to obtain the feature of environment voice signal Whether parameter, characteristic parameter is non-noise signal for environment-identification voice signal;Recognition unit be used for according to characteristic parameter with And whether default speech model environment-identification voice signal is non-noise signal, speech model includes feature in non-noise region The scope of parameter;If control unit is used for non-noise signal, current volume is reduced to preset value.
The method and terminal of a kind of automatic regulating volume disclosed in the embodiment of the present invention are by gathering environment voice signal Characteristic parameter come whether environment-identification voice is non-noise signal, if non-noise signal, then current volume is reduced to Preset value, can effectively prevent user from causing missing or omitting for information because the broadcast sound volume of video or audio is too high.
Brief description of the drawings
Technical scheme in order to illustrate more clearly the embodiments of the present invention, embodiment will be described below needed for be used Accompanying drawing is briefly described, it should be apparent that, drawings in the following description are some embodiments of the present invention, general for this area For logical technical staff, on the premise of not paying creative work, other accompanying drawings can also be obtained according to these accompanying drawings.
Fig. 1 is a kind of indicative flowchart of the method for automatic regulating volume provided in an embodiment of the present invention;
Fig. 2 is the schematic diagram of the sub-process of Fig. 1 provided in an embodiment of the present invention;
Fig. 3 a are a kind of function curve schematic diagrames of characteristic function formula provided in an embodiment of the present invention;
Fig. 3 b are the function curve schematic diagrames of another characteristic function formula provided in an embodiment of the present invention;
Fig. 4 is a kind of schematic block diagram of terminal provided in an embodiment of the present invention;
Fig. 5 is a kind of schematic block diagram of the recognition unit in Fig. 4 provided in an embodiment of the present invention;
Fig. 6 is the schematic block diagram of another terminal provided in an embodiment of the present invention.
Specific embodiment
Below in conjunction with the accompanying drawing in the embodiment of the present invention, the technical scheme in the embodiment of the present invention is carried out clear, complete Site preparation is described, it is clear that described embodiment is a part of embodiment of the invention, rather than whole embodiments.Based on this hair Embodiment in bright, the every other implementation that those of ordinary skill in the art are obtained under the premise of creative work is not made Example, belongs to the scope of protection of the invention.
It should be appreciated that when using in this specification and in the appended claims, term " including " and "comprising" instruction The presence of described feature, entirety, step, operation, element and/or component, but be not precluded from one or several further features, The presence or addition of entirety, step, operation, element, component and/or its set.It is also understood that in this description of the invention The term for being used is not intended to limit the present invention merely for the sake of the purpose of description specific embodiment.Such as in explanation of the invention As used in book and appended claims, unless context clearly indicates other situations, otherwise singulative " one ", " one " and " being somebody's turn to do " are intended to include plural form.
A kind of method for automatically adjusting provided in an embodiment of the present invention, the above method runs on terminal, and terminal is included still It is not restricted to any electronic product that man-machine interaction can be carried out with user, for example smart mobile phone (such as Android phone, ios Mobile phone, Windows Phone mobile phones etc.), flat board electric energy, palm electric energy, notebook computer, mobile internet device etc..More than Electronic equipment is only citing, and non exhaustive, and but terminal provided in an embodiment of the present invention includes is not restricted to above-mentioned electronic equipment. It should be noted that terminal provided in an embodiment of the present invention should have the element of collection environment voice signal, such as microphone.
1 is please see Figure, is a kind of indicative flowchart of the method for automatic regulating volume provided in an embodiment of the present invention, should Understand, the automatic regulating volume described in the embodiment of the present invention be based on current on the premise of audio or video resource etc. is played Carry out.As illustrated, a kind of method of automatic regulating volume provided in an embodiment of the present invention includes following S101~S105:
S101, detects current environment voice signal.
Specifically, the embodiment of the present invention detects environment voice signal preferably by microphone, in other feasible embodiments Can be used to detect environment voice signal using other electronic components, the embodiment of the present invention does not limit this specifically.
S102, obtains the characteristic parameter of environment voice signal, and whether characteristic parameter is non-noise for environment-identification voice Signal.
Specifically, the classification of environment voice signal includes noise signal and non-noise signal, noise signal in the present embodiment Including environmental noise signal and voice noise signal, environmental noise includes building site noise, vehicle noise, office's keyboard noise Include under dialogue noise and quiet environment under noisy environment Deng, voice noise other people talk with noise, such as subway traffic work Colleague's talk sound transmitted at a distance in dialogue noise, office noisy in tool etc..It should be appreciated that non-noise signal is on non- The voice signal of noise signal is stated, being depending on non-noise signal possible be with user-related dialogic voice signal.
Preferably, the characteristic parameter of the environment voice signal for getting includes fisrt feature parameter and second feature parameter. In other feasible embodiments, the characteristic parameter of the environment voice signal for getting can be a characteristic parameter or three or The characteristic parameter that person is more than three.The embodiment of the present invention will include that fisrt feature parameter and second feature parameter are with characteristic parameter Example is illustrated.
Preferably, fisrt feature parameter is the short-time zero-crossing rate of environment voice signal;Second feature parameter is environment voice The short-time average energy of signal, i.e. fisrt feature parameter are the short-time zero-crossing rates of the whole environment voice signal for obtaining, and second is special Levy the short-time average energy that parameter is the whole environment voice signal for obtaining.Short-time zero-crossing rate represents voice signal in a frame voice Number of times of the waveform through zero level;Short-time average energy represents the energy of n moment voice signals.Wherein short-time zero-crossing rate Zn and short When average energy computing formula it is as follows:
Wherein, N is window long, short-time energy average value be a frame sample value weighted average and;Sgn [] is sign function, x During (n)≤0, sgn [x (n)]=1;x(n)<When 0, sgn [x (n)]=- 1.
It should be appreciated that in the present embodiment, the short-time zero-crossing rate and short-time average energy of noise signal and non-noise signal It is different and then can to efficiently differentiate environment voice signal be noise signal or non-noise signal.For example, under quiet environment The short-time average energy of the voice noise that the colleague transmitted at a distance under voice noise such as office environment talks, its use with collection The short-time average energy of the related dialogic voice signal in family is different, this be based on it is general with user-related dialogue betide with Family is separated by relatively closely, and the voice signal that the colleague of distant place talks has been lost energy, therefore its short-time average energy by propagating Amount is different, and then can distinguish voice noise signal and non-noise signal.And voice noise and environment under noisy environment Noise is significantly different with the short-time zero-crossing rate and short-time average energy of non-noise, can more efficiently distinguish.
Whether S103, be non-noise signal according to characteristic parameter and default speech model environment-identification voice signal, Speech model includes the scope of characteristic parameter in non-noise region, if non-noise signal, performs S104, if not non-noise is believed Number, perform S105.
If specifically, obtain characteristic parameter in the range of characteristic parameter in non-noise region, be non-noise signal; It is noise signal if the characteristic parameter for obtaining is in the range of characteristic parameter in noise region.Due to noise signal and non-make an uproar The scope of the corresponding characteristic parameter of message number is different, and included in speech model characteristic parameter in noise region scope and The scope of characteristic parameter in non-noise region, therefore can identify that environment voice signal is noise according to default speech model Signal is also non-noise signal.It should be noted that speech model is trained according to some data samples.
Then be included in when characteristic parameter is only a parameter, in speech model in noise region and non-noise region this one The scope of individual characteristic parameter;If characteristic parameter is two or more, speech model is then included in noise region and non-makes an uproar The scope of two or more the characteristic parameter in sound region, such as in a kind of mode, first is special in non-noise region The scope of parameter is levied for (a1, a2), the scope (b1, b2) of second feature parameter, in another way, first in non-noise region The scope of characteristic parameter is (a1, a2), and when fisrt feature parameter is a1, the scope of second feature parameter is (b1, b2), now Each specific fisrt feature parameter corresponds to a scope for second feature parameter in non-noise region.
S104, preset value is reduced to by current volume, and flow terminates.
S105, holding current volume is constant, and flow terminates.
If specifically, non-noise signal, then it represents that environment voice be particularly likely that with user-related voice, then should Reducing current volume can in time hear voice to ensure user;Then it is noise signal if not non-noise signal, then can keeps Current volume is constant.
It should be appreciated that preset value can be a fixed value in certain embodiments, no matter that is, current volume is how many Current volume is reduced to a fixed value;In further embodiments, preset value can be the dynamic related to current volume Value, such as the 1/3 of current volume, the present invention are not limited specifically this.
In sum, the embodiment of the present invention by characteristic parameter come recognition detection to environment voice signal whether be non-making an uproar Message number determines whether to reduce current volume automatically, due to being depending on non-noise signal and user-related dialogic voice letter Number, therefore detect environment voice signal and can effectively prevent user because of current sound to reduce current volume after non-noise signal The broadcast sound volume of frequency or video is too high and miss voice messaging.On the other hand, embodiment of the present invention preferred feature parameter includes Fisrt feature parameter and second feature parameter, and fisrt feature parameter and second feature parameter are combined for recognizing that noise is believed Number and non-noise signal, improve identification accuracy.
Preferably, before current volume is reduced to preset value by S015, the above method also includes:
Whether there is default keyword in environment-identification voice signal, default keyword is related to user, for knowing Whether other environment voice is related to user;And if there is default keyword, then perform current volume is reduced to it is default Value;If not existing default keyword, keep current volume constant.
Specifically, default keyword is related to user, the name of such as user, the pet name and academic title etc., Ke Yiyong In the identity for representing user.Whether there is the default keyword can further environment-identification voice in environment-identification voice signal Whether be with it is user-related, if related to user, then reduce current volume;If not it is related to user, then keep current Volume is constant.For example can efficiently differentiate be sent in voice noise using which and be separated by close range still with user The dialogic voice unrelated with user and with user-related dialogic voice.Recognize whether that default keyword can be logical Keyword identification (KWR) system is crossed, that is, make use of the phonetic features such as energy, cepstrum coefficient in voice.
Preferably, default speech model includes the speech model of several different scenes, and different scenes can be expressed as In the scene of different location, such as office, family, on subway, the different time scene in same place is also denoted as, such as handled official business Different time sections in room:When go to work in the morning, during nooning, when go to work in the afternoon.
And then preferably, S103 according to the characteristic parameter and default speech model environment-identification voice signal whether be Before non-noise signal, the above method also includes:
Obtain current time;
Scene is currently located according to current time and the historical behavior information identifying user for prestoring;
The speech model that matches is searched from several speech models according to scene is currently located.
Specifically, the historical behavior information analysis user according to user is currently located scene, such as according to workaday difference Time period identifies user in office or family or way on and off duty.It should be appreciated that distinguishing the language of multiple different scenes Sound model is that, in order to more accurately identify noise signal and non-noise signal, this is based on the noise under the different scenes having There is larger difference in signal, similarly have the non-noise signal under a different scenes to there is also larger difference.
It is also understood that speech model is the data model obtained according to some data samples of collection.Data sample is The characteristic parameter gathered under noisy environment or under non-noisy environment.
Preferably, the characteristic parameter for being obtained in S102 includes fisrt feature parameter and second feature parameter.Now, please see Figure Whether according to characteristic parameter and default speech model environment-identification voice signal it is non-noise signal in 2, S103, including:
S201, the second spy of standard is calculated according to characteristic function formula in fisrt feature parameter and default speech model Parameter is levied, characteristic function formula is the relational expression of fisrt feature parameter and second feature parameter.
Specifically, preferred fisrt feature parameter is the short-time zero-crossing rate of environment voice signal in the present embodiment;Second feature Parameter is the short-time average energy of environment voice signal.Now characteristic function formula is the pass of short-time zero-crossing rate and short-time average energy It is formula, preferably polynomial function, because short-time zero-crossing rate is bent with polynomial function with the relational expression of short-time average energy Line is similar to.
In certain embodiments, characteristic function formula is by the fisrt feature parameter and second feature in the case of some noises What the data sample of fisrt feature parameter and second feature parameter in the case of the data sample of parameter and non-noise was obtained makes an uproar The boundary function of message number and non-noise signal, boundary function is specifically the first spy of noise signal and non-noise signal boundary Levy the relational expression of function and second feature function.For example, a kind of function curve diagram of characteristic function formula as shown in Figure 3 a is illustrated Figure, curve L1 is the coordinate curve schematic diagram of characteristic function.The region S1 of curve L1 L1 included below is expressed as non-noise region The scope of middle fisrt feature parameter and second feature parameter;Curve L1 area above S2 be noise region in fisrt feature parameter with The scope of second feature parameter.It should be appreciated that dividing non-noise region and noise region according to characteristic function formula in speech model Division rule be according to data sample derive.The second feature parameter of standard is now calculated according to fisrt feature parameter Afterwards, can be according to the model of the second feature parameter in the second feature parameter acquiring of standard to noise region and non-noise region Enclose.
In further embodiments, characteristic function formula is special by the fisrt feature parameter in the case of some noises and second Levy the relational expression or some non-noise situations of fisrt feature parameter that the data sample of parameter obtains and second feature parameter Under fisrt feature parameter and second feature parameter the fisrt feature parameter that obtains of data sample and second feature parameter pass It is formula.Now characteristic function formula be embodied as in noise signal the relational expression of fisrt feature parameter and second feature parameter or The relational expression of fisrt feature parameter and second feature parameter in non-noise signal.Divide non-according to characteristic function formula in speech model The division rule of the scope of corresponding fisrt feature parameter and second feature parameter is according to number in noise region and noise region According to sample derive, preferably division rule for keep fisrt feature functional value it is constant, second feature functional value increase or Preset value is reduced to generate the scope of second feature parameter in non-noise region and noise region.For example, as shown in Figure 3 b another A kind of function curve diagram schematic diagram of characteristic function formula, curve L2 is that the coordinate curve of characteristic function in non-noise scene is illustrated Figure.When fisrt feature parameter is Z1, corresponding second feature parameter is E0 in characteristic function formula, now second in non-noise region Characteristic parameter scope is (E0- △ E2, E0+ △ E1);Now, the scope of second feature parameter is less than E0- △ in noise region E2 and more than E0+ △ E1, wherein △ E2 and △ E1 are obtained according to sample data, be can be the same or different.
S202, second feature parameter in the non-noise region of second feature parameter and preset rules acquisition according to standard Scope.
Specifically, preset rules divide voice in representing speech model according to the second feature parameter of standard in the present embodiment The division rule in noise region and non-noise region in model.
As shown in Figure 3 a, if characteristic function formula is boundary function, the of standard is calculated according to fisrt feature parameter z1 After two characteristic parameter E0, second feature parameter E0 and division rule according to standard have got the second spy in non-noise region The scope of parameter is levied for S1 is less than or equal to E0.
S203, detect obtain second feature parameter whether in non-noise region second feature parameter scope, if In non-noise region in the range of second feature parameter, perform S204, if not in non-noise region second feature parameter model In enclosing, S205 is performed;
S204, environment voice signal is non-noise signal.
S205, environment voice signal is not non-noise signal.
Similarly, if characteristic parameter includes three or more than three, it is also applied for the above method and carrys out environment-identification voice letter Whether number it is non-noise signal, such as when including three characteristic parameters, two of which characteristic parameter is calculated according to preset rules Go out fourth feature parameter, retraining goes out fourth feature parameter with a remaining characteristic function formula for characteristic parameter.
When further, due to the method for performing automatic regulating volume, speech model has built up, in order to improve The accuracy of identification, to prevent speech model no longer where applicable from persistently occurring using the situation of the speech model, the above method is also Including:
If receiving the abnormal information of detection, whether the abnormal frequency of occurrences of recognition detection has exceeded particular value;
If having exceeded particular value, estimating for current detection error according to calculation of characteristic parameters when detection is abnormal every time Value;And the scope of characteristic parameter in the non-noise region of default speech model is adjusted according to discreet value;
If not less than particular value, not processed.
Specifically, the abnormal information of detection is to be operated according to user and generated, for example, adjust volume display reminding information Volume will be adjusted, if user does not receive to reduce the request of volume, is considered as one-time detection exception or is reduced user after volume Elevated volume is also considered as detection exception immediately.It should be understood that detecting whether that more than particular value be in order to the event for preventing detection abnormal is Once in a while speech model is have adjusted in the case of event.
It should be noted that the discreet value of detection error is the gap of the result and actual result drawn according to characteristic parameter Obtain.Feature ginseng in scope and noise region further according to characteristic parameter in non-noise region in error amount regulation speech model Several scopes, makes the speech model after adjustment be applied to current scene, improves detection accuracy.
For example, when characteristic parameter includes fisrt feature parameter and second feature parameter, the preferably discreet value of detection error It is the second feature parameter and the quadratic sum of the deviation of the second feature parameter of standard of acquisition when detection is abnormal every time, such as obtains Second feature parameter and standard second feature parameter difference quadratic sum.Now, it is preferably default according to discreet value regulation Speech model non-noise region in the scope of characteristic parameter be specifically:Fisrt feature parameter and the is adjusted according to discreet value The characteristic function formula of two characteristic parameters.
It should be appreciated that detection accuracy can be improved by monitor and detection exception and dynamic regulation speech model.
4 are please see Figure, is a kind of schematic block diagram of terminal provided in an embodiment of the present invention, the side of above-mentioned automatic regulating volume Method runs on terminal, as illustrated, the terminal 400 includes:Detection unit 401, collecting unit 402, recognition unit 403 and control Unit processed 404.
Wherein, detection unit 401, for detecting current environment voice signal.
Collecting unit 402, the characteristic parameter for obtaining environment voice signal, characteristic parameter is for environment-identification voice No is non-noise signal.
Specifically, the classification of environment voice signal includes noise signal and non-noise signal, noise signal in the present embodiment Including environmental noise signal and voice noise signal, environmental noise includes building site noise, vehicle noise, office's keyboard noise Include under dialogue noise and quiet environment under noisy environment Deng, voice noise other people talk with noise.It should be appreciated that non-noise Signal is the voice signal of non-above-mentioned noise signal, and being depending on non-noise signal possible be with user-related dialogic voice signal.
Preferably, the characteristic parameter of the environment voice signal for getting includes fisrt feature parameter and second feature parameter. In other feasible embodiments, the characteristic parameter of the environment voice signal for getting can be a characteristic parameter or three or The characteristic parameter that person is more than three.The embodiment of the present invention will include that fisrt feature parameter and second feature parameter are with characteristic parameter Example is illustrated.
Preferably, fisrt feature parameter is the short-time zero-crossing rate of environment voice signal;Second feature parameter is environment voice The short-time average energy of signal, i.e. fisrt feature parameter are the short-time zero-crossing rates of the whole environment voice signal for obtaining, and second is special Levy the short-time average energy that parameter is the whole environment voice signal for obtaining.Short-time zero-crossing rate represents voice signal in a frame voice Number of times of the waveform through zero level;Short-time average energy represents the energy of n moment voice signals.Wherein short-time zero-crossing rate Zn and short When average energy computing formula it is as follows:
Wherein, N is window long, short-time energy average value be a frame sample value weighted average and;Sgn [] is sign function, x During (n)≤0, sgn [x (n)]=1;x(n)<When 0, sgn [x (n)]=- 1.
It should be appreciated that in the present embodiment, the short-time zero-crossing rate and short-time average energy of noise signal and non-noise signal It is different and then can to efficiently differentiate environment voice signal be noise signal or non-noise signal.
Recognition unit 403, for according to characteristic parameter and default speech model environment-identification voice signal whether be Non- noise signal, speech model includes the scope of characteristic parameter in non-noise region.
Control unit 404, if for non-noise signal, current volume is reduced into preset value;And be additionally operable to, if It is not non-noise signal, control keeps current volume constant.
If specifically, obtain characteristic parameter in the range of characteristic parameter in non-noise region, be non-noise signal; It is noise signal if the characteristic parameter for obtaining is in the range of characteristic parameter in noise region.Due to noise signal and non-make an uproar The scope of the corresponding characteristic parameter of message number is different, and included in speech model characteristic parameter in noise region scope and The scope of characteristic parameter in non-noise region, therefore can identify that environment voice signal is noise according to default speech model Signal is also non-noise signal.It should be noted that speech model is trained according to some data samples.
Then be included in when characteristic parameter is only a parameter, in speech model in noise region and non-noise region this one The scope of individual characteristic parameter;If characteristic parameter is two or more, speech model is then included in noise region and non-makes an uproar The scope of two or more the characteristic parameter in sound region, such as in a kind of mode, first is special in non-noise region The scope of parameter is levied for (a1, a2), the scope (b1, b2) of second feature parameter, in another way, first in non-noise region The scope of characteristic parameter is (a1, a2), and when fisrt feature parameter is a1, the scope of second feature parameter is (b1, b2), now Each specific fisrt feature parameter corresponds to a scope for second feature parameter in non-noise region.
It should be appreciated that preset value can be a fixed value in certain embodiments, no matter that is, current volume is how many Current volume is reduced to a fixed value;In further embodiments, preset value can be the dynamic related to current volume Value, such as the 1/3 of current volume, the present invention are not limited specifically this.
Preferably, above-mentioned terminal 400 also includes discriminating unit 405.
Discriminating unit 405, in environment-identification voice signal whether there is default keyword, default keyword with Whether user is related, related to user for environment-identification voice.
Control unit 404, if being additionally operable to have default keyword, then is reduced to preset value by current volume;If no There is default keyword, control keeps current volume constant.
Specifically, default keyword is related to user, the name of such as user, the pet name and academic title etc., Ke Yiyong In the identity for representing user.Whether there is the default keyword can further environment-identification voice in environment-identification voice signal Whether be with it is user-related, if related to user, then reduce current volume;If not it is related to user, then keep current Volume is constant.
Preferably, default speech model includes the speech model of several different scenes, and different scenes can be expressed as In the scene of different location, such as office, family, on subway, the different time scene in same place is also denoted as, such as handled official business Different time sections in room:When go to work in the morning, during nooning, when go to work in the afternoon.And then preferably above-mentioned terminal 400 also includes:Obtain Take unit 406, matching unit 407 and searching unit 408.
Acquiring unit 406, for obtaining current time.
Matching unit 407, for being currently located field according to current time and the historical behavior information identifying user for prestoring Scape.
Searching unit 408, is currently located scene and the voice mould for matching is searched from several speech models for basis Type.
Specifically, the historical behavior information analysis user according to user is currently located scene, such as according to workaday difference Time period identifies user in office or family or way on and off duty.It should be appreciated that distinguishing the language of multiple different scenes Sound model is that, in order to more accurately identify noise signal and non-noise signal, this is based on the noise under the different scenes having There is larger difference in signal, similarly have the non-noise signal under a different scenes to there is also larger difference.
It is also understood that speech model is the data model obtained according to some data samples of collection.Data sample is The characteristic parameter gathered under noisy environment or under non-noisy environment.
Preferably, the characteristic parameter that collecting unit 402 is obtained includes fisrt feature parameter and second feature parameter.Now, 5 are please see Figure, recognition unit 403 includes:Computing unit 501, results unit 502 and judging unit 503.
Computing unit 501, for being calculated according to characteristic function formula in fisrt feature parameter and default speech model The second feature parameter of standard, characteristic function formula is the relational expression of fisrt feature parameter and second feature parameter.
Specifically, preferred fisrt feature parameter is the short-time zero-crossing rate of environment voice signal in the present embodiment;Second feature Parameter is the short-time average energy of environment voice signal.Now characteristic function formula is the pass of short-time zero-crossing rate and short-time average energy It is formula, preferably polynomial function, because short-time zero-crossing rate is bent with polynomial function with the relational expression of short-time average energy Line is similar to.
In certain embodiments, characteristic function formula is by the fisrt feature parameter and second feature in the case of some noises What the data sample of fisrt feature parameter and second feature parameter in the case of the data sample of parameter and non-noise was obtained makes an uproar The boundary function of message number and non-noise signal, boundary function is specifically the first spy of noise signal and non-noise signal boundary Levy the relational expression of function and second feature function.For example, a kind of curve map schematic diagram of characteristic function formula as shown in Figure 3 a, bent Line L1 is the coordinate curve schematic diagram of characteristic function.The region S1 of curve L1 L1 included below is expressed as first in non-noise region The scope of characteristic parameter and second feature parameter;Curve L1 area above S2 is fisrt feature parameter and the second spy in noise region Levy the scope of parameter.It should be appreciated that the division in non-noise region and noise region is divided in speech model according to characteristic function formula Rule is derived according to data sample.After the second feature parameter of standard is now calculated according to fisrt feature parameter, can be with The scope of second feature parameter acquiring according to standard to the second feature parameter in noise region and non-noise region.
In further embodiments, characteristic function formula is special by the fisrt feature parameter in the case of some noises and second Levy the relational expression or some non-noise situations of fisrt feature parameter that the data sample of parameter obtains and second feature parameter Under fisrt feature parameter and second feature parameter the fisrt feature parameter that obtains of data sample and second feature parameter pass It is formula.Now characteristic function formula be embodied as in noise signal the relational expression of fisrt feature parameter and second feature parameter or The relational expression of fisrt feature parameter and second feature parameter in non-noise signal.Divide non-according to characteristic function formula in speech model The division rule of the scope of corresponding fisrt feature parameter and second feature parameter is according to number in noise region and noise region According to sample derive, preferably division rule for keep fisrt feature functional value it is constant, second feature functional value increase or Preset value is reduced to generate the scope of second feature parameter in non-noise region and noise region.For example, as shown in Figure 3 b another A kind of curve map schematic diagram of characteristic function formula, curve L2 is the coordinate curve schematic diagram of characteristic function in non-noise scene.The When one characteristic parameter is Z1, corresponding second feature parameter is E0 in characteristic function formula, now second feature in non-noise region Parameter area is (E0- △ E2, E0+ △ E1);Now, in noise region the scope of second feature parameter be less than E0- △ E2 and More than E0+ △ E1, wherein △ E2 and △ E1 are obtained according to sample data, be can be the same or different.
Unit 502 is harvested, for obtaining in non-noise region the according to the second feature parameter and preset rules of standard The scope of two characteristic parameters.
Specifically, preset rules divide voice in representing speech model according to the second feature parameter of standard in the present embodiment The division rule in noise region and non-noise region in model.
As shown in Figure 3 a, if characteristic function formula is boundary function, the of standard is calculated according to fisrt feature parameter z1 After two characteristic parameter E0, second feature parameter E0 and division rule according to standard have got the second spy in non-noise region The scope of parameter is levied for S1 is less than or equal to E0.
For example, the curve map schematic diagram of characteristic function formula as shown in Figure 3.If shown curve L is showing for characteristic function formula It is intended to, the second feature parameter E2 of standard is calculated according to fisrt feature parameter Z1, the second feature parameter E2 of standard adds respectively Subtract default value △ E1 and △ E2 and generate the scope of the second feature parameter in non-noise region, other scopes are noise region The scope of second feature parameter, △ E1 and △ E2 can be with identical also different.
Judging unit 503, for detect obtain second feature parameter whether in non-noise region second feature parameter Scope, if in the range of second feature parameter in non-noise region, environment voice signal be non-noise signal, if not non- In noise region in the range of second feature parameter, environment voice signal is not non-noise signal.
Similarly, if characteristic parameter includes three or more than three, it is also applied for the above method and carrys out environment-identification voice letter Whether number it is non-noise signal, such as when including three characteristic parameters, two of which characteristic parameter is calculated according to preset rules Go out fourth feature parameter, retraining goes out fourth feature parameter with a remaining characteristic function formula for characteristic parameter.
When further, due to the method for performing automatic regulating volume, speech model has built up, in order to improve The accuracy of identification, to prevent speech model no longer where applicable from persistently occurring using the situation of the speech model, above-mentioned terminal 400 Also include identification unit 409, arithmetic element 410 and adjustment unit 411.
Wherein, unit 409 is identified, if the information abnormal for receiving detection, the abnormal frequency of occurrences of recognition detection is It is no to have exceeded particular value, wherein, if not less than particular value, not processed.
Arithmetic element 410, if for having exceeded particular value, being gone out currently according to calculation of characteristic parameters when detection is abnormal every time The discreet value of detection error.
Adjustment unit 411, for characteristic parameter in the non-noise region that default speech model is adjusted according to discreet value Scope.
Specifically, the abnormal information of detection is to be operated according to user and generated, for example saving volume display reminding information will Regulation volume, if user does not receive to reduce the request of volume, be considered as one-time detection it is abnormal or reduce after volume user with I.e. elevated volume is also considered as detection exception.It should be understood that detecting whether that more than particular value be in order to the event for preventing detection abnormal is even Speech model is have adjusted in the case of your event.
It should be noted that the discreet value of detection error is the gap of the result and actual result drawn according to characteristic parameter Obtain.Feature ginseng in scope and noise region further according to characteristic parameter in non-noise region in error amount regulation speech model Several scopes, makes the speech model after adjustment be applied to current scene, improves detection accuracy.
For example, when characteristic parameter includes fisrt feature parameter and second feature parameter, the preferably discreet value of detection error It is the second feature parameter and the quadratic sum of the deviation of the second feature parameter of standard of acquisition when detection is abnormal every time, such as obtains Second feature parameter and standard second feature parameter difference quadratic sum.Now, it is preferably default according to discreet value regulation Speech model non-noise region in the scope of characteristic parameter be specifically:Fisrt feature parameter and the is adjusted according to discreet value The characteristic function formula of two characteristic parameters.
It should be appreciated that detection accuracy can be improved by monitor and detection exception and dynamic regulation speech model.
It is the schematic block diagram of another terminal that embodiment of the present invention is provided referring to Fig. 6.The present embodiment as depicted In terminal 600 can include one or several processors 601, or several input units 602, or Several output devices 603 and memory 604.Above-mentioned processor 601, input unit 602, output device 603 and storage Device 604 is connected by bus.
Input unit 602 is used to receive the information of input.In implementing, the input unit 602 of the embodiment of the present invention can Including keyboard, mouse, light device of electrical input, acoustic input dephonoprojectoscope, touch input unit, scanner, microphone etc..
Output device 603 is used for external output information to user.In implementing, the output device of the embodiment of the present invention 603 may include display, loudspeaker, printer etc..
Memory 604 is used to store the routine data with various functions, in implementing, the storage of the embodiment of the present invention Device 604 can be system storage, such as, and volatile (such as RAM), and non-volatile (such as ROM, flash memory etc.), or Both combinations.In implementing, the memory 604 of the embodiment of the present invention can also be the external memory storage outside system, than Such as, disk, CD, tape etc..
Processor 601 is used for the instruction for calling the routine data stored in memory 604 to perform the storage of memory 604, And perform following operation:
The current environment voice signal of detection;The characteristic parameter of environment voice signal is obtained, characteristic parameter is used for identification ring Whether border voice is non-noise signal;Whether it is non-according to characteristic parameter and default speech model environment-identification voice signal Noise signal, speech model includes the scope of characteristic parameter in non-noise region, if non-noise signal, by current volume decrease As little as preset value, if not non-noise signal, keeps current volume constant.
Preferably, before current volume is reduced to preset value by processor 601, and following operation is specifically performed:
Whether there is default keyword in environment-identification voice signal, default keyword is related to user, for knowing Whether other environment voice is related to user;And if there is default keyword, then perform current volume is reduced to it is default Value;If not existing default keyword, keep current volume constant.
Preferably, processor 601 is according to the characteristic parameter and default speech model environment-identification voice signal It is no for non-noise signal before, and specifically perform following operation:
Obtain current time;Scene is currently located according to current time and the historical behavior information identifying user for prestoring; And the speech model that matches is searched from several speech models according to scene is currently located.
Preferably, the characteristic parameter for being obtained in S102 includes fisrt feature parameter and second feature parameter, processor 601 When according to characteristic parameter and default speech model environment-identification voice signal whether being non-noise signal, and specifically perform as follows Operation:
The second feature ginseng of standard is calculated according to characteristic function formula in fisrt feature parameter and default speech model Number, characteristic function formula is the relational expression of fisrt feature parameter and second feature parameter;Second feature parameter according to standard and Preset rules obtain the scope of second feature parameter in non-noise region;Detect the second feature parameter for obtaining whether in non-noise The scope of second feature parameter in region, if in the range of second feature parameter in non-noise region, environment voice signal is Non- noise signal, if not in the range of second feature parameter in non-noise region, environment voice signal is not non-noise signal.
When further, due to the method for performing automatic regulating volume, speech model has built up, in order to improve The accuracy of identification, to prevent speech model no longer where applicable from persistently occurring using the situation of the speech model, processor 601 is also Perform following operation:
If receiving the abnormal information of detection, whether the abnormal frequency of occurrences of recognition detection has exceeded particular value;
If having exceeded particular value, estimating for current detection error according to calculation of characteristic parameters when detection is abnormal every time Value;And the scope of characteristic parameter in the non-noise region of default speech model is adjusted according to discreet value;
If not less than particular value, not processed.
It should be appreciated that in embodiments of the present invention, alleged processor 601 can be central first processing units (Central Processing Unit, CPU), the processor 601 can also be other general processors, digital signal processor (Digital Signal Processor, DSP), application specific integrated circuit (Application Specific Integrated Circuit, ASIC), ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other FPGAs Device, discrete gate or transistor logic, discrete hardware components etc..General processor can be microprocessor or this at Reason device can also be any conventional processor etc..
It should be noted that the step in present invention method can according to actual needs carry out order adjustment, close And and delete.
Unit in embodiment of the present invention terminal can according to actual needs be merged, divides and deleted.
It is apparent to those skilled in the art that, for convenience of description and succinctly, the end of foregoing description End and the specific work process of unit, may be referred to the corresponding process in preceding method embodiment, will not be repeated here.In this Shen In the several embodiments for please being provided, it should be understood that disclosed terminal and method, can realize by another way. For example, device embodiment described above is only schematical, for example, the division of the unit, only a kind of logic Function is divided, and can have other dividing mode when actually realizing, such as several units or component can combine or can be with Another system is integrated into, or some features can be ignored, or do not perform.In addition, shown or discussed coupling each other Conjunction or direct-coupling or communication connection can be INDIRECT COUPLING or the communication connection by some interfaces, device or unit, also may be used To be electric, machinery or other forms are connected.
In addition, during each functional unit in each embodiment of the invention can be integrated in a first processing units, Can be that unit is individually physically present, or two or more units are integrated in a unit.Above-mentioned collection Into unit can both be realized in the form of hardware, it would however also be possible to employ the form of SFU software functional unit is realized.It is described integrated If unit is to realize in the form of SFU software functional unit and as independent production marketing or when using, can store in a meter In calculation machine read/write memory medium.Based on such understanding, technical scheme is substantially done to prior art in other words Go out the part of contribution, or all or part of the technical scheme can be embodied in the form of software product, the computer Software product is stored in a storage medium, including some instructions are used to so that a computer equipment (can be personal meter Calculation machine, server, or network equipment etc.) perform all or part of step of each embodiment methods described of the invention.And it is preceding The storage medium stated includes:USB flash disk, mobile hard disk, read-only storage (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disc or CD etc. are various can be with the medium of store program codes.
The above, is specific embodiment of the invention, but protection scope of the present invention is not limited thereto, any ripe Know those skilled in the art the invention discloses technical scope in, various equivalent modifications can be readily occurred in or replaced Change, these modifications or replacement should all be included within the scope of the present invention.Therefore, protection scope of the present invention should be with right It is required that protection domain be defined.

Claims (10)

1. a kind of method of automatic regulating volume, it is characterised in that including:
The current environment voice signal of detection;
The characteristic parameter of the environment voice signal is obtained, whether the characteristic parameter is non-making an uproar for environment-identification voice signal Message number;
Whether it is non-noise signal, institute's predicate according to the characteristic parameter and default speech model environment-identification voice signal Sound model includes the scope of characteristic parameter in non-noise region;
If non-noise signal, current volume is reduced to preset value.
2. method according to claim 1, it is characterised in that the characteristic parameter includes fisrt feature parameter and second feature Whether parameter, be non-noise signal according to the characteristic parameter and default speech model environment-identification voice signal, including:
The second feature ginseng of standard is calculated according to characteristic function formula in the fisrt feature parameter and default speech model Number, the characteristic function formula is the relational expression of fisrt feature parameter and second feature parameter, the second feature parameter of the standard Scope for obtaining second feature parameter in noise region and non-noise region in the speech model;
Second feature parameter and preset rules according to standard obtain the scope of second feature parameter in non-noise region;
Detect obtain the second feature parameter whether in non-noise region second feature parameter scope;
If in the range of second feature parameter in non-noise region, the environment voice signal is non-noise signal.
3. method according to claim 1, it is characterised in that methods described also includes:
If receiving the abnormal information of detection, whether the abnormal frequency of occurrences of recognition detection has exceeded particular value;
If having exceeded particular value, estimating for current detection error according to calculation of characteristic parameters when detection is abnormal every time Value;
The scope of characteristic parameter in the non-noise region of the default speech model is adjusted according to the discreet value.
4. method according to claim 1, it is characterised in that the default speech model includes several different scenes Speech model, whether described be non-noise according to the characteristic parameter and default speech model environment-identification voice signal Before signal, including:
Obtain current time;
Scene is currently located according to the current time and the historical behavior information identifying user for prestoring;
The speech model that matches is searched from several speech models according to the scene that is currently located.
5. method according to claim 1, it is characterised in that before current volume is reduced into preset value, the side Method also includes:
Whether there is default keyword in environment-identification voice signal, the default keyword is related to user, for knowing Whether other environment voice is related to user;
If there is default keyword, then perform and described current volume is reduced to preset value.
6. a kind of terminal, it is characterised in that the terminal includes:
Detection unit, for detecting current environment voice signal;
Collecting unit, the characteristic parameter for obtaining the environment voice signal, the characteristic parameter is used for environment-identification voice Whether signal is non-noise signal;
Recognition unit, for whether being non-making an uproar according to the characteristic parameter and default speech model environment-identification voice signal Message number, the speech model includes the scope of characteristic parameter in non-noise region;
Control unit, if for non-noise signal, current volume is reduced into preset value.
7. terminal according to claim 6, it is characterised in that the characteristic parameter includes fisrt feature parameter and second special Levy parameter, the recognition unit, including:
Computing unit, for calculating standard according to characteristic function formula in the fisrt feature parameter and default speech model Second feature parameter, the characteristic function formula is the relational expression of fisrt feature parameter and second feature parameter, the standard Second feature parameter is used to obtain the scope of second feature parameter in noise region and non-noise region in the speech model;
Unit is harvested, for according to second feature ginseng in the non-noise region of the second feature parameter and preset rules of standard acquisition Several scopes;
Judging unit, for detect obtain the second feature parameter whether in non-noise region second feature parameter model Enclose, if in the range of second feature parameter in non-noise region, the environment voice signal is non-noise signal.
8. terminal according to claim 6, it is characterised in that the terminal also includes:
Identification unit, if the information abnormal for receiving detection, it is specific whether the abnormal frequency of occurrences of recognition detection has exceeded Value;
Arithmetic element, if for having exceeded particular value, current inspection is gone out according to calculation of characteristic parameters when detection is abnormal every time Survey the discreet value of error;
Adjustment unit, for characteristic parameter in the non-noise region that the default speech model is adjusted according to the discreet value Scope.
9. terminal according to claim 6, it is characterised in that the default speech model includes several different scenes Speech model, the terminal also includes:
Acquiring unit, for obtaining current time;
Matching unit, for being currently located scene according to the current time and the historical behavior information identifying user for prestoring;
Searching unit, the speech model for matching is searched for being currently located scene according to from several speech models.
10. terminal according to claim 6, it is characterised in that the terminal also includes discriminating unit,
The discriminating unit, for whether there is default keyword, the default keyword in environment-identification voice signal It is related to user, it is whether related to user for environment-identification voice;If there is default keyword, described control unit ought Preceding volume is reduced to preset value.
CN201710124718.5A 2017-03-03 2017-03-03 The method and terminal of a kind of automatic regulating volume Withdrawn CN106936991A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710124718.5A CN106936991A (en) 2017-03-03 2017-03-03 The method and terminal of a kind of automatic regulating volume

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710124718.5A CN106936991A (en) 2017-03-03 2017-03-03 The method and terminal of a kind of automatic regulating volume

Publications (1)

Publication Number Publication Date
CN106936991A true CN106936991A (en) 2017-07-07

Family

ID=59424072

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710124718.5A Withdrawn CN106936991A (en) 2017-03-03 2017-03-03 The method and terminal of a kind of automatic regulating volume

Country Status (1)

Country Link
CN (1) CN106936991A (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107484000A (en) * 2017-09-29 2017-12-15 北京奇艺世纪科技有限公司 A kind of volume adjusting method of terminal, device and voice remote controller
CN107728990A (en) * 2017-09-30 2018-02-23 努比亚技术有限公司 A kind of audio frequency playing method, mobile terminal and computer-readable recording medium
CN108064007A (en) * 2017-11-07 2018-05-22 苏宁云商集团股份有限公司 Know method for distinguishing and microcontroller and intelligent sound box for the enhancing voice of intelligent sound box
CN108494975A (en) * 2018-06-27 2018-09-04 上海连尚网络科技有限公司 Incoming call response method and equipment
CN109408025A (en) * 2018-10-25 2019-03-01 北京小米移动软件有限公司 Audio frequency playing method, device and storage medium
CN110428835A (en) * 2019-08-22 2019-11-08 深圳市优必选科技股份有限公司 Voice equipment adjusting method and device, storage medium and voice equipment
WO2020224126A1 (en) * 2019-05-06 2020-11-12 平安科技(深圳)有限公司 Facial recognition-based adaptive adjustment method, system and readable storage medium
CN112887857A (en) * 2021-01-25 2021-06-01 湖南普奇水环境研究院有限公司 Hearing protection method and system for eliminating reception noise
CN112992153A (en) * 2021-04-27 2021-06-18 太平金融科技服务(上海)有限公司 Audio processing method, voiceprint recognition device and computer equipment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101227759A (en) * 2007-01-15 2008-07-23 上海杰得微电子有限公司 Earphone device with automatic volume control function
CN101415045A (en) * 2007-10-17 2009-04-22 北京三星通信技术研究有限公司 Method and apparatus for implementing intelligent automatic level control in communication network
CN102915753A (en) * 2012-10-23 2013-02-06 华为终端有限公司 Method for intelligently controlling volume of electronic device and implementation device of method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101227759A (en) * 2007-01-15 2008-07-23 上海杰得微电子有限公司 Earphone device with automatic volume control function
CN101415045A (en) * 2007-10-17 2009-04-22 北京三星通信技术研究有限公司 Method and apparatus for implementing intelligent automatic level control in communication network
CN102915753A (en) * 2012-10-23 2013-02-06 华为终端有限公司 Method for intelligently controlling volume of electronic device and implementation device of method

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107484000A (en) * 2017-09-29 2017-12-15 北京奇艺世纪科技有限公司 A kind of volume adjusting method of terminal, device and voice remote controller
CN107728990A (en) * 2017-09-30 2018-02-23 努比亚技术有限公司 A kind of audio frequency playing method, mobile terminal and computer-readable recording medium
CN107728990B (en) * 2017-09-30 2020-09-01 苏州红森软件有限公司 Audio playing method, mobile terminal and computer readable storage medium
CN108064007A (en) * 2017-11-07 2018-05-22 苏宁云商集团股份有限公司 Know method for distinguishing and microcontroller and intelligent sound box for the enhancing voice of intelligent sound box
CN108494975A (en) * 2018-06-27 2018-09-04 上海连尚网络科技有限公司 Incoming call response method and equipment
CN109408025A (en) * 2018-10-25 2019-03-01 北京小米移动软件有限公司 Audio frequency playing method, device and storage medium
CN109408025B (en) * 2018-10-25 2021-10-22 北京小米移动软件有限公司 Audio playing method, device and storage medium
WO2020224126A1 (en) * 2019-05-06 2020-11-12 平安科技(深圳)有限公司 Facial recognition-based adaptive adjustment method, system and readable storage medium
CN110428835A (en) * 2019-08-22 2019-11-08 深圳市优必选科技股份有限公司 Voice equipment adjusting method and device, storage medium and voice equipment
CN112887857A (en) * 2021-01-25 2021-06-01 湖南普奇水环境研究院有限公司 Hearing protection method and system for eliminating reception noise
CN112992153A (en) * 2021-04-27 2021-06-18 太平金融科技服务(上海)有限公司 Audio processing method, voiceprint recognition device and computer equipment
CN112992153B (en) * 2021-04-27 2021-08-17 太平金融科技服务(上海)有限公司 Audio processing method, voiceprint recognition device and computer equipment

Similar Documents

Publication Publication Date Title
CN106936991A (en) The method and terminal of a kind of automatic regulating volume
CN107731223B (en) Voice activity detection method, related device and equipment
WO2021208287A1 (en) Voice activity detection method and apparatus for emotion recognition, electronic device, and storage medium
CN109087670B (en) Emotion analysis method, system, server and storage medium
CN107644642B (en) Semantic recognition method and device, storage medium and electronic equipment
WO2018232233A1 (en) Data update method, client, and electronic device
CN107610698A (en) A kind of method for realizing Voice command, robot and computer-readable recording medium
CN103971681A (en) Voice recognition method and system
CN107193527A (en) A kind of audio control method, terminal and computer-readable recording medium
CN113707173B (en) Voice separation method, device, equipment and storage medium based on audio segmentation
CN112652306A (en) Voice wake-up method and device, computer equipment and storage medium
CN110246502A (en) Voice noise reduction method and device and terminal equipment
CN108924343A (en) Control method of electronic device, device, storage medium and electronic equipment
CN112951243A (en) Voice awakening method, device, chip, electronic equipment and storage medium
CN113593597A (en) Voice noise filtering method and device, electronic equipment and medium
CN113223554A (en) Wind noise detection method, device, equipment and storage medium
CN112669822A (en) Audio processing method and device, electronic equipment and storage medium
CN112669837B (en) Awakening method and device of intelligent terminal and electronic equipment
CN112581937A (en) Method and device for acquiring voice instruction
CN113658586A (en) Training method of voice recognition model, voice interaction method and device
CN111755029B (en) Voice processing method, device, storage medium and electronic equipment
CN112289311A (en) Voice wake-up method and device, electronic equipment and storage medium
CN105788590A (en) Speech recognition method, device, mobile terminal
CN114121022A (en) Voice wake-up method and device, electronic equipment and storage medium
CN114399992A (en) Voice instruction response method, device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication

Application publication date: 20170707

WW01 Invention patent application withdrawn after publication