WO2012137617A1 - 符号化方法、復号方法、符号化装置、復号装置、プログラム、記録媒体 - Google Patents
符号化方法、復号方法、符号化装置、復号装置、プログラム、記録媒体 Download PDFInfo
- Publication number
- WO2012137617A1 WO2012137617A1 PCT/JP2012/057685 JP2012057685W WO2012137617A1 WO 2012137617 A1 WO2012137617 A1 WO 2012137617A1 JP 2012057685 W JP2012057685 W JP 2012057685W WO 2012137617 A1 WO2012137617 A1 WO 2012137617A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- value
- error
- encoding
- decoding
- sample
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 88
- 238000001228 spectrum Methods 0.000 claims description 105
- 238000004364 calculation method Methods 0.000 claims description 30
- 238000013139 quantization Methods 0.000 abstract description 184
- 230000008569 process Effects 0.000 abstract description 31
- 239000013598 vector Substances 0.000 description 22
- 230000006870 function Effects 0.000 description 14
- 238000010606 normalization Methods 0.000 description 14
- 238000009499 grossing Methods 0.000 description 7
- 238000004458 analytical method Methods 0.000 description 6
- 238000012935 Averaging Methods 0.000 description 5
- 241000209094 Oryza Species 0.000 description 5
- 235000007164 Oryza sativa Nutrition 0.000 description 5
- 230000008707 rearrangement Effects 0.000 description 5
- 235000009566 rice Nutrition 0.000 description 5
- 238000006243 chemical reaction Methods 0.000 description 4
- 238000007906 compression Methods 0.000 description 4
- 230000006835 compression Effects 0.000 description 4
- 230000005484 gravity Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- 230000003595 spectral effect Effects 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000001174 ascending effect Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 238000013144 data compression Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000005315 distribution function Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/005—Correction of errors induced by the transmission channel, if related to the coding algorithm
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/008—Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/02—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
- G10L19/032—Quantisation or dequantisation of spectral components
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/16—Vocoder architecture
- G10L19/167—Audio streaming, i.e. formatting and decoding of an encoded audio signal representation into a data stream for transmission or storage purposes
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/16—Vocoder architecture
- G10L19/18—Vocoders using multiple modes
- G10L19/24—Variable rate codecs, e.g. for generating different qualities using a scalable representation such as hierarchical encoding or layered encoding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/038—Speech enhancement, e.g. noise reduction or echo cancellation using band spreading techniques
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/0017—Lossless audio signal coding; Perfect reconstruction of coded audio signal by transmission of coding error
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/02—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
- G10L19/0212—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders using orthogonal transformation
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/02—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
- G10L19/032—Quantisation or dequantisation of spectral components
- G10L19/035—Scalar quantisation
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/02—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
- G10L19/032—Quantisation or dequantisation of spectral components
- G10L19/038—Vector quantisation, e.g. TwinVQ audio
Definitions
- the present invention relates to an encoding technique for an acoustic signal and a decoding technique for a code string obtained by the encoding technique. More specifically, the present invention relates to encoding and decoding of a frequency domain sample sequence obtained by converting an acoustic signal into the frequency domain.
- Adaptive coding for orthogonal transform coefficients such as DFT (Discrete Fourier Transform) and MDCT (Modified Discrete Cosine Transform) is known as a coding method for low-bit (for example, about 10 kbit / s to 20 kbit / s) speech and acoustic signals. It has been.
- AMR-WB + Extended-Adaptive-Multi-Rate-Wideband
- TCX transform-coded-excitation
- coding based on TCX such as AMR-WB +
- TCX time domain coefficient based on periodicity
- coding with a large amount of variation reduces the coding efficiency.
- a series of MDCT coefficients which are discrete values obtained by quantizing the signal divided by gain, arranged from the lowest frequency is entropy such as arithmetic code.
- a short code is assigned if the amplitude is small, and a long code is assigned if the amplitude is large.
- the number of bits per frame is reduced on average, but if the number of assigned bits per frame is fixed, the reduced bits may not be used effectively.
- an object of the present invention is to provide an encoding / decoding technique that improves the quality of discrete signals, particularly audio-acoustic digital signals, by encoding with low bits with a low amount of computation. To do.
- An encoding method is an encoding method for encoding a frequency domain sample sequence derived from an acoustic signal in a predetermined time interval with a predetermined number of bits, and each of the frequency domain sample sequences
- An encoding step for generating a variable-length code by encoding an integer value corresponding to a sample value by variable-length coding, and an integer value corresponding to the value of each sample from the value of each sample in the frequency domain sample sequence Using the error calculation step of calculating the subtracted error value sequence and the surplus bits that are the number of bits obtained by subtracting the number of bits of the variable length code from the predetermined number, the error value sequence is encoded and the error code is obtained. Generating an error encoding step.
- a decoding method is a decoding method for decoding a code including a predetermined number of input bits, and generates a string of integer values by decoding a variable length code included in the code.
- the block diagram for demonstrating the structure of the encoding apparatus of embodiment The flowchart for demonstrating the process of the encoding apparatus of embodiment.
- This embodiment uses a predetermined time interval as a frame, and in the framework of quantizing a frequency domain sample sequence derived from an acoustic signal in the frame, the variable length of the sequence after weighted flattening of the frequency domain samples
- coding is performed and the distortion is reduced by quantizing the error signal by deciding the priority using the surplus bits saved by the variable length coding. I will. In particular, even when the number of assigned bits per frame is fixed, the advantage of variable length coding can be utilized.
- a sample sequence in the frequency domain derived from the acoustic signal in other words, as a sample sequence in the frequency domain based on the acoustic signal, for example, a DFT coefficient sequence obtained by converting the audio acoustic digital signal in frame units from the time domain to the frequency domain,
- the MDCT coefficient sequence include a coefficient sequence in which processing such as normalization, weighting, and quantization is applied to such a coefficient sequence.
- processing such as normalization, weighting, and quantization is applied to such a coefficient sequence.
- an embodiment will be described using an MDCT coefficient sequence as an example.
- the encoding device 1 includes a frequency domain transform unit 11, a linear prediction analysis unit 12, a linear prediction coefficient quantization coding unit 13, a power spectrum envelope calculation unit 14, a weighted envelope normalization unit 15, a normal
- a quantization gain calculation unit 16 a quantization unit 17, an error calculation unit 18, an encoding unit 19, and an error encoding unit 110 are provided.
- the encoding device 1 performs each process of the encoding method illustrated in FIG. Hereinafter, each process of the encoding device 1 will be described.
- Frequency domain transform unit 11 First, the frequency domain conversion unit 11 converts the audio-acoustic digital signal into N-point MDCT coefficient sequences in the frequency domain in units of frames (step S11).
- the encoding side quantizes the MDCT coefficient sequence, encodes the quantized MDCT coefficient sequence, transmits the obtained code sequence to the decoding side, and the decoding side quantizes the code sequence.
- the MDCT coefficient sequence can be reconstructed, and the time-domain audio-acoustic digital signal can be reconstructed by inverse MDCT transformation.
- the amplitude of the MDCT coefficient has approximately the same amplitude envelope (power spectrum envelope) as the power spectrum of a normal DFT. For this reason, by assigning information proportional to the logarithmic value of the amplitude envelope, the quantization distortion (quantization error) of the MDCT coefficients in all bands can be uniformly distributed, and the overall quantization distortion can be reduced. In addition, information compression is also realized. Note that the power spectrum envelope can be efficiently estimated using a linear prediction coefficient obtained by linear prediction analysis.
- a method for controlling such quantization error a method of adaptively assigning quantization bits of each MDCT coefficient (adjusting the quantization step width after flattening the amplitude), or weighted vector quantization is used. There is a method of adaptively weighting and determining a code.
- the quantization method performed in the embodiment of the present invention will be described, it should be noted that the present invention is not limited to the quantization method described.
- Linear prediction analysis unit 12 The linear prediction analysis unit 12 performs linear prediction analysis on the audio-acoustic digital signal in units of frames, and obtains and outputs linear prediction coefficients up to a predetermined order (step S12).
- Linear prediction coefficient quantization coding unit 13 calculates and outputs a code corresponding to the linear prediction coefficient obtained by the linear prediction analysis unit 12 and a quantized linear prediction coefficient (step S13). At that time, it is possible to convert the linear prediction coefficient to LSP (Line Spectral Pairs), obtain the code corresponding to the LSP and the quantized LSP, and convert the quantized LSP to the quantized linear prediction coefficient. Good.
- LSP Line Spectral Pairs
- the linear prediction coefficient code that is a code corresponding to the linear prediction coefficient is a part of the code transmitted to the decoding device 2.
- the power spectrum envelope calculation unit 14 obtains a power spectrum envelope by converting the quantized linear prediction coefficient output from the linear prediction coefficient quantization coding unit 13 into the frequency domain (step S14). The obtained power spectrum envelope is transmitted to the weighted envelope normalization unit 15. Further, as necessary, the error is transmitted to the error encoding unit 110 as indicated by a broken line in FIG.
- Each coefficient W (1),..., W (N) of the power spectrum envelope coefficient sequence corresponding to each coefficient X (1),..., X (N) of the N-point MDCT coefficient sequence is quantized.
- the obtained linear prediction coefficient can be obtained by converting it to the frequency domain.
- the time signal y (t) at the time t is a past value y (t ⁇ 1),..., Y ( tp), the prediction residual e (t), and the quantized linear prediction coefficients ⁇ 1 ,..., ⁇ p are expressed by Equation (1).
- each coefficient W (n) [1 ⁇ n ⁇ N] of the power spectrum envelope coefficient sequence is expressed by Expression (2).
- exp ( ⁇ ) is an exponential function with the Napier number as the base
- j is an imaginary unit
- ⁇ 2 is the predicted residual energy.
- the order p may be the same as the order of the quantized linear prediction coefficient output by the linear prediction coefficient quantization encoding unit 13 or the quantum output by the linear prediction coefficient quantization encoding unit 13. It may be less than the order of the normalized linear prediction coefficient.
- the power spectrum envelope calculation unit 14 may calculate an approximate value of the power spectrum envelope value or an estimated value of the power spectrum envelope value instead of the power spectrum envelope value.
- the power spectrum envelope value is each coefficient W (1),..., W (N) of the power spectrum envelope coefficient sequence.
- Weighting envelope normalization unit 15 normalizes each coefficient of the MDCT coefficient sequence based on the power spectrum envelope output from the power spectrum envelope calculation unit 14 (step S15).
- the weighted envelope normalization unit 15 calculates a weighted spectrum envelope coefficient obtained by smoothing a power spectrum envelope value series and a square root series thereof in the frequency direction. Use each frame to normalize each coefficient in the MDCT coefficient sequence.
- each coefficient x (1),..., X (N) of the weighted normalized MDCT coefficient sequence in units of frames is obtained.
- the weighted normalized MDCT coefficient sequence is transmitted to the normalized gain calculation unit 16, the quantization unit 17, and the error calculation unit 18.
- the weighted normalized MDCT coefficient sequence generally has a slightly larger amplitude in the low frequency region and has a fine structure due to the pitch period, but does not have as large an amplitude gradient and amplitude asperity as the original MDCT coefficient sequence.
- the normalized gain calculation unit 16 can quantize each coefficient x (1),..., X (N) of the weighted normalized MDCT coefficient sequence with a given total number of bits for each frame.
- a quantization step width is determined using the sum of amplitude values or energy values over all frequencies, and a coefficient g (hereinafter referred to as gain) for dividing each coefficient of the weighted normalized MDCT coefficient sequence so as to be the quantization step width.
- the gain information that is information representing the gain is a part of the code transmitted to the decoding device 2.
- the quantization unit 17 quantizes each coefficient x (1),..., X (N) of the weighted normalized MDCT coefficient sequence for each frame with the quantization step width determined in the process of step S16. (Step S17). That is, an integer value obtained by rounding off the decimals of the value of x (n) / g obtained by dividing each coefficient x (n) [1 ⁇ n ⁇ N] of the weighted normalized MDCT coefficient sequence by the gain g Let u (n) be a quantized MDCT coefficient.
- the quantized MDCT coefficient sequence for each frame is transmitted to the error calculation unit 18 and the encoding unit 19.
- a value obtained by rounding up or down the decimal point of the value of x (n) / g may be an integer value u (n).
- the integer value u (n) may be a value corresponding to the value of x (n) / g.
- the sequence x (n) / g corresponds to the sample sequence in the frequency domain in the claims.
- the x (n) / g sequence is an example of a sample sequence in the frequency domain.
- the quantized MDCT coefficient that is an integer value u (n) corresponds to an integer value corresponding to the value of each sample in the frequency domain sample sequence.
- Error calculation unit 18 The weighted normalized MDCT sequence obtained in the process of step S15, the gain g obtained in the process of step S16, and the quantized MDCT coefficient sequence in units of frames obtained in the process of step S17 are error calculation units 18. Will be input.
- the quantization error r (n) corresponding to (n) is assumed.
- the column of quantization error r (n) corresponds to the column of error values in the claims.
- the encoding unit 19 encodes the quantized MDCT coefficient sequence (sequence of quantized MDCT coefficients u (n)) output from the quantizing unit 17 for each frame, and the obtained code and the bit of the code The number is output (step S19).
- the encoding unit 19 can reduce the average code amount by, for example, variable length encoding to which a code having a length corresponding to the frequency of the value of the quantized MDCT coefficient sequence is assigned.
- variable length code include a Rice code, a Huffman code, an arithmetic code, and a run length code.
- the generated variable length code becomes a part of the code transmitted to the decoding device 2. What variable-length encoding method is executed is specified by the selection information. This selection information may be transmitted to the decoding device 2.
- the surplus bits that are not used for encoding the quantization error r (n) are used for other purposes, for example, for correcting the gain g. Since the quantization error r (n) is an error in rounding due to quantization, it is distributed almost uniformly from -0.5 to +0.5.
- the error encoding unit 110 calculates the number of surplus bits by subtracting the number of bits of the variable length code output from the encoding unit 19 from the number of bits set in advance as the code amount of the weighted normalized MDCT coefficient sequence. .
- the quantization error sequence obtained by the error calculation unit 18 is encoded with the number of surplus bits, and the obtained error code is output (step S110). This error code is a part of the code transmitted to the decoding device 2.
- Example 1 The operation of Example 1 is as follows.
- a code book for each possible value of the number of surplus bits is stored in advance.
- a vector having the same number of samples as the number of quantization error sequences that can be expressed by the number of surplus bits corresponding to each codebook, and a code corresponding to the vector are stored in advance. Yes.
- error encoding section 110 selects a codebook corresponding to the calculated number of surplus bits from the codebook stored in the codebook storage section, and uses the selected codebook to generate a vector Perform quantization.
- the encoding process after selecting a codebook is the same as general vector quantization. That is, a code corresponding to a vector that minimizes the distance between each vector of the selected codebook and the input quantization error sequence or maximizes their correlation is output as an error code.
- the vector stored in the codebook has the same number of samples as the quantization error sequence, but the number of vector samples stored in the codebook is 1 / integer of the quantization error sequence.
- the quantization error sequence may be vector quantized for each of a plurality of portions, and a plurality of codes obtained may be used as error codes.
- the power spectrum envelope value can be referred to.
- an approximate value of the power spectrum envelope value, an estimated value of the power spectrum envelope value, a value obtained by smoothing any of these values in the frequency direction, and a plurality of samples of any of these values Although an average value of the above and a value having the same magnitude relationship as at least one of these values can be referred to, only the case where the power spectrum envelope value is used will be described below.
- the amplitude tendency of the sample sequence in the frequency domain to be quantized corresponding to the spectrum envelope after weighted flattening in FIG.
- the weighted normalized MDCT coefficient x (n) is a very small value, that is, a value smaller than half the step width
- the weighted normalized MDCT coefficient x (n) is divided by the gain g.
- the result is 0, and the quantization error r (n) is much smaller than 0.5. That is, when the power spectrum envelope value is small to some extent, the effect on the auditory quality is small when the quantization error r (n) is encoded in addition to the weighted normalized MDCT coefficient x (n). You may exclude from the encoding object in the part 110. FIG. If the power spectrum envelope value is large to some extent, it is not known which quantization error of the sample is large.
- the quantization error sample r (n) is encoded by 1 bit for each surplus bit in descending order. Moreover, it is sufficient to exclude the case where the power spectrum envelope value is below a certain level.
- f (x) is a probability distribution function
- ⁇ is an absolute value of the reconstructed value in the decoding device.
- a quantization error sample to be encoded is selected, and the selected quantization error is selected.
- the position of the sample in the plurality of quantization error samples and the value of the selected quantization error sample may be encoded and transmitted to the decoding device 2 as an error code.
- a quantization error sample having the largest absolute value is selected from four quantization error samples in which the value of the corresponding quantization MDCT coefficient u (n) is 0, and the selected quantization error is selected.
- the value of the sample is quantized (for example, whether it is + or-) and the information is transmitted in 1 bit, and the position of the selected quantization error sample is transmitted in 2 bits. Since the code is not sent to the decoding device 2 for the quantization error samples that are not selected, the decoding value in the decoding device 2 is set to 0. In general, q bits are required to inform the decoding device of which position of 2 q samples is selected.
- [Mu] at this time may be the value of the center of gravity of the distribution of only the sample having the largest absolute value of the quantization error value in units of multiple samples.
- a sparse sample can be expressed by a combination of a plurality of sequences as shown in FIG.
- the first series only one of the four positions (two bits in the designation) is pulsed with + or-, and the other positions can be zero. That is, 3 bits are required to represent the first stream.
- the second sequence and the fifth sequence can be encoded with a total of 15 bits.
- U is the number of surplus bits
- T is the number of quantization error samples of the number of quantization error samples constituting the quantization error sequence
- the corresponding quantization MDCT coefficient u (n) is not 0, and the corresponding quantum
- encoding can be performed in the following procedure.
- the error encoding unit 110 corresponds to the corresponding power among the T quantization error samples in which the value of the corresponding quantization MDCT coefficient u (n) is not 0 in the quantization error sequence.
- U bits having a large spectral envelope value are selected, and a 1-bit code, which is information representing the sign of the quantization error sample, is generated for each selected quantization error sample. Is output as an error code. If the corresponding power spectrum envelope values are the same, for example, select according to a predetermined rule such as selecting a quantization error sample with a smaller position on the frequency axis (a quantization error sample with a lower frequency). To do.
- the error encoding unit 110 includes each of T quantization error samples in which the value of the corresponding quantization MDCT coefficient u (n) in the quantization error sequence is not 0. A 1-bit code that is information representing the positive / negative of the quantization error sample is generated.
- the error encoding unit 110 also encodes a quantization error sample in which the value of the corresponding quantization MDCT coefficient u (n) in the quantization error sequence is 0 with U-T bits.
- a quantization error sample in which the value of the corresponding quantization MDCT coefficient u (n) in the quantization error sequence is 0 with U-T bits.
- a plurality of quantization error samples having a corresponding power spectrum envelope value out of quantization error samples having a corresponding quantization MDCT coefficient u (n) value of 0 are extracted, and a vector is obtained for each of the plurality of quantization error samples.
- the error encoding unit 110 further outputs a combination of the generated U-bit code and U-T bit code as an error code.
- the error encoding unit 110 When T + S ⁇ U The error encoding unit 110 generates a 1-bit first-round code representing the positive / negative of the quantization error sample for each of all the quantization error samples included in the quantization error sequence. .
- the error encoding unit 110 further encodes the quantization error sample using the remaining U- (T + S) bits in the procedure (A) or (B).
- U- (T + S) is set as a new U
- the second round (A) is executed for the first round coding error. That is, as a result, at least some quantization error samples are quantized by 2 bits per quantization error sample.
- the value of the quantization error r (n) was uniform within the range of ⁇ 0.5 to +0.5, but the first round error value to be encoded in the second round. Is in the range of -0.25 to +0.25.
- the error encoding unit 110 among the quantization error samples constituting the quantization error sequence, the value of the corresponding quantization MDCT coefficient u (n) is not 0, and the quantization error r ( For a quantization error sample in which the value of n) is positive, a 1-bit second-round code representing the sign of the value obtained by subtracting the reconstruction value 0.25 from the quantization error sample value is used. Generate.
- the error encoding unit 110 has a value of the corresponding quantization MDCT coefficient u (n) that is not 0 among the error samples constituting the quantization error sequence, and the value of the quantization error r (n) is For a quantization error sample that is negative, a 1-bit second-round code representing the positive or negative value is generated for the value obtained by subtracting the reconstruction value ⁇ 0.25 from the quantization error sample value.
- the error encoding unit 110 has a value of the corresponding quantization MDCT coefficient u (n) of 0 among the error samples constituting the quantization error sequence and a value of the quantization error r (n).
- the value obtained by subtracting the reconstruction value A (A is a predetermined positive value smaller than 0.25) from the quantization error sample value is positive or negative. Is generated as a 1-bit second-round code.
- the error encoding unit 110 has a corresponding quantization MDCT coefficient u (n) value of 0 among the error samples constituting the quantization error sequence and a quantization error r (n) value of 0.
- u (n) the value obtained by subtracting the reconstruction value ⁇ A (A is a predetermined positive value smaller than 0.25) from the value of the quantization error sample.
- a 1-bit second-round code representing positive and negative is generated.
- the error encoding unit 110 outputs a combination of the generated first and second cycle codes as an error code.
- the quantization error sequence is encoded with fewer UU bits than U bits, so the condition of (C) may be T + S ⁇ UU.
- the weighted spectrum envelope coefficient obtained by the weighted envelope normalization unit 15 may be input to the error coding unit 110 and used, or may be calculated by the error coding unit 110.
- W (n) 1 [1 ⁇ n ⁇ N] an average value of approximate values of power spectrum envelope values and an average value of estimated values of power spectrum envelope values may be used.
- an average value of values obtained by smoothing the power spectrum envelope value, the approximate value of the power spectrum envelope value, or the estimated value of the power spectrum envelope value in the frequency direction may be used.
- the average value here is a value obtained by averaging the target values for a plurality of samples, that is, a value obtained by averaging the target values for the plurality of samples.
- a power spectrum envelope value, an approximate value of the power spectrum envelope value, an estimated value of the power spectrum envelope value, and any of these values A value having the same magnitude relationship as at least one of a value obtained by smoothing and a value obtained by averaging any one of these values for a plurality of samples may be used.
- a value having the same magnitude relationship is calculated by the error encoding unit 110 and used.
- the value having the same magnitude relationship is a square value or a square root.
- the power spectrum envelope value W (n) [1 ⁇ n ⁇ N] is the same value as the square of the power spectrum envelope value (W (n)) 2 [1 ⁇ n ⁇ N] (W (n)) 1/2 [1 ⁇ n ⁇ N], which is the square root of the power spectrum envelope value.
- the values obtained by the weighted envelope normalization unit 15 are input to the error encoding unit 110. May be used.
- a rearrangement unit 111 may be provided to rearrange the quantized MDCT coefficient sequences.
- the encoding unit 19 performs variable length encoding on the quantized MDCT coefficient sequence rearranged by the rearrangement unit 111.
- the number of bits can be greatly reduced by variable-length coding, and improvement by encoding errors can be expected.
- the rearrangement unit 11 includes, for each frame, (1) all samples of the quantized MDCT coefficient sequence, and (2) quantum so that samples having the same or similar index reflecting the sample size are collected.
- the rearranged sample sequence obtained by rearranging at least a part of samples included in the normalized MDCT coefficient sequence is output (step S111).
- the “index reflecting the sample size” is, for example, the absolute value or power (square value) of the amplitude of the sample, but is not limited thereto.
- Japanese Patent Application No. 2010-225949 PCT / JP2011 / 072752).
- the decoding device 2 MDCT coefficients are reconstructed by processing in the reverse order to the encoding processing by the encoding device 1.
- the code input to the decoding device 2 includes a variable length code, an error code, gain information, and a linear prediction coefficient code.
- selection information is output from the encoding device 1, this selection information is also input to the decoding device 2.
- the decoding device 2 includes a decoding unit 21, a power spectrum envelope calculation unit 22, an error decoding unit 23, a gain decoding unit 24, an addition unit 25, a weighted envelope inverse normalization unit 26, and a time domain conversion unit 27.
- the decoding device 2 performs each process of the decoding method illustrated in FIG. Hereinafter, each process of the decoding device 2 will be described.
- the decoding unit 21 decodes a variable length code included in an input code for each frame, and a sequence of decoded quantized MDCT coefficients u (n), that is, quantized MDCT coefficients u (n) of the encoding device. ) And the number of bits of the variable length code are output (step S21).
- a variable length decoding method corresponding to the variable length coding method executed to obtain the code string is executed.
- the details of the decoding process performed by the decoding unit 21 correspond to the details of the encoding process performed by the encoding unit 19 of the encoding device 1. Therefore, the description of the encoding process is incorporated herein and the decoding corresponding to the performed encoding is performed. Is a decoding process performed by the decoding unit 21, and this is a detailed description of the decoding process.
- sequence of decoded quantized MDCT coefficients u (n) corresponds to the sequence of integer values in the claims.
- variable length coding method is executed is specified by the selection information.
- the selection information includes, for example, information for specifying an application region and a rice parameter for Rice coding, information indicating an application region for run-length encoding, and information for specifying the type of entropy encoding.
- the decoding method corresponding to these encoding methods is applied to the corresponding region of the input code string. Since the decoding process corresponding to the Rice encoding, the decoding process corresponding to the entropy encoding, and the decoding process corresponding to the run length encoding are all well known, the description thereof is omitted (for example, refer to the above-mentioned Reference 1).
- the power spectrum envelope calculation unit 22 decodes the linear prediction coefficient code input from the encoding device 1 to obtain a quantized linear prediction coefficient, and converts the obtained quantized linear prediction coefficient into the frequency domain. A power spectrum envelope is obtained (step S22). The process for obtaining the power spectrum envelope from the quantized linear prediction coefficient is the same as that of the power spectrum envelope calculation unit 14 of the encoding device 1.
- an approximate value of the power spectrum envelope value and an estimated value of the power spectrum envelope value may be calculated as in the power spectrum envelope calculation unit 14 of the encoding device 1.
- the power spectrum envelope calculation unit 22 of the decoding device 2 also calculates the approximate value of the power spectrum envelope value.
- the power spectrum envelope may be calculated using the quantized linear prediction coefficient.
- the decoding device 2 may not include the power spectrum envelope calculation unit 22.
- the error decoding unit 23 calculates a number obtained by subtracting the number of bits output from the decoding unit 21 from the number of bits set in advance as the code amount of the quantized MDCT coefficient sequence as the number of surplus bits.
- the error code output from the error encoding unit 110 of the encoding device 1 is decoded by a decoding method corresponding to the error encoding unit 110 of the encoding device 1 to obtain a decoded quantization error q (n) (step).
- the number of bits given to the quantization error sequence in the encoding device 1 is obtained from the number of surplus bits based on the number of bits obtained by variable length encoding that can be understood by the decoding unit 21. Since the samples and the procedures are determined by the encoding device 1 and the decoding device 2 for each surplus bit number so that they can be uniquely decoded.
- sequence of decoding quantization errors corresponds to the sequence of error values in the claims.
- ⁇ Specific example 1 of error decoding> (corresponding to ⁇ specific example 1 of error encoding> of encoding apparatus 1)
- a code book for each value that the number of surplus bits can take is stored in advance in the code book storage unit in the error decoding unit 23.
- a vector having the same number of samples as the number of decoded quantization error sequences that can be expressed by the number of surplus bits corresponding to each codebook and a code corresponding to the vector are stored in advance. Is done.
- the error decoding unit 23 selects a code book corresponding to the calculated number of surplus bits from the code book stored in the code book storage unit, and performs vector inversion using the selected code book. Perform quantization.
- the decoding process after selecting the codebook is the same as general vector inverse quantization. That is, among the vectors of the selected codebook, a vector corresponding to the input error code is output as a decoded quantization error q (n).
- the vector stored in the codebook has the same number of samples as the sequence of the decoded quantization error, but the number of samples of the vector stored in the codebook is the integer number of the sequence of the decoded quantization error. It is also possible to perform vector inverse quantization on each of a plurality of codes included in an error code input for each of a plurality of portions of a decoding quantization error sequence.
- the error decoding unit 23 selects U samples having the corresponding power spectrum envelope value among the T samples whose decoded quantization MDCT coefficient u (n) is not 0, and For each selected sample, the 1-bit code included in the input error code is decoded to obtain the positive / negative information of the sample, and the obtained positive / negative information is converted into an absolute value 0.25 of the reconstructed value.
- the reconstructed value +0.25 or ⁇ 0.25 obtained as a result is output as a decoded quantization error q (n) corresponding to the decoded quantized MDCT coefficient u (n).
- a predetermined rule such as selecting a quantization error sample with a smaller position on the frequency axis (a quantization error sample with a lower frequency).
- a rule corresponding to the rule used in the error encoding unit 110 of the encoding device 1 is stored in the error decoding unit 23 in advance.
- the error decoding unit 23 is a 1-bit code included in the input error code for a sample whose decoded quantization MDCT coefficient u (n) is not 0.
- To obtain the positive / negative information of the decoded quantization error sample give the obtained positive / negative information to the absolute value 0.25 of the reconstructed value, and obtain the reconstructed value +0.25 or -0.25, which is the decoded quantization
- the decoded quantization error q (n) corresponding to the MDCT coefficient u (n) is output.
- the error decoding unit 23 also includes, in the input error code, the UT pieces of samples having the corresponding power spectrum envelope value that are large among the samples whose decoding quantization MDCT coefficient u (n) is 0. Decode the 1-bit code to obtain the positive / negative information of the decoded quantization error sample, and give the obtained positive / negative information to the absolute value A of the reconstructed value which is a predetermined positive value smaller than 0.25.
- the obtained reconstruction value + A or -A is output as a decoded quantization error q (n) corresponding to the decoded quantized MDCT coefficient u (n).
- the UT bit code included in the error code is vector-inverse-quantized for a plurality of samples having the corresponding power spectrum envelope value among the samples whose decoding quantization MDCT coefficient u (n) is 0.
- the absolute value of the reconstructed value when the value of the quantized MDCT coefficient u (n) and the value of the decoded quantized MDCT coefficient u (n) is not 0 is set to 0.25, for example, and the quantized MDCT coefficient u (n) And the absolute value of the reconstructed value when the value of the decoded quantized MDCT coefficient u (n) is 0 is A (0 ⁇ A ⁇ 0.25).
- the absolute values of these reconstruction values are examples, and the absolute values of the reconstruction values when the values of the quantized MDCT coefficient u (n) and the decoded quantized MDCT coefficient u (n) are not 0 are quantized.
- selection is made according to a predetermined rule such as selecting a sample having a smaller position on the frequency axis (a sample having a lower frequency).
- the 1-bit first-cycle code included in the input error code is decoded to obtain positive / negative information, and the obtained positive / negative information is given to the absolute value 0.25 of the reconstructed value + reconstructed value + 0.25 Alternatively, ⁇ 0.25 is set as the first round decoding quantization error q 1 (n) corresponding to the decoding quantization MDCT coefficient u (n). Furthermore, a reconstructed value obtained by decoding the 1-bit second-round code included in the input error code to obtain positive / negative information, and giving the obtained positive / negative information to the absolute value 0.125 of the reconstructed value Let +0.125 or -0.125 be the second round decoding quantization error q 2 (n). The first round decoding quantization error q 1 (n) and the second round decoding quantization error q 2 (n) are added to obtain a decoding quantization error q (n).
- the error decoding unit 23 performs the following processing for samples whose decoding quantization MDCT coefficient u (n) is 0.
- the 1-bit first-cycle code included in the input error code is decoded to obtain positive / negative information, and the obtained positive / negative information is given to the absolute value A of the reconstructed value which is a positive value smaller than 0.25.
- the reconstructed value + A or ⁇ A obtained in this way is set as the first round decoding quantization error q 1 (n) corresponding to the decoding quantization MDCT coefficient u (n).
- the 1-bit second-round code included in the input error code is decoded to obtain positive / negative information, and the obtained positive / negative information is given to the absolute value A / 2 of the reconstructed value.
- the configuration value + A / 2 or -A / 2 is set as the second round decoding quantization error q 2 (n).
- the first round decoding quantization error q 1 (n) and the second round decoding quantization error q 2 (n) are added to obtain a decoding quantization error q (n).
- the reconstructed value corresponding to the second cycle code The absolute value is set to 1 ⁇ 2 of the absolute value of the reconstructed value corresponding to the first cycle code.
- an approximate value of the power spectrum envelope value, an estimated value of the power spectrum envelope value, a value obtained by smoothing any of these values, Either a value obtained by averaging any of these values for a plurality of samples or a value having the same magnitude relationship as any of these values may be used.
- Gain Decoding Unit 24 The gain decoding unit 24 decodes the input gain information to obtain and output the gain g (step S24). The gain g is transmitted to the adding unit 25.
- the added value sequence generated by the adding unit 25 corresponds to a sample sequence in the frequency domain in the claims.
- "Weighting envelope inverse normalization unit 26” Next, the weighted envelope inverse normalization unit 26 obtains an MDCT coefficient sequence by dividing the power spectrum envelope value by each coefficient x ⁇ (n) of the decoded weighted normalized MDCT coefficient sequence for each frame (step S26). .
- time domain conversion unit 27 converts the MDCT coefficient sequence output from the weighted envelope inverse normalization unit 26 into the time domain for each frame to obtain a frame-based audio-acoustic digital signal (step S27).
- the sequence of the decoded quantized MDCT coefficients u (n) generated by the decoding unit 21 is the reordering unit of the decoding device 2.
- the rearranged sequence of decoded quantized MDCT coefficients u (n) is transmitted to the error decoding unit 23 and the adding unit 25.
- the error decoding unit 23 and the addition unit 25 replace the decoded quantized MDCT coefficient u (n) sequence generated by the decoding unit 21 with the rearranged decoded quantized MDCT coefficient u (n) sequence.
- the same processing as described above is performed.
- the encoding device 1 and the decoding device 2 include an input unit to which a keyboard or the like can be connected, an output unit to which a liquid crystal display or the like can be connected, a CPU (Central Processing Unit), and a RAM (Random) Access memory (ROM), ROM (read only memory), external storage device that is a hard disk, and a bus that connects the input unit, output unit, CPU, RAM, ROM, and external storage device so that data can be exchanged between them For example. If necessary, the encoding device 1 and the decoding device 2 may be provided with a device (drive) that can read and write a storage medium such as a CD-ROM.
- a device drive
- the external storage device of the encoding device 1 and the decoding device 2 stores a program for executing encoding and decoding, data necessary for processing of this program, and the like.
- the program may be stored in a ROM that is a read-only storage device. Data obtained by the processing of these programs is appropriately stored in a RAM or an external storage device.
- a storage device that stores data, addresses of storage areas, and the like is simply referred to as a “storage unit”.
- the storage unit of the encoding device 1 stores a program for encoding a frequency-domain sample sequence derived from a sound and audio signal, an error encoding, and the like.
- the storage unit of the decoding device 2 stores a program for decoding the input code.
- each program stored in the storage unit and data necessary for processing each program are read into the RAM as necessary, and are interpreted and executed by the CPU.
- the CPU implements predetermined functions (for example, the error calculation unit 18, the error encoding unit 110, and the encoding unit 19), thereby realizing encoding.
- each program stored in the storage unit and data necessary for processing each program are read into the RAM as necessary, and are interpreted and executed by the CPU.
- the decoding is realized by the CPU realizing a predetermined function (for example, the decoding unit 21).
- the quantization unit 17 of the encoding device 1 uses a value G (x (n) / g) obtained by expanding / contracting the value of x (n) / g by a predetermined function G instead of x (n) / g. Also good. Specifically, the quantization unit 17 uses the function G to obtain x (n) / g obtained by dividing each coefficient x (n) [1 ⁇ n ⁇ N] of the weighted normalized MDCT coefficient sequence by the gain g.
- This quantized MDCT coefficient is to be encoded by the encoding unit 19.
- sign (h) is a polarity sign function that outputs positive and negative signs of input h. For example, sign (h) outputs 1 if the input h is a positive number, and outputs -1 if the input h is a negative number.
- represents the absolute value of h.
- a is a predetermined number, for example, 0.75.
- the value G (x (n) / g) obtained by expanding / contracting the value of x (n) / g by a predetermined function G corresponds to the sample sequence in the frequency domain of the claims.
- the quantization error r (n) obtained by the error calculation unit 18 is G (x (n) / g) -u (n). This quantization error r (n) is to be encoded by the error encoding unit 110.
- G ⁇ 1 sign (h) ⁇
- an inverse function of the function G with respect to u (n) + q (n) obtained by the addition.
- G ⁇ 1 (h) sign (h) ⁇
- processing functions in the hardware entities (encoding device 1 and decoding device 2) described in the above embodiment are realized by a computer, the processing contents of the functions that the hardware entity should have are described by a program. Then, by executing this program on a computer, the processing functions in the hardware entity are realized on the computer.
- the program describing the processing contents can be recorded on a computer-readable recording medium.
- the computer-readable recording medium may be any recording medium such as a magnetic recording device, an optical disk, a magneto-optical recording medium, and a semiconductor memory.
- a magnetic recording device a hard disk device, a flexible disk, a magnetic tape or the like, and as an optical disk, a DVD (Digital Versatile Disc), a DVD-RAM (Random Access Memory), a CD-ROM (Compact Disc Read Only) Memory), CD-R (Recordable) / RW (ReWritable), etc.
- magneto-optical recording media MO (Magneto-Optical disc), etc., semiconductor memory, EEP-ROM (Electronically Erasable and Programmable-Read Only Memory), etc. Can be used.
- this program is distributed by selling, transferring, or lending a portable recording medium such as a DVD or CD-ROM in which the program is recorded. Furthermore, the program may be distributed by storing the program in a storage device of the server computer and transferring the program from the server computer to another computer via a network.
- a computer that executes such a program first stores a program recorded on a portable recording medium or a program transferred from a server computer in its own storage device.
- the computer reads the program stored in its own recording medium and executes the process according to the read program.
- the computer may directly read the program from a portable recording medium and execute processing according to the program, and the program is transferred from the server computer to the computer.
- the processing according to the received program may be executed sequentially.
- the program is not transferred from the server computer to the computer, and the above-described processing is executed by a so-called ASP (Application Service Provider) type service that realizes a processing function only by an execution instruction and result acquisition. It is good.
- the program in this embodiment includes information that is used for processing by an electronic computer and that conforms to the program (data that is not a direct command to the computer but has a property that defines the processing of the computer).
- the hardware entity is configured by executing a predetermined program on the computer.
- a predetermined program on the computer.
- at least a part of these processing contents may be realized in hardware.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Signal Processing (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Computational Linguistics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Quality & Reliability (AREA)
- Mathematical Physics (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
- Signal Processing For Digital Recording And Reproducing (AREA)
Abstract
Description
最初に図1から図4を参照して符号化処理を説明する。
まず、周波数領域変換部11がフレーム単位で音声音響ディジタル信号を周波数領域のN点のMDCT係数列に変換する(ステップS11)。
線形予測分析部12は、フレーム単位で音声音響ディジタル信号を線形予測分析し、予め定めた次数までの線形予測係数を求めて出力する(ステップS12)。
線形予測係数量子化符号化部13は、線形予測分析部12が求めた線形予測係数に対応する符号と、量子化済み線形予測係数とを求めて出力する(ステップS13)。
その際、線形予測係数をLSP(Line Spectral Pairs)に変換し、LSPに対応する符号と量子化済みLSPとを求め、量子化済みLSPを量子化済み線形予測係数に変換する処理を行ってもよい。
パワースペクトル包絡算出部14は、線形予測係数量子化符号化部13が出力した量子化済み線形予測係数を周波数領域に変換してパワースペクトル包絡を求める(ステップS14)。求まったパワースペクトル包絡は、重み付け包絡正規化部15に送信される。また、必要に応じて、図1に破線で示すように誤差符号化部110に送信される。
重み付け包絡正規化部15は、パワースペクトル包絡算出部14が出力したパワースペクトル包絡によって、MDCT係数列の各係数を正規化する(ステップS15)。ここでは聴覚的に歪が小さくなるような量子化の実現のために、重み付け包絡正規化部15は、パワースペクトル包絡値の系列やその平方根の系列を周波数方向で平滑化した重み付けスペクトル包絡係数を用いてフレーム単位でMDCT係数列の各係数を正規化する。この結果、フレーム単位の重み付け正規化MDCT係数列の各係数x(1),・・・,x(N)が得られる。重み付け正規化MDCT係数列は、正規化利得計算部16、量子化部17及び誤差算出部18に送信される。重み付け正規化MDCT係数列は一般には低周波領域にやや大きな振幅を持ち、ピッチ周期に起因する微細構造をもつものの、元のMDCT係数列ほどの大きな振幅傾きや振幅の凹凸を持たない。
次に、正規化利得計算部16が、フレームごとに、重み付け正規化MDCT係数列の各係数x(1),・・・,x(N)を与えられた総ビット数で量子化できるように、全周波数に亘る振幅値の和またはエネルギー値を用いて量子化ステップ幅を決定し、この量子化ステップ幅になるように重み付け正規化MDCT係数列の各係数を割り算する係数g(以下、利得という。)を求める(ステップS16)。この利得を表す情報である利得情報は、復号装置2に送信される符号の一部とされる。
次に、量子化部17が、フレームごとに、重み付け正規化MDCT係数列の各係数x(1),・・・,x(N)をステップS16の処理で決定された量子化ステップ幅で量子化する(ステップS17)。すなわち、重み付け正規化MDCT係数列の各係数x(n)[1≦n≦N]を利得gで割り算して得られるx(n)/gの値の小数点以下を四捨五入して得られる整数値u(n)を量子化MDCT係数とする。フレーム単位の量子化MDCT係数列は、誤差算出部18及び符号化部19に送信される。なお、x(n)/gの値の小数点以下を切り上げ又は切り下げして得られる値を、整数値u(n)としてもよい。このように、整数値u(n)はx(n)/gの値に対応する値としてもよい。
ステップS15の処理で得られた重み付け正規化MDCT系列と、ステップS16の処理で得られた利得gと、ステップS17の処理で得られたフレーム単位の量子化MDCT係数列とが、誤差算出部18の入力となる。量子化による誤差はr(n)=x(n)/g-u(n) [1≦n≦N]で求められる。すなわち、重み付け正規化MDCT係数列の各係数x(n)を利得gで除算した値からその各係数x(n)に対応する量子化MDCT係数u(n)を減算した値をその各係数x(n)に対応する量子化誤差r(n)とする。
次に、符号化部19が、フレームごとに、量子化部17が出力した量子化MDCT係数列(量子化MDCT係数u(n)の列)を符号化し、得られた符号とその符号のビット数を出力する(ステップS19)。
(参考文献1)David Salomon, "Data Compression : The Complete Reference," 3rd edition, Springer-Verlag, ISBN-10: 0-387-40697-2, 2004.
整数値である量子化MDCT係数列の各係数u(1),・・・,u(N)の可変長符号化の結果、量子化MDCT係数列を表現するに必要なビット数がわかり、想定されたビット数から圧縮によって得られた余剰ビットがわかる。もし、フレームをまたがってビットをやりくりできるならば、余剰ビットは次のフレーム以降で有効利用できる。もし、フレーム内では固定のビット数が割り当てられる場合には別の符号化で有効利用する必要があり、そうでなければ可変長符号化による平均ビット数の削減の意味がなくなる。
量子化誤差の値を符号化する際に、複数のサンプルをまとめてベクトル量子化を行ってもよい。ただし、一般にテーブル(符号帳)中に符号系列を蓄積し、入力と符号系列の距離計算が必要で、メモリ量と演算量が多くなる。また任意のビット数に対応するために個別の符号帳が必要になるなど、構成が煩雑になる。
量子化誤差列に含まれる量子化誤差を1サンプルずつ符号化する場合には、量子化誤差列に含まれる量子化誤差サンプルについての優先順位をきめて、優先順位の高い量子化誤差サンプルから余剰ビット数で符号化可能なもののみを符号化する。例えば、量子化誤差の絶対値又はエネルギーが大きい量子化誤差サンプルから優先的に符号化する。
誤差符号化部110は、量子化誤差列のうちの対応する量子化MDCT係数u(n)の値が0でないT個の量子化誤差サンプルのうち、対応するパワースペクトル包絡値が大きいものからU個選択して、選択されたそれぞれの量子化誤差サンプルに対して、量子化誤差サンプルの正負を表わす情報である1ビットの符号を生成して、生成したUビットの符号を誤差符号として出力する。なお、対応するパワースペクトル包絡値が同一である場合は、例えば、より周波数軸上での位置の小さい量子化誤差サンプル(周波数が低い量子化誤差サンプル)を選択する等の予め定めた規則に従って選択する。
誤差符号化部110は、量子化誤差列のうちの対応する量子化MDCT係数u(n)の値が0でないT個の量子化誤差サンプルのそれぞれについて、量子化誤差サンプルの正負を表わす情報である1ビットの符号を生成する。
誤差符号化部110は、量子化誤差系列に含まれるすべての量子化誤差サンプルのそれぞれについて、量子化誤差サンプルの正負を表わす1ビットの1巡目符号を生成する。
続いて図5から図6を参照して復号処理を説明する。
まず、復号部21が、フレームごとに、入力された符号に含まれる可変長符号を復号して、復号量子化MDCT係数u(n)の列、すなわち符号化装置の量子化MDCT係数u(n)と同一のものと、可変長符号のビットの個数とを出力する(ステップS21)。当然であるが、符号列を得るために実行された可変長符号化方法に対応する可変長復号方法が実行される。復号部21による復号処理の詳細は符号化装置1の符号化部19による符号化処理の詳細に対応するので、当該符号化処理の説明をここに援用し、実行された符号化に対応する復号が復号部21の行う復号処理であることを明記し、これをもって復号処理の詳細な説明とする。
パワースペクトル包絡算出部22は、符号化装置1から入力された線形予測係数符号を復号して量子化済み線形予測係数を得て、得られた量子化済み線形予測係数を周波数領域に変換してパワースペクトル包絡を求める(ステップS22)。量子化済み線形予測係数からパワースペクトル包絡を求める処理は符号化装置1のパワースペクトル包絡算出部14と同じである。
誤差復号部23は、まず、量子化MDCT係数列の符号量として予め設定されたビット数から復号部21が出力したビット数を減算した数を余剰ビット数として算出する。次に符号化装置1の誤差符号化部110が出力した誤差符号を、符号化装置1の誤差符号化部110に対応する復号方法で復号して復号量子化誤差q(n)を得る(ステップS23)。符号化装置1において量子化誤差列に与えられたビット数は、復号部21でわかる可変長符号化によるビット数に基づく余剰ビット数から求める。余剰ビット数ごとに符号、復号でサンプルや手順を符号化装置1と復号装置2とで対応するように決めておくので、一意に復号できる。
誤差復号部23内の符号帳記憶部に、余剰ビット数の取り得る値ごとの符号帳を予め格納しておく。各符号帳には、その各符号帳に対応する余剰ビット数で表現可能な個数の復号量子化誤差の系列と同じサンプル数のベクトルと、そのベクトルに対応する符号、とが対応付けて予め格納される。
余剰ビット数をU、復号部21が出力した復号量子化MDCT係数u(n)の値が0でないもののサンプル数をT、復号部21が出力した復号量子化MDCT係数u(n)の値が0であるサンプル数をSとするとき、次のような復号手順が好ましい。
誤差復号部23は、復号量子化MDCT係数u(n)の値が0でないT個サンプルのうち、対応するパワースペクトル包絡値が大きいものからU個選択して、選択されたそれぞれのサンプルに対して、入力された誤差符号に含まれる1ビットの符号を復号してサンプルの正負の情報を得て、得られた正負の情報を再構成値の絶対値0.25に与えて得られる再構成値+0.25又は-0.25を、その復号量子化MDCT係数u(n)に対応する復号量子化誤差q(n)として出力する。なお、対応するパワースペクトル包絡値が同一である場合は、例えば、より周波数軸上での位置の小さい量子化誤差サンプル(周波数が低い量子化誤差サンプル)を選択する等の予め定めた規則に従って選択する。例えば、符号化装置1の誤差符号化部110で用いた規則に対応する規則を誤差復号部23に予め保持しておく。
誤差復号部23は、復号量子化MDCT係数u(n)の値が0でないサンプルに対して、入力された誤差符号に含まれる1ビットの符号を復号して復号量子化誤差サンプルの正負の情報を得て、得られた正負の情報を再構成値の絶対値0.25に与えて得られる再構成値+0.25または-0.25を、その復号量子化MDCT係数u(n)に対応する復号量子化誤差q(n)として出力する。
誤差復号部23は、復号量子化MDCT係数u(n)の値が0でないサンプルについては、以下の処理を行う。
利得復号部24は、入力された利得情報を復号して利得gを得て出力する(ステップS24)。利得gは、加算部25に送信される。
加算部25は、フレームごとに、復号部21が出力した復号量子化MDCT係数列の各係数u(n)と、誤差復号部23が出力した復号量子化誤差系列の対応する各係数q(n)とを加算した加算値を計算する。そして、この加算値に、利得復号部24が出力した利得gを乗じた系列を生成し復号重み付け正規化MDCT係数系列とする(ステップS25)。復号重み付け正規化MDCT係数系列の各係数をx^(n)と表記する。x^(n)=(u(n)+q(n))*gである。
「重み付け包絡逆正規化部26」
次に、重み付け包絡逆正規化部26が、フレームごとに、復号重み付け正規化MDCT係数列の各係数x^(n)にパワースペクトル包絡値を除算することでMDCT係数列を得る(ステップS26)。
次に、時間領域変換部27が、フレームごとに、重み付け包絡逆正規化部26が出力したMDCT係数列を時間領域に変換してフレーム単位の音声音響ディジタル信号を得る(ステップS27)。
上述の実施形態に関わる符号化装置1、復号装置2は、キーボードなどが接続可能な入力部と、液晶ディスプレイなどが接続可能な出力部と、CPU(Central Processing Unit)、メモリであるRAM(Random Access Memory)やROM(Read Only Memory)と、ハードディスクである外部記憶装置と、これらの入力部、出力部、CPU、RAM、ROM、外部記憶装置間のデータのやり取りが可能なように接続するバスとを例えば備えている。また、必要に応じて、符号化装置1、復号装置2に、CD-ROMなどの記憶媒体を読み書きできる装置(ドライブ)などを設けてもよい。
符号化装置1の量子化部17では、x(n)/gではなく、x(n)/gの値を所定の関数Gで伸縮させた値G(x(n)/g)を用いてもよい。具体的には、量子化部17は、重み付け正規化MDCT係数列の各係数x(n)[1≦n≦N]を利得gで割り算して得られるx(n)/gを関数Gで伸縮させた値G(x(n)/g)に対応する整数値、例えばG(x(n)/g)の小数点以下を四捨五入又は切り上げ又は切り下げして得られる整数値u(n)を量子化MDCT係数とする。この量子化MDCT係数が符号化部19の符号化の対象となる。
Claims (20)
- 所定の時間区間の音響信号に由来する周波数領域のサンプル列を所定の個数のビットで符号化する符号化方法であって、
上記周波数領域のサンプル列の各サンプルの値に対応する整数値を可変長符号化により符号化して可変長符号を生成する符号化ステップと、
上記周波数領域のサンプル列の各サンプルの値からその各サンプルの値に対応する整数値を減算した誤差値の列を算出する誤差算出ステップと、
上記所定の個数から上記可変長符号のビットの個数を減算した個数のビットである余剰ビットを用いて、上記誤差値の列を符号化して誤差符号を生成する誤差符号化ステップと、
を含む符号化方法。 - 請求項1の符号化方法において、
上記誤差符号化ステップは、上記余剰ビットを用いて、上記誤差の列を構成する誤差サンプルのうち対応する上記整数値が0でない誤差サンプルを優先的に符号化する、
符号化方法。 - 請求項1の符号化方法において、
上記誤差符号化ステップは、上記余剰ビットを用いて、上記誤差値の列を構成する誤差サンプルのうち対応するパワースペクトル包絡値、パワースペクトル包絡値の概算値又はパワースペクトル包絡値の推定値が大きい誤差サンプルを優先的に符号化する、
符号化方法。 - 請求項1から3の何れかの符号化方法において、
上記誤差符号化ステップは、上記誤差値の列を構成する誤差サンプルのうち符号化の対象となる各誤差サンプルの値の正負の情報を1ビットで符号化する、
符号化方法。 - 請求項4の符号化方法において、
整数値に応じて定まる値を再構成値の絶対値とし、上記再構成値の絶対値を誤差サンプルの値が正のときに対応する再構成値とし、0から上記再構成値の絶対値を減算した値を誤差サンプルの値が負のときに対応する再構成値として、
上記誤差符号化ステップは、上記余剰ビットの数が上記誤差値の列を構成する誤差サンプルの数よりも大きい場合には、上記各誤差サンプルに対応する再構成値を上記各誤差サンプルの値から減算した値の正負の情報を別の1ビットを用いて更に符号化する、
符号化方法。 - 請求項4又は5の符号化方法において、
整数値が0でない場合の再構成値の絶対値は、整数値が0である場合の再構成値の絶対値よりも大きい、
符号化方法。 - 入力された所定の個数のビットで構成される符号を復号する復号方法であって、
上記符号に含まれる可変長符号を復号して整数値の列を生成する復号ステップと、
上記所定の個数から上記可変長符号のビットの個数を減算した個数のビットである余剰ビットで構成される、上記符号に含まれる誤差符号を復号して誤差値の列を生成する誤差復号ステップと、
上記整数値の列の各サンプルと上記誤差値の列の対応する誤差サンプルとを加算する加算ステップと、
を含む復号方法。 - 請求項7の復号方法において、
上記誤差復号ステップは、上記余剰ビットで表現された、上記誤差の列を構成する誤差サンプルのうち対応する上記整数値が0でない誤差サンプルを復号する、
復号方法。 - 請求項7の復号方法において、
上記誤差復号ステップは、上記余剰ビットで表現された、上記誤差値の列を構成する誤差サンプルのうち対応するパワースペクトル包絡値、パワースペクトル包絡値の概算値又はパワースペクトル包絡値の推定値が大きい誤差サンプルを復号する、
復号方法。 - 請求項7から9の何れかの復号方法において、
整数値に応じて定まる値を再構成値の絶対値として、
上記誤差復号ステップは、上記誤差値の列の各誤差サンプルの値を、上記誤差符号を復号することにより得られた上記各誤差サンプルに対応する1ビットの情報で定まる正負を上記各誤差サンプルに対応する整数値に基づく再構成値の絶対値に対して反映させた値とする、
復号方法。 - 請求項10の復号方法において、
上記誤差復号ステップは、上記各誤差サンプルの値に対応する別の1ビットの情報がある場合には、上記各誤差サンプルの値を、上記反映させた値と、上記別の1ビットの情報で定まる正負を上記各誤差サンプルに対応する整数値に基づく再構成値の絶対値の1/2の値に対して反映させた値とを加算した値とする、
復号方法。 - 請求項10又は11の復号方法において、
整数値が0でない場合の再構成値の絶対値は、整数値が0である場合の再構成値の絶対値よりも大きい、
復号方法。 - 所定の時間区間の音響信号に由来する周波数領域のサンプル列を所定の個数のビットで符号化する符号化装置であって、
上記周波数領域のサンプル列の各サンプルの値に対応する整数値を可変長符号化により符号化して可変長符号を生成する符号化部と、
上記周波数領域のサンプル列の各サンプルの値からその各サンプルの値に対応する整数値を減算した誤差値の列を算出する誤差算出部と、
上記所定の個数から上記可変長符号のビットの個数を減算した個数のビットである余剰ビットを用いて、上記誤差値の列を符号化して誤差符号を生成する誤差符号化部と、
を含む符号化装置。 - 請求項13の符号化装置において、
上記誤差符号化部は、上記余剰ビットを用いて、上記誤差の列を構成する誤差サンプルのうち対応する上記整数値が0でない誤差サンプルを優先的に符号化する、
符号化装置。 - 請求項13の符号化装置において、
上記誤差符号化部は、上記余剰ビットを用いて、上記誤差値の列を構成する誤差サンプルのうち対応するパワースペクトル包絡値、パワースペクトル包絡値の概算値又はパワースペクトル包絡値の推定値が大きい誤差サンプルを優先的に符号化する、
符号化装置。 - 入力された所定の個数のビットで構成される符号を復号する復号装置であって、
上記符号に含まれる可変長符号を復号して整数値の列を生成する復号部と、
上記所定の個数から上記可変長符号のビットの個数を減算した個数のビットである余剰ビットで構成される、上記符号に含まれる誤差符号を復号して誤差値の列を生成する誤差復号部と、
上記整数値の列の各サンプルと上記誤差値の列の対応する誤差サンプルとを加算する加算部と、
を含む復号装置。 - 請求項16の復号装置において、
上記誤差復号ステップは、上記余剰ビットで表現された、上記誤差の列を構成する誤差サンプルのうち対応する上記整数値が0でない誤差サンプルを復号する、
復号装置。 - 請求項16の復号装置において、
上記誤差復号部は、上記余剰ビットで表現された、上記誤差値の列を構成する誤差サンプルのうち対応するパワースペクトル包絡値、パワースペクトル包絡値の概算値又はパワースペクトル包絡値の推定値が大きい誤差サンプルを復号する、
復号装置。 - 請求項1から12に記載された方法の各ステップをコンピュータに実現させるためのプログラム。
- 請求項1から12に記載された方法の各ステップをコンピュータに実現させるためのプログラムを記録したコンピュータ読み取り可能な記録媒体。
Priority Applications (12)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PL16195433T PL3154057T3 (pl) | 2011-04-05 | 2012-03-26 | Dekodowanie sygnału akustycznego |
EP16195433.4A EP3154057B1 (en) | 2011-04-05 | 2012-03-26 | Acoustic signal decoding |
ES12767213.7T ES2617958T3 (es) | 2011-04-05 | 2012-03-26 | Codificación de una señal acústica |
EP18196322.4A EP3441967A1 (en) | 2011-04-05 | 2012-03-26 | Decoding method, decoder, program, and recording medium |
US14/007,844 US10515643B2 (en) | 2011-04-05 | 2012-03-26 | Encoding method, decoding method, encoder, decoder, program, and recording medium |
EP12767213.7A EP2696343B1 (en) | 2011-04-05 | 2012-03-26 | Encoding an acoustic signal |
CN201280015955.3A CN103460287B (zh) | 2011-04-05 | 2012-03-26 | 音响信号的编码方法、解码方法、编码装置、解码装置 |
RU2013143624/08A RU2571561C2 (ru) | 2011-04-05 | 2012-03-26 | Способ кодирования, способ декодирования, кодер, декодер, программа и носитель записи |
KR1020137025380A KR101569060B1 (ko) | 2011-04-05 | 2012-03-26 | 부호화 방법, 복호 방법, 부호화 장치, 복호 장치, 프로그램, 기록 매체 |
JP2013508811A JP5603484B2 (ja) | 2011-04-05 | 2012-03-26 | 符号化方法、復号方法、符号化装置、復号装置、プログラム、記録媒体 |
US16/687,176 US11074919B2 (en) | 2011-04-05 | 2019-11-18 | Encoding method, decoding method, encoder, decoder, program, and recording medium |
US16/687,144 US11024319B2 (en) | 2011-04-05 | 2019-11-18 | Encoding method, decoding method, encoder, decoder, program, and recording medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2011083740 | 2011-04-05 | ||
JP2011-083740 | 2011-04-05 |
Related Child Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/007,844 A-371-Of-International US10515643B2 (en) | 2011-04-05 | 2012-03-26 | Encoding method, decoding method, encoder, decoder, program, and recording medium |
US16/687,144 Continuation US11024319B2 (en) | 2011-04-05 | 2019-11-18 | Encoding method, decoding method, encoder, decoder, program, and recording medium |
US16/687,176 Continuation US11074919B2 (en) | 2011-04-05 | 2019-11-18 | Encoding method, decoding method, encoder, decoder, program, and recording medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012137617A1 true WO2012137617A1 (ja) | 2012-10-11 |
Family
ID=46969018
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2012/057685 WO2012137617A1 (ja) | 2011-04-05 | 2012-03-26 | 符号化方法、復号方法、符号化装置、復号装置、プログラム、記録媒体 |
Country Status (10)
Country | Link |
---|---|
US (3) | US10515643B2 (ja) |
EP (3) | EP3154057B1 (ja) |
JP (1) | JP5603484B2 (ja) |
KR (1) | KR101569060B1 (ja) |
CN (1) | CN103460287B (ja) |
ES (2) | ES2704742T3 (ja) |
PL (1) | PL3154057T3 (ja) |
RU (1) | RU2571561C2 (ja) |
TR (1) | TR201900411T4 (ja) |
WO (1) | WO2012137617A1 (ja) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015166693A1 (ja) * | 2014-05-01 | 2015-11-05 | 日本電信電話株式会社 | 符号化装置、復号装置、符号化方法、復号方法、符号化プログラム、復号プログラム、記録媒体 |
JP2016508617A (ja) * | 2013-01-22 | 2016-03-22 | フラウンホーファー−ゲゼルシャフト・ツール・フェルデルング・デル・アンゲヴァンテン・フォルシュング・アインゲトラーゲネル・フェライン | 隠しオブジェクトを信号混合操作に使用する空間オーディオオブジェクト符号化の装置及び方法 |
JP2017528751A (ja) * | 2014-07-28 | 2017-09-28 | サムスン エレクトロニクス カンパニー リミテッド | 信号符号化方法及びその装置、並びに信号復号方法及びその装置 |
US11367455B2 (en) | 2015-03-13 | 2022-06-21 | Dolby International Ab | Decoding audio bitstreams with enhanced spectral band replication metadata in at least one fill element |
Families Citing this family (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5997592B2 (ja) * | 2012-04-27 | 2016-09-28 | 株式会社Nttドコモ | 音声復号装置 |
CN107369455B (zh) * | 2014-03-21 | 2020-12-15 | 华为技术有限公司 | 语音频码流的解码方法及装置 |
KR101826237B1 (ko) * | 2014-03-24 | 2018-02-13 | 니폰 덴신 덴와 가부시끼가이샤 | 부호화 방법, 부호화 장치, 프로그램 및 기록 매체 |
US10332533B2 (en) * | 2014-04-24 | 2019-06-25 | Nippon Telegraph And Telephone Corporation | Frequency domain parameter sequence generating method, encoding method, decoding method, frequency domain parameter sequence generating apparatus, encoding apparatus, decoding apparatus, program, and recording medium |
EP3594946B1 (en) * | 2014-05-01 | 2020-10-28 | Nippon Telegraph And Telephone Corporation | Decoding of a sound signal |
CN106415715B (zh) * | 2014-05-01 | 2019-11-01 | 日本电信电话株式会社 | 编码装置、编码方法、记录介质 |
CN106537500B (zh) * | 2014-05-01 | 2019-09-13 | 日本电信电话株式会社 | 周期性综合包络序列生成装置、周期性综合包络序列生成方法、记录介质 |
WO2016121826A1 (ja) * | 2015-01-30 | 2016-08-04 | 日本電信電話株式会社 | 符号化装置、復号装置、これらの方法、プログラム及び記録媒体 |
US10276186B2 (en) * | 2015-01-30 | 2019-04-30 | Nippon Telegraph And Telephone Corporation | Parameter determination device, method, program and recording medium for determining a parameter indicating a characteristic of sound signal |
JP6712643B2 (ja) * | 2016-09-15 | 2020-06-24 | 日本電信電話株式会社 | サンプル列変形装置、信号符号化装置、信号復号装置、サンプル列変形方法、信号符号化方法、信号復号方法、およびプログラム |
EP3637418B1 (en) * | 2017-06-07 | 2022-03-16 | Nippon Telegraph And Telephone Corporation | Encoding device, decoding device, smoothing device, reverse-smoothing device, methods therefor, and program |
CN110771045B (zh) * | 2017-06-22 | 2024-03-29 | 日本电信电话株式会社 | 编码装置、解码装置、编码方法、解码方法、以及记录介质 |
EP3761313B1 (en) * | 2018-03-02 | 2023-01-18 | Nippon Telegraph And Telephone Corporation | Encoding device, encoding method, program, and recording medium |
CN118038881A (zh) | 2018-04-05 | 2024-05-14 | 瑞典爱立信有限公司 | 支持生成舒适噪声的方法和设备 |
WO2019198383A1 (ja) * | 2018-04-13 | 2019-10-17 | 日本電信電話株式会社 | 符号化装置、復号装置、符号化方法、復号方法、プログラム、および記録媒体 |
JP7322620B2 (ja) * | 2019-09-13 | 2023-08-08 | 富士通株式会社 | 情報処理装置、情報処理方法および情報処理プログラム |
CN114913863B (zh) * | 2021-02-09 | 2024-10-18 | 同响科技股份有限公司 | 数字音信数据编码方法 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH03191628A (ja) * | 1989-12-21 | 1991-08-21 | Toshiba Corp | 可変レート符号化方式 |
JPH0470800A (ja) * | 1990-07-11 | 1992-03-05 | Sharp Corp | 音声情報圧縮装置 |
JPH09214348A (ja) * | 1996-01-30 | 1997-08-15 | Sony Corp | 信号符号化方法 |
JP2006011170A (ja) * | 2004-06-28 | 2006-01-12 | Sony Corp | 信号符号化装置及び方法、並びに信号復号装置及び方法 |
JP2010225949A (ja) | 2009-03-25 | 2010-10-07 | Kyocera Corp | 発熱体の放熱構造 |
WO2012046685A1 (ja) | 2010-10-05 | 2012-04-12 | 日本電信電話株式会社 | 符号化方法、復号方法、符号化装置、復号装置、プログラム、記録媒体 |
Family Cites Families (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6091460A (en) * | 1994-03-31 | 2000-07-18 | Mitsubishi Denki Kabushiki Kaisha | Video signal encoding method and system |
JP3170193B2 (ja) * | 1995-03-16 | 2001-05-28 | 松下電器産業株式会社 | 画像信号の符号化装置及び復号装置 |
US5956674A (en) * | 1995-12-01 | 1999-09-21 | Digital Theater Systems, Inc. | Multi-channel predictive subband audio coder using psychoacoustic adaptive bit allocation in frequency, time and over the multiple channels |
US20030039648A1 (en) * | 1998-09-16 | 2003-02-27 | Genentech, Inc. | Compositions and methods for the diagnosis and treatment of tumor |
US6677262B2 (en) * | 2000-07-05 | 2004-01-13 | Shin-Etsu Chemical Co., Ltd. | Rare earth oxide, basic rare earth carbonate, making method, phosphor, and ceramic |
US7136418B2 (en) * | 2001-05-03 | 2006-11-14 | University Of Washington | Scalable and perceptually ranked signal coding and decoding |
US7599835B2 (en) * | 2002-03-08 | 2009-10-06 | Nippon Telegraph And Telephone Corporation | Digital signal encoding method, decoding method, encoding device, decoding device, digital signal encoding program, and decoding program |
US7275036B2 (en) * | 2002-04-18 | 2007-09-25 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Apparatus and method for coding a time-discrete audio signal to obtain coded audio data and for decoding coded audio data |
JP4296753B2 (ja) * | 2002-05-20 | 2009-07-15 | ソニー株式会社 | 音響信号符号化方法及び装置、音響信号復号方法及び装置、並びにプログラム及び記録媒体 |
DE10236694A1 (de) * | 2002-08-09 | 2004-02-26 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung und Verfahren zum skalierbaren Codieren und Vorrichtung und Verfahren zum skalierbaren Decodieren |
US7502743B2 (en) * | 2002-09-04 | 2009-03-10 | Microsoft Corporation | Multi-channel audio encoding and decoding with multi-channel transform selection |
KR100477699B1 (ko) | 2003-01-15 | 2005-03-18 | 삼성전자주식회사 | 양자화 잡음 분포 조절 방법 및 장치 |
US8107535B2 (en) * | 2003-06-10 | 2012-01-31 | Rensselaer Polytechnic Institute (Rpi) | Method and apparatus for scalable motion vector coding |
DE10345996A1 (de) * | 2003-10-02 | 2005-04-28 | Fraunhofer Ges Forschung | Vorrichtung und Verfahren zum Verarbeiten von wenigstens zwei Eingangswerten |
US7668712B2 (en) * | 2004-03-31 | 2010-02-23 | Microsoft Corporation | Audio encoding and decoding with intra frames and adaptive forward error correction |
US7587254B2 (en) * | 2004-04-23 | 2009-09-08 | Nokia Corporation | Dynamic range control and equalization of digital audio using warped processing |
US7895034B2 (en) * | 2004-09-17 | 2011-02-22 | Digital Rise Technology Co., Ltd. | Audio encoding system |
JP4328358B2 (ja) * | 2004-12-07 | 2009-09-09 | 日本電信電話株式会社 | 情報圧縮符号化装置、その復号化装置、これらの方法、およびこれらのプログラムとその記録媒体 |
ATE521143T1 (de) * | 2005-02-23 | 2011-09-15 | Ericsson Telefon Ab L M | Adaptive bitzuweisung für die mehrkanal- audiokodierung |
KR100818268B1 (ko) * | 2005-04-14 | 2008-04-02 | 삼성전자주식회사 | 오디오 데이터 부호화 및 복호화 장치와 방법 |
US7617436B2 (en) * | 2005-08-02 | 2009-11-10 | Nokia Corporation | Method, device, and system for forward channel error recovery in video sequence transmission over packet-based network |
KR20070046752A (ko) * | 2005-10-31 | 2007-05-03 | 엘지전자 주식회사 | 신호 처리 방법 및 장치 |
TWI276047B (en) * | 2005-12-15 | 2007-03-11 | Ind Tech Res Inst | An apparatus and method for lossless entropy coding of audio signal |
JP4548348B2 (ja) | 2006-01-18 | 2010-09-22 | カシオ計算機株式会社 | 音声符号化装置及び音声符号化方法 |
US8036903B2 (en) * | 2006-10-18 | 2011-10-11 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Analysis filterbank, synthesis filterbank, encoder, de-coder, mixer and conferencing system |
KR101471978B1 (ko) | 2007-02-02 | 2014-12-12 | 삼성전자주식회사 | 오디오 신호의 음질 향상을 위한 데이터 삽입 방법 및 그장치 |
JP4871894B2 (ja) * | 2007-03-02 | 2012-02-08 | パナソニック株式会社 | 符号化装置、復号装置、符号化方法および復号方法 |
CN101308657B (zh) * | 2007-05-16 | 2011-10-26 | 中兴通讯股份有限公司 | 一种基于先进音频编码器的码流合成方法 |
WO2009004727A1 (ja) * | 2007-07-04 | 2009-01-08 | Fujitsu Limited | 符号化装置、符号化方法および符号化プログラム |
US7937574B2 (en) * | 2007-07-17 | 2011-05-03 | Advanced Micro Devices, Inc. | Precise counter hardware for microcode loops |
EP2063417A1 (en) * | 2007-11-23 | 2009-05-27 | Deutsche Thomson OHG | Rounding noise shaping for integer transform based encoding and decoding |
WO2009075326A1 (ja) * | 2007-12-11 | 2009-06-18 | Nippon Telegraph And Telephone Corporation | 符号化方法、復号化方法、これらの方法を用いた装置、プログラム、記録媒体 |
KR101452722B1 (ko) * | 2008-02-19 | 2014-10-23 | 삼성전자주식회사 | 신호 부호화 및 복호화 방법 및 장치 |
US8386271B2 (en) * | 2008-03-25 | 2013-02-26 | Microsoft Corporation | Lossless and near lossless scalable audio codec |
US20100191534A1 (en) * | 2009-01-23 | 2010-07-29 | Qualcomm Incorporated | Method and apparatus for compression or decompression of digital signals |
WO2010084951A1 (ja) * | 2009-01-23 | 2010-07-29 | 日本電信電話株式会社 | パラメータ選択方法、パラメータ選択装置、プログラム及び記録媒体 |
EP2523189B1 (en) * | 2010-01-08 | 2014-09-03 | Nippon Telegraph And Telephone Corporation | Encoding method, decoding method, encoder apparatus, decoder apparatus, program and recording medium |
-
2012
- 2012-03-26 ES ES16195433T patent/ES2704742T3/es active Active
- 2012-03-26 EP EP16195433.4A patent/EP3154057B1/en active Active
- 2012-03-26 CN CN201280015955.3A patent/CN103460287B/zh active Active
- 2012-03-26 EP EP12767213.7A patent/EP2696343B1/en active Active
- 2012-03-26 WO PCT/JP2012/057685 patent/WO2012137617A1/ja active Application Filing
- 2012-03-26 US US14/007,844 patent/US10515643B2/en active Active
- 2012-03-26 RU RU2013143624/08A patent/RU2571561C2/ru active
- 2012-03-26 PL PL16195433T patent/PL3154057T3/pl unknown
- 2012-03-26 TR TR2019/00411T patent/TR201900411T4/tr unknown
- 2012-03-26 KR KR1020137025380A patent/KR101569060B1/ko active IP Right Grant
- 2012-03-26 JP JP2013508811A patent/JP5603484B2/ja active Active
- 2012-03-26 EP EP18196322.4A patent/EP3441967A1/en active Pending
- 2012-03-26 ES ES12767213.7T patent/ES2617958T3/es active Active
-
2019
- 2019-11-18 US US16/687,176 patent/US11074919B2/en active Active
- 2019-11-18 US US16/687,144 patent/US11024319B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH03191628A (ja) * | 1989-12-21 | 1991-08-21 | Toshiba Corp | 可変レート符号化方式 |
JPH0470800A (ja) * | 1990-07-11 | 1992-03-05 | Sharp Corp | 音声情報圧縮装置 |
JPH09214348A (ja) * | 1996-01-30 | 1997-08-15 | Sony Corp | 信号符号化方法 |
JP2006011170A (ja) * | 2004-06-28 | 2006-01-12 | Sony Corp | 信号符号化装置及び方法、並びに信号復号装置及び方法 |
JP2010225949A (ja) | 2009-03-25 | 2010-10-07 | Kyocera Corp | 発熱体の放熱構造 |
WO2012046685A1 (ja) | 2010-10-05 | 2012-04-12 | 日本電信電話株式会社 | 符号化方法、復号方法、符号化装置、復号装置、プログラム、記録媒体 |
Non-Patent Citations (3)
Title |
---|
DAVID SALOMON: "Data Compression: The Complete Reference", 2004, SPRINGER-VERLAG |
ETSI TS 126 290 V6.3.0, June 2005 (2005-06-01) |
See also references of EP2696343A4 |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10482888B2 (en) | 2013-01-22 | 2019-11-19 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Apparatus and method for spatial audio object coding employing hidden objects for signal mixture manipulation |
JP2016508617A (ja) * | 2013-01-22 | 2016-03-22 | フラウンホーファー−ゲゼルシャフト・ツール・フェルデルング・デル・アンゲヴァンテン・フォルシュング・アインゲトラーゲネル・フェライン | 隠しオブジェクトを信号混合操作に使用する空間オーディオオブジェクト符号化の装置及び方法 |
EP3696812A1 (en) * | 2014-05-01 | 2020-08-19 | Nippon Telegraph and Telephone Corporation | Encoder, decoder, coding method, decoding method, coding program, decoding program and recording medium |
CN112820305A (zh) * | 2014-05-01 | 2021-05-18 | 日本电信电话株式会社 | 编码装置、编码方法、编码程序、记录介质 |
EP3139380A4 (en) * | 2014-05-01 | 2017-11-01 | Nippon Telegraph And Telephone Corporation | Encoding device, decoding device, encoding method, decoding method, encoding program, decoding program, and recording medium |
JP2018013795A (ja) * | 2014-05-01 | 2018-01-25 | 日本電信電話株式会社 | 符号化装置、復号装置、符号化方法、復号方法、符号化プログラム、復号プログラム、記録媒体 |
EP3509063A3 (en) * | 2014-05-01 | 2019-08-07 | Nippon Telegraph and Telephone Corporation | Encoder, decoder, coding method, decoding method, coding program, decoding program and recording medium |
JPWO2015166693A1 (ja) * | 2014-05-01 | 2017-04-20 | 日本電信電話株式会社 | 符号化装置、復号装置、符号化方法、復号方法、符号化プログラム、復号プログラム、記録媒体 |
WO2015166693A1 (ja) * | 2014-05-01 | 2015-11-05 | 日本電信電話株式会社 | 符号化装置、復号装置、符号化方法、復号方法、符号化プログラム、復号プログラム、記録媒体 |
EP3703051A1 (en) * | 2014-05-01 | 2020-09-02 | Nippon Telegraph and Telephone Corporation | Encoder, decoder, coding method, decoding method, coding program, decoding program and recording medium |
CN112820305B (zh) * | 2014-05-01 | 2023-12-15 | 日本电信电话株式会社 | 编码装置、编码方法、编码程序、记录介质 |
JP2017528751A (ja) * | 2014-07-28 | 2017-09-28 | サムスン エレクトロニクス カンパニー リミテッド | 信号符号化方法及びその装置、並びに信号復号方法及びその装置 |
US11616954B2 (en) | 2014-07-28 | 2023-03-28 | Samsung Electronics Co., Ltd. | Signal encoding method and apparatus and signal decoding method and apparatus |
US10827175B2 (en) | 2014-07-28 | 2020-11-03 | Samsung Electronics Co., Ltd. | Signal encoding method and apparatus and signal decoding method and apparatus |
US11367455B2 (en) | 2015-03-13 | 2022-06-21 | Dolby International Ab | Decoding audio bitstreams with enhanced spectral band replication metadata in at least one fill element |
US11417350B2 (en) | 2015-03-13 | 2022-08-16 | Dolby International Ab | Decoding audio bitstreams with enhanced spectral band replication metadata in at least one fill element |
US11664038B2 (en) | 2015-03-13 | 2023-05-30 | Dolby International Ab | Decoding audio bitstreams with enhanced spectral band replication metadata in at least one fill element |
US11842743B2 (en) | 2015-03-13 | 2023-12-12 | Dolby International Ab | Decoding audio bitstreams with enhanced spectral band replication metadata in at least one fill element |
US12094477B2 (en) | 2015-03-13 | 2024-09-17 | Dolby International Ab | Decoding audio bitstreams with enhanced spectral band replication metadata in at least one fill element |
Also Published As
Publication number | Publication date |
---|---|
EP2696343A4 (en) | 2014-11-12 |
US11074919B2 (en) | 2021-07-27 |
KR101569060B1 (ko) | 2015-11-13 |
ES2617958T3 (es) | 2017-06-20 |
KR20130133854A (ko) | 2013-12-09 |
PL3154057T3 (pl) | 2019-04-30 |
EP3154057B1 (en) | 2018-10-17 |
CN103460287A (zh) | 2013-12-18 |
US11024319B2 (en) | 2021-06-01 |
RU2013143624A (ru) | 2015-05-10 |
US20200090665A1 (en) | 2020-03-19 |
TR201900411T4 (tr) | 2019-02-21 |
JP5603484B2 (ja) | 2014-10-08 |
EP3154057A1 (en) | 2017-04-12 |
US10515643B2 (en) | 2019-12-24 |
US20140019145A1 (en) | 2014-01-16 |
US20200090664A1 (en) | 2020-03-19 |
EP3441967A1 (en) | 2019-02-13 |
EP2696343B1 (en) | 2016-12-21 |
EP2696343A1 (en) | 2014-02-12 |
CN103460287B (zh) | 2016-03-23 |
ES2704742T3 (es) | 2019-03-19 |
JPWO2012137617A1 (ja) | 2014-07-28 |
RU2571561C2 (ru) | 2015-12-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5603484B2 (ja) | 符号化方法、復号方法、符号化装置、復号装置、プログラム、記録媒体 | |
RU2554554C2 (ru) | Способ кодирования, кодер, способ определения величины периодического признака, устройство определения величины периодического признака, программа и носитель записи | |
US10083703B2 (en) | Frequency domain pitch period based encoding and decoding in accordance with magnitude and amplitude criteria | |
JP5612698B2 (ja) | 符号化方法、復号方法、符号化装置、復号装置、プログラム、記録媒体 | |
JP6867528B2 (ja) | 周期性統合包絡系列生成装置、周期性統合包絡系列生成方法、周期性統合包絡系列生成プログラム、記録媒体 | |
JP6595687B2 (ja) | 符号化方法、符号化装置、プログラム、および記録媒体 | |
JP5694751B2 (ja) | 符号化方法、復号方法、符号化装置、復号装置、プログラム、記録媒体 | |
WO2013180164A1 (ja) | 符号化方法、符号化装置、プログラム、および記録媒体 | |
JP5579932B2 (ja) | 符号化方法、装置、プログラム及び記録媒体 | |
JP5714172B2 (ja) | 符号化装置、この方法、プログラムおよび記録媒体 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201280015955.3 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12767213 Country of ref document: EP Kind code of ref document: A1 |
|
REEP | Request for entry into the european phase |
Ref document number: 2012767213 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012767213 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2013508811 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20137025380 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14007844 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2013143624 Country of ref document: RU Kind code of ref document: A |