CA2186433C - Speech coding apparatus having amplitude information set to correspond with position information - Google Patents

Speech coding apparatus having amplitude information set to correspond with position information Download PDF

Info

Publication number
CA2186433C
CA2186433C CA002186433A CA2186433A CA2186433C CA 2186433 C CA2186433 C CA 2186433C CA 002186433 A CA002186433 A CA 002186433A CA 2186433 A CA2186433 A CA 2186433A CA 2186433 C CA2186433 C CA 2186433C
Authority
CA
Canada
Prior art keywords
excitation
coding apparatus
pulse
pulses
amplitude
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CA002186433A
Other languages
French (fr)
Other versions
CA2186433A1 (en
Inventor
Kazunori Ozawa
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NEC Corp
Original Assignee
NEC Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NEC Corp filed Critical NEC Corp
Publication of CA2186433A1 publication Critical patent/CA2186433A1/en
Application granted granted Critical
Publication of CA2186433C publication Critical patent/CA2186433C/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/08Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
    • G10L19/10Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters the excitation function being a multipulse excitation

Abstract

The invention provides a speech coding apparatus by which good sound quality can be obtained even when the bit rate is low. The speech coding apparatus includes an excitation quantization circuit which quantizes an excitation signal using a plurality of pulses. The position of at least one of the pulses is represented by a number of bits determined in advance, and the amplitude of the pulse is determined in advance depending upon the position of the pulse.

Description

Speech Coding Apparatus having Amplitude ;
Information Set to Correspond with Position Information This invention relates to a speech coding apparatus, and more particularly to a speech coding apparatus which codes a speech~signal at a low bit rate and a high quality.
Various methods which code a speech signal in a high-efficiency manner are already known, and a representative one of the known methods is CELP (Code-Excited Linear-Predictive Coding) disclosed, for example, in M.
Schroeder and B. Atal, "Code-excited linear prediction:
High quality speech at low bit rates", Proc. ICASSP, 1985, pp.937-940 (hereinafter referred to as document 1) or Kleijn et al., "Improved speech quality and efficient vector quantization in SELP", Proc. ICASSP, 1988, pp.155-158 (hereinafter referred to as document 2). In those prior art methods, on the transmission side, spectrum parameters representative of a spectrum characteristic of a speech signal are extracted from the speech signal for each frame (for example, 20 ms) using a linear predictive (LPC) analysis. Each frame is r v divided into subframes (for example, of 5 ms), and for each subframe, parameters for an adaptive codebook (a delay parameter and a gain parameter corresponding to a pitch period) are extracted based on the excitation signal in the past, and then the speech signal of the subframe is pitch-predicted using the adaptive codebook.
Then, based on a residue signal' obtained by the pitch prediction, an optimum excitation code vector is selected from within an excitation codebook (vector quantization codebook) which includes predetermined kinds of noise signals, and an optimum gain is calculated to quantize the excitation signal. The selection of an excitation code vector is performed so as to minimize an error power between a signal synthesized based on the selected noise signal and the residue signal. Then, an index and a gain representative of the kind of the selected code vector as well as the spectrum parameter and the parameters of the adaptive codebook are combined and transmitted by a multiplexer section. Description of operation of the reception side is omitted herein.
The prior art coding described above is disadvantageous in that a large quantity of calculation is required for selection of an optimum excitation I
Y

code vector from within an excitation codebook. This arises from the fact that, with the coding methods of the documents 1 and 2, in order to select an excitation code vector, filtering or convolution calculation is performed once for code vectors, and such calculation is repeated a number of times equal to the number of code vectors stored in the codebook. For example, when the bit number of the codebook is B and the number of elements is N, if the filter or impulse response length upon filtering or convolution calculation is K, then the quantity of calculation required is N x K x 2g x 8,000/N
per one second. As an example, where B = 10, N = 40 and k = 10, 81,920,000 calculations are required. In this manner, the prior art coding is disadvantageous in that a very large quantity of calculation is required.
Various methods which achieve remarkable reduction in calculation quantity required for searching of an excitation codebook have been disclosed. One of the methods is an ACELP (Algebraic Code Excited Linear Prediction) method, which is disclosed, for example, in C. Laflamme et al., "16 kbps wideband speech coding technique based on algebraic CELP", Proc. ICASSP, 1991, pp.l3-16 (hereinafter referred to as document 3).
According to the method disclosed in the document 3, an excitation signal is represented by and transmitted as a plurality of pulses whose positions are represented by predetermined bit numbers. Here, since the amplitude of each pulse is limited to +1.0 or -1.0, no amplitude needs to be transmitted except the polarity of each pulse.
The polarity of each pulse is determined one-by-one from the speech signal, and fixed before searching for pulse positions. Consequently, the calculation quantity for searching of pulses can be reduced remarkably.
Further, while the method of the document 3 can reduce the calculation quantity remarkably, it is disadvantageous in that it does not provide a sufficiently high speech quality. The reason is that, since each pulse only has the polarity of positive or negative and its absolute amplitude is always 1.0 irrespective of the position of the pulse, the amplitudes of the pulses are quantized, but very roughly, resulting in low speech quality.
It is an object of the present invention to provide a speech coding apparatus which can code a speech signal with a comparatively small quantity of calculation and does not suffer from much deterioration , 2186+33 in picture quality even when the bit rate is low.
In order to attain the object described above, according to an aspect of the present invention, there is provided a speech coding apparatus for calculating a spectral parameter from a speech signal inputted thereto, quantizing an excitation signal of the speech signal using the spectral parameter and outputting the quantized excitation signal, comprising an excitation quantization section for quantizing the excitation signal using a plurality of pulses such that a position of at least one of the pulses is represented by a number of bits determined in advance and an amplitude of the pulse is determined in advance depending upon the position of the pulse.
In the speech coding apparatus, when the excitation quantization section forms M pulses for each fixed-interval of-time to quantize an excitation signal, where the amplitude and the position of the ith pulse are represented by q. and m., respectively, the excitation signal can be represented by the following equation (1):
M
v(n) = G E g. 8(n - m; ) (1) =t where G is the gain representative of the entire level.
For at least one pulse, for example, for two pulses, an amplitude value is determined in advance for each of the combinations of the positions of them, depending upon the positions of the pulses.
Preferably, the position which can be assumed by each pulse is limited in advance. The position of each pulse may be, for example, an even-numbered sample position, an odd-numbered sample position or every Lth sample position.
According to another aspect of the present invention, there is provided a speech coding apparatus for calculating a spectral parameter from a speech signal inputted thereto, quantizing an excitation signal of the speech signal using the spectral parameter and outputting the quantized excitation signal, comprising an excitation quantization section for quantizing the excitation signal using a plurality of pulses such that a position of at least one of the pulses is represented by a number of bits determined in advance and amplitudes of the plurality of pulses are quantized simultaneously.
In the speech coding apparatus, amplitude patterns representative of amplitudes of a plurality of pulses (for example, 2 pulses) for B bits (2~ amplitude J
patterns) in the equation (1) above are prepared as an amplitude codebook_in advance, and an optimum amplitude pattern is selected from among the amplitude patterns.
Also with thepresent speech coding apparatus, preferably the position which can be assumed by each pulse is limited in advance.-According to a further aspect of the present invention, there is a speech coding apparatus for calculating a spectral parameter from a speech signal inputted thereto, quantizing an excitation signal of the speech signal using the spectral parameter and outputting the quantized excitation signal, comprising a mode discrimination section for discriminating a mode from the speech signal inputted thereto and outputting discrimination information, and an excitation quantization section for quantizing the excitation signal using a plurality of pulses when the discriminationinformation from the mode discrimination section represents a specific mode such that a position of at least one of the pulses is represented by a number of bits determined in advance and an amplitude of the pulse is determined in advance depending upon the position of the pulse.
In the speech coding apparatus, an input signal _7_ is divided into frames, and a mode is discriminated for each frame using a characteristic amount. For example, four modes of 0to 3 may be used. The modes generally correspond to the following portions of the speech signal. In particular, mode 0: a silent/consonant portion, mode 1: a transition portion, mode 2: a weak steady portion of a vowel, and mode 3: a strong steady portion of a vowel. Then, when a frame is in a predetermined mode, for at least one pulse, for example, for two pulses, an amplitude value is determined for each of combinations of positions of them depending upon the positions of the pulses.
According to a still further aspect of the present invention, there is provided a speech coding apparatus for calculating a spectral parameter from a speech signal inputted thereto, quantizing an excitation signal of the speech signal using the spectral parameter and outputting the quantized excitation signal, comprising a mode discrimination section for discriminating a mode from the speech signal inputted thereto and outputting discrimination information, and an excitation quantization section for-quantizing the excitation signal using a plurality of pulses when the discrimination information from the mode discrimination _g_ 218b433 section represents a specific mode such that a position of at least one of the pulses is represented by a number of bits determined in advance and amplitudes of the plurality of pulses are quantized simultaneously.
In the speech coding apparatus, an input signal is divided into frames, and a mode is discriminated for each frame using a characteristic amount. Then, when a frame is in a predetermined mode, ampli-tude patterns representative of amplitudes of a plurality of pulses (for example, 2 pulses) for B bits (2g amplitude patterns) are prepared as an amplitude codebook in advance, and an optimum amplitude pattern is selected from among the amplitude patterns.
In summary, with the speech coding apparatus of the present invention, since the excitation quantization section quantizes the excitation signal using a plurality of pulses such that a position of at least one of the pulses is represented by a number of bits determined in advance and an amplitude of the pulse is determined in advance depending upon the position of the pulse or the amplitude of the pulse is learned in advance using a speech signal depending upon the position of the pulse, the speech quality is improved comparing with that obtained by the conventional methods while suppressing the amount of calculation for searching for an excitation low.
Further, with the speech coding apparatus, since it includes a codebook in order to quantize amplitudes of a plurality of pulses simultaneously, it is advantageous in that the speech quality is further improved compared with that obtained by the conventional methods while the amount of calculation for searching for an excitation low is suppressed.
The above and other objects, features and advantages of the present invention will become apparent from the following description and the appended claims, taken in conjunction with the accompanying drawings in which like parts or elements are denoted by like reference characters.
FIG. 1 is a block diagram of a speech coding apparatus showing a preferred embodiment of the present invention;
FIGS. 2 and 3 are similar views to FIG.1, but showing modifications to the speech coding apparatus of FIG. 1;
FIG. 4 is a similar view to FIG. l, but showing a further modification to the speech coding apparatus of FIG. 1;

FIGS. 5 to 7 are similar views to FIG. 4, but showing modifications to the modified speech coding apparatus of FIG. 4;
FIG. 8 is a similar view to FIGS. 4 to 7, but showing a speech coding apparatus according to another preferred embodiment of the present invention; and FIGS. 9 to 13 are similar views to FIGS. 4 to 8, but showing modifications to the speech coding apparatus of FIG. 8.
Referring to FIG. 1, there is shown in block diagram a speech coding apparatus according to a preferred embodiment of the present invention. The speech coding apparatus shown includes a framing circuit 110, a subframing circuit 120, a spectrum parameter calculation circuit 200, a spectrum parameter quantization circuit 210, an LSP codebook 211, a perceptual weighting circuit 230, a subtraction circuit 235, an adaptive codebook circuit 500, an excitation quantization circuit 350, a gain quantization circuit 365, a response signal calculation circuit 240, a weighting signal calculation circuit 360, an impulse response calculation circuit 310, a gain codebook 390 and a multiplexer 400.

When a speech signal is inputted from an input terminal 100, it is divided into frames (for example, of ms) by the framing circuit 110 and is further divided into subframes (for example, of 2 ms) shorter than the frames by the subframing circuit 120.
The spectrum parameter calculation circuit 200 applies a window (for example,''of 24 ms) longer than the subframe length to the speech signal of at least one subframe to cut out the speech signal, and calculates a predetermined order number (for example, P = 10 orders) of spectrum parameters. Here, for the calculation of spectrum parameters, an LPC analysis, a Burg analysis and so forth which are well-known in the art can be used. Here, the Burg analysis is used. Details of the Burg analysis are disclosed, for example, in T.
Nakamizo, "Signal Analysis and System Identification", Corona, 1988, pp.82-87 (hereinafter referred to as document 4), and since the Burg analysis is a known technique, description of it is omitted herein.
' Further, the spectrum parameter calculation circuit 200 converts linear predictive coefficients a.
(i - 1, ..., 10) calculated using the Burg method into LSP parameters suitable for quantization and interpolation. Such conversion from linear predictive coefficients into LSP parameters is disclosed in N.
Sugamura et al., "Speech Data Compression by LSP Speech Analysis-Synthesis Technique", Journal of the Electronic Communications Society of Japan, J64-A, 1981, pp.599-606 (hereinafter referred to as document 5). For example, linear predictive coefficients calculated for the second and fourth subframes based on the Burg method are converted into LSP parameters, whereas LSP parameters of the first and third subframes are determined by linear interpolation, and the LSP parameters of the first and third subframes are inversely converted back into linear predictive coefficients. Then, the linear predictive coefficients a~~ (i - 1, ..., 10, 1 = 1, ..., 5) of the first to fourth subframes are outputted to the perceptual weighting circuit 230. The LSP parameters of the fourth subframe are outputted to the spectrum parameter quantization circuit 210.
The spectrum parameter quantization circuit 210 efficiently quantizes the LSP parameters of a predetermined subframe and outputs a quantization value which minimizes the distortion of the following equation (2):
F
D; - E W(i) [LSP(i) - QLSP(i); ]2 (2) where LSP(i), QLSP(i); and W(i) are the LSP parameter of the ith-order before quantization, the ~th result after the quantization, and the weighting coefficient, respectively.
In the following description, it is assumed that vector quantization is used as a quantization method, and LSP parameters of the fourth subframe are quantized.
Any known technique can be employed as the technique for vector quantization of LSP parameters. Particularly, a technique disclosed in, for example, Japanese Patent Laid-Open Application No. Heisei 4-171500 (hereinafter referred to as document 6), Japanese Patent Laid-Open Application No. Heisei 4-363000 (hereinafter referred to as document 7), Japanese Patent Laid-Open Application No. Heisei 5-6199 (hereinafter referred to as document 8), T. Nomura et al., "LSP Coding VQ-SVQ with Interpolation in 4.075 kbps M-LCELP Speech Coder", Proc.
Mobile Multimedia Communications, 1993, pp.B.2.5 (hereinafter referred to as document 9) or the like can be used. Accordingly, description of details of the technique is omitted herein.
The spectrum parameter quantization circuit 210 regenerates the LSP parameters of the first to fourth subframes based on the LSP parameters quantized with the V~
fourth subframe. Here, linear interpolation of the quantization LSP parameters of the fourth subframe of the current frame and the quantization LSP parameters of the fourth subframe of the directly-preceding frame is performed to regenerate LSP parameters of the first to third subframes. Here, after a code vector which minimizes the error power between the LSP parameters before quantization and the LSP parameters after quantization is selected, the LSP parameters of the first to fourth subframes are regenerated by linear interpolation. In order to further improve the performance, after a plurality of candidates are first selected as a code vector which minimizes the error power, the accumulated distortion may be evaluated with regard to each of the candidates to select a set of a candidate and an interpolation LSP parameter which exhibit a minimum-accumulated distortion. Details are disclosed, for example, in Japanese Patent Laid-Open Application No. Heisei 6-222797 (hereinafter referred to as document 10).
The LSP parameters of the first to third subframes regenerated in such a manner as described above and the quantization LSP parameters of the fourth subframe are converted into linear predictive coefficients a'~~ (i - 1, ..., 10, 1 = l, ... 5) for each subframe, and the linear predictive coefficients a'~~ are outputted to the impulse response calculation circuit 310. Further, an index representative of the code vector of the quantization LSP parameters of the fourth subframe is outputted to the multiplexer 400.
The perceptual weighting circuit 230 receives the linear predictive coefficients a'm (i = 1, ..., 10, 1 = 1, ..., 5) before quantization for each subframe from the spectrum parameter calculation circuit 200, performs perceptual weighting for the speech signal of the subframe based on the technique of the document 1, and outputs a resulting perceptual weighting signal.
The response signal calculation circuit 240 receives the linear predictive coefficients a~~ for each subframe from the spectrum parameter calculation circuit 200, receives the linear predictive coefficients a'~~
regenerated by quantization and interpolation for each subframe from the spectrum parameter quantization circuit 210, calculates, for one subframe, a response signal with which the input signal is reduced to zero (d(n) - 0) using a value of a filter memory stored therein, and outputs the response signal to the .
218b433 subtraction circuit 235. Here, the response signal xE(n)is represented by the following equation (3):
Xz ~17~ _ ;o ;o ;a d Cn) -~a;d Cn-i) +Ea;y'y (n-i) +~a' ,y'xZ (n-i) (3) 3-> ;-1 ;-;
where, when n-i _< 0, y(n-i) = p(N + (n - i)) (4) x~ (n-i) = sw (N + (n - i) ) (5) where N is the subframe length, Y is the weighting coefficient for controlling the perceptual weighting amount and has a value equal to the value of an equation (7) given hereinbelow, and sw(n) and p(n) are an output signal of the weighting signal calculation circuit 360 and an output signal of the term of the denominator of a filter of the first term of the right side of the equation (7), respectively.
The subtraction circuit 235 subtracts response signals for one subframe from the perceptual weighting signal based on the following equation (6):
x'w(n) = xw(n) - xz(n) (6) and outputs the signal x'u(n) to the adaptive codebook circuit 500.
The impulse response calculation circuit 310 calculates a predetermined number L of impulse responses hW(n) of a perceptual weighting filter whose z conversion is represented by the following equation (7):
Io 1-~aiz-i HW (z) _ ~ I 1 (7) Io Io 1-~ai7~z ' 1-~ai7iz-i i-I i-1 and outputs them to the adaptive codebook circuit 500 and the excitation quantization circuit 350.
The adaptive codebook circuit 500 receives the excitation signal v(n) in the past from the gain quantization circuit 365, receives the output signal x'W(n) from the subtraction circuit 235 and the impulse responses hW(n) from the impulse response calculation circuit 310. Then, the adaptive codebook circuit 500 calculates a delay T corresponding to the pitch so that the distortion of the following equation (8) may be minimized, and outputs an index representative of the delay to the multiplexer 400.

DT=~xW2 (n) - C~xW Cn) yW (n-T) 7 2 / C~yW2 (n-T) 7 (8) n-0 n-0 n-0 where yW (n - T) - v(n - T)~hW (n) (9) where the symbol * signifies a convolution calculation.

a=~xw (n) yw (n-T) /Eyw2 Cn-T) (i 0) o-a n-o Here, in order to improve the extraction accuracy of a delay with regard to voice of a woman or a child, the delay may be calculated not as an integer sample value but a decimal fraction sample value. A
detailed method is disclosed, for example, in P. Kroon, "Pitch predictors with high terminal resolution", Proc.
ICASSP, 1990, pp.661-664 (hereinafter referred to as document 11).
Further, the adaptive codebook circuit 500 performs pitch prediction based on the following equation (J.1) and outputs a resulting predictive residue signal ew(n) to the excitation quantization circuit 350.
eu (n) = xw (n) - (3V (n - T)~hw (n) (11) The excitation quantization circuit 350 forms M pulses as described hereinabove. The excitation quantization circuit 350 quantizes the position of at least one pulse with a predetermined number of bits, and outputs an index representative of the position to the multiplexer 400. As a method of searching for the position of a pulse, various methods wherein the positions of pulses are searched for sequentially, pulse-by-pulse, have been proposed; one of the methods is disclosed, for example, in K. Ozawa et al, "A study on pulse search algorithms for multipulse excited speech coder realization" (hereinafter referred to as document 12). Therefore, description of details of the method is omitted herein. Also the method disclosed in the document 3 or a method which will be hereinafter described in connection with equations (16) to (21) may be employed instead.
In this instance, the amplitude of at least one pulse is determined depending upon the position of it.
Here, it is assumed that, as an example, the amplitudes of two pulses from among M pulses are determined in advance depending upon a combination of the positions of the two pulses.. If it is assumed now that if the first and second pulses can assume two different positions, four combinations of the positions of the pulses, that is, (1, 1), (1, 2), (2, 1) and (2, 2), are available, and corresponding to the combinations of the positions, available combinations of the amplitudes of the two pulses are, for example, (1.0, 1.0), (1.0, 0.1), (0.1, 1.0) and (0.1, 0.1). Since the amplitudes are determined in accordance with the combinations of the positions in advance, information for representation of the amplitudes need not be transmitted.
It is to be noted that the pulses other than the two pulses may have, for simplified operation, an amplitude such as, for example, 1.0 or -1.0 determined in advance without depending upon the positions.
The information of the amplitudes and the positions is outputted to the gain quantization circuit 365.
The gain quantization circuit 365 reads out gain code vectors from the gain codebook 390 and selects one of the gain code vectors so that, for the selected excitation code vector, the following equation (12) may be minimized. Here, it is assumed that both the adaptive codebook gain and the excitation gain are vector-quantized simultaneously.
N_1 M
Dk= ~ CxW (n) -a' Kv (n-T) *hW (n) -G' K ~g~ hW (n-m;) J 2 ( 1 2) n_0 where a'k and G'k are kth code vectors in a two-dimensional gain codebook stored in the gain codebook 390. An index representative of the selected gain code vector is outputted to the multiplexer 400.
The weighting signal calculation circuit 360 2 ~ ~s~3~
receives output parameters of the spectrum parameter calculation circuit 200 and the individual indices, and reads out code vectors corresponding to the indices.
Then, the weighting signal calculation circuit 360 calculates an excitation signal v(n) based on the following equation (13):
M
v(n) = p'kv(n - T) + G'k E 8(n - m. ) (13) The excitation signal v(n) is outputted to the adaptive codebook circuit 500.
Then, the weighting signal calculation circuit 360 calculates the response signal sw(n) for each subframe based on the following equation (14) using the output parameters of the spectrum parameter calculation circuit 200 and the output parameters of the spectrum parameter quantization circuit 210, and outputs the response signal sw(n) to the response-signal calculation circuit 240.
Sa (n) _ ;o ;o io v (n) -~a;v (n-i) +~a,y'p Cn-i) +~,a' ;y'sw (n-i) (14) ;-; ;
FIG. -2 shows in block diagram a modification to the speech coding apparatus of the first embodiment of , the present invention described hereinabove with reference to FIG. 1. Referring to FIG. 2, the modified speech coding apparatus is different from the speech coding apparatus of the first embodiment only in that it includes, in place of the excitation quantization circuit 350, an excitation quantization circuit 355 which operates in a somewhat different manner from the excitation quantization circuit 350, and additionally includes an amplitude pattern storage circuit 359. In the modified speech coding apparatus, amplitude values of pulses are stored as amplitude patterns in the amplitude pattern storage circuit 359, and position information of a pulse is inputted to the amplitude pattern storage circuit 359 to read out one of the amplitude patterns. Those patterns are learned using a data base of a large amount of speech data depending upon a combination of positions of pulses and is determined decisively depending upon positions.
FIG. 3 shows in block diagram another modification to the speech coding apparatus of the first embodiment of the present invention described hereinabove with reference to FIG. 1. Referring to FIG.
3, the modified speech coding apparatus shown is different from the speech coding apparatus of the first embodiment only in that it includes an excitation quantization circuit 357 in place of the excitation quantization circuit 350. In the modified speech coding apparatus, the position which may be assumed by each pulse is limited in advance by the excitation quantization circuit 357. The position of each pulse may be, for example, an even-numbered sample position, an odd-numbered sample position or every Lth sample position. Here, it is assumed that every Lth sample position is assumed, and the value of L is selected in accordance with the following equation:
L = N/M (15) where N and M are the subframe length and the number of pulses, respectively.
It is to be noted that the amplitude of at least one pulse may be determined in advance depending upon the position of the pulse.
FIG. 4 shows in block diagram a further modification to the speech coding apparatus of the first embodiment of the present invention described hereinabove with reference to FIG. 1. Referring to FIG.
4, the modified speech coding apparatus is different from the speech coding apparatus of the first embodiment only in that It includes an excitation quantization v circuit 450 in place of the excitation quantization circuit 350 and additionally includes a pulse amplitude codebook 451. In the modified speech coding apparatus, the excitation quantization circuit 450 calculates the positions of pulses by the same method as in the speech coding apparatus of the first embodiment, and quantizes and outputs the pulse positions~to the multiplexer 400 and the gain quantization circuit 365.
Further, the excitation quantization circuit 450 vector-quantizes the amplitudes of a plurality of pulses simultaneously. In particular, the excitation quantization circuit 450 reads out pulse amplitude code vectors from the pulse amplitude codebook 451 and selects one of the amplitude code vectors which minimizes the distortion of the following equation (16):

Dk= ~ LeW (n) -G Wig' ;khW (n-mi) J 2 (1 6) -o a-~
where G is the optimum gain, and g'~k is the ith pulse . amplitude of the kth amplitude code vector.
The minimization of the equation (16) can be formulated in the following manner. If the equation (16) is partially differentiated with the amplitude g'.
of a pulse and then set to 0, then Dk=~ Cew (n) /G] 2 - CEø (n) vk (n) ] 2 / [fig' Jkhw (n-mJ) 7 2 (1 7) n-~ n-~ 3-1 where N-I
~(n) _ ~ eu(i) hW(i-n), n = 0, ..., N-1 (18) i = n M
Vk (n) _ ~ g'; k S(n - mi ) (19) i = 1 Accordingly, the minimization of the equation (16) is equivalent to maximization of the second term of the right side of the equation (17).
The denominator of the second term of the right side of the equation (17) can be transformed into the following equation (20):
M
CEg' Jkhw (n-mi) ] 2 =
i-I
M / ~I c~ M-1 M
~g' ik2 ~ Ulli , mi) TG ~ ~g' ikg' Jk'Y Ulli > mJ) (L ~) J-J i-I J-i+I
where N-I
~(m~ , m; ) _ ~ h(n - i) h(n - ,~), ,)? i (21) n=j Accordingly, by calculating g'ik2 and g'ikg';k of the equation (20) for each amplitude code vector k in advance and storing them in a codebook, the quantity of calculation required can be reduced remarkably.
Further, if ~ and ~ are calculated once for each subframe, then the quantity of calculation can be further reduced. -The number of product sum calculations necessary for amplitude quantization in this instance is approximately N2 + [(M - 1)~ + M]2g + NL + M2B per subframe where M is the number of pulses per subframe, N
the subframe length, L the impulse response length, and B the bit number of the amplitude codebook. When B =
10, N = 40, M = 4 and L = 20, the quantity of product sum calculation is 3,347,200 per one second. Further, in searching for the position of a pulse, if the method 1 disclosed in the document 12 is used, then since no calculation quantity is produced newly with respect to the calculation quantity described above, the calculation quantity is reduced to approximately 1/24 comparing with those of the conventional methods of the documents 1 and 2.
Accordingly, it can be seen that, where the method of the present invention is-employed, the quantity of calculation required for searching for the amplitude and the position of a pulse is very small comparing with those of the conventional methods.

a The excitation quantization circuit 450 outputs an index of the amplitude code vector selected by the method described above to the multiplexer 400.
Further, the excitation quantization circuit 450 outputs the position of each pulse and the amplitude of each pulse by an amplitude code vector to the gain quantization circuit 365. ' The pulse amplitude codebook 451 can be replaced by a pulse polarity codebook. In that case, polarities of plural pulses are vector-quantized simultaneously.
FIG. 5 shows in block diagram a modification to the modified speech coding apparatus described hereinabove with reference to FIG. 4. Referring to FIG.
5, the modified speech coding apparatus is different from the modified speech coding apparatus of FIG. 4 in that it includes a single excitation and gain quantization circuit 550 in place of the excitation quantization circuit 450 and the gain quantization circuit 365. In the modified speech coding apparatus, the excitation and gain quantization circuit 550 performs both quantization of gains and quantization of amplitudes of pulses. The excitation and gain quantization circuit 550 calculates the positions of pulses and quantizes them using the same methods as s those employed in the excitation quantization circuit 450. The amplitude and the gain of a pulse are quantized simultaneously selecting a pulse amplitude code vector and a gain code vector from within the pulse amplitude codebook 451 and the gain codebook 390, respectively, so that the following equation (22) may be minimized. ' Dk= ~ CxW Cn) -a' kv (n-T) *hW (n) -G' k Wig' ,khW (n-m;) ] 2 (2 2) n-o ;-~
where g'~k is the ith pulse amplitude of the kth pulse amplitude code vector, a'k and G'k are kth code vectors of the two dimensional gain codebook stored in the gain codebook 390. From all combinations of pulse amplitude vectors and gain code vectors, one optimum combination can be selected so that the equation (22) above may be minimized.
Further, pre-selection may be introduced in order to reduce the searching calculation quantity. For example, a plurality of pulse amplitude code vectors are preliminarily selected in an ascending order of the distortion of the~equation (16) or (17), and a gain codebook is searched for each candidate, whereafter, from the thus-searched-out gain codebooks, one combination of a pulse amplitude code vector and a gain code vector which minimizes the equation (22) is selected.
Then, an index representative of the selected pulse amplitude code vector and gain code vector is outputted to the multiplexer 400.
The pulse amplitude codebook 451 can be replaced by a pulse polarity codebook. ,In that case, polarities of plural pulses are vector-quantized simultaneously.
FIG. 6 shows in block diagram another modification to the modified speech coding apparatus described hereinabove with reference to FIG. 4.
Referring to FIG. 6, the modified speech coding apparatus is different from the modified speech coding apparatus of FIG. 4 only in that it includes a pulse amplitude trained codebook 580 in place of the pulse amplitude codebook 451. The pulse amplitude trained codebook 580 is produced by training in advance, using a speech signal, a codebook for simultaneous quantization of the amplitudes or polarities of a plurality of pulses. A training method for the codebook is disclosed, for example, in Linde et al., "An algorithm for vector quantization design", IEEE Trans. Commun., January 1980, pp.84-95 (hereinafter referred to as document 13).

It is to be noted that the modified speech coding apparatus of FIG. 6 may be further modified such that a gain is quantized with a gain codebook while a pulse amplitude is quantized with a pulse amplitude codebook similarly as in the speech coding apparatus of FIG. S-.
FIG. 7 shows in block diagram a further modification to the modified speech coding apparatus described hereinabove with reference to FIG. 4.
Referring to FIG. 7, the modified speech coding apparatus is different from the modified speech coding apparatus of FIG. 4 only in that it includes an excitation quantization circuit 470 in place of the excitation quantization circuit 450. In particular, the position which can be assumed by each pulse is limited in advance. The position of each pulse may be, for example, an even-numbered sample position, an odd-numbered sample position or every Lth sample position.
Here, it is assumed that every Lth sample position is used, and the value of L is selected in accordance with the equation (13) given hereinabove.
It is to be noted that the amplitudes or polarities of a plurality of pulses may be quantized simultaneously using a codebook.

f FIG. 8 shows in block diagram a speech coding apparatus according to another preferred embodiment of the present invention. Referring to FIG. 8, the speech coding apparatus is a modification to the speech coding apparatus of the first embodiment described hereinabove with reference to FIG. 1. The speech coding apparatus of the present embodiment is different from the speech coding apparatus of the first embodiment in that it includes an excitation quantization circuit 600 in place of the excitation quantization circuit 350 and additionally includes a mode discrimination circuit 800.
The mode discrimination circuit 800 receives a perceptual weighting signal in units of a frame from the perceptual weighting circuit 230 and outputs mode discrimination information. Here, a characteristic amount of a current frame is used for discrimination of a mode. The characteristic amount may be, for example, a pitch-predictive gain averaged in a frame. For calculation of the pitch-predictive gain, for example, the following equation (23) is used:
L
G = 101ogi o [ (1/L) E (P. /E. ) ] (23) =i where L is the length of subframes included in the frame, and P. and E. are the speech power and the pitch predictive error power of the ith subframe, respectively.
N-I
P;=~,Xw3R Cn) C24) OHO
N_1 N_I
E;=P,- C~xw; Cn) xw; Cn-T) 7 2 / C~xw;2 Cn-T) 7 C25) n~0 n~0 where T is the optimum delay which maximizes the predictive gain.
The frame average pitch predictive gain G is compared with a plurality of threshold values to classify it into a plurality of different modes.
The number of modes may be, for example, 4. The mode discrimination circuit 800 outputs the mode identification information to the excitation quantization circuit 600 and the multiplexer 400.
The excitation quantization circuit 600 performs the following processing when the mode identification information represents a predetermined mode.
Where M pulses are to be determined as seen from the equation (1) given hereinabove, the excitation quantization circuit 600 quantizes the position of at least one pulse with a predetermined number of bits and outputs an index representative of the position to the i multiplexer 400. In this instance, the amplitude of the at least one pulse is determined depending upon the position in advance.
Here, it is assumed that, as an example, the amplitudes of two pulses from among M pulses are determined in advance depending upon a combination of the positions of the two pulses. If it is assumed now that if the first and second pulses can assume two different positions, four combinations of the positions of the two pulses, that is, (1, 1), (1, 2), (2, 1) and (2, 2), are available, and corresponding to the combinations of the positions, available combinations of the amplitudes of the two pulses are, for example, (1.0, 1.0), (1.0, 0.1), (0.1, 1.0) and (0..1, 0.1). Since the amplitudes are determined in accordance with the combinations of the positions in advance, information for representation of the amplitudes need not be transmitted.
It is to be noted that the pulses other than the two pulses may have, for simplified operation, an amplitude such as, for example, 1.0 or -1.0 determined in advance without depending upon the positions.
The information of the amplitudes and the positions is outputted to the gain quantization circuit 218b433 365. -FIG. 9 shows in block diagram a modification to the speech coding apparatus of the embodiment described hereinabove with reference to FIG. 8. Referring to FIG.
9, the modified speech coding apparatus is different from the speech coding apparatus of FIG. 8 only in that it includes an excitation quantization circuit 650 in place of the excitation quantization circuit 600 and additionally includes an amplitude pattern storage circuit 359. The excitation quantization circuit 650 receives discrimination information from the mode discrimination circuit 800 and, when the discrimination information represents a predetermined mode, the excitation quantization circuit 650 receives position information of a pulse to read out one of patterns of amplitude values of pulses from the amplitude pattern storage circuit 359.
Those patterns are trained using a data base of a large amount of speech data depending upon a combination of positions of pulses and is determined decisively depending upon positions. The training method disclosed in the document 13 mentioned hereinabove can be used as the training method in this instance.

FIG. 10 shows in block diagram another modification to the speech coding apparatus of the embodiment described hereinabove with reference to FIG.
8. Referring to FIG. 10, the modified speech coding apparatus is different from the speech coding apparatus of FIG. 8 only in that it includes an excitation quantization circuit 680 in place of the excitation quantization circuit 600. The excitation quantization circuit 680 receives discrimination information from the mode discrimination circuit 800 and, when the discrimination information represents a predetermined mode, the position which can be assumed by each pulse is limited in advance. The position of each pulse may be, for example, an even-numbered sample position, an odd-numbered sample position or every Lth sample position.
Here, it is assumed that every Lth sample position is assumed, and the value of L is selected in accordance with the equation (15) given hereinabove.
It is to be noted that the amplitude of at least one pulse may be learned as an amplitude pattern in advance depending upon the position of the pulse.
FIG. 11 shows in block diagram a further modification to the speech coding apparatus of the embodiment described hereinabove with reference to FIG.

8. Referring to FIG. 11, the modified speech coding apparatus is different from the speech coding apparatus of FIG. 8 only in that it includes an excitation quantization circuit 700 in place of the excitation quantization circuit 600 and additionally includes a pulse-amplitude codebook 451. The excitation quantization circuit 700 receives discrimination information from the mode discrimination circuit 800 and, when the discrimination information represents a predetermined mode, the excitation quantization circuit 700 quantizes the position of at least one pulse with a predetermined number of bits and outputs an index to the gain quantization circuit 365 and the multiplexer 400. Then, the excitation quantization circuit 700 vector quantizes the amplitudes of a plurality of pulses simultaneously. Then, the excitation quantization circuit 700 reads out pulse-amplitude code vectors from the pulse amplitude codebook 451 and selects one of the amplitude code vectors which minimizes the distortion of the equation (14) given hereinabove. Then, the excitation quantization circuit 700 outputs an index of the selected amplitude node vector to the gain quantization circuit 365 and the multiplexer 400.
It is to be noted that the modified speech t coding apparatus of FIG. 11 may be further modified such that a gain is quantized with a gain codebook while a pulse amplitude is quantized with a pulse amplitude codebook using the equation (17) given hereinabove.
FIG. 12 shows in block diagram a still further modification to the speech coding apparatus of the embodiment described hereinabove with reference to FIG.
8. Referring to FIG. 12, the modified speech coding apparatus is different from the speech coding apparatus of FIG. 8 only in that it includes an excitation quantization circuit 750 in place of the excitation quantization circuit 600 and additionally includes a pulse-amplitude trained codebook 580. The excitation quantization circuit 750 receives discrimination information from the mode discrimination circuit 800 and, when the discrimination information represents a predetermined mode, the excitation quantization circuit 750 quantizes the position of at least one pulse with a predetermined number of bits and outputs an index to the gain quantization circuit 365 and the multiplexer 400. Then, the excitation quantization circuit 750 vector quantizes the amplitudes of a plurality of pulses simultaneously. Then, the excitation quantization circuit 750 reads out pulse-amplitude code vectors trained in advance from the pulse-amplitude training codebook 580 and selects one of the amplitude code vectors which minimizes the distortion of the equation (14) given hereinabove. Then, the excitation quantization circuit 750 outputs an index of the selected amplitude code vector to the gain quantization circuit 365 and the multiplexes 400.
It is to be noted that the modified speech coding apparatus of FIG. 12 may be further modified such that a gain is quantized with a gain codebook while a pulse amplitude is quantized with a pulse-amplitude codebook using the equation (22) given hereinabove.
FIG. 13 shows in block diagram a yet further modification to the speech coding apparatus of the embodiment described hereinabove with reference to FIG.
8. Referring to FIG. 13, the modified speech coding apparatus is different from the speech coding apparatus of FIG. 8 only in that it includes an excitation quantization circuit 780 in place of the excitation quantization circuit 600 and additionally includes a pulse-amplitude codebook 451. The excitation quantization circuit 780 receives discrimination information from the mode discrimination circuit 800 and, when the discrimination information represents a r predetermined mode, the excitation quantization circuit 700 quantizes the position of at least one pulse with a predetermined number of bits and outputs an index to the gain quantization circuit 365 and the multiplexes 400. Here, the position which can be assumed by each pulse is limited in advance. The position of each pulse may be, for example, an even-numbered sample position, an odd-numbered sample position or every Lth sample position. Here, it is assumed that every Lth sample position is assumed, and the value of L is selected in accordance with the equation (15) given hereinabove.
Then, the excitation quantization circuit 780 outputs an index to the gain quantization circuit 365 and the multiplexes 400.
It is to be noted that the modified speech coding apparatus of FIG. 13 may be further modified such that a gain is quantized with a gain codebook while a pulse amplitude is quantized with a pulse-amplitude codebook using the equation (22) given hereinabove.
It is to be noted that such a codebook trained in advance as described hereinabove in connection with the modified speech coding apparatus of FIG. 11 may be used as the pulse-amplitude codebook 451 in any of the speech coding apparatus of the embodiments described r y hereinabove which include such pulse-amplitude codebook 451.
It is to be noted that the speech coding apparatus of the embodiment of FIG. 8 and the modifications to it may be modified such that the mode discrimination information from the mode discrimination circuit is used to change over'the adaptive codebook circuit or the gain codebook.
Having now fully described the invention, it will be apparent to one of ordinary skill in the art that many changes and modifications can be made thereto without departing from the spirit and scope of the invention as set forth herein.

Claims (12)

1. A speech coding apparatus for calculating a spectral parameter from a speech signal inputted thereto, quantizing an excitation signal of the speech signal using the spectral parameter and outputting the quantized excitation signal, comprising:
an excitation quantization section for quantizing the excitation signal using a plurality of pulses such that at least one of the pulses has a position represented by a number of bits determined in advance and wherein an amplitude corresponding to the position is determined in advance.
2. A speech coding apparatus as claimed in claim 1, wherein said excitation quantization section trains the amplitude of the at least one pulse in advance using the speech signal depending upon the position of the at least one pulse.
3. A speech coding apparatus as claimed in claim 1 or 2, wherein said excitation quantization section is constructed such that the position which can be assumed by the at least one pulse is limited in advance.
4. A speech coding apparatus for calculating a spectral parameter from a speech signal inputted thereto, quantizing an excitation signal of the speech signal using the spectral parameter and outputting the quantized excitation signal, comprising:
an excitation quantization section for quantizing the excitation signal using a plurality of pulses such that at least one of the pulses has a position represented by a number of bits determined in advance and wherein an amplitude or polarity of the at least one pulse is quantized simultaneously with amplitudes and polarities of the remaining plurality of pulses.
5. A speech coding apparatus as claimed in claim 4, further comprising a code book determined by designing using the speech signal in advance, said excitation quantization section using said code book to quantize the amplitudes or polarities of the plurality of pulses simultaneously.
6. A speech coding apparatus as claimed in claim 4 or 5, wherein said excitation quantization section is constructed such that the position which can be assumed by the at least one pulse is limited in advance.
7. A speech coding apparatus for calculating a spectral parameter from a speech signal inputted thereto, quantizing an excitation signal of the speech signal using the spectral parameter and outputting the quantized excitation signal, comprising:
a mode discrimination section for discriminating a mode from the speech signal inputted thereto and outputting discrimination information; and an excitation quantization section for quantizing the excitation signal using a plurality of pulses when the discrimination information from said mode discrimination section represents a specific mode such that at least one of the pulses has a position represented by a number of bits determined in advance and wherein an amplitude corresponding to the position is determined in advance.
8. A speech coding apparatus as claimed in claim 7, wherein said excitation quantization section trains and determines the amplitude of the at least one pulse in advance using the speech signal depending upon the position of the at lease one pulse.
9. A speech coding apparatus as claimed in claim 7 or 8, wherein said excitation quantization section is constructed such that the position which can be assumed by the at least one pulse is limited in advance.
10. A speech coding apparatus for calculating a spectral parameter from a speech signal inputted thereto, quantizing an excitation signal of the speech signal using the spectral parameter and outputting the quantized excitation signal, comprising:
a mode discrimination section for discriminating a mode from the speech signal inputted thereto and outputting discrimination information; and an excitation quantization section for quantizing the excitation signal using a plurality of pulses when the discrimination information from said mode discrimination section represents a specific mode such that at least one of the pulses has a position represented by a number of bits determined in advance and wherein an amplitude or polarity of the at least one pulse is quantized simultaneously with amplitudes and polarities of the remaining plurality of pulses.
11. A speech coding apparatus as claimed in claim 10, further comprising a code book determined by training using the speech signal in advance, said excitation quantization section using said code book to quantize the amplitudes or polarities of the plurality of pulses simultaneously.
12. A speech coding apparatus as claimed in claim 10 or 11, wherein said excitation quantization section is constructed such that the position which can be assumed by the at least one pulse is limited in advance.
CA002186433A 1995-09-27 1996-09-25 Speech coding apparatus having amplitude information set to correspond with position information Expired - Fee Related CA2186433C (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP24988995A JP3196595B2 (en) 1995-09-27 1995-09-27 Audio coding device
JP249889/1995 1995-09-27

Publications (2)

Publication Number Publication Date
CA2186433A1 CA2186433A1 (en) 1997-03-28
CA2186433C true CA2186433C (en) 2001-03-27

Family

ID=17199725

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002186433A Expired - Fee Related CA2186433C (en) 1995-09-27 1996-09-25 Speech coding apparatus having amplitude information set to correspond with position information

Country Status (5)

Country Link
US (1) US5826226A (en)
EP (1) EP0766232B1 (en)
JP (1) JP3196595B2 (en)
CA (1) CA2186433C (en)
DE (1) DE69636209T2 (en)

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6393391B1 (en) 1998-04-15 2002-05-21 Nec Corporation Speech coder for high quality at low bit rates
JP3094908B2 (en) * 1996-04-17 2000-10-03 日本電気株式会社 Audio coding device
DE69712539T2 (en) * 1996-11-07 2002-08-29 Matsushita Electric Ind Co Ltd Method and apparatus for generating a vector quantization code book
US5970444A (en) * 1997-03-13 1999-10-19 Nippon Telegraph And Telephone Corporation Speech coding method
CA2233896C (en) * 1997-04-09 2002-11-19 Kazunori Ozawa Signal coding system
CA2684452C (en) * 1997-10-22 2014-01-14 Panasonic Corporation Multi-stage vector quantization for speech encoding
JP3180762B2 (en) 1998-05-11 2001-06-25 日本電気株式会社 Audio encoding device and audio decoding device
JP3998330B2 (en) * 1998-06-08 2007-10-24 沖電気工業株式会社 Encoder
JP3319396B2 (en) * 1998-07-13 2002-08-26 日本電気株式会社 Speech encoder and speech encoder / decoder
US6480822B2 (en) 1998-08-24 2002-11-12 Conexant Systems, Inc. Low complexity random codebook structure
US6823303B1 (en) * 1998-08-24 2004-11-23 Conexant Systems, Inc. Speech encoder using voice activity detection in coding noise
US6493665B1 (en) * 1998-08-24 2002-12-10 Conexant Systems, Inc. Speech classification and parameter weighting used in codebook search
JP3824810B2 (en) * 1998-09-01 2006-09-20 富士通株式会社 Speech coding method, speech coding apparatus, and speech decoding apparatus
ATE328407T1 (en) * 1998-09-11 2006-06-15 Motorola Inc METHOD FOR CODING INFORMATION SIGNALS
CA2348659C (en) * 1999-08-23 2008-08-05 Kazutoshi Yasunaga Apparatus and method for speech coding
US6782360B1 (en) * 1999-09-22 2004-08-24 Mindspeed Technologies, Inc. Gain quantization for a CELP speech coder
US6959274B1 (en) 1999-09-22 2005-10-25 Mindspeed Technologies, Inc. Fixed rate speech compression system and method
US6842733B1 (en) 2000-09-15 2005-01-11 Mindspeed Technologies, Inc. Signal processing system for filtering spectral content of a signal for speech coding
US6850884B2 (en) * 2000-09-15 2005-02-01 Mindspeed Technologies, Inc. Selection of coding parameters based on spectral content of a speech signal
WO2002071395A2 (en) * 2001-03-02 2002-09-12 Matsushita Electric Industrial Co., Ltd. Apparatus for coding scaling factors in an audio coder
JP3582589B2 (en) * 2001-03-07 2004-10-27 日本電気株式会社 Speech coding apparatus and speech decoding apparatus
TW564400B (en) * 2001-12-25 2003-12-01 Univ Nat Cheng Kung Speech coding/decoding method and speech coder/decoder
AU2003211229A1 (en) * 2002-02-20 2003-09-09 Matsushita Electric Industrial Co., Ltd. Fixed sound source vector generation method and fixed sound source codebook
ES2305852T3 (en) * 2003-10-10 2008-11-01 Agency For Science, Technology And Research PROCEDURE FOR CODING A DIGITAL SIGNAL IN A SCALABLE BINARY FLOW, PROCEDURE FOR DECODING A SCALABLE BINARY FLOW.
AU2008222241B2 (en) * 2007-03-02 2012-11-29 Panasonic Intellectual Property Corporation Of America Encoding device and encoding method
US20110026581A1 (en) * 2007-10-16 2011-02-03 Nokia Corporation Scalable Coding with Partial Eror Protection
KR101747524B1 (en) * 2015-06-22 2017-06-16 송태원 Fish trap

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA1229681A (en) * 1984-03-06 1987-11-24 Kazunori Ozawa Method and apparatus for speech-band signal coding
CA1255802A (en) * 1984-07-05 1989-06-13 Kazunori Ozawa Low bit-rate pattern encoding and decoding with a reduced number of excitation pulses
NL8500843A (en) * 1985-03-22 1986-10-16 Koninkl Philips Electronics Nv MULTIPULS EXCITATION LINEAR-PREDICTIVE VOICE CODER.
GB8630820D0 (en) * 1986-12-23 1987-02-04 British Telecomm Stochastic coder
CA1337217C (en) * 1987-08-28 1995-10-03 Daniel Kenneth Freeman Speech coding
JP2903533B2 (en) * 1989-03-22 1999-06-07 日本電気株式会社 Audio coding method
JP2940005B2 (en) * 1989-07-20 1999-08-25 日本電気株式会社 Audio coding device
JP3114197B2 (en) * 1990-11-02 2000-12-04 日本電気株式会社 Voice parameter coding method
JP3151874B2 (en) * 1991-02-26 2001-04-03 日本電気株式会社 Voice parameter coding method and apparatus
JP3143956B2 (en) * 1991-06-27 2001-03-07 日本電気株式会社 Voice parameter coding method
JP3053464B2 (en) * 1991-07-03 2000-06-19 富士通株式会社 Audio coding device
JP2746039B2 (en) * 1993-01-22 1998-04-28 日本電気株式会社 Audio coding method
DE69426860T2 (en) * 1993-12-10 2001-07-19 Nec Corp Speech coder and method for searching codebooks
US5602961A (en) * 1994-05-31 1997-02-11 Alaris, Inc. Method and apparatus for speech compression using multi-mode code excited linear predictive coding
FR2720850B1 (en) * 1994-06-03 1996-08-14 Matra Communication Linear prediction speech coding method.

Also Published As

Publication number Publication date
CA2186433A1 (en) 1997-03-28
JP3196595B2 (en) 2001-08-06
JPH0990995A (en) 1997-04-04
US5826226A (en) 1998-10-20
EP0766232A2 (en) 1997-04-02
DE69636209T2 (en) 2007-04-05
EP0766232B1 (en) 2006-06-07
DE69636209D1 (en) 2006-07-20
EP0766232A3 (en) 1998-06-03

Similar Documents

Publication Publication Date Title
CA2186433C (en) Speech coding apparatus having amplitude information set to correspond with position information
US6023672A (en) Speech coder
CA2271410C (en) Speech coding apparatus and speech decoding apparatus
EP1005022B1 (en) Speech encoding method and speech encoding system
EP0834863B1 (en) Speech coder at low bit rates
US7680669B2 (en) Sound encoding apparatus and method, and sound decoding apparatus and method
CA2336360C (en) Speech coder
US6751585B2 (en) Speech coder for high quality at low bit rates
US5884252A (en) Method of and apparatus for coding speech signal
EP1154407A2 (en) Position information encoding in a multipulse speech coder
EP1100076A2 (en) Multimode speech encoder with gain smoothing
JP3144284B2 (en) Audio coding device
JP3299099B2 (en) Audio coding device
JP3471542B2 (en) Audio coding device
JPH09319399A (en) Voice encoder

Legal Events

Date Code Title Description
EEER Examination request
MKLA Lapsed