CA2177422C - Voice/unvoiced classification of speech for use in speech decoding during frame erasures - Google Patents

Voice/unvoiced classification of speech for use in speech decoding during frame erasures Download PDF

Info

Publication number
CA2177422C
CA2177422C CA002177422A CA2177422A CA2177422C CA 2177422 C CA2177422 C CA 2177422C CA 002177422 A CA002177422 A CA 002177422A CA 2177422 A CA2177422 A CA 2177422A CA 2177422 C CA2177422 C CA 2177422C
Authority
CA
Canada
Prior art keywords
signal
speech
decoder
scale
factor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
CA002177422A
Other languages
French (fr)
Other versions
CA2177422A1 (en
Inventor
Peter Kroon
Yair Shoham
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
AT&T Corp
Original Assignee
AT&T IPM Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by AT&T IPM Corp filed Critical AT&T IPM Corp
Publication of CA2177422A1 publication Critical patent/CA2177422A1/en
Application granted granted Critical
Publication of CA2177422C publication Critical patent/CA2177422C/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B7/00Radio transmission systems, i.e. using radiation field
    • H04B7/24Radio transmission systems, i.e. using radiation field for communication between two or more posts
    • H04B7/26Radio transmission systems, i.e. using radiation field for communication between two or more posts at least one of which is mobile
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/93Discriminating between voiced and unvoiced parts of speech signals
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L2019/0001Codebooks
    • G10L2019/0003Backward prediction of gain
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/93Discriminating between voiced and unvoiced parts of speech signals
    • G10L2025/932Decision in previous or following frames

Abstract

A speech decoder includes a first portion comprising an adaptive codebook and a second portion comprising a fixed codebook. The decoder generates a speech excitation signal selectively based on output signals from said first and second portions when said decoder fails to receive reliably at least a portion of a current frame of compressed speech information. The decoder does this by classifying the speech signal to be generated as periodic or non-periodic and then generating an excitation signal based on this classification. If the speech signal is classified as periodic, the excitation signal is generated based on the output signal from the first portion and not on the output signal from the second portion. If the speech signal is classified as non-periodic, the excitation signal is generated based on the output signal from said second portion and not on the output signal from said first portion.

Description

1 ~177~22 VOICE/UNVOICED CLASSIFICATION OF SPEECH FOR USE IN SPEECH
DECODING DURING FRAME ERASURES

S Field of the Invention The present invention relates generally to speech coding arrangements for use inco~ lunication systems, and more particularly to the ways in which such speech coders function in the event of burst-like errors in tr~3n~mi~ion.

Ba~k~round of the Invention Many collllllunication systems, such as cellular telephone and personal communications systems, rely on wireless channels to collllllunicate information. In the course of collllllullicating such information, wireless colllll~unication channels can suffer from several sources of error, such as multipath fading. These error sources can cause, among other things, the problem offrame erasure. Erasure refers to the total loss or whole or partial corruption of a set of bits communicated to a receiver. A frame is a predetermined fixed number of bits which may be collllllunicated as a block through a communication channel. A frame may therefore represent a time-segment of a speech signal.
If a frame of bits is totally lost, then the receiver has no bits to inte~ t. Under such ch.;ulllsLallces, the receiver may produce a me~ningless result. If a frame of received bits is corrupted and therefore unreliable, the receiver may produce a severely distorted result. In either case, the frame of bits may be thought of as "erased" in that the frame is unavailable or unusable by the receiver.
As the demand for wireless system capacity has increased, a need has arisen to make the best use of available wireless system bandwidth. One way to enhance theefficient use of system bandwidth is to employ a signal compression technique. For wireless systems which carry speech signals, speech col~ ssion (or speech coding) techniques may be employed for this purpose. Such speech coding techniques include 2177~22
- 2 analysis-by-synthesis speech coders, such as the well-known Code-Excited Linear Prediction (or CELP) speech coder.
The problem of packet loss in packet-switched networks employing speech coding arrangements is very similar to frame erasure in the wireless context. That is, due to 5 packet loss, a speech decoder may either fail to receive a frame or receive a frame having a significant number of missing bits. In either case, the speech decoder is presented with the same essential problem -- the need to synthesize speech despite the loss of compressed speech information. Both "frame erasure" and "packet loss" concern a communication channel (or network) problem which causes the loss of transmitted bits.
10 For purposes of this description, the term "frame erasure" may be deemed to include "packet loss."
Among other things, CELP speech coders employ a codebook of excitation signals to encode an original speech signal. These excitation signals, scaled by an excitation gain, are used to "excite" filters which synthesize a speech signal (or some 15 precursor to a speech signal) in response to the excitation. The synthesized speech signal is compared to the original speech signal. The codebook excitation signal is identified which yields a synthesized speech signal which most closely matches the original signal.
The identified excitation signal's codebook index and gain representation (which is often itself a gain codebook index) are then communicated to a CELP decoder (depending upon 20 the type of CELP system, other types of information, such as linear prediction (LPC) filter coefficients, may be colllllll..lic~t~d as well). The decoder contains codebooks identical to those of the CELP coder. The decoder uses the transmitted indices to select an excitation signal and gain value. This selected scaled excitation signal is used to excite the decoder's LPC filter. Thus excited, the LPC filter of the decoder generates a decoded (orquantized) 25 speech signal -- the same speech signal which was previously determined to be closest to the original speech signal.
Some CELP systems also employ other components, such as a periodicity model (e.g., a pitch-predictivefilter or an adaptive codebook). Such a model sim~ s the periodicity of voiced speech. In such CELP systems, parameters relating to these components must also be sent to the decoder. In the case of an adaptive codebook, signals representing a pitch-period (delay) and adaptive codebook gain must also be sent to the decoder so that the decoder can recreate the operation of the adaptive codebook in the speech synthesis process.
Wireless and other systems which employ speech coders may be more sensitive to the problem of frame erasure than those systems which do not compress speech. This sensitivity is due to the reduced redundancy of coded speech (compared to uncoded speech) making the possible loss of each tr~n.~mitted bit more significant. In the context of a CELP speech coders experiencing frame erasure, excitation signal codebook indices and other signals representing speech in the frame may be either lost or substantially corrupted preventing proper synthesis of speech at the decoder. For example, because of the erased frame(s), the CELP decoder will not be able to reliably identify which entry in its codebook should be used to synthesize speech. As a result, speech coding system performance may degrade significantly.
Because frame erasure causes the loss of excitation signal codebook indicies, LPC
coefficients, adaptive codebook delay information, and adaptive and fixed codebook gain information, normal techniques for synthesizing an excitation signal in a speech decoder are ineffective. Therefore, these normal techniques must be replaced by alternative measures.
SummarY of the Invention In accordance with the present invention, a speech decoder includes a first portion co~ lising an adaptive codebook and a second portion comprising a fixed codebook. The decoder generating a speech excitation signal selectively based on output signals from said first and second portions when said decoder fails to receive reliably at least a portion of a current frame of colllylcssed speech information. The decoder does this by classifying the speech signal to be generated as periodic or non-periodic and then generating an excitation signal based on this classification.

2177~22 If the speech signal is classified as periodic, the excitation signal is generated based on the output signal from the first portion and not on the output signal from the second portion. If the speech signal is classified as non-periodic, the excitation signal is generated based on the output signal from said second portion and not on the output signal from said 5 first portion.
See sections II.B. l . and 2. of the Detailed Description for a discussion relating to the present invention.

Brief Des~ Jti- ~ of the Drawin~
Figure 1 presents a block diagram of a G.729 Draft decoder modified in accordance with the present invention.
Figure 2 presents an illustrative wireless communication system employing the embodiment of the present invention presented in Figure 1 .

Detailed Descri~tion I. Introduction The present invention concerns the operation of a speech coding system experiencing frame erasure -- that is, the loss of a group of consecutive bits in the compressed bit-stream, which group is ordinarily used to synthesize speech. The description which follows concerns features of the present invention applied illustratively to an 8 kbit/s CELP speech coding system proposed to the 11 U for adoption as its international standard G.729. For the convenience of the reader, a preliminary draft recommendation for the G.729 standard is attached hereto as an Appendix (the draft will be referred to herein as the "G.729 Draft"). The G.729 Draft includes detailed descriptions of the speech encoder and decoder (see G.729 Draft, sections 3 and 4, respectively). The illustrative embodiment of the present invention is directed to modifications of normal G.729 decoder operation, as detailed in G.729 Draft section 4.3.
No modifications to the encoder are required to implement the present invention.

The applicability of the present invention to the proposed G.729 standard - notwithstanding, those of ordinary skill in the art will appreciate that features of the present invention have applicability to other speech coding systems.
Knowledge of the erasure of one or more frames is an input signal, e, to the S illustrative embodiment of the present invention. Such knowledge may be obtained in any of the conventional ways well-known in the art. For example, whole or partially corrupted frames may be detected through the use of a conventional error detection code. When a frame is determined to have been erased, e = 1 and special procedures are initiated as described below. Otherwise, if not erased (e = O) normal procedures are used.
10 Conventional error protection codes could be implemented as part of a conventional radio tr~n~mi~.~ion/reception subsystem of a wireless collullunication system.
In addition to the application of the full set of remedial measures applied as the result of an erasure (e = 1), the decoder employs a subset of these measures when a parity error is detected. A parity bit is computed based on the pitch delay index of the first of two subframes of a frame of coded speech. See G.729 Draft Section 3.7.1. This parity bit is computed by the decoder and checked against the parity bit received from the encoder.
If the two parity bits are not the same, the delay index is said to be corrupted (PE = 1, in the embodiment) and special proces~ing of the pitch delay is invoked.
For clarity of explanation, the illustrative embodiment of the present invention is 20 presented as comprising individual functional blocks. The functions these blocks represent may be provided through the use of either shared or dedicated hardware, including, but not limited to, haldware capable of executing software. For example, the blocks presented in Figure 1 may be provided by a single shared processor. (Use of the term "processor" should not be construed to refer exclusively to hardware capable of executing 25 software.) Illustrative embodiments may comprise digital signal processor (DSP) hardware, such as the AT&T DSP16 or DSP32C, read-only memory (ROM) for storing software performing the operations ~licc~ e~l below, and random access memory (RAM) for storing DSP results. Very large scale integration (VLSI) hardware embodiments, as well 6 2177~22 as custom VLSI ch~;ui~ly in combination with a general purpose DSP circuit, may also be provided.

II. An Illustraffve Elllbo l;.. t S Figure 1 presents a block diagram of a G.729 Draft decoder modified inaccordance with the present invention (Figure 1 is a version of figure 3 of the G.728 standard draft which has been augmented to more clearly illustrate features of the claimed invention). In normal operation (i.e., without experiencing frame erasure) the decoder operates in accordance with the G.729 Draft as described in sections 4.1 - 4.2.
During frame erasure, the operation of the embodiment of Figure 1 is augmented by special processing to make up for the erasure of information from the encoder.

A. NormalD~co~rOperation The encoder described in the G.729 Draft provides a frame of data representing compressed speech every 10 ms. The frame comprises 80 bits and is detailed in Tables 1 and 9 of the G.729 Draft. Each 80-bit frame of compressed speech is sent over a col,ul,unication channel to a decoder which synthesizes a speech (representing two subframes) signals based on the frame produced by the encoder. The channel over which the frames are col,ulwnicated (not shown) may be of any type (such as conventional telephone networks, packet-based networks, cellular or wireless networks, ATM
networks, etc.) and/or may comprise a storage medium (such as m~gnetic storage, semiconductor RAM or ROM, optical storage such as CD-ROM, etc.).
The illustrative decoder of Figure 1 includes both an adaptive codebook (ACB) portion and a fixed codebook (FCB) portion . The ACB portion includes ACB 50 and a gain amplifier 55. The FCB portion includes a FCB 10, a pitch predictive filter (PPF) 20, and gain amplifier 30. The decoder decodes tr~n~mittecl parameters (see G.729 Draft Section 4.1) and performs synthesis to obtain reconstructed speech.
The FCB 10 operates in response to an index, I, sent by the encoder. Index I is received through switch 40. The FCB 10 generates a vector, c(n), of length equal to a - 7 2177~22 subframe. See G.729 Draft Section 4.1.2. This vector is applied to the PPF 20. PPF 20 operates to yield a vector for application to the FCB gain amplifier 30. See G.729 Draft Sections 3.8 and 4.1.3. The amplifier, which applies a gain, g c, from the channel, generates a scaled version of the vector produced by the PPF 20. See G.729 Draft Section 4.1.3. The output signal of the amplifier 30 is supplied to summer 85 (through switch 42).
The gain applied to the vector produced by PPF 20 is determined based on information provided by the encoder. This information is communicated as codebook indices. The decoder receives these indicies and synthesi7es a gain correction factor,~ .
See G.729 Draft Section 4.1.4. This gain correction factor, ~, is supplied to code vector prediction energy (E-) processor 120. E-processor 120 determines a value of the code vector predicted error energy, R, in accordance with the following expression:
R (n) = 20 log ~ [dB]

The value of R is stored in a processor buffer which holds the five most recent (successive) values of R . R (n) represents the predicted error energy of the fixed code vector at subframe n. The predicted mean-removed energy of the codevector is formed as a weighted sum of past values of R:
E (n) = ~,b j R (n-i) where b = [0.68 0.58 0.34 0.19] and where the past values of R are obtained from the buffer. This predicted energy is then output from processor 120 to a predicted gain processor 125.
Processor 125 determines the actual energy of the code vector supplied by codebook 10. This is done according to the following expression:

E = 10 log (40~Ci ) ' where i indexes the samples of the vector. The predicted gain is then computed as follows:

- ~ 8 2t77422 - (E(~)+E---E)/20 where E is the mean energy of the FCB (e.g., 30 dB) Finally, the actual scale factor (or gain) is computed by multiplying the received gain correction factor, ~, by the predicted gain, gc at multiplier 130. This value is then 5 supplied to amplifier 30 to scale the fixed codebook contribution provided by PPF 20.
Also provided to the summer 85 is the output signal generated by the ACB portionof the decoder. The ACB portion comprises the ACB 50 which generates a excitation signal, v(n), of length equal to a subframe based on past excitation signals and the ACB
pitch-period, M, received (through switch 43) from encoder via the channel. See G.729 Draft Section 4.1.1. This vector is scaled by amplifier 250 based on gain factor, g p, received over the channel. This scaled vector is the output of the ACB portion.
Summer 85 generates an excitation signal, u(n), in response to signals from the FCB and ACB portions of the decoder. The excitation signal, u(n), is applied to an LPC
synthesis filter 90 which synthesizes a speech signal based on LPC coefficients, ai, received over the channel. See G.729 Draft Section 4.1.6.
Finally, the output of the LPC synthesis filter 90 is supplied to a post processor 100 which pel~olllls adaptive postfiltering (see G.729 Draft Sections 4.2.1 4.2.4), high-pass filtering (see G.729 Draft Section 4.2.5), and up-scaling (see G.729 Draft Section 4.2.5).
B. Exr;t~ti- l Signal Synthesis During Frame Erasure In the presence of frame erasures, the decoder of Figure 1 does not receive reliable information (if it receives anything at all) from which an excitation signal, u(n), may be synthP.~i7P-I As such, the decoder will not know which vector of signal samples should be 25 extracted from codebook 10, or what is the proper delay value to use for the adaptive codebook 50. In this case, the decoder must obtain a substitute excitation signal for use in synthesizing a speech signal. The generation of a substitute excitation signal during periods of frame erasure is dependent on whether the erased frame is classified as voiced 2177~2~
g (periodic) or unvoiced (aperiodic). An indication of periodicity for the erased frame is obtained from the post processor 100, which classifies each properly received frame as periodic or aperiodic. See G.729 Draft Section 4.2.1. The erased frame is taken to have the same periodicity classification as the previous frame processed by the postfilter. The 5 binary signal representing periodicity, v, is determined according to postfilter variable gpit.
Signal v = 1 if gpit > 0; else, v = 0. As such, for example, if the last good frame was classified as periodic, v = 1; otherwise v = 0.

1. Erasure of Frames Repr~s~..t;.~ Periodic Speech For an erased frame (e = 1) which is thought to have represented speech which isperiodic (v = 1), the contribution of the fixed codebook is set to zero. This isaccomplished by switch 42 which switches states (in the direction of the arrow) from its normal (biased) opelaling position coupling amplifier 30 to summer 85 to a position which decouples the fixed codebook contribution from the excitation signal, u(n). This switching of state is accomplished in accordance with the control signal developed by AND-gate 110 (which tests for the condition that the frame is erased, e = 1, and it was a periodic frame, v = 1). On the other hand, the contribution of the adaptive codebook is maintained in its normal ope,dlhlg position by switch 45 (since e = 1 but not_v = O).
The pitch delay, M, used by the adaptive codebook during an erased frame is determined by delay processor 60. Delay processor 60 stores the most recently received pitch delay from the encoder. This value is overwritten with each successive pitch delay received. For the first erased frame following a "good" (correctly received) frame, delay processor 60 generates a value for M which is equal to the pitch delay of the last good frame (i.e., the previous frame). To avoid excessive periodicity, for each successive erased frame processor 60 increllænls the value of M by one (1). The processor 60 restricts the value of M to be less than or equal to 143 samples. Switch 43 effects the application of the pitch delay from processor 60 to adaptive codebook 50 by ch~nging state from its normal opeldling position to its "voiced frame erasure" position in response to an indication of an erasure of a voiced frame (since e = 1 and v = 1).

_, - 10 2177422 The adaptive codebook gain is also synthesized in the event of an erasure of a voiced frame in accordance with the procedure discussed below in section C. Note that switch 44 operates identically to switch 43 in that it effects the application of a synthesized adaptive codebook gain by changing state from its normal opeldting position to its "voiced frame erasure" position.

2. Erasure of Frames Re~ s~ Aperiodic Speech For an erased frame (e = 1) which is thought to have represented speech which isaperiodic (v = 0), the contribution of the adaptive codebook is set to zero. This is accomplished by switch 45 which switches states (in the direction of the arrow) from its normal (biased) opeldting position coupling amplifier 55 to summer 85 to a position which decouples the adaptive codebook contribution from the excitation signal, u(n). This ~wilching of state is accomplished in accordance with the control signal developed by AND-gate 75 (which tests for the condition that the frame is erased, e = 1, and it was an aperiodic frame, not_v = 1). On the other hand, the contribution of the fixed codebook is m~int~ined in its normal O~ldtil~g position by switch 42 (since e = 1 but v = 0).
The fixed codebook index, I, and codebook vector sign are not available do to the erasure. In order to synthesize a fixed codebook index and sign index from which a codebook vector, c(n), could be determined, a random number generator 45 is used. The output of the random number generator 45 is coupled to the fixed codebook 10 through switch 40. Switch 40 is normally is a state which couples index I and sign information to the fixed codebook. However, gate 47 applies a control signal to the switch which causes the switch to change state when an erasure occurs of an aperiodic frame (e = 1 and not_v =1).
The random number generator 45 employs the function:
seed=seed* 31821 + 13849 to generate the fixed codebook index and sign. The initial seed value for the generator 45 is equal to 21845. For a given coder subframe, the codebook index is the 13 least significant bits of the random number. The random sign is the 4 least significant bits of the next random number. Thus the random number generator is run twice for each fixedcodebook vector needed. Note that a noise vector could have been generated on a sample-by-sample basis rather than using the random number generator in combination with the FCB.
The fixed codebook gain is also synthesized in the event of an erasure of an aperiodic frame in accordance with the procedure discussed below in section D. Note that switch 41 operates identically to switch 40 in that it effects the application of a synthesized fixed codebook gain by ch~n~ing state from its normal opelatillg position to its "voiced frame erasure" position.
Since PPF 20 adds periodicity (when delay is less than a subframe), PPF 20 should not be used in the event of an erasure of an aperiodic frame. Therefore switch 21 selects either the output of FCB 10 when e = 0 or the output of PPF 20 when e = 1.

C. LPC Filter Coefficients for Erased Frames The excitation signal, u(n), syn~hesi7e~1 during an erased frame is applied to the LPC synthesis filter 90. As with other components of the decoder which depend on data from the encoder, the LPC synthesis filter 90 must have substitute LPC coefficients, ai, during erased frames. This is accomplished by l~peatillg the LPC coefficients of the last good frame. LPC coefficients received from the encoder in a non-erased frame arestored by memory 95. Newly received LPC coefficients overwrite previously received coefficients in memory 95. Upon the occurrence of a frame erasure, the coefficients stored in memory 95 are supplied to the LPC synthesis filter via switch 46. Switch 46 is normally biased to couple LPC coefficients received in a good frame to the filter 90.
However, in the event of an erased frame (e = 1), the switch changes state (in the direction of the arrow) coupling memory 95 to the filter 90.

D. Attenl~qt;on of Adaptive and Fixed Codebook Gains As discussed above, both the adaptive and fixed codebooks 50, 10 have a corresponding gain amplifier 55, 30 which applies a scale factor to the codebook output signal. Ordinarily, the values of the scale factors for these amplifiers is supplied by the encoder. However, in the event of a frame erasure, the scale factor information is not available from the encoder. Therefore, the scale factor information must be synthesized.
For both the fixed and adaptive codebooks, the synthesis of the scale factor is accomplished by attenuation processors 65 and 115 which scale (or attenuate) the value of the scale factor used in the previous subframe. Thus, in the case of a frame erasure following a good frame, the value of the scale factor of the first subframe of the erased frame for use by the amplifier is the second scale factor from the good frame multiplied by an attenuation factor. In the case of successive erased subframes, the later erased subframe (subframe n) uses the value of the scale factor from the former erased subframe (subframe n- 1 ) multiplied by the attenuation factor. This technique is used no matter how many successive erased frames (and subframes) occur. Attenuation processors 65, 115 store each new scale factor, whether received in a good frame or synthesi7e~1 for an erased frame, in the event that the next subframe will be en erased subframe.
Specifically, attenuation processor 11 5 synthesizes the fixed codebook gain, gc, for erased subframe n in accordance with:
gc( = 0.98 gc( ) .
Attenuation processor 65 synthesizes the adaptive codebook gain, gp, for erased subframe n in accordance with:
g (n) = 0 g g (n-l) In addition, processor 65 limits (or clips) the value of the synthesized gain to be less than 0.9. The process of attenuating gains is performed to avoid undesired pelceptual effects.

E. Attenuaffon of Gain Predictor Memory As discussed above, there is a buffer which forms part of E-Processor 120 which stores the five most recent values of the prediction error energy. This buffer is used to predict a value for the predicted energy of the code vector from the fixed codebook.
However, due to frame erasure, there will be no information communicated to the decoder from the encoder from which new values of the prediction error energy.

_ - 13 2177422 Therefore, such values will have to be synthesized. This synthesis is accomplished by E-processor 120 according to the following expression:
R (n) = (O 25 ~, R (n)) _ 4Ø
, Thus, a new value for R (n) is computed as the average of the four previous values of R
5 less 4dB. The attenuation of the value of R is performed so as to ensure that once a good frame is received undesirable speech distortion is not created. The value of thesynthesized R is limited not to fall below -14dB.

F. An Illustrative Wireless System As stated above, the present invention has application to wireless speech communication systems. Figure 2 presents an illustrative wireless co~ lul~ication system employing an embodiment of the present invention. Figure 2 includes a transmitter 600 and a receiver 700. An illustrative embodiment of the transmitter 600 is a wireless base station. An illustrative embodiment of the receiver 700 is a mobile user terminal, such as a 15 cellular or wireless telephone, or other personal collllllullications system device.
(Naturally, a wireless base station and user terminal may also include receiver and transmitter Cil-;Ui~ , respectively.) The transmitter 600 includes a speech coder 610, which may be, for example, a coder according to the G.729 Draft. The transmitter further includes a conventional channel coder 620 to provide error detection (or detection and 20 correction) capability; a conventional modulator 630; and conventional radio tr~n~mi~sion circuitry; all well known in the art. Radio signals transmitted by tldnc"~it~l 600 are received by receiver 700 through a tr~n~miccion channel. Due to, for example, possible destructive interference of various multipath components of the transmitted signal, receiver 700 may be in a deep fade preventing the cledr reception of transmitted bits.
25 Under such circllm~t~nces, frame erasure may occur.
Receiver 700 includes conventional radio receiver circuitry 710, conventional demodulator 720, channel decoder 730, and a speech decoder 740 in accordance with the present invention. Note that the channel decoder generates a frame erasure signal ~177422 _ - 14 - whenever the channel decoder determines the presence of a substantial number of bit errors (or unreceived bits). Alternatively (or in addition to a frame erasure signal from the channel decoder), demodulator 720 may provide a frame erasure signal to the decoder 740.
s G. D;~ on Although specific embodiments of this invention have been shown and described herein, it is to be understood that these embodiments are merely illustrative of the many possible specific arrange~ which can be devised in application of the principles of the 10 invention. Numerous and varied other arrangements can be devised in accordance with these principles by those of ordinary skill in the art without departing from the spirit and scope of the invention.
In addition, although the illustrative embodiment of present invention refers tocodebook "amplifiers," it will be understood by those of ordinary skill in the art that this 15 term encomp~c~es the scaling of digital signals. Moreover, such scaling may be accomplished with scale factors (or gains) which are less than or equal to one (including negative values), as well as greater than one.

2~ 77~22 P. K-oon ~10 INTERN~TIONAL TELECO!~I~IUNICATION Ui! ION
TELECO~i~IUNICATIONS ST.~.~'DARDIZ.~TION SECTOR

Date: June l99.i Origioal: E

STUDY GROUP 15 CO~-TRIBUTION- Q. 12/15 Dra~t Recommendation G., 29 Coding of Speech at 8 kbit/s using Conjugate-Structure-Algebraic-Code-E~cited Linear-Predicti~e (CS-ACELP) Coding June 7. 199~, version ~.0 Notc: Until th~ RCCG cndation ~1 approrcd ~y thc ITU, ncithcr thc C codc nor thctest vcctors will ~c availa~lc from thc ITU. ~o o~tatn thc C ~O~KC codc, contact:
~ir. Gerhard Schroeder, Rapporteur SG15/Q.12 Deutsche Telekom AG, Postfach 100003, 64216 Darmstadt, Germany Phone +49 6151 83 3973, Fax: +49 6151 837828. Email: gerhard.schroedersfzl3.fz.dbp.de P. I;roon 5-10 Contents 1 Introduction 19 2 General description of the coder 20 2.1 Encoder . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 2.2 Decoder ........................................... 2-2 2.3 Delay ............................................. 23 2.4 Speech coder description .......................... 23 2.5 Notational conventions ............................ 2
3 F~ l ee r ;ption of the encoder 28 3.1 Pre-processing .................................... 2d 3.2 Linear prediction analysis and q~sr.~i~cti-q ..... 2d 3.2.1 Windowingand autocorrelation computation ..... 29 3.2.2 Levinson-Durbin algorithm .................... 30 3.2.3 LP to LSP c~ ~ ............................ 303.2.4 Q us~;7~ of the LSP coefficient~ ......... 323.2.5 T ~ n of the LSP coeflficients ............. .33.2.6 LSP to LP c~ ~ r ........................... 3~3.3 Perceptual weighting .............................. 3.;
3.4 Open-loop pitch analysis .......................... 36 3.5 Computation of the impulse r~on~ ................ 31 2177~22 P Kroon 5-10 3 6 Computation of the target si~Dai ~ ...... ........ 38 3 7 Adaptive-codebook search 38 3 7 1 Generation of the adapti-e codebook vector 40 3 7 2 Codeword computation for adapti-e codebool~ delays 40 3 7 3 Computation of the adaptive-codebook gain 41 3 8 Fixed codebook structure and search 41 3 8 1 Fixed-codebook search procedure 42 3 8 2 Codeword computation of the fixed codebook 44 3 9 Qu~n~i7~ion of the gains 44 39 1 Gain prediction 45 3 9 2 Codebook search for gain quantization 46 3 9 3 Codeword computation for gain quantizer 47 3 10 . lemory update 47 3 11 Encodcr and Decoder initi.~ tj; n 47
4 F~-r - -~ description of the tecoder 49 4 1 Parameter deco~ing pr~cedure 49 4 1 1 D~eo~ g of LP filter parameters 50 4 1 2 ~o~ling of the adaptive codebook vector 50 4 1 3 Decoding of the fi~ed codebook ector 51 4 1 4 Decoding of the adapti-e and fixed codebook gains 51 4 1 5 Computation of the parity bit 51 P. I;roon 5-10 4.1.6 Computing the reconstructed speech ................. 51 4.2 Post-processing ......................................... 52 4 .2.1 Pitch postfilter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .i2 4.2.2 Short-term postfilter .............................. 53 4.2.3 Tilt compensation .................................. 53 4.2.4 Adaptive gain control .............................. 54 4.2.5 High-pass filtering and up-scaling ................. 54 4.3 ConceAIm~nt of frame erasures and parity errors ......... 55 4.3.1 Repetition of LP filter parameters ................. 56 4.3.2 Att~nu2~ion of adaptive and fixed codcbaol~ gains .. 56 4.3.3 Attenl~ti-- of the memory of the gain predictor ... 56 4.3.4 Generation of the repl~ . excitation ............... 56
5 Bit-esact te~cription of the CS-ACELP coter 58 5.1 Use of the ~im~ iq- soft~lra,re ......................... 58 5.2 Orga~ tion of the ~im~ ti-n software .................... 58 2177~22 P. ~;roon ~10 1 Introduction Tbis Recommend7tion contains the description of an algorithm for the coding of speech signals at 8 kbit/s using Conjugate-Structure-Algebraic-Code-E!~cited Linear-Predicti~e (C~ACELP) coding.
This coder is designed to operate with a digital signal obtained by first performing telephone bandwidth filtering (ITU Rec.G.710) of the analog input signal, then sampling it at 8000 Hz.
followed by con~e~sion to 16 bit linear PCM for the input to the encoder. The output of the decoder should be converted back to an analog signal by similar means. Other input/output characteristics, such as those specified by ITU Rec.G.711 for 64 kbit/s PCM data, should be coo~erted to 16 bit linear PCM before encoding, or from 16 bit linear PCM to the appropriate format after decoding.
The bitstream from the encoder to the decoder is defined within this standard.

This Recommendation is organized as follows: Sectioo 2 gi~es a general outline of the CS-ACELP algoritbm. In Sections 3 and 4, the C~ACELP encoder and decoder principlff are dis-cussed, rffpecti-ely. Section ~ describes the software that definff this coder in 16 bit fixed point ?-rj--T~ Tn~

- 21774~2 P. I;roon 5-10 2 General description of the coder The C~ACELP codet is based on the code-excited linear-predicti~e (CELP) coding model. The coder operates on speech frames of 10 ms corresponding to 80 samples at a sampling rate of 8000 samples/sec. For every 10 msec frame, the speech signal is analyzed to extract the parameters of the CELP model (LP filter coefficients, adaptive and fixed codebook indices and gains). These parameters are encoded and transmitted, The bit allocation of the coder parameters is shown in Table 1. At the decoder, these parameters are used to retrie-e the excitation and synthesis filter Table 1: Bit allocation of the 8 kbit/s CS-ACELP algotithm (10 msec frame).
Paramctcr Codcword Subframc ISubframc 27otalpcrframc LSP L0, Ll, L2, L3 18 Adapti~re codebook delay Pl, P2 8 5 13 Delay paritJ P0 FL~ed codeboolc inde~c Cl, C2 13 13 26 Fi~ced codeboolc ~i8~ Sl, S2 4 4 8 Co~l~L g~.ins (st~ge 1~ GAI, GA2 3 3 6 Codeboolc 8 ins (sta~e 2) GBI, GB2 4 4 8 Tot~l 80 parameters. The speech is reco,,~.t,~cted by filtering this excitation through the LP synthesis filter, as is sho~n in Figure 1. The short-term synthesis filter is based on a 10th order linear prediction ourP~JT
EXCIT~TlON _ LONt~TERI~ SHO~T-TER~ FILTER SPEECH

P~fUMETER GE; ~G

RECEIYED SITSTREA~

Figure 1: Block diagram of conceptual CELP s,~-nthesis model.

(LP) filter. The long-term. or pitch synthesis filter is implemented using the so-called adapti~e codebool~ approach for delays less than the subframe length. After computing the reconstructed speech, it is further enhanced by a postfilter.

P Kroon 5-10 2. 1 Encoder The signal flow at the encoder is shown in Figure 2. The input signal is high-pass filtered and scaled INPUT
SPEECH

LP ANALYSIS
QUAhTlZATlON
t3cINTERPOI ATION
~ col~rr~ LPC inb Gp (~ ~ FILTER
_ AD~PTIYE
CoDFr~

LPC hb ~~~~~~~~~~~~~~~~~~~~~~~~~ PITCH
_ . __, . . ANALYSIS ~
IAL
WEIGHT NG ~~

--------------------'------'--------'----------' ~SE
,. _ ___........... SEARCH ~ ~

~~ ~ ' ~ ~ ~ ~ TRANSAllrrED
~~~~~~~~~~~~ PARAMtltR _ ~TSTREA~
~UANTIZAT~---------------- . E~;C50:N~ ~~-LPC nh Figure 2: Signal flo~r at the CS-ACELP encoder.

in the pr~p~ bloclc. The pre-proces#d signal #rves as thc input signal tor all subsequent analysis. LP analysis is done once per 10 ms frame to compute the LP filter coefflcients. These coefficients are converted to line spectrum pairs (LSP) and quanti2ed using predicti~e two-stage ~-ector~quantization (VQ) with 18 bits. The excitation #quence is chosen by using an analysis-b~-synthesis #arch procedure in which the error bet~ een the original and synthesized speech is mini-ni7~d according to a perceptually weighted distortion measure. This is done by filtering the 2177~22 P. I;roon 5-10 error signal with a perceptual weighting filter, whose coefficient~ are derived from tbe unquantized LP filter. The arnount of perceptual weighting is made adapti-e to impro~e the performance for input sicnals with a flat frequency-response.

The e.~citation parameters (fflced and adapti~,e codebook parameters) are determined per sub-frame of 5 ms (40 samples) each. The quantized and unquantized LP filter coefflcients are used for the second subframe, while in the first subframe interpolated LP filter coefflcients are used (both quantized and unquantized). An open-loop pitch delay is: :im~ed once per 10 rns frame based on the perceptuaUy weighted speech signal. Then the following operations are repeated for each subframe. The target signal z(n) i9 computed by filtering the LP residual through the weighted synthesis filter W(z)/A(z). The initial states of these filters are updated by filtering the error betwee~ LP residual and ~x~ita~ jon. This is equivalent to the common approach of subtracting the zero-input response of the weighted synthesis filter from the weighted speech signa1. The impulse response. h(n), of the weighted synthesis filter is computed. Closed-loop pitch analysis is then done (to find the adaptive codebeolt dela~ and gain), using the target z(n) ant impulse responge h(n), by searching around the value of the open-loop pitch delay. A rr r ~I pitch delay with 1/3 resolutioD is used. The pitch delay is encoded with 8 bits in the first subframe and differentially encoded Irith 5 bits in the second subfrarne. The target signal r(n) is updated by removing the adaptive codebool~ contribution (filtered adaptive codevector), aDd this new target, z2(n), is used in the fixed algebraic codebc ~1~ #arch (to find the optimum excitation). An algebraic codcblocl~
with 17 bits is used for the fixed code~bool~ excitation. The gains of the adaptive and fixed code-book are Yector qu~nti7ed with 7 bits, (with MA prediction applied to the fixed codebook gain).
Finally~ tne filter m~moriP< are updated using thc determined excitation signal.

2.2 Decoder The siual fiow at the decoder ia shown in Figure 3. First, the parameters indices are extracted from the received b ~l - The~e indices are decoded to obtain the coder pararneters corresponding to a 10 ms speee~ frune. The~e parameters are the LSP coefflcieDts. the 2 rr~ pitch delays, the 2 fi~ed cod bu'- vectors, ant the 2 sets of adaptive and fi~ed codebooli gains. The LSP
coefficients are interpolated and converted to LP filter coefflcients for each subframe. Then, for each 40-sample subframe the following steps are done:

the excitation is con~lru~ted by adding the adaptive and fixed codeboolt vectors scaled by their respective gains, P. E;roon 5-10 GC

Gp ~) ~ S FNITHEER~ ~ P~S~iG
~DAPT VE ~

Figure 3: Signal flow at the CS-ACELP decoder.

the speech is ~ ted by filtering the excitation through the LP synthffis filter, the reconstructed speech signal is passed through a post-processing stage, which comprises of an adapti~e postfilter based on the long-term and short-term synthesis filters, followed by a high-pass filkr and scaling operation.

2.3 Delay This coder encodes speech and other audio signals with 10 ms frames. In addition. there is a look-ahead of 5 ms, resulting in a total algorithmic delay of IS m9. All additional delays in a practical implPm~nt~ n of this coder are due to:

processing time needed for encoding and decoding operations.
trancmi--i~n time on the commnnic~tion linl~, multiplexing telay when comhining audio data with other data.

2.4 Speecb coder de~cription The dexriptioo of the speech coding algorithm of this F~commendation is made in terms of bit-exact, fixed-point I -th~motical operations. The ANSI C code indicated in Section 5. which constitutes an integral part of thi~ Recommendation, reflects this bit-exact, fixed-point descripti~-e approach. The ma~ -tical descriptions of the encoder (Section 3), and decoder (Section 4), can be implemented in se~reral other fA~hion~ possibly leading to a codec implementation not complying with this Recomm~n~ on Therefore, the algorithm description of the C code of Section 5 shall - 21774~2 P. ~;roon 5-10 tal~e precedence ove- the m~thematical descriptions of Sections 3 and 4 whcnever discrepancies are found. A non-~ c set of te9t sequences which can be used in conjunction with the ~ code are available from the ITU.

2.5 Notational conventions Throughout this document it is tried to maintain the following notational conventions.

Codebooks are denoted by caligraphic characters (e.g. C).
Time signals are denoted by the symbol and the sample time inde~c bet~een parenthesis (e.g.
s(n)). The symbol n is used as sample instant inde~.
Superscript time indices (e.g 9(m)) refer to that variable corrffpon.iing to subframe m.
Superscripts identify a particular element in a coefflcient array.
A identifies a quantized version of a parameter.
Range notations are done using square brac}ets, where the boundaries are included (e.g.
[0.6, 0.9]).
Iog denotes a logarithm ~ith base 10.

Table 2 lists the mo~t relevant symbob used throughout this doe, - .~ glossary of the most Table 2: Glossary of symbols.
Namc Rcfcrcncc ~CJC. ,~t _ n 1/A(-) Eq- (2) LP synthesu filter ~I(s) Eq- (1) input high-p~s filter H~(s) Eq. (77) pitch postfilter Eq- (83) short-term postfilter H~(s) Eq. (85) tilt-comF t )rl filter Hb2(z) Eq. (90) output high-pass filter P(~) Eq. (46) pitch filter W(z) Eq. (27) ~._;oLti..g filter relevant signals is gi~en in Table 3. Table 4 summarizes relevant ariablff and their ~lim~

~ - 2177~22 P I;roon 5-10 Constant parametets are listed in Table S The acronyms used in this Recommendation are sum-marized in Table 6 Table 3 Glossary of signals .~arnc Dc~crlptton h(n) impulse response of weighting and s~nthesis filters r(~) auto-correlation sequence r'(~) modified auto-correlation sequence R(~) correlation sequeoce ~w(n) -eighted speech signal ~(n) spfflh signal ~'(n) .. i~ d speech signal ~f(n) pos~filt~red output ~f'(n) gain-scaled postfiltered output i(n) ~ ed speech signal r(n) residual signal (n) target signal z2(n) second target signal v(n) adapti~re cod~ t c(n) fi~ced s ~ ~ba o 1~ ~ 1 o y(n) v(n) ~ h(n) ~(n) c(n) ~ h(n) u(n) ~ a~i - to LP synthesi~ filter d(n) .u.~ i - between target signal aod h(n) cw(n) error signal 2177~22 P. I;ro~n 5-10 Table 4: Glossary of variables.
lVamc Sizc OCJcnption gp 1 adapti-~e cod~bool~ ~ain gc I fi~ced codeboolc gain gO 1 modificd 8ain for pitch postfilter gp" 1 pitch g~ for pitch postfi~tcr g~ 1 gain term short-term postfilter g, 1 gain tcrm tilt postfiltcr Top I opcn-loop pitch dclay a, 10 LP c~ ~ ,w 1~ 1 0 t~ r ~ _ -o, 2 LAR ~ -L~ 10 LSF c~- ~1;7~d t qi 10 LSP r~ ~
r(t) 11 ..,..
wj 10 LSP ~
l, 10 LSP quantizcr output . .
P ~roon ~10 Table 5 Glossary of constants Nomc VcJuc De~crlption f, 8000 samplins frequency fO 60 bandwidth ~p~ r 0 94/0 98 weight factor perceptual ~eighting filter 0 60/[0 4-0 7] ~veight factor perceptual ~eighting filter ~" 0 55 weight factor post filter 7~t 0 70 weight factor post filter ~p o so weight factor pitch post filter ~, 0 90/0 2 wcight factor tilt post filter C Table 7 fi~ced (algebraic) codeb,oolt C0 Section 3 2 ~ moving average predictor codcbD~k Cl Stio 3 2 4 First stage LSP codcb olc C2 Stion 3 2 4 Sond ~tage LSP cuicb~ool~ (lo part) C3 Stion 3 2 4 Sond stage LSP codcb~l~ (high part) ~A Stio 3 9 First stage gun .ud~b D:IC
~7B Sectian 3 9 Sond 5 a8e 8ain cod~bD~lc ~n 9 Eq (6) cu ~'~ti lag window u~lp Eq (3) LPC analy~i~ window Table 6 Glossar~r of acronym8 Acron~ DcJ~ n CELP code-e%cited linear-prediction MA moving average MSB most ~;~D r "t bit LP lillear pl "

LSP line spectral pair LSF line spectral L~
VQ vtor quantization 2~77422 P. I;roon 5-10 3 Functional description of the encoder In this section we describe the different functions of the encoder represented in the blocks of Figure 1.

3. 1 Pre-processing As stated in Section 2, the input to the speech encoder is assumed to be a 16 bit PCM signal.
Two pre-processing functions are applied before the encoding process: 1) signal scaling, and 2) hi~h-pass filterin8.

The scaling consists of dividing the input by a factor 2 to reduce the p~ ' ty of overflows in the fixed-point imple -nta~ion The high-pass filter ser-es as a precaution against undesired low-frequency components. A second order pole/zero filter ~ith a cutoff frequency of 140 ~8 i used. Loth the scaling and high-pass filtering are combined by dividing the c~ll;. tc at the numera or of this filter by 2. The resulting filter is given by 0.46363718 - 0.92~24705z-1 + 0.~6363~18z-2 H~l(z) = l - 1.905946~z-1 + 0.911~024z-3 ( ) The input signal filtered through Hhl(z) is referred to as s(n). and ~ill be used in all subsequent coder operations.

3.2 Linear prediction analysi~ and quantization The short-term analysis and s~ ~ filters are based on 10th order linear prelict; (LP) filters.
The LP synthesis filter is defined as A(z) l + ~il~ ajz-i (2) where ai, i = 1, . . . ,10, are the (qn~ ed) linear prediction (LP) coefficients. Short-term predic-tion~ or linear prediction analysis is pe.fulllled once per speech ~rame using the autocorrelation approach ~ith a 30 ms asymmetric window. Every 80 samplff ( 10 ms), the autocorrelation coeffl-cients of v,indowed speech are computed and converted to the LP CoC~ h.L~:s using the Levinson algorithm. Then the LP co.?lfici~n~ are transformed to the LSP domaio for ~, -nti7~tisv and interpolation purpose~. The interpolated quantized and unquantized filters are converted back to the LP filter coefflcients (to cor.;.~,r. the synthesis and weighting filters at each subframe).

.
P. I;roon a-10 3.2.1 Windowing and autocorrelatio~ computatiorl The LP analysis window consists of two parts: the first part is half a Hammiog windo~ and the second part is a quarter of a cosine function cycle. The window is given by:

'P( ) CO (2r(n-200)) n--200 239 (3) There i9 a 5 ms lookahead in the LP analysis which means that 40 samples are needed from the future speech frame. This translates into an extra delay of 5 ms at the encoder stage. The LP
analysis window applies to 120 samples from pa~t speech frames~ 80 samples from the present speech &ame, and 40 samples from the future frame. The windo~ing in LP analysis is illustrated in Figure 4.
r//~; ", 7~ 2 \~\\\\`1 U WINDOWS

SU~FRAMES

Figure 4: Windowing in LP anal~sis. The different shading patterns identify corresponding exci-tation and LP analysis frames.

The autocorrelation coefficients of the windowed spfflh s'(n) = wlp(n) s(n), n = O, . . ., 239, (4) are computed by ~39 r(~ s'(n)s'(n ~ = O,. . ., 10, (a) To avoid ar;~ pr~bl .. ~ for low-level input signals the value of r(0) has a lo~er boundary of r(O) = 1Ø A 60 lIz bantwidth ~Yp~- -n is applied, by multiplying the autocorrelation coeflicients with wlal(t) = exp --I ( fo ) , ~ = 1, . . ., 10. (6) where fO = 60 Hz is the bandwidth expansion and f, = 8000 ~z is the sampling frequency. Further, r(0) is multiplied by the wbite noise correction factor 1.0001, which is equi~alent to adding a noise ~oor at -40dB.

217742~

P. I;roon ~10 3.2.2 L~ ~ o - Durbin algorithm The modified autocorrelation coefflcients r'(O) = L 000 l r(O) r'(k) = w~g(k)r(~), t = 1.... 10 are used to obtain the LP filter coeflicients ai, t = 1, .... 10. b- sol-ing the set of equations ~ aj r'(~ r'(l~ = 1 ................ ,10. (8) i=l The set of equations in (8) is solved using the Levinson-Durbin algoritbm. This algorithm uses the following recursion:
E(O) = r'(O) for i = 1 to 10 a('~l) = 1 Icj = - [~,J_I a(i~l)r'(i--j)¦ /E(i -1) aji) = ~j -forj = 1 to i--1 a(i) = a(i~l) + ~, cnd E;(i) = (1 - ~j2)E(i - 1), ~f E(i) < O tAcn E(i) = 0.01 end The final solution is given as aj = a(l), j = 1, . . . ,10.

3.2.3 LP to LSP co~

The LP filter coP~-' ~ aj, i = 1, . . . ,10 are converted to the line spectral pair (LSP) repre,senta-tion for c,uP ti and intetpolation purposes. For a 10th order LP filter, the LSP coeflicients are defined as tbc roots of the sum and difference polynomials Fl(z) = A(z) + z-11,4(z-l), (9) and F2(z) = A(z)--z-llA(:-I). (10) respecti-ely. The pol-nomial Fl(z) is symmetric, and F:( ) is antis-mmetric. It can be pro-en that aU roots of these polynomials are on the unit circle and they alternate each other. Fl(:) has P f;roon 5-10 a root z = -I (;.. = J) and F2(:) has a root z = I (~ = 0). To elimin~te these two roots, we define the ne~ polynomials Fl(z) = F~( )/(l + z-l), (11) and F2(z) = F2(z)/( l - z ~ l ) ( 12) Each polynomial has 5 conjugate roots on the unit circle (e~ ), therefore, the polynomials can be ~ritten as Fl(z) = ~ 2~iz-1 + z-2) (13 i= I ,3,. . .,9 and F2(z) = II (1- 2qiz~l + z ), i=2,~ o where ~j = cos(~j) with ~i being the line spectral frequencies (LSF) and they satisfy the ordering property 0 < ~1 < w2 < . . < ~lo < ~. We refer to gi as the LSP coefficients in the cosine domain.
Since both pol~ - ~' Fl(z) and F2(z) are symmetric only the first 5 CoPf~ s of each polynomial need to be computed. The coefficients of these polynomials are found by the recursive relations fi (i + 1 ) = ai+l + a10_;--fi (i), i = 0, . . ., 4, f2(i+1)= aj+l-alo_;+f2(i), i=O,...,4, (1~5) where fl(0) = f2(0) = 1Ø The LSP coefflcients are fount by evaluating the polynomials Fl(z) and F~(z) at 60 points equally spaced between 0 and ~r and checking for sign changes. A sign change signifies the ~-- t~ -e of a root and the sign change interval is then divided 4 times to better tracl~ the root. The Chcb~ I polrl.or.lials are used to evaluate Fl(z) and F2(z). In this method thc roots are found dirtly in the cosine domain {gj}. The polynomialJ Fl(z) or F,(z), evaluated at z = cJ~, ca~ be written aa F(~) = 2c-J5~C(z), (16) with C(z) = Ts(z) + f(l)T4(z) + f(2)T3(r) + f(3)T2(~) + f(4)TI(l) + f(5)/2, (17) where Tm(~) = cc~s(m~) is the mth order Chebyshev polynomial, and f(i), i = 1, . . ., 5. are the coefficients of either Fl(z) or F2(z), computed using the equations in (15). The polynomial C(~) is e~aluated at a certain value of ~ = cos(~) using the recursive relation:
for ~ = 4 downto 1 P. ~;roon 5-10 b~ = 2~b~+~ 2 + f(5 ~ J~) end C(~ b~ +f(S)/2 with initial vaJues 1~5 = 1 and J6 = -3.2.4 Qllontj7~ of the LSP coefflcients The LP filter co~ffi~i~nts are quantized using the LSP repr~lltation in tbe frequency domain; thatis -= arccos(qj), i = 1,...,10. (18) ~here ~j are the line spectral frequencies (LSF) in the normalized frequency domain ~0, ~]. Aswitched 4th order MA prediction is used to predict the current set of LSF coefflcients. The difference between the computed and predicted set of coefficients is quantised using a t~o-stage ector gu~-ti70r~ Tbe first stage is a 10-dimensional VQ using codebook ~:1 with 128 entries (.7 bits). The second stage i9 a 10 bit VQ which has been implemented as a split VQ using two ~d ---' cod~booL, ~:2 and r3 containing 32 entries (5 bits) each.
To e~plain the qu~nti7~ti~n process, it is convenient to first describe the decoding process.
Each co~ is obtained from the sum of 2 codebooks:
~lj(Ll) + 1~2j(L2) i = 1, .... 5, (19) rlj(Ll) + ~`3(i-s)(L3) i = 6~ ~10~
~here Ll, L2, and L3 are the codebook indices. To avoid sharp resonances in the quantized LP
synthesis filters, tbe co~: Ij are arranged such that adjacent coll;~ic,~ts have a minimllm distance of J. The rearr~n~-- ~ routine is shown below:
fori=2,...10 if (li_l > l --J) 1,--1 = (li + /i-l--J)/2 4=(lf +li_l+J)/2 cn~
cml This rearr~ngem~r~t process is e~cecuted twice. First with a value of J = 0.0001, then with a ~alue of J = 0.000095.

After this real. g~m~nt process, the quantized LSF coefflcients 6.îj(m) for the current frame n, are obtained from the weighted sum of previous quantizer outputs l('n-i~. and the current quantizer P. I;roon 5-10 output l(m) mt)l(m~ + ~ mjtl( ), i = 1, . . ., 10. (20) t=l 1~=1 where m~ are the coefficients of the switched MA predictor. Which ~A predictor to use is defined by a separate bit ro. At startup the initial values of l(~) are given by l~ r/ll for all 1; < 0.

After computins ~j, the correspooding filter is checked for stability. This is done as follows:

l. Order the coefficient ~j in increasing value, 2. If ~1 < 0.005 then ~1 = 0.005, 3. If ~j+, - ~j < 0.0001, then ~,îi+l = ~j + 0.0001 i = 1, . . . ,9, 4 If ~lo > 3.135 theo wlO = 3.135.

The procedure for encoding the LSF parameters can be outlinet as follo~s. For each of the two MA predictors the best appr~Yim~tier to the current LSF ~rector has to be found. The best approYim~tion is defined as the one that minimi7~s a weighted mean-squared error ELPC = ~ Wj(~"j _ ~j)2 (21) i=l The- weights w; are made adaptive as a function of the unqu~ I LSF coefficients, 1.0 'f ~2--0.04~r--1 > 0.
wl = ~
10(~2--0.04~ - 1)2 + 1 othcrwi~c wj 2 < i ~ 9 = 1.0 if ~i+l--~i-l--1 > ' (22) 10(1./j+~ --1)2 ~ 1 othcrwiJc 1.0 if --~9 + 0.92~--1 > 0, wlO = ~
10~-~9 + 0.92~--1)2 + 1 otherwi~c In addition, the ~ghts W5 and w~ are multiplied by 1.2 each.

The vector to be ~ d for the current frame i9 obtained from 1 = [~(m)--~ mjtlj(m~t)] /(1--~ mjt)~ i = 1, . . .,10. (23) ~=1 ~=1 The first codeboolc 1'l is searched and the entry Ll that minimi7~ the (unweighted) mean-squared error is selected. This is followed by a search of the second codebooli ~2, which defines P t;roon ~l0 the lower part of ~he second stage. For each pos.sible candidate~ the partial ector df7;, i = 1...., 5 ij reconstructed usin6 Eq. (20), and rearranged to guarantee a minimllrn distance of 0.0001. The Yector with inde~c L2 ~hich after addition to the first stage candidate and rearranging, approximates the lower pan of tbe corresponding target best in the ~eighted ~SE sense is selected. L'sing the selected fi~,t stage ector Ll and the lower part of the ;econd stage (L2), the higher part of the second stage is ~.earched from codeboo,~ 1~3. Again the rearr~n~f~mf~nt procedure is used to ~uarantee a minim~m distance of 0.0001. The ector L3 that minimi7~ the o~erall weighted ~SE
i, selected.

This process is done for each of the two .~ predictors defined by ro, and the MA predictor L0 that produces the lowest weighted MSE is selected.

3.2.5 Interpolation of the LSP coefflcients The quAn~i7~d (and unquantized) LP cof~ffici7~n~s are used for the second subframe. For the first subframe, the qu~- ,cd (and unquantized) LP co~ nts are obtained from linear interpolation of the CO~ in~ parameters in the adjacent subframes. The interpolation is done on the LSP
cc~ s in the q domain. Let qj(m) be the LSP Co'"~iCif'o~' at the 2nd subframe of frame m, and ~j(m~l) the LSP c ~ at the 2nd subframe of the past frame (m - 1). The (unquantized) interpolated LSP co l~ nts iQ each of the 2 subframes are gi~en by Su~f~amc 1: Q1; = 0.5Qj ) + 0..jyj , i = 1, .... 10, Sullframc 2: q2i = q(m) i = 1, . . .,10. (24) The same interpolation procell... is used for the interpolation of tbe quantized LSP co~lEici~nts by s h~;t~.~ing q, by jj in Eq. (24).

3.2.6 LSP tO LP C-J~

Once the LSP coefflcients are quantized and interpolated. the- are con-erted bacl~ to LP coefficients ~a;}. The ccn~ n to the LP domain is done as follo~s. The co~fl;cif~nts of Fl(;) and ~2(Z) are found b~ f~p2~in~ Eqs. (13) and (14) knowing the quantized and interpolated LSP coefflcients.
The followin8 recursi~e relation is used to compute fl(i). i = 1...., 5, &om q for i = 1 to 5 fi(i) =--2 q2i-1 fl(i--1) + 2fl(i--2) for j = i--1 downlo I

~177 422 P ~;roon 5-10 fl(j) = f~ 2 Q~ f~(j--1) + fl(j--2) cn~
end ~rith initial values fl(0) = I and fl(--1) = 0. The coefficients f2(i) are computed similarl~ by replacin~ q~j_, by q2i.

Once the coefficients fl(i) and f2(i) are found, ~I(z) and F2(z) are multiplied by I + z~~ and I - z~~, respectively, to obtain Fl(z) and F2(z); that i9 fl(i) = fl(i) + fl(i - 1), = 1, . . ., 5, f2(i) = f2(~)--f2(i--1), ~ = 1,.. ,5. (25) Finally the LP coefficients are found by aj = ~ 0.5f;(i) + 0.5f2(i), i = 1, . . ., 5, (26) 0.5fl(i--5) - 0.5f2(i--5), i = 6, . . ., 10.
This is directly derived from ehe relatioo A(z) = (Fl(z) + F2(z))/2, and bccause F~(z) and F2(z) are symmetric and antisymmetric polynomials, respectively.

3.3 P~rceptual weighting The perceptual .~ g' g filter is based on the unquantized LP filter roPfli-iPnts and is given by W(z) A(~/~2) 1 + ~10l r2a z~i (27) The valu~ of 71 and 72 detprmir- the frequency response of the filter W(z). By proper adjustment of these ~ariables it is possible to malce the weightilrg more effective. This is accomplished by malcing ~rl and 'r2 a functioD of the spectral shape of the input signal. This ~laFt~tion is done once per 10 rns frame, but arl interpolation procedure for each first subframe i5 used to smooth this adn"ta'i process. The sptral shape is obtained from a 2nd-order linear prediction filter~
obtained as a by product from the LeYinson-Durbin re.~ ;on (Section 3.2.2). The reflection co. ~.- ~j, are con~rerted to Log Area Ratio (LAR) coPfl;~iPn~- oj by oj=log(l0+~i) i=1,2. (28) These LAR coefficients are used for the second subframe. The LAR coeffi~ nt- for the first subframe are obtained through linear interpolation with the LAR parameters from the previous 217742~

P. ~roon ~-10 frame, and are given by:
Subframc 1: olj = 0.5ojm~l) +0.5Ojm), i = 1, . . .,2, Subframe 2: o2j = im)~ i = 1,. . .,2. (29) The spectral envelope is characterized as being either flat (flat = 1) or tilted (flat = O). For each subframe this characterization is obtained by applying a threshold function to the LAR coefflcients.
To avoid rapid changes, a hysteresis is used by taking into account the value of flat in the previous subfrarne (m- 1), 0 if ol < -1.74 and 2 > 0.65 and flat(m~l) = 1, flat(m) = ~ 1 if l > -1.52 and 2 < 0.43 and flat(m~l) = 0, (30) flat(m~l) otherwi#.
If the interpolated spectrum for a subframe is classified as flat (flat(m) = 1), the weight factors are set to 71 = 0.94 and 72 = 0.6. If the spectrum i~ rlPc~if ed as tilted (flat(m) = 0), the value of 71 is set to 0.98, and the value of 77 is adapted to the strength of the r~. ~r^- in the LP
synthesis filter, but is bounded bet~een 0.4 and 0.7. If a strong r~nsn~ is present, the value of ~q is set closer to the upperbound. This adaptation i9 achieved by a criterion based on thc minimr~-n distance between 2 SU:L~ LSP co~fl;ri~n~ for the current subfrarne. The minim m distance i9 given by dmin = min~j+l--~j] i = 1,. . .,9. (31) The following linear relation i~ u#d to compute 72:
- 72 =--6.0 * dmin + 1.0, and 0.4 < 72 ~ 0 7 (32) The weighted speech signal iD a subframe i~ given by sw(n) = J(n) + ~ aj71s(n ~ , aj72sW(n--i), n = O, . . ., 39. (33) i=l i=l The ~eighted sp~ signal Jw(n) is used to find an ~tirr sti/~n of the pitch delay in the speech &ame.

3.4 Open-loop pitch analysis To reduce the complexity of the search for the best adaptive codebook delay, the search range is limited around a csn~lid~te delay Top~ obtained from an open-loop pitch analysis. This open-loop P. '~roon 5-lO
pitch analysis i~ done once per frame ( 10 ms.). The open-loop pitch estimation uses the weighted speech signal Jw(n) ot Eq. (33), and is done as follows: In the first step, 3 maximaof the correlatioo R(l~ su(n)sw(n - t) (34) are found in the fol1Owing three ranges i= l: ~0,...,143, i = 2: 40,. . .,79, i = 3: 20, . . . , 39.

The retained maxima R(t,), i = 1, . . ., 3, are normalized through ~/n sw;(n_ t,) The ~inner among the three normalized correlationJ i9 selected by fa~oring the delay~ with the values in the lower range. Thi~ is done by weighting the normalized correlations co..~"onding to the longer delays. The best open-loop delay Top is determined as follows:
Top = tl R~(Top) = f~(tl) f ~(t2) > 0.85R~(Top) ~P(Top~ = R'(t~) Top = t2 cnd if ~(t3) > 0.85R~(Top) R~(Top) = R~(t3) Top = t3 cnd This procedure of di~iding the delay range into 3 sections and favoring the lower sections is used to a.roid choosing pitch ' . ' 3.5 Computation of the impulse response The impulse responYc, h(n), of the weighted synthesi~ filter W(z)/A(z) is computed for each subframc. This impube response is needed for the search of adaptive and fixed codebookc. The impulse response h(n) is computed b- filtering the ~rector of co~ffi~i~n~c of the filter A(z/71) extended by zeros through the two filteK l/A(z) and 1/A(z/~2).

P. I;roon ~10 3.B Computation of the target signal The target signal ~(n) for the adaptive codeboolc search is usually computed b- subtracting the zero-input response of the weighted synthesis filter W(z)/A(z) = A( /~l)/[A(:)~4(:l-r2)] from the weighted speech signal sw(n) of Eq. (33). This is done on a subframe basis.

An equivalent procedure for computing the target signal, which is used in this Recommendation, is the filtering of the LP residual signal r(n) through the combination of synthesis filter l/A(z) and the weighting filter A(z/7l)/A(z/^r2). After determining the e~ccitation for the subframe, the initial states of these filters are updated by filtering the difference between the LP residual ant e~citation. The memory update of these filters is explained in Section 3.10.

The residual signal r(n), which is needed for finding the target ~ector is also used in the adaptive codebook search to extend the past excitation buffer. This simplifies the adaptive codeboolc searcb procedure for delays less than the subframe size of 40 as will be e~plained in the nest #ction. The LP residual is given by r(n) = s(n) + ~ ajs(n--i), n = 0...., 39. (36) ~=1 3.7 Adaptive-codebook search The adaptive-codeboolc parameters (or pitch parameters) are the delay and gain. In the adaptive ~odfbook approach for imp! C the pitch filter, the excitation is repeated for delays less than the subframe length. In the #areh sta~e, the excitation i~ extended by the LP residual to simplify the closed-loop search. The ?~ p1,'~e codeba~~ search is done every (3 ms) subframe. In the first subfrarne, a fractional pitch delay Tl i~ u~ed with a resolution of 1/3 in the range [193, 843] and integers only in the range [85, 1~31. For the second subframe, a delay T2 with a resolution of 1/3 is alway~ used in the range [(int)TI - 53, (int)TI + 432], where (int)TI is the nearest integer to the tr -r~l pitch dely Tl of the first subframe. This range is adapted for the cases where T
straddles the b _ '- of the delay range.

For each subframe the optimal delay is determined using clo~d-loop analysis that minimi7~
the weighted mean-squared error. In the first subframe the delav Tl is found be xarching a small range (6 samples) of delay values around the open-loop delay 1~ (see Section 3.4). The search boundaries tmjn and tmaS are defined by t",in = Top--3 P E;roon ~ io If tmin < 20 tllcn tmi" = 20 tma,~ = tmir~ + 6 if tm~ > 143 ~hcn tm~S = 143 tm~ = tm3r ~ 6 ent For the second subframe, closed-loop pitch analysis is done around the pitch selected in the first subframe to find the optimal delay T2. The search boundaries are between tmir~ - 3 and tm~s + 3 where tmi" and tmaS are derived from Tl as follows:
tmjn = ( ~nt)TI--5 If tmjn < 20 tltcr tmjn = 20 tm~S = tm-n + 9 f tm"r > 143 thcn tmar = 143 tm~ = tmas ~ 9 cnt The closed-loop pitch search minimi7~ the mean-squated ~eighted error between the original and s.~ d speech. This is achieved by m~Yimi7ing the term R(~) ~ 3n9 0 z(n)yt(n) (37) " 0yt(n)y~(n) where ~(n) is the target signal and y~(n) is the past filtered excitation at delay t (past excitation convolved with h(n)). Note that the æarch range is limited around a preselected ~alue, which is the open-loop pitch Tol, for the first subframe, and Tl for the second subframe.
The con~olution y~(n) i9 compukd for the delay tmin, and for the other integer delays in the xarch range ~ = tmjn + l,....tm"5, it is updated using the r~cu.,i~_ relation y~(n) = y~_l(n--I) + u(-~)h(n), n = 39,. . .,0, (38) where u(n), n = -143, . . ., 39, i9 the excitation buffer, and y~_l(--1) = 0. Note that in the search stage, the samples u(n), n = O, . . ., 39 are not known, and they are needed for pitch delays less than 40. To simplify the search, the LP residual is copied to u(n) to make the relation in Eq. (38 ~alid for all delays.

For the determi~tion of T~, and 1-1 if the optimum integer closed-loop delay is less than 84.
the fractions around the optimum integer delay ha~e to be tested. The fractional pitch search is done b~r interpolating the normalized correlation in Eq. (37) and searching for its m~Yiml.m - 2177g22 P. i~roon 5-10 The interpolation is dooe using a FIR filter bl~ based on a Hamming windowed sinc function with tbe sicc truncated at ill and padded with zeros at 112 (b~,(12) = 0). The filter has its cut-off tr~ (-3dB) at 3600 Hz in the oversampled domain. The interpolated values of R(lr) ror the fractioos - 3, --, O, -, and - are obtained u ing the interpolation formula , R(lc - i)bl2(t + i.3) + ~ R(l; + 1 + i)b,2(3 - t + i.3), t = 0,1, 2, (39) i=o .=o where t = 0,1, 2 corresponds to the fractions 0, 3, and 3, respectively. Note that it is necessary to compute correlation terms in Eq. (37) using a range tm,n - 4, tmar + 4, to allow for the proper interpoiation.

3.7.1 Generation of the adaptive codebool~ vector Once the noninteger pitch delay has been determined, the adaptive codebook vector v(n) is com-puted by interpolating the past excitation signal u(n) at the gi~en integer delay 1~ and fraction t t(n)=~u(n-~+i)b3c(t+i.3)+~u(n-~+l+i)b30(3-t+i.3), n=0,.. ,39, t=0,1,2.
.=o i=o (40) The interpolation filter b30 is based on a ~mming windowed sinc functions with the sinc truncated at ~29 and padded with zeros at 1 30 (b30(30) = 0). The filters has a cut-off frequency (-3 dB) at 3600 Hz in the oversarnpled domain.

3.7.2 Codeword rQm, - tatio~ for adapti~e codeboolt delays The pitch delay T~ is encoded with 8 bits in the first subframe and the relative delay in the xcond subf2 - is encoded ~rith 5 bit~. A fi_ ti~nsi delay T is re?r~nted by its integer part (int)T, and a ~ ~ -' pa1t frac/3, frac = -1,0,1. Ihe pitch index Pl is now encoded as Pl ((int)TI - 19) * 3 + frac - 1, if Tl = [19,...,851, frac = [-1,0, 1] (41) ((int)TI--85) + 197, if Tl = [86, ...,143], frac = 0 The value of the pitch delay T2 is encoded relative to the value of Tl. Using the same interpre-tation ;IS before, the fractional delay T2 represented by its integer part (int)~2, and a fractional part fracl3, froc = -1. 0,1, is encoded as P2 = ((int)T2 - tmjn) ~ 3 + frac + 2 (42) - ~177~22 P. I;roon ~10 ~here tm,n is derived from Tl as before.

To make the coder more robust against random bit errors, a parity bit P0 is computed on the delay index of the first subframe. The parity bit is generated through an XOR operation on the
6 most significant bits of Pl. At the decoder this parity bit is reeomputed and if the recomputed value does not agree with the transmitted value, an error concealment procedure is applied.

3.7.3 Computatioll of the adapti~e-codeboolc gain Once the adaptive-codebook delay is determined, the adaptive-codebook gain gp is computed as 9p = n o (n)3/(n) bounded by 0 ~ gp < 1.2, (43) ~here y(n) is the filtered adaptive codebook vector (zero-state response of W(z)/.i(z) to v(n)).
This vector is obtained by con~ul~ing u(n) with h(n) y(n) = ~ v(i)h(n - i) n = 0, . . ., 39- (44) .=o .~ok tha~ by m~Yimi7ing the term in Eq. (37) in most cases gp > O. In case the signal contains only negative cG~le!~tions~ the value of gp is æt to 0.

3.8 Fixed cotebook: structure and search The fixed cod~bool~ is based oo an algebraic codebool~ structure using an interlea ed single-pulse permut~ n (ISPP) design. In thi~ c~debaclt, each codeb,ool~ ector contains 4 noo-zero pulses.
Each puLse can have either the arnplitudes +1 or -1, and can assume the position~ given in Table 7.

The - ~ b- ' lrector c(n) is constructed by talcing a zero vector~ and putting the 4 unit pulses at the found 1~ ~ tiplied with their corle;~yc~~ing sign.
c(n) = JO ~(n - io) + Jl ~(n--il) + s2 ~(n--i2) + s3 ~(n--i3), n = 0, .... 39. (45) ~here ~(0) is a unit pulse. A special feature incorporated in the codeboolt is that the selected code-bool~ vector is filtered through ao adaptive pre-filter P(z) which enhances harmonic components to improve the synt~ - d speech quality. Here the filter P(z) = 1/(1--~z ) (46) P. I;roon .~10 Table 7: Structure of fixed codebooli C.
PUlJC S~gn Posit~on~
iO sO 0, 5, 10, 15, 20, 25, 30, 35 il sl 1,6,11,16,21,26,31,36 i2 ~2 2,7, 12, 17,22,27,32,37 i3 s3 3, 8, 13, 18, 23, 28, 33. 38 4,9, 14, 19,24,29,3~,39 is used, where T i8 the integer component of the pitch delay of the current subframe, and ~ is a pitch gain. The value of ,B is made adaptive by using the quantized adaptive codebook gain from the previous subframe bounded by 0.2 and 0.8.
~=gp~ 1), 0.2~<0.8. (4~) This filter enhances the harmonic structure for delays less than the subframe size of 40. This modification is incorporated in the fixed codebook search by modifying the impulse response h(n).
according to h(n) = h(n) + ,~h(n - T), n = T, ...39. (48) 3.8.1 Fixed-codeboolc search procedure The fi~ced codeboolt is searched by minimi7ing the mean-squared error between ~he weighted input speech sw(n) of Eq. (33), and the weighted re~ .rted speech. The target signal used in the closed-loop pitch search is updated by subtracting the adaptive codebook contribution. That is z2(n) = z(n) - gpy(n), n = 0, ....39. (49) where y(n) is the filtered adaptive codebook vector of Eq. (44).

The matrix H u defined as the lower triangular Toepliz con-olution matrix with diagonal h(0) and lower ~iiae~ nr'- h(l), . . ., h(39). If c~ is the algebraic codevector at index ~! then the codeboo~
is searched by m~Yimi7ing the term C~2 (~3n9 o dttl)Ct(n))2 (_0) E~ c~ ~ct where d(n) is the correlation between the target signal ~q(n) and the impulse response h(n), and = H'H is the matrix of correlations of h(n). The signal d(n) and the matrLlt ~ are computed P. I;roon 5-10 before the codebool~ #arch. The elements of d(n) are computed from d(n) = ~, z(i)h(i - n), n = O, . . ., 39, (51) i=n and the elements of the symmetric matrix ~ are computed by ~(i, j) = ~ h(n--i)h(n - j), (j > i). (52) n=J

Note that only the ele.,l.nt~ actually needed are computed and an efflcient storage procedure has been designet to speed up the #arch procedure.
The algebraic structure of the codebook C allows for a fast search procedure since the codebool~
vector c~ contains only four nonzero pulses. The correlation in the numerator of Eq. (50) for a given vector c~ is given by C = ~ a,d(m,), (53) ~here rn, is the position of the ith pUI# and ai is its amplitude. The energy in the d~nomin~tor of Eq. (50) is gi--en by 3 ~ 3 E = ~, ~(Tnj, mj) + 2 ~ ~ ajaj~(mj, mj). (54) i=O i=O j=i+l To simplify the #arch procelL.~;, the pulse amplitudes are predetr rminr~d by qu~n~i7irl~ the signal d(n). This is done by #ttint the amplitude of a pulse at a certain position equal to the sign of d(n) at tbat position. Bcfore the codeba ~k #arch, thc following steps are done. First, the signal d(n) is d~Yom~ ~d into two si6nals: the absolute signal d'(n) = It(n)l and the sign signal sign[d(n)~. Secood, the matri~t ~ is mo~lifir~d by including the sign information; that is, o'(i, j) = sign [d(i)~ sign[t(;)] ~(i, j), i = 0, . . ., 39, j = i, . . ., 39. (55) To remo~re the f ctor 2 in Eq. (54) ~'(i, i) = 0.5~(i, i), i = 0, . . ., 39. (56) The correlation in Eq. (53) is now given by C = d'(mO) + d'(ml ) + t'(m2) + t'(m3), (57) and the energy in Eq. (54) is given by E = ~'(mo. mo) ~177~22 P. E;roon 5-10 ~.
+ ~'(ml, ml ) + ~'(mO, ml ) + ~'(m2, m2) + ~'(mO, m2) + ~'(ml, m2) + ~'(m3, m3) + ~'(mO, mS) + O'(ml, m3) + ~ (m2, m3) (58) A focused search approach is used to further simplify the search procedure. In this approach a precomputed threshold is tested before entering the last loop, and the loop is entered only if this threshold i9 exceeded. The m~Yiml-m number of times the loop can be entered is fixed so that a low pe~c: ~ge of the codebool~ is searched. The threshold i9 computed based on the correlation C. The m~ m absolute correlation and the average correlation due to the contribution of the first three pulses, ma~S and av3' are found before the codeboolc search. The threshold i9 given by thr3 = at~3 + h'3(ma~3 - atJ3). (59) The fourth loop is entered only if the absolute correlation (due to three pulses) exceeds thr3, where O < K3 < 1. The value of ~3 controls the percentage of code~ook search and it is set here to 0.4.
~ote that this results in a variable search time, and to further control the search the number-of times the last loop is entered (for the 2 subframes) cannot exceed a certain m~i , which is set here to 180 (the average worst case per subframe i9 90 times).

3.8.2 Codeword computation of the fi~ced codeboolc The pulse positions of the pulses iO, i1, ant i2, are encoded witb 3 bits each, whik the position of i3 is encoded with 4 bits. Each pulse amplitude i9 encoded with 1 bit. This gives a total of 17 bits for the 4 pulses. By defining J = 1 if the sign is positive and J = O is the sign is negative, the sign codeword is obtained from S=~0+2~81+4~2+8~J3 (60) and the fi~ed ro~b~lt codc..~ i5 obtained from C = (iO/5) + 8 ~ (il/5) + 64 ~ (i2/5) + 512 ~ (2 ~ (i3/5) + j~) (61) where j~=Oifi3=3,8,..,and j~= l if i3=4,9,...

3.9 Quantization of the gain~

The adaptive-codebook gain (pitch gain) and the fixed (algebraic) codebook gain are vector quan-tized using 7 bits. The gain codebool~ search is dooe by minimi7ing the mean-squared weighted 2177~22 P. ~;roon 5-10 error between original and reconsttucted speech which is given by E = ~c~x + gpy~y + 92ztZ _ 2gpx~y--2gcx~z + ~gFg~ytz (62) where ~ i8 the target vector (see Section 3.6). r is the filtered adapti~-e codebook ector of Eq. (44), and z is the fixed codebook vector convolved with h(n), ~(n) = ~ c(i)h(n - i) n = 0, . . ., 39. (63) 3.9.1 Gain pre.l;cti~,r.

The fixed codebool~ gain gc can be exptessed as 9e = rgc~ (64) ~vhere gc is a predicted ~ain based on previous fixed codebook energies, and ~ i8 a correction factor.

The mean euerur of the fixed codebool~ contribution i8 given by g (40 ~ ') (65) After scaling the vector cj with the fixed codebc ~1; gain 9~, the energy of the scaled fixed codebook is g,iven by 2010ggc + E. Let E(m) be the mean-removed energy (in dB) of tbe (scaled) fixed codebool~ contribution at subframe m, given by E(m) = 2010ggc + E--E, (66) ~here E = 30 dB is the mean energy of the fixed codeboolc excitation. The gain gc can be expressed as a funetion of E(m), E, and E by gc = lo(E )+E-E)/20 (67) The predicted gun 9'c iJ found by predicting the log cn_.c-,~ of the current fixed codebool~
contribution from the l-g ~ n~.g~ of previous fixed codebool~ contributions. The 4th order MA
prediction i8 done a~ follows. The predicted energy is given by ~(m) = ~bj~(m-i), (68) i=l u-here [bl ~2 b3 b~] = [0.68 0.58 0.34 0.19] are the MA prediction coefficients, and R(m) is the quantized ~rersion of the prediction error R(m) at sub&ame m, defined by R(m) = E~m)--E(m) (69) P. ~;roon 5-10 The predicted gain gc i9 found by replacing E(m) by its predicted value in Eq (67).
9' = lolE~'+~-The correctiOD factor ~ ia related to the gain-prediction error by (rn ) ( m) ~ (~ ) ( 71 ) 3.9.2 Codeboo} search for gain quantization The adapti~re-codebook gain, gp, and the factor ~ are vector quantized using a 2-stage conjugate akuctut~d codebool~. The first stage consists of a 3 bit two--lim~ ql codebook 5A, and the second stage consists of a 4 bit two-.limencional codebook 5B. The first element in each codebook represents the quantized adaptive codebook gain gp, and the second element represents the quan-tized fixed coclebool~ gain correction factor i Given codeboolc indices m and n for 5A and 5B, respecti~ely, the qrq-~;7~d adaptive-codebook gain is given by gp = 5A1(m) +5Bl(n). ~72) and the qnq~i7ed fi~ced-cvdcb~olt gain by 9c = g'c i = gc (5A2(m) + 5B,(n)).

Thisa conjugate ak~ 1 ~e simplifies the cotebool~ search, by applying a p~e sele:~ n process.
The optimum pikh gain gp, and fixed-codeboolt gain, gc, are derived from Eq. (62), and are used for the pre-~i~cti~r The codeb~sk ~A contains 8 entries in which the second element (co--~~"onding to gc) has in general larger values than the fir~at element (corresponding to gp). This bias allows a pre sele~lion using the value of gc. ln this pre sel,e t; n process, a cluster of 4 vectors whose second elernent are close to g2c, ~here g2c is derived from gc and gp. Similarly, the codebool~
5B contains 16 entries in which havc a bias towards the first element (corresponding to gp). A
cluster of 8 vectors who~e first ~I~.. t~ are close to gp are selted. Hence for each codebool~
the best 50 % . ~id - vectors are selected. This is follo~ed b- an exhaustive search over the r~m~ining 4 * 8 = 32 possibilitiea, such that the combination of the two indices minimiq~ the weighted mean-squared error of Eq. (62).

P l;roon 5-10 3.9.3 Codeword computatio~l for gain qu~nti~er The codewords GA aod CB for the gain quantizer are obtained from the indices corresponding to the best choice. To reduce the impact of single bit errors the codebook indices are mapped.

3.10 Memory update An update of the states of the synthesis and weighting filters is needed to compute tbe target signal in the next subframe. After the two gains are qu~nti7e~ the excitation signal, u(n). in the present subframe is found by u(n) = gpu(n) + gcc(n), n = O, . . ., 39, (74) where gp and 9~ are the quantized adaptive and fixed codeboolt gains. respecti-ely. r (n) the adaptive codebool~ vector (intpolated past excitation), and c(n) is the fixed codeboolc vtor (algebraic codevtor including pitch sharpening). The states of the filters can be updated by filtering the signal r(n)--u(n) (difference between residual and excitation) through the filters l/A(z) and A(z/-rl)/Atz/r2) for the 40 sample subframe and saving the states of the filters. This would require 3 filter opctal -- A simpler approach, which requires only one filtering is as follows.
The local synthesis speech, s(n), is computed by filtering the excitation signal through l/A(z).
The output of the filter due to the input r(n)--u(n) is equivalent to c(n) = s(n) - s(n). So the states of the synthesis filter l/A(z) are given by c(n), n = 30, . . ., 39. Updating tbe states of the filter A(Z/ tl )/A(z/-r2) can be done by filtering the error signal c(n) through this filter to find the perceptually weighted error cw(n). However, the signal cw(n) can be equi-alentl-- found by ~w(n) = r(n)--gpy(n) + gcz(n). (7~) Since the signals ~(n), y(n), and z(n) are a-ailable, the states of the weighting filter are updated by computing cw(n) ~ in Eq. (75) for n = 30, .... 39. This saves two filter operations.

3.11 Encoder and Decoder initialization All static encoder variables should be initialized to 0, except the variables listed in table 8. These variables need to be initialized for the decoder as well.

P ~;roon ~lo Table 8: Descriptioo of parameters with nonzero i~itializatioo.
Variablc Refc~cncc lnitial ualuc Section 3.8 0.8 ii Sfftion 3.2.4 i~/ll Sfftion 3.2.4 0.9595, R~l') Stion 3.9.1 -14 ~177422 P. ~;roon .5-10 4 Functional description of the decoder The signal flow at the decoder ~as shov n in Section 2 (Figure 3). First the parameters are decoded (LP coefflcients, adaptive codebool~ ~ector, fixed codebook ~ector~ and gains). These decoded parameters are used to compute the reconstructed speech signal. This process is described iD
Section 4.1. This reconstructed signal is enhanced by a post-processing operation consisting of a postfilter and a high-pass filter (Section 4.2). Section 4.3 describes the error concealment procedure used when either a parity error has occurred, or when the frame erasure flag has been set.

4.1 Parameter decoding procedure The transmitted parameters are listed in Table 9. At startup all static encoder variables should be Table 9: Description of transmitted parameters indicea The bitstream ordering is reflected by the order in the table. For each parameter the most significant bit (!.fSB) i5 transmitted first.
Sy n~ol Dc~ n H-tJ
LO S-ritched predictor inde~ of LSP qu~ntizer Ll First stage vector of LSP quantizer 7 L2 Sond stage lower vector of LSP qDantizer 5 L3 Second stage hig~er vector of LSP quantizer 5 Pl Pitcb dela~ 1st subfrarne 8 PO Parit~ bit for pitch Sl Siglls of pul~es ld subfrarne 4 Cl Fi~ced .udeb o ol~ 1st subframe 13 GAl Gain .~ieboolt (stage 1) l~t subfr~me 3 GBl G , codcbool~ (~tage 2) 1st subfr~me 4 P2 Pitch dela~ 2nd subframe 5 S2 Sigus of pulses 2nd subfrarne 4 C2 FLsed ~ debo-l~ 2nd subframe 13 G~2 Gain rr~eboolc (stage 1) 2nd subfra~ne 3 GB2 G~ codcboolc (stage 2) 2nd sub~me 4 initialized to 0, except the ariables listed in Table 8. The decoding process is done in the following order:

217742~
P. I;roon 5-lO
4.1.1 Decoding of LP filter paralneter~

The received indices L0, Ll, L2, and L3 of the LSP quantizer are used to reconstruct the quan-tized LSP coefflcieuts using the procedure described in Section 3.2.4. The interpolation procedure described in Section 3.2.5 is used to obtain 2 interpolated LSP ~-ectors (corresponding to 2 sub-frames). For each subframe, the interpolated LSP vector is con-~erted to LP filter coefflcients aj, which are used for synthesizing the reconstructed speech in the subframe.

The following steps are repeated for each subframe:

1. decoding of the adaptive codebook vector, 2. decoding of the fixed codeboolc vector, 3. decoding of the adaptive and fixed codebook gains, 4. computation of the reconstructed speech, 4.1.2 Decoding of the adapti~e codebook ~ector The received adaptive codeb~-lc index is used to find the integer and f. t; ~~1 parts of the pitch ~elay. The integer part (int)TI and fractional part frac of Tl are obtained from Pl as follows:
if Pl < 197 (int)TI = (Pl+2)/3 + 19 frac = Pl - (int)TI*3 + 58 C~JC
(int)TI = P1- 112 frac = O
en~
The integet and fractional part of T2 are obtained &om P2 and t~an~ ~here tmjn is derived from Pl as follo~
tmjn = (int)TI--5 If tmjn < 20 tllcn tmjn = 20 tmaS = tmin + 9 if tmJS > 143 thcn tm~S = 143 tmjn = tm~s ~ 9 end 2177~22 P. I~;roon ~10 Now T2 i9 obtained from (int)T~ = (P2+2)/3-1 + tmjn frac = P2-2- ((P2+2)/3-1)*3 The adaptivc codeboolc vector v(n) is found by interpolating the past excitation u(n) (at the pitch delay) using Eq. (40).

4.1.3 Decoding of the fi~ced codeboolc vector The received fixed codebook index C is used to extract the positions of the excilation pulses. The pul# signs are obtained from S. Once the pulse positions and signs are decoded the fixed codebool~
vector c(n), can be constructed. If the integer part of the pitch delay, T, is less than the subframe size 40, the pitch enhancement procedure is applied which modifies c(n) according to Eq. (48).

4.1.4 Decoding of the adapti~re and fi~ced codeboo} gainO

The received gain codebook index gives the adaptive cod~book gain jp and ~he fixed codeboo~
gain correction factor j. This procedure is dffcribet in detail in Sectioo 3.9. The ~tim~ted fixed codeboolt gain g' is found using Eq. (70). The fixed codeboolc vector is obtained from the product of the quantized gain correction factor vith this predicted gain (Eq. (64)). The adaptive codeboo~
gain is reconstructed using Eq. (72).

4.1.5 Computation of the parity bit Before the speecb is r~ ~o- ~rl~, the parity bit is reCom~ uted from the adaptive codebool~ delay (Section 3.7.2). If this bit io not identical to the transmitted parity bit P0, it is lil~ely that bit errors occurred during tr ~ and the error conce~ procedure of Stion 4.3 is used.

4.1.6 Cc.~p. - e the r~co~Okucted speech The excitation u(n) at the input of the synthesis filter (see Eq. (74)) is input to the LP s-nthffis filter. The rec~r.Jlructed speech for the subframe is given by s(n) = u(n)--~ajs(n - i), n = 0....,39. (~6) i=l - ~177422 P. ~;roon 5-lO
where aj are the interpolated LP filter coefflcients.

The reconstructed speech s(n)is then processed by a post processor which is described in the next section.

4.2 Post-processing Post-processing consists of three functions: adaptive postfiltering. high-pass filtering, and signal up-scaling. The adaptive postfilter is the cascade of three filters: a pitch postfilter Hp( ), a short-term postfilter H~(:), and a tilt compensation filter Ht(;), followed by an adaptive gain control procedure. The postfilter is updated every subframe of ~ ms. The postfiltering process is organized as follows. First, the synthesis speech s(n) is inverse filtered through i(z/rn) to produce the residual signal r(n). The signal r(n) is used to compute the pitch delay T and gain gpj~. The signal r(n) is filtered through the pitch postfilter Hp(z) to produce the signal r'(n) which, in its turn, is filtered by the synthe.~is filter l/[gfA(z/^~d)]. Finally, the signal at the output of the synthesis filter l/[g~A(z/-rd)~ is passed to the tilt compeLsation filter H~(z) resulting in the postfiltered synthesis speech signal sf(n). Adaptive gain controle is then applied between sf(n) and s(n) resulting in the signal sf'(n). The high-pass filtering and scaling operation operate on the postfiltered signal sf'(n).

4.2.1 Pitch po~tfilter The pitch, or harmonic, p~s "' er is given by Hp(z) = 1 + (1 +gOZ-T), (77) where ~ is the pitch delay and 9o is a gain factor given by go = ~pgpi~. (78) v.here 9pi1 is the pitch gain. Both the pitch delay and gain are determined from the decoder output signal. ~ote that gp" is bounded by 1, and it is set to zero if the pitch prediction gain is less that 3 dB. The factor ~p cootrols the amount of harmonic postfiltering and has the value -tp = 0.5. The pitch delay and gain are computed from the residual signal r(n) obtained by filtering the speech s(n) through .i(z/~n), which is the numerator of the short-term postfilter (see Section 4.2.2) r(n) = s(n) + ~ ~na,s(n - ~). (79) =1 P. I;roon i-10 Tbe pitch delay i9 computed using a two pass procedure. The first pass selects the best integer To in the ran8e [Tl - I,TI + 1], where Tl is the integer part of the (transmitted) pitch delay in the first subframe. The bfft integer delay is the one that maximizes the correlation R(l~ r(n)r(n - k) (80) n=O
The second pass chooses the best fractional delay T with resolution 1/8 around To. This is done by finding the delay witb the highest normalized correlation.
R'(le) = ~n_o r(n)it(n) (81) 3n9 0 ~t(n)it(n) where it(n) is the residual signal at delay /r. Once the optimal delay T is found, the corresponding correlation value is compared against a threshold. If R'(T) < 0.5 then the harmonic postfilter is disabled by setting gp" = 0. Otherwise the value of 9pj~ is computed from:
9Pjt = ~39~ r (n)i ( )' bounded by 0 ~ gpi, < 1Ø (82) The noninteger delayed signal ;t(n) is first computed using an interpolation filter of length 33.
ARer the selection of T, rt(n) is recomputed ~ith a longer interpolation filter of length 129. The nevr signal replaces the prerious one only if the longer filter increases the value of R'(T).

4.2.2 Short-term post1ilter The short-term postfilter is given by Hl(2) = _ ( / ;~ lol ~n i , (83) where d(z) i9 the received qn~n~ -!i LP inver# filter (LP analysis is not done at the decoder), and the factors ~" and 'rd control the amount of short-term postfiltering, and are #t to 'rn = 0 55, and ^~d = 0.7. The gain term gJ i~ c~lc~ ted on the truncated impulse respocse, hl(n), of the filter A(z/-tn)/A(z/-~d) and giren by 19 n=O (84) 4.2.3 Tilt compensation Finally, the filter H~(z) compensates for the tilt in the short-term postfilter Hl(z) and is give~ b~
H~(z) = g (1 + ~tA~Iz 1), (85) .53 .., P. I;roon 5-10 .
here ~ i9 a tilt factor, ~ being the first reflection coefficient calculated on ht(n) with rh(l) rh(i) = ~ h~(j)h~(j + i) (86) Tbe gain term g, = 1- ¦ rt~ compensates for the decreasing effect of 9t in H~(:). Furthermore.
it has been shown that the product filter Hl(:)Ht(z) has generally no gain.

Two values for ~, are used depending on the sign of ~1. If kl is negative, ~t = 0.9, and if kl is positive, ~ = 0.2.

4.2.4 Adaptive gain control Adaptive gain control is used to compensate for gain differences between the reconstructed speech signal s(n) and the postfiltered signal sf(n). The gain scaling factor G for the present subframe is computed by ~n_o l8f(n)l (8~) The gain-scaled postfiltered signal sf'(n) is gi-en by sf'(n) = g(n)sf(n), n = 0, . . ., 39, (8~, where g(n) is updated on a sampl~by-sample basis and given by g(n) = 0.85g(n - 1) + 0.15G, n = 0, . . .,39. (89) The initial value of g(-l) = 1Ø

4.2.5 High-pass filtering and up ~cD~ g A high-pass filter at a cuto~ frequency of 100 lIz is applied to the reconstructed and postfiltered speech ~f'(n). The filter is gi-en by 0.93980581 - 1.8795834z-1 + 0.939~0581z-2 Hh2(2) = 1--1.9330~35z-1 + 0.93589199z-~ ( ) Up-scaling consists of multiplying the high-pass filtered output by a factor 2 to retrieve the input signal le~el.

- 217742~
P. Kroon .;-10 4.3 Concealment of frame erasures and parity errors .~n error concealment procedure has been incorporated in the decoder to reduce the degradations in the reconstructed speech because of frame erasures or random errors in the bitstream. This error conc~ r~ process is functional when either i) the frame of coder parameters (corresponding to a 10 rns frame) has been identified as being erased, or ii~ a checl;sum error occurs on the parity bit for the pitch delay index Pl. The latter could occur wben the bitstream has been corrupted by random bit errors.

If a parity error occurs on Pl, the delay value Tl is set to the value of the delay of the previous frame. Thc value of T2 is derived with the procedure outlined in Section 4.1.2, using this new value of Tl. If consecutive parity errors occur, the previous value of T~, incremented by l, is used.
The mecb~ni~n for detecting frame erasures is not defined in the Recommendation, and will depend on the ~pplicati~n The concealment strateW has to reconstruct the current frame, based on pr~ received information. The method used replaces the missing P-r~it3~i~n signal with one of similar characteristics, while gradually decaying its enerW. Thi~ is done by using a voicing classifier b~ed on the long-term prediction gain, which is computed as part of the long-term postfilter analysis. The pitch postfilter (see Section 4.2.1) finds the long-term predictor for which the pKdiction gain is more than 3 dB. This is done by setting a threshold of 0.5 on the normalized correlation f~(/c) (Eq. (81)). For the error conce~lmPnt process, these frames will be classified as periodic. Otherwise the frame is declared nonperiodic. An erased frame inherits its class from the preceding (r~con;.lr.~ted) speecb frame. Note that the voicing classification is continuously updated based on this rec~ ed speech signal. Hence, for many conxcuti-e erased frames the ~lr ~~ ti )r might change. Typically, this only happens if tbe original classification was periodic.

The spific steps tal~en for an eraxd frame are:

1. Kpetition of the LP filter parameters, 2. ^-t~ ti dadaptive and fixed codebook gains, 3. attrnu?~ion of the memory of the gain predictor, 4. generation of the replacement excitation.

217742~
P. I;roon j-lO
4.3.1 Repetition of LP filter parameters The LP paramekr~ of the last good frame are used. The states Or the LSF predictor contain the values of the received codewords lj. Since the current codeword is not a-ailable it is computed from the repeated LSF parameter ~i and the predictor memory from [ (m) ~ m~l(m--t~] /(1 _ ~ mjc), i = 1, . . ., 10. (91) ~=1 ~=1 4.3.2 AttcLl ~ ~; of adaptive and fixed codebook gains An attenuated version of the previous ~xed codebook gain is used.
9(m) = 0 989(m--1) (92) The same is done for the adaptive codebook gain. In addition a clipping operatioo is used to keep its value below 0.9.
9(m) = o gg(m-l) and 9(m) ~ 0 9 4.3.3 Att~ ~ ~ of the ~ ~ of the gain predictor The gain predictor uses the energy of pre--iously selected codebooks. To allo~ for a smooth continuation of the coder once good frames are received, the memory of the ~in predictor is updated with an att~n~ ed version of the codebook energy. The value Of ~(m~ for the current subframe n is set to the averaged ~ nt;~ed gain prediction error, attenuated by 4 dB.

~(m) = (0.25 ~ ~(m-i)) _ 4.0 and R(m) ~--14.
i=l 4.3.4 C~ of the ~ t excitation The excitation uxd depends orl the periodicity classification. If the last correctl~ received &ame ~as cla~ssified as periodic. the current frame is considered to be periodic as well. In that case only the adaptive codebook is used, and the fixed codebook contribution is set to zero. The pitch delay is based on the last correctly received pitch delay and is repeated for each successive frame. To avoid excessive periodicit- the delay is increased by one for each next subframe but bounded by 143. The adaptive codebook gain is based on an attenuated value according to Eq. (93).

2177~22 P. Kroon 5-10 If the last correctly received frame was classified as nonperiodic, the current frame i~ considered to be nonperiodic as well, and the adaptive codebook contribution is set to zero. The fixed codebook contribution is generated by randomly selecting a codebook index and sign index. The random generator is based on the function seed = seed * 31821 + 13849, (95) with the initial seed value of 21845. The random codebook index is derived from the 13 least significant bits of the next random number. The random sign is derived from the 4 least significant bits of the next random number. The fixed codebook gain is attenuated according to Eq. (92).

s~l77922 -P. ~;roon 5-10 5 Bit-exact description of the CS-ACELP coder A.~rSI C code simulating the C~ACELP coder in 16 bit fi.Yed-point is a~ailable from ITU-T. The following sections summarize thc use of thi9 simulation code. and how the soft~ are is organized.

5.1 Use of the simulation software The C code consists of two main programs coder.c, which simulates the encoder, and dccotor.c, ~hich simulate_ the decoder. The encoder is run as follo~s:
coder inputSilc bstrea filc The inputfile and outputfile are sampled data files containing l~bit PC~ signals. The bitstream file contains 81 16-bit words, where the first word can be used to indicate frame erasure, and the ren~ining 80 words contain one bit each. The decoder taliff this bitstream file and produces a postfiltered output file containing a 16-bit PCM signal.
dccod-r bstrea ~ile output~

5.2 Org~ni~stion of the simulation software In the fixed-point ANSI C sim~ tion~ only two types of fi~ced-point data are used as is sho~n in Table 10. To facilit~te the impl ~.t -- of the cim ~ qn code, loop indices, Boolean values and Table 10: Data types used in A~SI C cimlll~ti~r Typc .Ua~. ualuc A~n. ualuc Dc~ A
Wordl6 0%7fff 0~8000 siRncd ~'s~ 16 bit word Word32 0~7fffffffL OA~OOOOCOOL signed 2 ~ .' Jt 32 bit word flags use the type F14~, ~hich woult be either 16 bit or 32 bits depending on the target platform.

All the computations are done using a predefined set of basic operators. The description of these operators is given in Table 11. The tables used by the simulation coder are summarized in Table 12. These main programs use a library of routines that are summarized in Tables 13. 14, and 15.

2177~2~

P. ~;roon 5-10 Table 11: B~ic operations used in .~'SI C simulation.
Opera~lon Descriptlon Vordl6 sature~Vord32 L_Yarl) Lirnit to 16 bits Vordl6 add(Vordl6 varl, Vortl6 var2) Short addition Vordl6 sub(Vordl6 Yarl, Vordl6 ar2) Sbort subtraction Vordl6 ab~_s(Yordl6 Yarl) Short abs Vordl6 shl(Vordl6 Yarl, Yordl6 ar2) Short shift left Vortl6 shr(Vordl6 varl, Vordl6 rar2) Short shift right Vordl6 ault(Vordl6 arl, Vordl6 ar2) Short multiplication Vord32 L_-ult(Vordl6 arl, Vordl6 ar2) Long multiplicati~~
Vordl6 n-gat-(Vordl6 arl) Short Degate Vordl6 e~tract_h(Vord32 L_Yarl) E~ctract high Vordl6 e~tract_l(Vord32 L_Yarl) E~tract low Vordl6 round(Vord32 L_-arl) Round Vord32 L_aac(Vord32 L_var3, Vordl6 arl, Vordl6 ar2) Mac Vord32 L_aJu(Vord32 L_-ar3, Vordl6 arl, Vordl6 ar2) Msu Vord32 L_-ac~s(Vord32 L_var3, Vordl6 arl, Vordl6 ar2) Mac without sat Yord32 L_~suV-(Vord32 L_rar3, Vordl6 arl, Vortl6 ar2) Msu without sat Vord32 L_add(Vord32 L_-arl, Vord32 L_-ar2) Lon6 addition Vord32 L_sub(Vord32 L_-arl, Vord32 L_-ar2) Long subtraction Vord32 L_add_c (Vord32 L_varl, Vord32 L_Yar2) Lon6 dd rith c Vord32 L_-ub_c(Vord32 L_-arl, Vord32 L_-ar2) Long sub with c Vord32 L_n-gat-(VorU2 L_-arl) Lon8 ne6~te Vortl6 ault_r(Vordl6 arl, Vordl6 ar2) M, ' 'i ~ -q ith rount Vord32 L_shl(Vord32 L_- rl, Vordl6 ar2) Lon6 shift left Vord32 L_shr(Vord32 L_-arl, Vordl6 ar2) Long shift riRht Vordl6 shr_r(Vordl6 arl, Vordl6 ar2) Shift ri6ht ~ith round Vordl6 ac_r(Vord32 L_-ar3, Vordl6 arl, Vordl6 ar2) M~L with roundin6 Vordl6 a-u_r(Vord32 L_-ar3, Vordl6 arl, Vordl6 ar2) Msu wilh rounding Vord32 L_d po-it_h(Vordl6 arl) 16 bit varl -" ~fSB
Vord32 L_d-po-it_l(Vordl6 arl) 16 bit varl -, LSB
Vord32 L_shr_r(Vord32 L_-arl, Vordl6 ar2) Long shift right with round Vord32 L_abs(Vord32 L_-arl) Long abs Vord32 L_sat(Vord32 L_-arl) Long saturation Vordl6 noru_s(Vordl6 arl) Short norm Vordl6 di-_s(Vordl6 arl, Vordl6 ~ar2) Short di~ision Vordl6 nor _l(Vord32 L_-arl) Long norm 217742~

P. ~;roon 5-10 Table 12: Summary of tables.
Flle Ta~le namc Siz~ De~crtptlon tab_hup.c tab_hup_c 28 upsampling filter for postfilter tab_hup.c tab_hup_1 112 upsampling filter for po~tfilter inter_3.c inter_3 13 FIR filter for interpolating the correlation pred_lt3.c int-r_3 31 FIR filter for interpolating past e%citation lspcb.tab lspcbl 128xlO LSP quantizer (first stage) l~pcb.tab l~pcb2 32xlO LSP quantizer (~econd stage) l~pcb.tab fg 2x 4 xlO !IA pI~ :or~l in LSP VQ
l~pcb.tab fg_su 2xlO used in LSP VQ
l~pcb.tab f~_sun_in 2xlO used in LSP VQ
qua_gain.tab gb}1 8x2 .odebv~k GA in 8ain VQ
qua_gain.~ab gbl~2 16x2 codcbv~l~ GB in 8ain VQ
qua_gain.tab yl 8 used in 8ain VQ
qua_gain.tab i apl 8 used in 8ain VQ
qua_gain.tab uy2 16 used in 8ain VQ
qua_gain.tab i-a21 16 used in 8ain VQ
indo-.tab indo- 240 LP analysi~ window 1ag_~ind.tab lag_h 10 laB wintow for baDtwittb ~p~r (high part) lag_-ind.tab la~_l 10 lag window for bantwidth ~ p~r ' A (IOW part) grid.tab grid 61 8rid point in LP to LSP ~
inr_sqrt.tab tabl- 49 lool~up tabb in inverse square root computatioD
log2.tab tabl- 33 lool~up table in base 2 hg;-ntl-- computation l~p_l-t.tab tabl- 65 loolcup table in LSF to LSP, L-~ ' and vice versa l--p_l~f.tab ~lop- 64 line slope~ in LSP to LSF c~
po-2.tab t bl- 33 loolcup table in 25 computation acelp.~ prototypes for fi~ed cv~bvvlc search ld8}.h prototype~ and constants typedef . h type d~fi- t- -~177~2~
.
P I;roon 5-10 Table 13 Summar- of encoder specific routines F~lcnamc De~cription acelp_co c Search fi~ed codebook autocorr c Compute autocorrelation ror LP analysis az_lsp c compute LSPs &om LP co ~ ts cod_ld8lr c encoder routine conrolr-.c convolution operation corr_~y2 c compute cur.~' ~i terms for gain qlan~i7~^cG
enc_lag3 c encode adaptive codebook inde~
g_pitch c compute adaptive codebool~ gain gainpr-d c gain predictor int_lpc c interpolation of LSP
inter_3 c tl_ -; ' delay interpolation lag_rind c lab-windo~nn6 le-inson.c le~inson rursion lsp-nc c LSP encodin6 routinc lspg-tq c LSP quanti er l-plS-tt c computc LSP quantizer distortion lspg-t- c compute LSP weights lspla t c select LSP ~IA predictor l-ppr- c p ~ ~ e first LSP codeb~t l-ppr-- c LSP pre~ictor routines lsp--ll c first stage LSP quantizer lsp--l2 c ~econd stage LSP quanti_er lsp tab c stability test ~or LSP quanti_er pitch_ir c clo#d-loop pitch search pitch_ol c open-loop pitch search pr-_proc c pre-pr. e (HP filterin8 and scaling) p~ c computation of perceptual ~ ;bhti g ~ s qua_~ain c gain quanti_er qua_lsp c LSP quantizer relsp-- c LSP quanti_er 217742~

P Kroon 5-10 Table 1~ Summary of decoder specific rout ines Filenamc Dc~crip~lon d_lsp.c decode LP inlo~ n de_acelp.c decode algebraic codebool~
dec_gain . c decode gains d-c_lag3.c decode adapti~e codebook inde~c dec_ld8~ . c decoder routine l~pd-c.c LSP decoding routine po~t_pro.c post processing (HP filtering and scaling) pred_lt3.c generatioo of adaptive co~ebook pct.c postfilter ~outincs Table 15: Summary of general routines.
Filcnamc D~c. .r t J .-ba-icop2 . c b~ic oper~'~
bit~.c bit mani~ routines gainpr~d.c gain predictor int_lpc.c inkrpolation of LSP
int-r_3.c L 1 delay interpolation l~p_az.e compute LP from LSP c - 1 -l~p_l-S.c ~ between LSP and LSF
l-p_l-f2.c hi61~ preci ion c~r-- bet-veen LSP and LSF
18p-~p.c ~p~r ' of LSP ~ t~
l~p-tab.c stability test for LSP quantizer p_parit~. c compute pitch pa ity pred_lt3.c generation of adaptive col~boolc rando-.c random generator re~idu.c compute residual signal s~n_f ilt . c srnthis filter ight_a.c bandwidth e-p-- n LP co lfic;~nts

Claims (20)

1. A method for use in a speech decoder which includes a first portion comprising an adaptive codebook and a second portion comprising a fixed codebook, said decoder generating a speech excitation signal selectively based on output signals from said first and second portions when said decoder fails to receive reliably at least a portion of a current frame of compressed speech information, the method comprising:
classifying a speech signal to be generated by the decoder as periodic or non-periodic;
based on the classification of the speech signal, either generating said excitation signal based on the output signal from said first portion and not on the output signal from said second portion if the speech signal is classified as periodic, or generating said excitation signal based on the output signal from said second portion and not on the output signal from said first portion if the speech signal is classified as non-periodic.
2. The method of claim 1 wherein the step of classifying is performed based on information provided by an adaptive post-filter.
3. The method of claim 1 wherein the classification of the speech signal is based on compressed speech information received in a previous frame.
4. The method of claim 1 wherein the output signal from said first portion is generated based on a vector signal from said adaptive codebook, the method further comprising:
determining an adaptive codebook delay signal based on a measure of a speech signal pitch-period received by the decoder in a previous frame; and selecting the vector signal with use of the adaptive codebook delay signal.
5. The method of claim 4 wherein the step of determining the adaptive codebook delay signal comprises incrementing the measure of speech signal pitch-period by one or more speech signal sample intervals.
6. The method of claim 1 wherein the first portion further comprises an amplifier for generating an amplified signal based on a vector signal from the adaptive codebook and a scale-factor, the method further comprising determining the scale-factor based on scale-factor information received by the decoder in a previous frame.
7. The method of claim 6 wherein the step of determining the scale-factor comprises attenuating a scale-factor corresponding to scale-factor information of said previous frame.
8. The method of claim 1 wherein the output signal from said second portion is based on a vector signal from said fixed codebook, the method further comprising:
determining a fixed codebook index signal with use of a random number generator;and selecting the vector signal with use of the fixed codebook index signal.
9. The method of claim 1 wherein the second portion further comprises an amplifier for generating an amplified signal based on a vector signal from the fixed codebook and a scale-factor, the method further comprising determining the scale-factor based on scale-factor information received by the decoder in a previous frame.
10. The method of claim 9 wherein the step of determining the scale-factor comprises attenuating a scale-factor corresponding to scale factor information of said previous frame.
11. A speech decoder for generating a speech signal based on compressed speech information received from a communication channel, the decoder comprising:

an adaptive codebook memory;
a fixed codebook memory;
means for classifying the speech signal to be generated by the decoder as periodic or non-periodic;
means for forming an excitation signal, said means comprising first means for forming an excitation signal when said decoder fails to receive reliably at least a portion of a current frame of compressed speech information, said first means forming said excitation signal based on a vector signal from said adaptive codebook memory and not based on a vector signal from said fixed codebook memory, when the speech signal to be generated is classified as periodic, and based on a vector signal from said fixed codebook memory and not on a vector signal from said adaptive codebook memory, when said speech signal to be generated is classified as non-periodic; and a linear predictive filter for synthesizing a speech signal based on said excitation signal.
12. The decoder of claim 11 wherein the means for classifying comprises a portion of an adaptive post-filter.
13. The decoder of claim 11 wherein the means for classifying classifies the speech signal based on compressed speech information received in a previous frame.
14. The decoder of claim 11 further comprising:
means for determining an adaptive codebook delay signal based on a measure of a speech signal pitch-period received by the decoder in a previous frame; and means for selecting the vector signal from the adaptive codebook memory with useof the adaptive codebook delay signal.
15. The decoder of claim 14 wherein the means for determining the adaptive codebook delay signal comprises means for incrementing the measure of speech signal pitch-period by one or more speech signal sample intervals.
16. The decoder of claim 11 further comprising:
an amplifier for generating an amplified signal based on a vector signal from the adaptive codebook and a scale-factor; and means for determining the scale-factor based on scale-factor information received by the decoder in a previous frame.
17. The decoder of claim 16 wherein the means for determining the scale-factor comprises means for attenuating a scale-factor corresponding to said previous frame.
18. The decoder of claim 11 further comprising a random number generator, said generator for determining a fixed codebook index signal for use in selecting the fixed codebook vector signal.
19. The decoder of claim 11 further comprising:
an amplifier for generating an amplified signal based on the vector signal from said fixed codebook and a scale-factor; and means for determining the scale-factor based on scale-factor information received by the decoder in a previous frame.
20. The decoder of claim 19 wherein the means for determining the scale-factor comprises menas for attenuating a scale-factor corresponding to said previous frame.
CA002177422A 1995-06-07 1996-05-27 Voice/unvoiced classification of speech for use in speech decoding during frame erasures Expired - Lifetime CA2177422C (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US482,708 1995-06-07
US08/482,708 US5732389A (en) 1995-06-07 1995-06-07 Voiced/unvoiced classification of speech for excitation codebook selection in celp speech decoding during frame erasures

Publications (2)

Publication Number Publication Date
CA2177422A1 CA2177422A1 (en) 1996-12-08
CA2177422C true CA2177422C (en) 2000-08-22

Family

ID=23917119

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002177422A Expired - Lifetime CA2177422C (en) 1995-06-07 1996-05-27 Voice/unvoiced classification of speech for use in speech decoding during frame erasures

Country Status (8)

Country Link
US (1) US5732389A (en)
EP (1) EP0747883B1 (en)
JP (1) JPH09120298A (en)
KR (1) KR100389178B1 (en)
AU (1) AU5464996A (en)
CA (1) CA2177422C (en)
DE (1) DE69613908T2 (en)
ES (1) ES2163589T3 (en)

Families Citing this family (86)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH08179796A (en) * 1994-12-21 1996-07-12 Sony Corp Voice coding method
FR2729247A1 (en) * 1995-01-06 1996-07-12 Matra Communication SYNTHETIC ANALYSIS-SPEECH CODING METHOD
FR2729246A1 (en) * 1995-01-06 1996-07-12 Matra Communication SYNTHETIC ANALYSIS-SPEECH CODING METHOD
EP0773533B1 (en) * 1995-11-09 2000-04-26 Nokia Mobile Phones Ltd. Method of synthesizing a block of a speech signal in a CELP-type coder
TW307960B (en) * 1996-02-15 1997-06-11 Philips Electronics Nv Reduced complexity signal transmission system
JP2940464B2 (en) * 1996-03-27 1999-08-25 日本電気株式会社 Audio decoding device
GB2312360B (en) * 1996-04-12 2001-01-24 Olympus Optical Co Voice signal coding apparatus
JPH1020891A (en) * 1996-07-09 1998-01-23 Sony Corp Method for encoding speech and device therefor
DE69737012T2 (en) * 1996-08-02 2007-06-06 Matsushita Electric Industrial Co., Ltd., Kadoma LANGUAGE CODIER, LANGUAGE DECODER AND RECORDING MEDIUM THEREFOR
ES2373968T3 (en) * 1997-02-10 2012-02-10 Koninklijke Philips Electronics N.V. COMMUNICATION NETWORK TO TRANSMIT VOICE SIGNALS.
US5970444A (en) * 1997-03-13 1999-10-19 Nippon Telegraph And Telephone Corporation Speech coding method
KR100198476B1 (en) 1997-04-23 1999-06-15 윤종용 Quantizer and the method of spectrum without noise
IL120788A (en) * 1997-05-06 2000-07-16 Audiocodes Ltd Systems and methods for encoding and decoding speech for lossy transmission networks
US6014623A (en) * 1997-06-12 2000-01-11 United Microelectronics Corp. Method of encoding synthetic speech
JP3206497B2 (en) * 1997-06-16 2001-09-10 日本電気株式会社 Signal Generation Adaptive Codebook Using Index
US5966688A (en) * 1997-10-28 1999-10-12 Hughes Electronics Corporation Speech mode based multi-stage vector quantizer
JP3618217B2 (en) * 1998-02-26 2005-02-09 パイオニア株式会社 Audio pitch encoding method, audio pitch encoding device, and recording medium on which audio pitch encoding program is recorded
FI113571B (en) * 1998-03-09 2004-05-14 Nokia Corp speech Coding
JP3554567B2 (en) * 1998-05-27 2004-08-18 株式会社エヌ・ティ・ティ・ドコモ Audio decoder and audio decoding method
US6714908B1 (en) * 1998-05-27 2004-03-30 Ntt Mobile Communications Network, Inc. Modified concealing device and method for a speech decoder
US6810377B1 (en) * 1998-06-19 2004-10-26 Comsat Corporation Lost frame recovery techniques for parametric, LPC-based speech coding systems
JP3319396B2 (en) * 1998-07-13 2002-08-26 日本電気株式会社 Speech encoder and speech encoder / decoder
US7072832B1 (en) 1998-08-24 2006-07-04 Mindspeed Technologies, Inc. System for speech encoding having an adaptive encoding arrangement
US6330533B2 (en) * 1998-08-24 2001-12-11 Conexant Systems, Inc. Speech encoder adaptively applying pitch preprocessing with warping of target signal
US6188980B1 (en) * 1998-08-24 2001-02-13 Conexant Systems, Inc. Synchronized encoder-decoder frame concealment using speech coding parameters including line spectral frequencies and filter coefficients
US6188981B1 (en) * 1998-09-18 2001-02-13 Conexant Systems, Inc. Method and apparatus for detecting voice activity in a speech signal
FI114833B (en) * 1999-01-08 2004-12-31 Nokia Corp A method, a speech encoder and a mobile station for generating speech coding frames
EP1039442B1 (en) * 1999-03-25 2006-03-01 Yamaha Corporation Method and apparatus for compressing and generating waveform
US6393394B1 (en) * 1999-07-19 2002-05-21 Qualcomm Incorporated Method and apparatus for interleaving line spectral information quantization methods in a speech coder
US6457036B1 (en) * 1999-08-24 2002-09-24 Avaya Technology Corp. System for accurately performing an integer multiply-divide operation
US6775649B1 (en) * 1999-09-01 2004-08-10 Texas Instruments Incorporated Concealment of frame erasures for speech transmission and storage system and method
US6959274B1 (en) 1999-09-22 2005-10-25 Mindspeed Technologies, Inc. Fixed rate speech compression system and method
US6782360B1 (en) * 1999-09-22 2004-08-24 Mindspeed Technologies, Inc. Gain quantization for a CELP speech coder
USRE43209E1 (en) 1999-11-08 2012-02-21 Mitsubishi Denki Kabushiki Kaisha Speech coding apparatus and speech decoding apparatus
JP3594854B2 (en) 1999-11-08 2004-12-02 三菱電機株式会社 Audio encoding device and audio decoding device
US6826527B1 (en) * 1999-11-23 2004-11-30 Texas Instruments Incorporated Concealment of frame erasures and method
US6792405B2 (en) * 1999-12-10 2004-09-14 At&T Corp. Bitstream-based feature extraction method for a front-end speech recognizer
US7110947B2 (en) * 1999-12-10 2006-09-19 At&T Corp. Frame erasure concealment technique for a bitstream-based feature extractor
US7574351B2 (en) * 1999-12-14 2009-08-11 Texas Instruments Incorporated Arranging CELP information of one frame in a second packet
WO2002003377A1 (en) * 2000-07-05 2002-01-10 Koninklijke Philips Electronics N.V. Method of calculating line spectral frequencies
FR2813722B1 (en) * 2000-09-05 2003-01-24 France Telecom METHOD AND DEVICE FOR CONCEALING ERRORS AND TRANSMISSION SYSTEM COMPRISING SUCH A DEVICE
JP3558031B2 (en) * 2000-11-06 2004-08-25 日本電気株式会社 Speech decoding device
KR100910282B1 (en) * 2000-11-30 2009-08-03 파나소닉 주식회사 Vector quantizing device for lpc parameters, decoding device for lpc parameters, recording medium, voice encoding device, voice decoding device, voice signal transmitting device, and voice signal receiving device
EP1235203B1 (en) * 2001-02-27 2009-08-12 Texas Instruments Incorporated Method for concealing erased speech frames and decoder therefor
US6996522B2 (en) * 2001-03-13 2006-02-07 Industrial Technology Research Institute Celp-Based speech coding for fine grain scalability by altering sub-frame pitch-pulse
US6766289B2 (en) * 2001-06-04 2004-07-20 Qualcomm Incorporated Fast code-vector searching
JP2003044098A (en) * 2001-07-26 2003-02-14 Nec Corp Device and method for expanding voice band
US6754203B2 (en) * 2001-11-27 2004-06-22 The Board Of Trustees Of The University Of Illinois Method and program product for organizing data into packets
US7236928B2 (en) * 2001-12-19 2007-06-26 Ntt Docomo, Inc. Joint optimization of speech excitation and filter parameters
US7177304B1 (en) * 2002-01-03 2007-02-13 Cisco Technology, Inc. Devices, softwares and methods for prioritizing between voice data packets for discard decision purposes
US20030135374A1 (en) * 2002-01-16 2003-07-17 Hardwick John C. Speech synthesizer
CA2388439A1 (en) * 2002-05-31 2003-11-30 Voiceage Corporation A method and device for efficient frame erasure concealment in linear predictive based speech codecs
JP2004151123A (en) * 2002-10-23 2004-05-27 Nec Corp Method and device for code conversion, and program and storage medium for the program
US6961696B2 (en) * 2003-02-07 2005-11-01 Motorola, Inc. Class quantization for distributed speech recognition
US20040176950A1 (en) * 2003-03-04 2004-09-09 Docomo Communications Laboratories Usa, Inc. Methods and apparatuses for variable dimension vector quantization
US7024358B2 (en) * 2003-03-15 2006-04-04 Mindspeed Technologies, Inc. Recovering an erased voice frame with time warping
US20050049853A1 (en) * 2003-09-01 2005-03-03 Mi-Suk Lee Frame loss concealment method and device for VoIP system
KR101008022B1 (en) * 2004-02-10 2011-01-14 삼성전자주식회사 Voiced sound and unvoiced sound detection method and apparatus
BRPI0418838A (en) * 2004-05-17 2007-11-13 Nokia Corp method for supporting an audio signal encoding, module for supporting an audio signal encoding, electronic device, audio encoding system, and software program product
EP1775717B1 (en) * 2004-07-20 2013-09-11 Panasonic Corporation Speech decoding apparatus and compensation frame generation method
KR100647290B1 (en) * 2004-09-22 2006-11-23 삼성전자주식회사 Voice encoder/decoder for selecting quantization/dequantization using synthesized speech-characteristics
CN1763844B (en) * 2004-10-18 2010-05-05 中国科学院声学研究所 End-point detecting method, apparatus and speech recognition system based on sliding window
US7630886B2 (en) * 2005-01-14 2009-12-08 Nokia Corporation Hochwald construction of unitary matrix codebooks via eigen coordinate transformations
JP5009910B2 (en) * 2005-07-22 2012-08-29 フランス・テレコム Method for rate switching of rate scalable and bandwidth scalable audio decoding
JP5142727B2 (en) * 2005-12-27 2013-02-13 パナソニック株式会社 Speech decoding apparatus and speech decoding method
US8255207B2 (en) * 2005-12-28 2012-08-28 Voiceage Corporation Method and device for efficient frame erasure concealment in speech codecs
JP3981399B1 (en) * 2006-03-10 2007-09-26 松下電器産業株式会社 Fixed codebook search apparatus and fixed codebook search method
AU2011202622B2 (en) * 2006-03-10 2011-08-11 Iii Holdings 12, Llc Fixed codebook searching apparatus and fixed codebook searching method
EP2228789B1 (en) * 2006-03-20 2012-07-25 Mindspeed Technologies, Inc. Open-loop pitch track smoothing
KR100900438B1 (en) * 2006-04-25 2009-06-01 삼성전자주식회사 Apparatus and method for voice packet recovery
US20070282601A1 (en) * 2006-06-02 2007-12-06 Texas Instruments Inc. Packet loss concealment for a conjugate structure algebraic code excited linear prediction decoder
EP2040251B1 (en) * 2006-07-12 2019-10-09 III Holdings 12, LLC Audio decoding device and audio encoding device
CN101583995B (en) 2006-11-10 2012-06-27 松下电器产业株式会社 Parameter decoding device, parameter encoding device, and parameter decoding method
US8521519B2 (en) * 2007-03-02 2013-08-27 Panasonic Corporation Adaptive audio signal source vector quantization device and adaptive audio signal source vector quantization method that search for pitch period based on variable resolution
CN102576541B (en) * 2009-10-21 2013-09-18 杜比国际公司 Oversampling in a combined transposer filter bank
CN102687199B (en) * 2010-01-08 2015-11-25 日本电信电话株式会社 Coding method, coding/decoding method, code device, decoding device
US8542766B2 (en) * 2010-05-04 2013-09-24 Samsung Electronics Co., Ltd. Time alignment algorithm for transmitters with EER/ET amplifiers and others
KR101847213B1 (en) * 2010-09-28 2018-04-11 한국전자통신연구원 Method and apparatus for decoding audio signal using shaping function
KR20120032444A (en) 2010-09-28 2012-04-05 한국전자통신연구원 Method and apparatus for decoding audio signal using adpative codebook update
US9047863B2 (en) * 2012-01-12 2015-06-02 Qualcomm Incorporated Systems, methods, apparatus, and computer-readable media for criticality threshold control
TWI591620B (en) 2012-03-21 2017-07-11 三星電子股份有限公司 Method of generating high frequency noise
US9728200B2 (en) * 2013-01-29 2017-08-08 Qualcomm Incorporated Systems, methods, apparatus, and computer-readable media for adaptive formant sharpening in linear prediction coding
US9418671B2 (en) * 2013-08-15 2016-08-16 Huawei Technologies Co., Ltd. Adaptive high-pass post-filter
MX355258B (en) * 2013-10-18 2018-04-11 Fraunhofer Ges Forschung Concept for encoding an audio signal and decoding an audio signal using deterministic and noise like information.
SG11201603000SA (en) * 2013-10-18 2016-05-30 Fraunhofer Ges Forschung Concept for encoding an audio signal and decoding an audio signal using speech related spectral shaping information
FR3013496A1 (en) * 2013-11-15 2015-05-22 Orange TRANSITION FROM TRANSFORMED CODING / DECODING TO PREDICTIVE CODING / DECODING

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5091945A (en) * 1989-09-28 1992-02-25 At&T Bell Laboratories Source dependent channel coding with error protection
JP3102015B2 (en) * 1990-05-28 2000-10-23 日本電気株式会社 Audio decoding method
US5495555A (en) * 1992-06-01 1996-02-27 Hughes Aircraft Company High quality low bit rate celp-based speech codec
SE502244C2 (en) * 1993-06-11 1995-09-25 Ericsson Telefon Ab L M Method and apparatus for decoding audio signals in a system for mobile radio communication

Also Published As

Publication number Publication date
MX9602142A (en) 1997-09-30
JPH09120298A (en) 1997-05-06
ES2163589T3 (en) 2002-02-01
KR970004467A (en) 1997-01-29
EP0747883A3 (en) 1997-09-17
EP0747883B1 (en) 2001-07-18
DE69613908T2 (en) 2002-04-04
AU5464996A (en) 1996-12-19
KR100389178B1 (en) 2003-09-02
US5732389A (en) 1998-03-24
CA2177422A1 (en) 1996-12-08
EP0747883A2 (en) 1996-12-11
DE69613908D1 (en) 2001-08-23

Similar Documents

Publication Publication Date Title
CA2177422C (en) Voice/unvoiced classification of speech for use in speech decoding during frame erasures
AU709754B2 (en) Pitch delay modification during frame erasures
CA2177414C (en) Improved adaptive codebook-based speech compression system
JP5412463B2 (en) Speech parameter smoothing based on the presence of noise-like signal in speech signal
JP5166425B2 (en) Method and device for encoding transition frames in speech signals
EP1576585B1 (en) Method and device for robust predictive vector quantization of linear prediction parameters in variable bit rate speech coding
US5060269A (en) Hybrid switched multi-pulse/stochastic speech coding technique
US5778335A (en) Method and apparatus for efficient multiband celp wideband speech and music coding and decoding
EP1979895A1 (en) Method and device for efficient frame erasure concealment in speech codecs
JPH09127991A (en) Voice coding method, device therefor, voice decoding method, and device therefor
JPH10124094A (en) Voice analysis method and method and device for voice coding
JP3062226B2 (en) Conditional stochastic excitation coding
EP0747884B1 (en) Codebook gain attenuation during frame erasures
JPH1063297A (en) Method and device for voice coding
Teague Harmonic pre-emphasis for improving the performance of FS-1016 CELP
MXPA96002142A (en) Speech classification with voice / no voice for use in decodification of speech during decorated by quad
MXPA96002143A (en) System for speech compression based on adaptable codigocifrado, better

Legal Events

Date Code Title Description
EEER Examination request
MKEX Expiry

Effective date: 20160527