CA2490064A1 - Audio coding method and apparatus using harmonic extraction - Google Patents

Audio coding method and apparatus using harmonic extraction Download PDF

Info

Publication number
CA2490064A1
CA2490064A1 CA002490064A CA2490064A CA2490064A1 CA 2490064 A1 CA2490064 A1 CA 2490064A1 CA 002490064 A CA002490064 A CA 002490064A CA 2490064 A CA2490064 A CA 2490064A CA 2490064 A1 CA2490064 A1 CA 2490064A1
Authority
CA
Canada
Prior art keywords
audio data
pcm audio
harmonic components
pcm
harmonic
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
CA002490064A
Other languages
French (fr)
Inventor
Ho-Jin Ha
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Ho-Jin Ha
Samsung Electronics Co., Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ho-Jin Ha, Samsung Electronics Co., Ltd. filed Critical Ho-Jin Ha
Publication of CA2490064A1 publication Critical patent/CA2490064A1/en
Abandoned legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/02Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
    • HELECTRICITY
    • H03ELECTRONIC CIRCUITRY
    • H03MCODING; DECODING; CODE CONVERSION IN GENERAL
    • H03M7/00Conversion of a code where information is represented by a given sequence or number of digits to a code where the same, similar or subset of information is represented by a different sequence or number of digits
    • H03M7/30Compression; Expansion; Suppression of unnecessary data, e.g. redundancy reduction

Abstract

A method and apparatus for effectively encoding an audio signal into a Moving Picture Experts Group (MPEG)-1 layer III audio signal of a low-speed bitrate.
In the audio encoding method, harmonic components are extracted using fast Fourier transformation (FFT) result information that is obtained by applying psycho-acoustic model 2 to received pulse code modulation (PCM) audio data.
Then, the extracted harmonic components are removed from the received PCM
audio data. Thereafter, the PCM audio data from which the extracted harmonic components are removed is subjected to a modified discrete cosine transform (MDCT) and quantization. Accordingly, efficient encoding can be achieved even using a small number of allocated bits.

Description

AUDIO CODING METHOD AND APPARATUS
USING HARMONIC EXTRACTION
Technical Field s The present invention relates to a method of compressing an audio signal, and more particularly, to a method of and apparatus for efficiently compressing an audio signal into an MPEG-1 layer-3 audio signal with a low-speed bit rate.
Io Background Art Moving Picture Experts Group-1 (MPEG-1 ) establishes standards regarding digital video compression and digital audio compression and is supported by the International Standardization Organization (ISO).
The MPEG-1 audio standard is used to compress 16-bit audio that is is sampled at a 44.1 Khz sampling rate and stored on a 60-minute or 72-minute CD and is classified into 3 layers according to a compression method and the complexity of a codec.
Layer III is the most complex, uses much more filters than layer II, and adopts Huffman coding. Upon encoding at 112 Kbps, an 2o excellent-quality sound can be heard. Upon encoding at 128 Kbps, a sound nearly the same as the original sound is obtained. Upon encoding at 160 Kbps or 192 Kbps, an excellent sound that a human ear cannot distinguish from the original sound can be heard. In general, MPEG-1 layer-3 audio is referred to as MP3 audio.
2s MP3 audio is produced through a discrete cosine transform (DCT), bit allocation based on psycho-acoustic model 2, quantization, and the like. To be more specific, while the number of bits used to compress audio data is kept to be minimal, modified DCT (MDCT) is performed using the result of psycho-acoustic model 2.

In audio compression techniques, the ear of a human is the most important. The human ear cannot hear if the intensity of a sound is at or below a predetermined level. If someone talks loudly in the office, it can be easily recognized who is talking. However, if an airplane s passes at that moment, the talking cannot be heard. Even after the airplane has passed, the talking still cannot be heard because of a lingering sound. Accordingly, in psycho-acoustic model 2, data having a volume equal to or greater than a masking threshold is sampled among data having a volume equal to or greater than the minimum to audible limit corresponding to when it is quiet. The sampling is performed on each sub-band.
However, when a sound signal is compressed at a low-speed bit rate of no more than 64Kbps, psycho-acoustic model 2 is not suitable because the number of bits used to quantize a signal such as a pre-echo is signal is limited. Hence, in order to overcome this problem caused by low-speed MP3 audio, the present invention provides a method of effectively processing an audio signal at a low speed by removing a harmonic component from an original signal using a fast Fourier transform (FFT) adopted in psycho-acoustic model 2 and compressing 20 only a transient component using MDCT.
In an FFT process adopted in a conventional psycho-acoustic model, only signal analysis is performed, and the result of the FFT is not used. Since the result of the FFT is not used for signal compression, it can be considered to be a waste of resources.
2s Korean Patent Publication No. 1995-022322 discloses a bit allocation method employing a psycho-acoustic model. However, the disclosed method is different from a method of the present invention for increasing compression efficiency by removing a harmonic component from an original signal using the result of an FFT adopted in a 3o psycho-acoustic model.
2 Korean Patent Publication No. 1998-072457 discloses a signal processing method and apparatus in the psycho-acoustic model 2, by which the amount of computation is significantly reduced by reducing computation overload while compressing an audio signal. That is, the s disclosed signal processing method includes a step of obtaining an individual masking boundary value using an FFT result, a step of selecting a global masking boundary value, and a step of shifting to the next frequency position. This method is the same as the present invention in that an FFT result value is used but different in that it uses a to different quantization method.
U.S. Patent No. 5,930,373 discloses a method for enhancing the quality of a sound signal using the residue harmonics of a low frequency signal. However, the disclosed method and the quantization method according to the present invention are different in that they use different is techniques of using residue harmonics.
Disclosure of the Invention To solve the above and other problems, it is an aspect of the present invention to provide a method of effectively processing an audio 2o signal at a low speed by removing a harmonic component from an original audio signal using the result of a fast Fourier transform (FFT) used in psycho-acoustic model 2 and compressing only a residue transient using a modified discrete cosine transform (MDCT).
The above and other aspects of the present invention are 2s achieved by an audio coding method using harmonic components. In this method, first, pulse code modulation (PCM) audio data are received, and harmonic components are extracted from the received PCM audio data by applying psycho-acoustic model 2. Next, a modified discrete cosine transform (MDCT) is performed on the received PCM audio data 3o from which the extracted harmonic components are removed.
3 Thereafter, the MDCTed audio data is quantized, and an audio packet is produced from quantized audio data and the extracted harmonic components.
The above and other aspects of the present invention are also .
s achieved by an audio coding method using harmonic components, in which PCM audio data is first received and stored. Then, psycho-acoustic model 2 based on the audible limit characteristics of a human is applied to the stored data to obtain fast Fourier transformation (FFT) result, perceptual energy information regarding received data, and to bit allocation information used for quantization. Thereafter, harmonic components are extracted from the received PCM audio data using the FFT result information. Next, the extracted harmonic components are encoded, and the encoded harmonic components are decoded. Then, a MDCT is performed on a number of samples of the received PCM
Is audio data from which the extracted harmonic components are removed, which depends on the value of the perceptual energy information.
Thereafter, the MDCTed audio data is quantized by allocating bits according to the bit allocation information. Finally, an audio packet is produced from the quantized, MDCTed audio data and the encoded 2o harmonic components.
The above and other aspects of the present invention are still achieved by an audio coding apparatus using harmonic components. In the apparatus, a PCM audio data storage unit receives and stores PCM
audio data. A psycho-acoustic model 2 performing unit receives the 25 PCM audio data from the PCM audio data storage unit and performs psycho-acoustic model 2 to obtain FFT result information, perceptual energy information regarding received data, and bit allocation information used for quantization. A harmonic extraction unit extracts harmonic components from the received PCM audio data using the FFT
3o result information. A harmonic encoding unit encodes the extracted
4
5 PCT/KR02/02348 harmonic components outputting encoded harmonic components. A
harmonic decoding unit decodes the encoded harmonic components.
An MDCT unit performs a MDCT on the stored PCM audio data from which the decoded harmonic components are removed, according to the s perceptual energy information. A quantization unit quantizes the MDCTed audio data according to the bit allocation information. An MPEG layer III bitstream production unit transforms the quantized, MDCTed audio data and the encoded harmonic components output from the harmonic encoding unit into an MPEG audio layer III packet.
to To achieve the above and other aspects, the present invention provides a computer readable recording medium which stores a computer program for executing the above methods.
Brief Description of the Drawings is FIG. 1 shows the format of an MPEG-1 layer III audio stream;
FIG. 2 is a block diagram of an apparatus for producing an MPEG-1 layer III audio stream;
FIG. 3 is a flowchart illustrating a computation process in a psycho-acoustic model;
2o FIG. 4 is a block diagram of an apparatus according to the present invention for producing a low-speed MPEG-1 layer III audio stream;
FIG. 5 is a flowchart illustrating harmonic extraction, harmonic encoding, and harmonic decoding based on psycho-acoustic model 2;
2s FIGS. 6A, 6B, 6C, and 6D illustrate harmonic component samples extracted in stages in order to extract harmonic components using an FFT result in psycho-acoustic model 2;
FIG. 7 is a table showing limited frequency ranges varying according to K values; and FIG. 8 is a flowchart illustrating a process according to the present invention for producing an audio stream by removing a harmonic component.
s Best mode for carryina out the Invention Referring to FIG. 1, a moving picture experts group (MPEG)-1 layer III audio stream is composed of audio access units (AAUs) 100.
The AAU 100 is a minimal unit that can be independently accessed, and compresses and stores data with a fixed number of samples. The AAU
l0 100 includes a header 110, a cyclic redundancy check (CRC) 120, audio data 130, and auxiliary data 140.
The header 110 stores a syncword, ID information, layer information, information regarding whether a protection bit exists, bitrate index information, sampling frequency information, information regarding Is whether a padding bit exists, a private bit, mode information, mode extension information, copyright information, information regarding whether an audio stream is an original one or a copy, and information on emphasis characteristics.
The CRC 120 is optional. The presence or absence of the CRC
20 120 is defined in the header 110, and the length of the CRC 120 is 16 bits.
The audio data 130 is a portion into which compressed audio data is inserted.
The auxiliary data 140 is data which is filled into a space 2s remaining when the end of the audio data 130 does not reach the end of an AAU. Arbitrary data other than MPEG audio can be inserted into the auxiliary data 140.
FIG. 2 is a block diagram of an apparatus for producing an MPEG-1 layer III audio stream. A pulse code modulation (PCM) audio 3o signal input unit 210 has a buffer in which PCM audio data is stored.
6 Here, the PCM audio signal input unit 210 receives, as the PCM audio data, granules, each composed of 576 samples.
A psycho-acoustic model 2 performing unit 220 receives the PCM
audio data from the buffer of the PCM audio signal input unit 210 and s performs psycho-acoustic model 2. A discrete cosine transforming (DCT) unit 230 receives the PCM audio data in units of granules and performs a DCT operation at the same time when psycho-acoustic model 2 is performed.
A modified DCT (MDCT) unit 240 performs an MDCT using the io result of the application of psycho-acoustic model 2 and the result of the DCT performed by the DCT unit 230. If perceptual energy is greater than a predetermined threshold, the MDCT is performed using a short window. If the perceptual energy is smaller than the predetermined threshold, the MDCT is performed using a long window.
is In perceptual coding, which is an audio signal compression technique, a reproduced signal is different to an original signal. That is, detailed information that people cannot perceive using the characteristics of the human ear can be omitted. Perceptual energy denotes energy that a human can perceive.
2o A quantization unit 250 performs quantization using bit allocation information generated as a result of the application of psycho-acoustic model 2 and using the result of the MDCT operation. An MPEG-1 layer III bitstream producing unit 260 transforms the quantized data into data to be inserted into an audio data area of an MPEG-1 bitstream, using 2s Huffman coding.
FIG. 3 is a flowchart illustrating a computation process in a psycho-acoustic model. First, PCM audio data is received in granules, each composed of 576 samples, in step 310. Next, long windows, each composed of 1024 samples, or short windows, each composed of 256
7 samples, are formed using the received PCM audio data, in step 320.
That is, one packet is constituted of multiple samples.
Thereafter, in step 330, a fast Fourier transform (FFT) is performed one window at a time on the windows formed in step 320.
s Then, psycho-acoustic model 2 is applied, in step 340.
In step 350, a perceptual energy value is obtained through the application of psycho-acoustic model 2 and applied to a MDCT unit and the MDCT unit selects a window to be applied. A signal to masking ratio (SMR) value.for each threshold bandwidth is calculated and applied to to a quantization unit to determine the number of bits to be allocated.
Finally, MDCT and quantization are performed using the perceptual energy value and the SMR value, in step 360.
FIG. 4 is a block diagram of an apparatus for producing a low-speed MPEG-1 layer III audio stream, according to the present is invention. A PCM audio signal storage unit 410 has a buffer in which it stores PCM audio data. A psycho-acoustic model 2 performing unit 420 performs an FFT on 1024 samples or 256 samples at a time and outputs perceptual energy information and bit allocation information.
As described above with reference to FIG. 3, when 2o psycho-acoustic model 2 is applied, the perceptual energy information and the bit allocation information that depends on an SMR are output.
Since the psycho-acoustic model 2 performing unit 420 performs an FFT, a harmonic extraction unit 430 extracts a harmonic component from the result of the FFT. This will be described later with reference to FIG. 6.
25 A harmonic encoding unit 440 encodes the extracted harmonic component and transmits the encoded harmonic component to an MPEG-1 layer III bitstream producing unit 480. The encoded harmonic component forms MPEG-1 audio, together with quantized audio data.
The encoding process of a harmonic component will be described later 3o in detail.
8 A harmonic decoding unit 450 decodes the encoded harmonic component to obtain PCM data in the time domain. A MDCT unit 460 subtracts the decoded harmonic component from the original input PCM
signal and performs a MDCT on the result of the subtraction. To be s more specific, if the perceptual energy information value received from the psycho-acoustic model 2 unit 420 is greater than a predetermined threshold, a MDCT is performed on 18 samples at a time. If the perceptual energy information value received from the psycho-acoustic model 2 performing unit 420 is equal to or smaller than the to predetermined threshold, a MDCT is performed on 36 samples at a time.
The harmonic component extraction is performed on data arranged in a frequency domain using a tonal/non-tonal decision condition and auditory limit characteristics that are defined in psycho-acoustic model 2. This will be described later in detail.
Is A quantization unit 470 performs quantization using the bit allocation information obtained by the psycho-acoustic model 2 performing unit 420. The MPEG-1 layer III bitstream producing unit 480 packetizes the harmonic component data made by the harmonic encoding unit 440 and quantized audio data obtained by the 2o quantization unit 470 to obtain compressed audio data.
FIG. 5 is a flowchart illustrating a harmonic extraction step 510, a harmonic encoding step 520, and a harmonic decoding step 530 based on psycho-acoustic model 2. The steps performed in psycho-acoustic model 2 in FIG. 5 are the same as the steps performed in 2s psycho-acoustic model 2 in FIG. 3. The result of the FFT performed based on the psycho-acoustic model 2 performing unit is used in step 510 of extracting a harmonic component. The extracted harmonic component is encoded to an MPEG-1 bitstream in step 520. The harmonic extraction step 510 will now be described in greater detail with 3o reference to FIGS. 6A through 6D.
9 FIGS. 6A, 6B, 6C, and 6D illustrate samples extracted in stages when harmonic components are extracted using the result of the FFT
performed in psycho-acoustic model 2. If PCM audio data as shown in FIG. 6A are input, an FFT is first performed on the received data in order s to determine sound pressure for each datum. One of the plurality of received PCM audio data whose sound pressure has been obtained is selected. If the values of the PCM audio data on the left and right sides of the selected data are smaller than the selected PCM audio data value, only the selected PCM audio data is extracted. This process is applied to to all of the received PCM audio data.
Sound pressure is the energy value of a sample in a frequency domain. In the present invention, only samples having sound pressures that are greater than a predetermined level are determined to be harmonic components. Accordingly, the samples shown in FIG. 6B are is extracted. Thereafter, only samples having sound pressures that are greater than a predetermined level are extracted. For example, if the predetermined level is set to be 7.0 dB, samples having sound pressures smaller than 7.0 dB are not selected, and only the samples shown in FIG.
6C remain. The remaining samples are not all considered to be 2o harmonic components, and some samples are extracted from the remaining samples according to the table of FIG. 7. Hence, finally, the samples shown in FIG. 6D remain.
FIG. 7 is a table showing a limited frequency range that varies according to a K value. Given that K is a value representing the Zs location of a sample in a frequency domain, if the K value is smaller than 3 or greater than 500, the values of samples present within the limited frequency range of 0 are 0 and accordingly not selected. Likewise, as shown in FIG. 7, if the K value is equal to or greater than 3 and smaller than 63, a corresponding range value is set to be 2. If the K value is 3o equal to or greater than 63 and smaller than 127, a corresponding range value is set to be 3. If the K value is equal to or greater than 127 and smaller than 255, a corresponding range value is set to be 6. If the K
value is equal to or greater than 255 and smaller than 500, a corresponding range value is set to be 12.
s Setting 500 as the limit was made in consideration of the limit of the audible frequency of a human and was based on an assumption that there is no difference in the quality of reproduced sound between when sample values corresponding to a frequency equal to or greater than 500 are considered and when they are not considered.
to Consequently, only the sample values of FIG. 6D are extracted and determined to be harmonic components.
Harmonic encoding 520 includes amplitude encoding, frequency encoding, and phase encoding. These three encoding methods use Equations 1 and 2:
Enc_peak_AmpMax = integer((28 -1 )loglo(AmpMax +10) ) ...(1 ) loglo 213 Enc Amp =integer((2 -1)loglo(Amp+10)) ...(2) loglo (AmpMax +10) Is wherein AmpMax denotes a peak amplitude, Enc peak-AmpMax denotes a result value obtained by encoding the value AmpMax, and Amp denotes amplitudes other than the peak amplitude.
In the amplitude encoding, when a peak amplitude is set as the value AmpMax, the peak amplitude is first encoded in a 8-bit log scale to Zo obtain Enc peak AmpMax as shown in Equation 1, and the other amplitudes Amp are encoded in a 5-bit log scale to obtain Enc-Amp as shown in Equation 2.
In the frequency encoding, only samples corresponding to values K ranging from 58 (2498Hz) to 372 (16KHz) are encoded in 2s consideration of the human's auditory characteristics. Since 314 is obtained by subtracting 58 from 372, the samples are encoded using 9 bits.
The phase encoding is achieved using 3 bits.
After such harmonic extraction and harmonic encoding, encoded s harmonic components are decoded and then undergo MDCT.
FIG. 8 is a flowchart illustrating a process for producing an audio stream by removing harmonic components, according to the present invention. First, in step 810, PCM audio data is received and stored.
Then, in step 820, psycho-acoustic model 2 using the audible limit to characteristics of a human being is applied to the stored data in order to obtain FFT result information, perceptual energy information regarding the received data, and bit allocation information used for quantization.
Thereafter, in step 830, harmonic components are extracted from the received PCM audio data using the FFT result information.
is The harmonic components are extracted in the following process.
First, sound pressure for each of the plurality of received PCM audio data is obtained using the FFT result information. Next, one of the plurality of received PCM audio data whose sound pressures are obtained is selected. If the values of the PCM audio data on the left 2o and right sides of the selected data are smaller than the value of the selected PCM audio data, only the selected PCM audio data is extracted.
This process is applied to all of the received PCM audio data.
Thereafter, only PCM audio data that each have sound pressure greater than a predetermined value of 7.0 dB are extracted from the PCM audio 2s data extracted in the previous step. Finally, harmonic components are extracted by not selecting PCM audio data in a predetermined frequency range among the audio data extracted in the previous step.
After the harmonic extraction in step 830, the extracted harmonic components are encoded and output in step 840. Then, encoded 3o harmonic components are decoded in step 850.

Next, in step 860, the received PCM audio data from which the decoded harmonic components are removed is subject to MDCT
according to the perceptual energy information. To be more specific, if a perceptual energy value is greater than a predetermined threshold, s MDCT is performed using a short window, for example, on 18 samples at a time. If the perceptual energy value is smaller than the predetermined threshold, MDCT is performed using a long window, for example, on 36 samples at a time.
Thereafter, in step 870, the MDCT result values are quantized by to allocating bits according to the bit allocation information.
Finally, in step 880, the quantized audio data and the encoded harmonic components are subject to Huffman coding to obtain an audio packet.
The embodiments of the present invention can be written as is computer programs and can be implemented in general-use digital computers that execute the programs using a computer readable recording medium. Examples of computer readable recording media include magnetic storage media (e.g., ROM, floppy disks, hard disks, etc.), optical recording media (e.g., CD-ROMs, or DVDs), and a storage 2o medium such as a carrier wave (e.g., transmission through the Internet).
While the present invention has been particularly shown and described with reference to preferred embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit 2s and scope of the present invention as defined by the following claims.
Hence, disclosed embodiments must be considered not restrictive but explanatory. The scope of the present invention is not presented in the above description but in the following claims, and all difference in the equivalent scope to the scope of the claims must be interpreted as being 3o included in the present invention.

Industrial Applicability As described above, in the present invention, the number of quantization bits generated upon production of a low-speed MPEG-1 s layer III audio stream is minimized. Using FFT results used in psycho-acoustic model 2, harmonic components are simply removed from an input audio signal, and only a transient portion is compressed using MDCT. Therefore, the input audio signal can be effectively compressed at a low-speed bitrate.

Claims (10)

What is claimed is:
1. An audio coding method using harmonic components, the method comprising:
(a) receiving pulse code modulation (PCM) audio data and extracting harmonic components from the received PCM audio data by applying psycho-acoustic model 2;
(b) performing a modified discrete cosine transform (MDCT) on the received PCM audio data from which the extracted harmonic components are removed; and (c) quantizing the MDCTed audio data and producing an audio packet from quantized audio data and the extracted harmonic components.
2. An audio coding method using harmonic components, the method comprising:
(a) receiving and storing PCM audio data and applying psycho-acoustic model 2 based on the audible limit characteristics of a human to the stored data to obtain fast Fourier transformation (FFT) result, perceptual energy information regarding received data, and bit allocation information used for quantization;
(b) extracting harmonic components from the received PCM audio data using the FFT result information;
(c) encoding the extracted harmonic components, outputting encoded harmonic components, and decoding the encoding harmonic components;
(d) performing a MDCT on a number of samples of the received PCM audio data from which the extracted harmonic components are removed, which depends on the value of the perceptual energy information;

(e) quantizing the MDCTed audio data by allocating bits according to the bit allocation information; and (f) producing an audio packet from the quantized, MDCTed audio data and the encoded harmonic components.
3. The audio coding method of claim 2, wherein step (b) comprises:
(b1) obtaining sound pressures for the plurality of received PCM
audio data using the FFT result information;
(b2) selecting a data value from the plurality of PCM audio data whose sound pressure is obtained and firstly extracting only the selected PCM audio datum if the value of PCM audio data on the right and left sides of the selected PCM audio data value are smaller than the selected PCM audio data value;
(b3) applying step (b2) to all of the received PCM audio data;
(b4) secondly extracting only the PCM audio data whose sound pressures are greater than a predetermined sound pressure, from the firstly-extracted PCM audio data; and (b5) not selecting PCM audio data that exist within a predetermined frequency range depending on a frequency location, among the PCM audio data secondly extracted in step (b4).
4. The audio coding method of claim 3, wherein the predetermined sound pressure in step b5 is 7.0 dB.
5. The audio coding method of claim 2, wherein in step (d), if the value of the perceptual energy information is greater than a predetermined threshold, MDCT is performed on 18 samples at a time, or if the value of the perceptual energy information is smaller than the predetermined threshold, MDCT is performed on 36 samples at a time.
6. An audio coding apparatus using harmonic components, the apparatus comprising:
a PCM audio data storage unit receiving and storing PCM audio data;
a psycho-acoustic model 2 performing unit receiving the PCM
audio data from the PCM audio data storage unit and performing psycho-acoustic model 2 to obtain FFT result information, perceptual energy information regarding received data, and bit allocation information used for quantization;
a harmonic extraction unit extracting harmonic components from the received PCM audio data using the FFT result information;
a harmonic encoding unit encoding the extracted harmonic components outputting encoded harmonic components;
a harmonic decoding unit decoding the encoded harmonic components;
an MDCT unit performing MDCT on the stored PCM audio data from which the decoded harmonic components are removed, according to the perceptual energy information;
a quantization unit quantizing the MDCTed audio data according to the bit allocation information; and an MPEG layer III bitstream production unit transforming the quantized, MDCTed audio data and the encoded harmonic components output from the harmonic encoding unit into an MPEG audio layer III
packet.
7. The audio coding apparatus of claim 6, wherein the harmonic extraction unit performs harmonic extraction through the following steps of:

obtaining sound pressures for the plurality of received PCM audio data using the FFT result information, selecting a datum from the plurality of PCM audio data whose sound pressures are obtained, and firstly extracting only the selected PCM audio datum if the value of PCM
audio data on the right and left sides of the selected PCM audio datum are smaller than the value of the selected PCM audio datum;
applying the firstly-extraction to all of the received PCM audio data and secondly extracting only the PCM audio data whose sound pressures are greater than a predetermined sound pressure, from the firstly-extracted PCM audio data; and deserting PCM audio data that exist within a predetermined frequency range depending on a frequency location, from the secondly-extracted PCM audio data.
8. The audio coding apparatus of claim 6, wherein the MDCT
unit performs MDCT on 18 samples at a time if the value of the perceptual energy information is greater than a predetermined threshold, or performs MDCT on 36 samples at a time if the value of the perceptual energy information is smaller than the predetermined threshold.
9. A computer readable recording medium which stores a computer program for executing claim 1.
10. A computer readable recording medium which stores a computer program for executing claim 2.
CA002490064A 2002-06-27 2002-12-12 Audio coding method and apparatus using harmonic extraction Abandoned CA2490064A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
KR2002/36310 2002-06-27
KR10-2002-0036310A KR100462611B1 (en) 2002-06-27 2002-06-27 Audio coding method with harmonic extraction and apparatus thereof.
PCT/KR2002/002348 WO2003063135A1 (en) 2002-06-27 2002-12-12 Audio coding method and apparatus using harmonic extraction

Publications (1)

Publication Number Publication Date
CA2490064A1 true CA2490064A1 (en) 2003-07-31

Family

ID=27607091

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002490064A Abandoned CA2490064A1 (en) 2002-06-27 2002-12-12 Audio coding method and apparatus using harmonic extraction

Country Status (9)

Country Link
US (1) US20040002854A1 (en)
JP (1) JP2005531014A (en)
KR (1) KR100462611B1 (en)
CN (1) CN1262990C (en)
CA (1) CA2490064A1 (en)
DE (1) DE10297751B4 (en)
GB (1) GB2408184B (en)
RU (1) RU2289858C2 (en)
WO (1) WO2003063135A1 (en)

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080120097A1 (en) * 2004-03-30 2008-05-22 Guy Fleishman Apparatus and Method for Digital Coding of Sound
WO2005096509A1 (en) 2004-03-31 2005-10-13 Intel Corporation Multi-threshold message passing decoding of low-density parity check codes
WO2007075098A1 (en) 2005-12-26 2007-07-05 Intel Corporation Generalized multi-threshold decoder for low-density parity check codes
WO2006046587A1 (en) 2004-10-28 2006-05-04 Matsushita Electric Industrial Co., Ltd. Scalable encoding apparatus, scalable decoding apparatus, and methods thereof
WO2006073324A1 (en) 2004-12-29 2006-07-13 Intel Corporation Channel estimation and fixed thresholds for multi-threshold decoding of low-density parity check codes
KR100707186B1 (en) * 2005-03-24 2007-04-13 삼성전자주식회사 Audio coding and decoding apparatus and method, and recoding medium thereof
JP4720302B2 (en) * 2005-06-07 2011-07-13 トヨタ自動車株式会社 Automatic transmission clutch device
KR100684029B1 (en) * 2005-09-13 2007-02-20 엘지전자 주식회사 Method for generating harmonics using fourier transform and apparatus thereof, method for generating harmonics by down-sampling and apparatus thereof and method for enhancing sound and apparatus thereof
KR100788706B1 (en) * 2006-11-28 2007-12-26 삼성전자주식회사 Method for encoding and decoding of broadband voice signal
US8706480B2 (en) 2007-06-11 2014-04-22 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Audio encoder for encoding an audio signal having an impulse-like portion and stationary portion, encoding methods, decoder, decoding method, and encoding audio signal
RU2464540C2 (en) * 2007-12-13 2012-10-20 Квэлкомм Инкорпорейтед Fast algorithms for computation of 5-point dct-ii, dct-iv, and dst-iv, and architectures
US8631060B2 (en) 2007-12-13 2014-01-14 Qualcomm Incorporated Fast algorithms for computation of 5-point DCT-II, DCT-IV, and DST-IV, and architectures
CN101552005A (en) * 2008-04-03 2009-10-07 华为技术有限公司 Encoding method, decoding method, system and device
PL3246918T3 (en) 2008-07-11 2023-11-06 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Audio decoder, method for decoding an audio signal and computer program
EP4224471A3 (en) 2008-07-11 2023-09-06 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Audio encoder and audio decoder
CN101751928B (en) * 2008-12-08 2012-06-13 扬智科技股份有限公司 Method for simplifying acoustic model analysis through applying audio frame frequency spectrum flatness and device thereof
EP2491553B1 (en) * 2009-10-20 2016-10-12 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Audio encoder, audio decoder, method for encoding an audio information, method for decoding an audio information and computer program using an iterative interval size reduction
BR112013020588B1 (en) * 2011-02-14 2021-07-13 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. APPARATUS AND METHOD FOR ENCODING A PART OF AN AUDIO SIGNAL USING A TRANSIENT DETECTION AND A QUALITY RESULT
TWI562133B (en) * 2011-05-13 2016-12-11 Samsung Electronics Co Ltd Bit allocating method and non-transitory computer-readable recording medium
RU2464649C1 (en) 2011-06-01 2012-10-20 Корпорация "САМСУНГ ЭЛЕКТРОНИКС Ко., Лтд." Audio signal processing method
CN103516440B (en) 2012-06-29 2015-07-08 华为技术有限公司 Audio signal processing method and encoding device

Family Cites Families (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5023910A (en) * 1988-04-08 1991-06-11 At&T Bell Laboratories Vector quantization in a harmonic speech coding arrangement
JPH0364800A (en) * 1989-08-03 1991-03-20 Ricoh Co Ltd Voice encoding and decoding system
JP3266920B2 (en) * 1991-09-25 2002-03-18 三菱電機株式会社 Audio encoding device, audio decoding device, and audio encoding / decoding device
US5285498A (en) * 1992-03-02 1994-02-08 At&T Bell Laboratories Method and apparatus for coding audio signals based on perceptual model
KR100458969B1 (en) * 1993-05-31 2005-04-06 소니 가부시끼 가이샤 Signal encoding or decoding apparatus, and signal encoding or decoding method
JP3721582B2 (en) * 1993-06-30 2005-11-30 ソニー株式会社 Signal encoding apparatus and method, and signal decoding apparatus and method
JPH0736486A (en) * 1993-07-22 1995-02-07 Matsushita Electric Ind Co Ltd Speech encoding device
JP3131542B2 (en) * 1993-11-25 2001-02-05 シャープ株式会社 Encoding / decoding device
TW321810B (en) * 1995-10-26 1997-12-01 Sony Co Ltd
JP2778567B2 (en) * 1995-12-23 1998-07-23 日本電気株式会社 Signal encoding apparatus and method
JPH09246983A (en) * 1996-03-08 1997-09-19 Nec Eng Ltd Digital signal processor
JPH10178349A (en) * 1996-12-19 1998-06-30 Matsushita Electric Ind Co Ltd Coding and decoding method for audio signal
KR19980072457A (en) * 1997-03-05 1998-11-05 이준우 Signal processing method and apparatus therefor in psychoacoustic sound when compressing audio signal
US5930373A (en) * 1997-04-04 1999-07-27 K.S. Waves Ltd. Method and system for enhancing quality of sound signal
DE19742201C1 (en) * 1997-09-24 1999-02-04 Fraunhofer Ges Forschung Method of encoding time discrete audio signals, esp. for studio use
US6704705B1 (en) * 1998-09-04 2004-03-09 Nortel Networks Limited Perceptual audio coding
KR100300887B1 (en) * 1999-02-24 2001-09-26 유수근 A method for backward decoding an audio data
JP2000267700A (en) * 1999-03-17 2000-09-29 Yrp Kokino Idotai Tsushin Kenkyusho:Kk Method and device for encoding and decoding voice
JP2000276194A (en) * 1999-03-25 2000-10-06 Yamaha Corp Waveform compressing method and waveform generating method
US6377916B1 (en) * 1999-11-29 2002-04-23 Digital Voice Systems, Inc. Multiband harmonic transform coder
DE10000934C1 (en) * 2000-01-12 2001-09-27 Fraunhofer Ges Forschung Device and method for determining an encoding block pattern of a decoded signal
JP4055336B2 (en) * 2000-07-05 2008-03-05 日本電気株式会社 Speech coding apparatus and speech coding method used therefor
KR100348899B1 (en) * 2000-09-19 2002-08-14 한국전자통신연구원 The Harmonic-Noise Speech Coding Algorhthm Using Cepstrum Analysis Method
KR100378796B1 (en) * 2001-04-03 2003-04-03 엘지전자 주식회사 Digital audio encoder and decoding method
US6732071B2 (en) * 2001-09-27 2004-05-04 Intel Corporation Method, apparatus, and system for efficient rate control in audio encoding
KR100472442B1 (en) * 2002-02-16 2005-03-08 삼성전자주식회사 Method for compressing audio signal using wavelet packet transform and apparatus thereof

Also Published As

Publication number Publication date
CN1262990C (en) 2006-07-05
RU2004138088A (en) 2005-06-27
CN1639769A (en) 2005-07-13
DE10297751B4 (en) 2005-12-22
DE10297751T5 (en) 2005-07-07
KR20040001184A (en) 2004-01-07
RU2289858C2 (en) 2006-12-20
GB0427660D0 (en) 2005-01-19
JP2005531014A (en) 2005-10-13
KR100462611B1 (en) 2004-12-20
WO2003063135A1 (en) 2003-07-31
GB2408184A (en) 2005-05-18
US20040002854A1 (en) 2004-01-01
GB2408184B (en) 2006-01-04

Similar Documents

Publication Publication Date Title
EP1715476B1 (en) Low-bitrate encoding/decoding method and system
CA2490064A1 (en) Audio coding method and apparatus using harmonic extraction
KR100571824B1 (en) Method for encoding/decoding of embedding the ancillary data in MPEG-4 BSAC audio bitstream and apparatus using thereof
US7835907B2 (en) Method and apparatus for low bit rate encoding and decoding
JP2006048043A (en) Method and apparatus to restore high frequency component of audio data
EP1960999A1 (en) Method, medium, and apparatus encoding and/or decoding an audio signal
JP5587599B2 (en) Quantization method, encoding method, quantization device, encoding device, inverse quantization method, decoding method, inverse quantization device, decoding device, processing device
JP2003523535A (en) Method and apparatus for converting an audio signal between a plurality of data compression formats
KR100750115B1 (en) Method and apparatus for encoding/decoding audio signal
JP3353868B2 (en) Audio signal conversion encoding method and decoding method
KR100378796B1 (en) Digital audio encoder and decoding method
US20080133250A1 (en) Method and Related Device for Improving the Processing of MP3 Decoding and Encoding
KR100754389B1 (en) Apparatus and method for encoding a speech signal and an audio signal
JP3348759B2 (en) Transform coding method and transform decoding method
KR100928966B1 (en) Low bitrate encoding/decoding method and apparatus
Cavagnolo et al. Introduction to Digital Audio Compression
KR100940532B1 (en) Low bitrate decoding method and apparatus
JP2001094432A (en) Sub-band coding and decoding method
Reyes et al. A new perceptual entropy-based method to achieve a signal adapted wavelet tree in a low bit rate perceptual audio coder
Reyes et al. On the coding gain of dynamic Huffman coding applied to a wavelet-based perceptual audio coder
Mandal et al. Digital Audio Compression

Legal Events

Date Code Title Description
EEER Examination request
FZDE Discontinued