US20040005065A1 - Sound event detection system - Google Patents

Sound event detection system Download PDF

Info

Publication number
US20040005065A1
US20040005065A1 US10/428,451 US42845103A US2004005065A1 US 20040005065 A1 US20040005065 A1 US 20040005065A1 US 42845103 A US42845103 A US 42845103A US 2004005065 A1 US2004005065 A1 US 2004005065A1
Authority
US
United States
Prior art keywords
sound
sound event
input channel
signal
events
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/428,451
Inventor
David Griesinger
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harman International Industries Inc
Original Assignee
Harman International Industries Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harman International Industries Inc filed Critical Harman International Industries Inc
Priority to US10/428,451 priority Critical patent/US20040005065A1/en
Assigned to HARMAN INTERNATIONAL INDUSTRIES, INCORPORATED reassignment HARMAN INTERNATIONAL INDUSTRIES, INCORPORATED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GRIESINGER, DAVID H.
Publication of US20040005065A1 publication Critical patent/US20040005065A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/78Detection of presence or absence of voice signals
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S3/00Direction-finders for determining the direction from which infrasonic, sonic, ultrasonic, or electromagnetic waves, or particle emission, not having a directional significance, are being received
    • G01S3/80Direction-finders for determining the direction from which infrasonic, sonic, ultrasonic, or electromagnetic waves, or particle emission, not having a directional significance, are being received using ultrasonic, sonic or infrasonic waves
    • G01S3/802Systems for determining direction or deviation from predetermined direction
    • G01S3/808Systems for determining direction or deviation from predetermined direction using transducers spaced apart and measuring phase or time difference between signals therefrom, i.e. path-difference systems
    • G01S3/8083Systems for determining direction or deviation from predetermined direction using transducers spaced apart and measuring phase or time difference between signals therefrom, i.e. path-difference systems determining direction of source
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01VGEOPHYSICS; GRAVITATIONAL MEASUREMENTS; DETECTING MASSES OR OBJECTS; TAGS
    • G01V1/00Seismology; Seismic or acoustic prospecting or detecting
    • G01V1/001Acoustic presence detection
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0272Voice signal separating

Definitions

  • the invention relates generally to detecting and localizing sound. More particularly, this invention relates to detecting and/or localizing sounds that include sound events in a complex sound field
  • sound generally consists of a complicated combination of many different sounds occurring over time to form a complex sound field.
  • the sounds in a sound field can overlap, occur one after the other, or in any combination.
  • One way in which the individual sounds in a sound field are classified is according to whether a sound has emanated or originated from a particular location. Sounds that can be detected as emanating or originating from a particular direction are referred to as directional sounds, while sounds detected as emanating or originating from no particular direction at all are referred to as non-directional sounds.
  • Another way of classifying sounds is according to whether a sound is a transient or a steady-state sound.
  • Steady-state sounds are those that have a generally constant level of power over time, such as a sustained musical note.
  • Steady-state sounds can be either directional or non-directional sounds.
  • Transient sounds are sounds that have an initial energy spike, such as a shout or a drum hit. Transients can also be either directional or non-directional sounds.
  • An example of a non-directional transient sound is speech in a reverberant space where the direct sound is blocked by an object. In this case if the reverberation time is less than one second the time characteristics of the signal are preserved, but information about its direction is lost.
  • syllables Two types of sound events are syllables and impulsive sounds.
  • Syllables include phonemes and notes.
  • Phonemes are transient sounds that are characteristic of phones in human speech and can be particularly useful in detecting and localizing syllables in human speech. Notes are the individual notes created by a musical instrument.
  • Syllables generally have the following characteristics: a finite duration of about at least 50 ms up to about 200 ms, but typically around 150 ms; rise times of about 33 ms; generally occur no more frequently than about once every 0.2 ms to about 0.5 ms; and may have low or high volume (amplitude).
  • impulsive sounds are transients of very short duration such as a drum hit or fricatives and explosives in speech.
  • Impulsive sounds generally have the following characteristics: a short duration of about 5 ms to about 50 ms, rise times of about 1 ms to about 10 ms, and high volume.
  • the sound field need only be generated in one input or “input channel.”
  • the sound field needs to be generated in at least two inputs or input channels.
  • the archetype for sound localization is natural hearing, where the azimuth of the sound is detected primarily by the arrival time difference between the two input channels represented by the two ears.
  • the azimuth of a sound source is determined primarily by the amplitude and phase relationships between the signals generated by two or more input channels.
  • the directions of the source of these sounds are described in terms of an angle between each corresponding pair of channels (each a “channel pair”). If sounds are generated in only two channels, the directions of the sounds are given in terms of an angle for that channel pair, generally a left/right angle “lr.” If the sounds are generated in four channels, the directions of the sounds are given in terms of an angle for each channel pair, generally, a left/right angle lr, and a front/back angle “cs.” For example, when the sound field is generated in two channels, the average direction of the sounds in the sound field is given in terms of lr only.
  • the value for lr ranges from about ⁇ 45 degrees to about 45 degrees, with ⁇ 45 degrees indicating that the sound field originates from the left input channel, 45 degrees indicating that the sound field o riginates from the right input channel and 0 degrees indicating that the sound field originates from a position in the middle precisely between the right and left input channels (a position often referred to as “center”).
  • a second directional component is specified. Additionally, even if the sound field is generated in only one channel pair, a second directional component may also be specified because it is often possible to derive an additional channel pair from the one channel pair.
  • One known technique for determining these angles is used in reproducing recorded sound.
  • this known technique determines the intended direction of sounds by comparing the amplitudes of the signals in one input channel of a input channel pair with the signals in the corresponding input channel of the input channel pair (generally, the left with the right, and the center with the surround). More specifically, this ratio of amplitudes is used to determine what is generally referred to an “ordinary steering angle” or “OSA” for each input channel pair.
  • OSA ordinary steering angle
  • the voltage signals in each input channel of an input channel pair are rectified and the logarithms of the rectified voltages are taken.
  • the filter is made slow enough not to distinguish the fluctuations characteristic of the non-directional signals, the filter is generally too slow to distinguish the fluctuations of certain sound events, particularly impulsive sounds. As a result, many sound events are not properly localized. No matter how these filters are designed, they generally work well on only one type of music but not on all. For example, the fast filter will work well on complex popular music, which is full of rapid changes, but will reflect false directional changes (steer too greatly) when a highly-reverberant classical piece is reproduced.
  • sounds recorded in a given number of input channels and then reproduced over a different number of channels are stereo and surround.
  • Sounds recorded for reproduction in stereo are intended to be perceived as originating only from the front.
  • Sounds recorded for reproduction in surround are intended to be perceived as originating from all around, generally with one or two input channels used to reproduce sounds from the rear.
  • the techniques used to record sounds intended for reproduction in stereo are generally different from those used to record sounds intended for reproduction in surround.
  • sounds recorded for reproduction in surround generally need to be capable of high-quality reproduction in stereo.
  • the Dolby Surround® system which records sounds for reproduction in surround, adds a negative phase to the sounds intended for reproduction from behind the listener (the rear). This negative phase is generally undetected by stereo reproduction systems and is transparent to the listener.
  • the negative phase is detected by a surround reproduction system that then reproduces the associated sounds in the rear input channels.
  • Properties of the human hearing mechanism have been modeled and used to create systems with these features because many of the problems encountered by known sound detection and systems are not experienced by the human hearing mechanism.
  • the human hearing mechanism can distinguish transient from non-transient sounds and sound events from other transient sounds even in the presence of a background signal by relying on the properties of beginning detection and accommodation.
  • Beginning detection results from human neurology which is highly evolved to determine the starting points and end points of sounds and makes the human hearing mechanism particularly sensitive to the rise and fall times of sounds.
  • Accommodation is the property that allows the human hearing mechanism to distinguish transients from steady-state sounds by gradually ignoring the presence of the steady state sounds.
  • the human hearing mechanism relies more heavily on certain frequencies and the apparent direction of the beginnings of sounds. The human hearing mechanism relies most heavily on frequencies between about 500 Hz and about 4000 Hz for direction determination.
  • This frequency bias is largely a result of the frequency response of the external ear (the pinnae, concha and ear canal) and by the frequency transfer function of the middle ear.
  • the human hearing mechanism also relies on the direction indicated in the rise time of a sound more than the direction as indicated in other portions of the sound. This reliance is advantageous because the direction indicated in the rise time of a sound is less likely to be corrupted by sound reflections or reverberations even in a highly reflective environment.
  • the properties of the human hearing mechanism have also been used to create sound event detection and localization methods for determining the location of sound events that occur in the presence of a background signal. These sound event detection and localization methods produce at least one steering angle indicating the direction of the sound event whenever a trigger signal indicates that a sound event has occurred (“differential steering angles”). Sound event detection and localization methods use differential signals to determine the differential steering angles by comparing the relative power of the sound event in each input channel of each input channel pair after any steady-state sounds have been removed. By comparing the power in each input channel of an input channel pair instead of the voltage in each input channel of an input channel pair as done in known methods, individual sound events can be identified and separately localized.
  • the properties of the human hearing mechanism may also be used to determine the location of all sounds within a sound field.
  • sound localization methods have been developed that localize the individual sounds in the presence of background sounds for any combination of sounds more accurately.
  • the sound localization methods divide the sound field into sound events and non-sound events and localize the sound events in terms of a true differential steering angle or a true ordinary steering angle and the non-sound events in terms of a filtered ordinary steering angle. These methods essentially treat the non-sound events as a separate sound for which the direction is separately determined.
  • the sound localization methods, as well as the sound event detection and localization methods and the sound event detection methods can be implemented for any number and combination of sound event types in a sound field generated in any number of input channel pairs.
  • the properties of the human hearing mechanism have been modeled to create electronic circuitry that detects sound events in the presence of a background signal.
  • These sound event detectors may be used for a variety of applications for which the detection of specific types of sound events is helpful.
  • a sound event detector that detects syllables may be used as part of a phonic detector or as part of a speech recognition or speech coding system.
  • a sound event detector that detects syllables may be used in conjunction with an audio amplification device, such as a microphone. This allows the microphone to remain off until a syllable is detected from a speaker, thus preventing the microphone from amplifying undesired sounds and feedback through the microphone itself when the speaker is silent.
  • Sound event detectors like the sound event detection methods, are based on models of the beginning detection property and the accommodation property and generally produce a trigger signal that includes some flag or marker, such as a pulse, that indicates the occurrence of a sound event.
  • Sound event detectors may be used for a variety of applications for which the detection of specific types of sound events is helpful.
  • a sound event detector that detects syllables may be used as part of a phonic detector or as part of a speech recognition or speech coding system.
  • a sound event detector that detects syllables may be used in conjunction with an audio amplification device, such as a microphone.
  • FIG. 1 is a flow chart of a sound event detection method for a single input channel and single sound event type.
  • FIG. 2 is a flow chart of an accommodation method.
  • FIG. 3 is a series of time-domain plots of a sample segment of a sound field before, during and after the onsets of transients have been distinguished and the transients have been separated from background sounds.
  • FIG. 4 is a flow chart of a beginning detection method for a single input channel.
  • FIG. 5 is a flow chart of a sound event detection method for multiple input channels.
  • FIG. 6 is a flow chart of a beginning detection method for multiple input channels.
  • FIG. 7 is a flow chart of a sound event detection method for a single input channel and multiple sound event types.
  • FIG. 8 is a flow chart of a sound event detection method for multiple input channels and multiple sound event types.
  • FIG. 9 is a flow chart of a sound event detection and localization method for a single input channel pair and a single sound event type.
  • FIG. 10 is a flow chart of a sound event detection and localization method for multiple input channel pairs and a single sound event type.
  • FIG. 11 is a flow chart of a sound event detection and localization method for multiple input channel pairs and a single sound event type.
  • FIG. 12 is a flow chart of a sound event detection and localization method for a single input channel pair and multiple sound event types.
  • FIG. 13 is a flow chart of a sound localization method.
  • FIG. 14 is a flow chart of a direction selection method for a single input channel pair and a single sound event type.
  • FIG. 15 is a flow chart of a direction selection method for a single channel pair and a single short-duration sound event.
  • FIG. 16 is a flow chart of a direction selection method for multiple input channel pairs and a single sound event type.
  • FIG. 17 is a flow chart of a differential steering angle determination method.
  • FIG. 19 is block diagram of a sound event device.
  • FIG. 20 is a block diagram of a sound event detector for a single input channel and a single sound event type.
  • FIG. 21 is a circuit diagram of an accommodation circuit for a single input channel.
  • FIG. 23 is a circuit diagram of a sound event detector for a single input channel and a short-duration sound event.
  • FIG. 24 is a block diagram of a sound event detector for multiple input channels and a single sound event type.
  • FIG. 25 is a circuit diagram of an alternate trigger generation circuit for multiple input channel pairs.
  • FIG. 26 is a block diagram of a sound event detector for a single input channel and multiple sound event types.
  • FIG. 27 is a block diagram of a sound event detector for multiple input channels and multiple sound event types.
  • FIG. 28 is a block diagram of a sound event detector and localizer for a single input channel pair and a single sound event type.
  • FIG. 29 is a circuit diagram of a sound event localization circuit for a single input channel pair and a single sound event type.
  • FIG. 30 is a block diagram of a sound event detector and localizer for multiple input channel pairs and a single sound event type.
  • FIG. 31 is a circuit diagram of a sound event localization circuit for multiple input channel pairs and a single sound event type.
  • FIG. 32 is a block diagram of a sound event detector and localizer for a single input channel pair and multiple sound event types.
  • FIG. 33 is a block diagram of a sound event detector and localizer for multiple input channel pairs and multiple sound event types.
  • FIG. 34 is a block diagram of a stereo/surround detector for a single sound event type.
  • FIG. 35 is a block diagram of a stereo/surround detector for multiple sound event types.
  • FIG. 36 is a block diagram of a sound localizer for a single input channel pair and a single sound event type.
  • FIG. 37 is a circuit diagram of a sound localization circuit for a single input channel pair and a single sound event type.
  • FIG. 38 is a circuit diagram of a sound localization circuit for a single input channel pair and a short-duration sound event.
  • FIG. 39 is a block diagram of a sound localizer for multiple input channel pairs and a single sound event type.
  • FIG. 40 is a circuit diagram of a sound localization circuit for multiple input channel pairs and a single sound event type.
  • FIG. 41 is a block diagram of a sound localization circuit for multiple input channel pairs and a short-duration sound event.
  • FIG. 42 is a block diagram of a sound localizer for a single input channel pair and multiple sound event types.
  • FIG. 43 is a block diagram of a sound localizer for multiple input channel pairs and multiple sound event types.
  • Sound detection and localization systems have been developed that detect and separately determine the direction of sound events in a complex sound field based on the properties of the human hearing mechanism.
  • the human hearing mechanism which includes the ears, auditory system and related portions of the brain, was chosen as a model for two basic reasons. The first is that the human hearing mechanism detects sounds and determines their direction very effectively and without the shortcomings of the known systems. The second reason is that many of the mistakes made by systems modeled after the human hearing mechanism will generally be the same mistakes made by the human hearing mechanism, thus making these mistakes inaudible.
  • the human hearing mechanism was studied and used as a model from which to create sound detection and localization systems.
  • extremely effective, high-quality methods have been created for the detection of sound events in the presence of steady-state sounds, for the detection and localization of sound events in the presence of steady-state sounds, and for the detection and localization of all sounds in a sound field.
  • These methods can also be implemented in software and hardware to create programs, devices and even more complicated systems applicable to a wide range of applications.
  • the human hearing mechanism In breaking down a sound field into individual sounds, the human hearing mechanism has the ability to distinguish transient sounds from other sounds and to distinguish sound events from non-directional transients. Furthermore, the human hearing mechanism can distinguish transient from non-transient sounds and sound events from other transient sounds even in the presence of a background signal by relying on the properties of beginning detection and accommodation. Beginning detection results from human neurology which is highly evolved to determine the starting points and end points of sounds and makes the human hearing mechanism particularly sensitive to the rise and fall times of sounds. Accommodation is the property that allows the human hearing mechanism to distinguish transients from steady-state sounds by gradually ignoring the presence of the steady state sounds. The properties of beginning detection and accommodation and their effect on localization were discovered and demonstrated experimentally.
  • the azimuth (as an indication of direction) of a tone can only be detected when the tone starts or stops. If the tone is continuous, it becomes very difficult to determine its direction. Furthermore, if new sounds are added to a steady state tone, the true direction of the new sounds is easy to determine. This demonstrates that some sounds have both a transient and a steady-state component. It also demonstrates that the human hearing mechanism is particularly sensitive to the transient components, which include the beginnings and ends of sounds and uses the beginnings and ends of sounds for localization (the beginning detection property). It also demonstrates that the human hearing mechanism ignores steady-state sounds over time and generally does not rely on them for localization (the accommodation property).
  • the rate at which the human hearing mechanism gradually ignores the steady-state sounds is independent of the sound event type being detected. It was also found that the accommodation rate may not be constant and may adjust to acoustic conditions and to the rapidity of speech. However, a reasonable average value for the accommodation rate was found to be about 300 ms.
  • the human hearing mechanism relies more heavily on certain frequencies and the apparent direction of the beginnings of sounds.
  • the human hearing mechanism relies most heavily on frequencies between about 500 Hz and about 4000 Hz for direction determination. This frequency bias is largely a result of the frequency response of the external ear (the pinnae, concha and ear canal) and by the frequency transfer function of the middle ear.
  • the human hearing mechanism also relies on the direction indicated in the rise time of a sound more than the direction as indicated in other portions of the sound. This reliance is advantageous because the direction indicated in the rise time of a sound is less likely to be corrupted by sound reflections or reverberations even in a highly reflective environment.
  • the human hearing mechanism After detecting the directions of the individual sounds, the human hearing mechanism reassembles related sounds into streams and separately determines the direction of each stream.
  • the direction of each stream is generally the average direction of all sounds within the stream.
  • sounds are reassembled into a foreground stream and a background stream.
  • the foreground stream may consist of dialog and the background stream may consist of environmental sounds being produced around the source of the dialog.
  • the process of reassembling related sounds into streams and determining the direction of each stream is performed unconsciously and automatically by the human hearing mechanism.
  • Sound event detection methods produce a signal (referred to in this application as a “trigger signal”), which includes a flag or marker, such as a pulse, that indicates the occurrence of a sound event.
  • the sound event detection methods can be implemented to detect any number and combination of sound events in any number of input channels. In the following description, the sound event detection methods are discussed in order of increasing complexity, with each subsequent sound event detection method incorporating the steps of the prior methods, except as indicated.
  • FIG. 1 An example of a method for detecting sound events in the presence of a background signal that only detects a single sound event type in a sound field that is generated in a single input channel (“sound event detection methods for a single input channel and a single sound event type”), is shown in FIG. 1 and indicated by reference number 100 .
  • the sound event detection method for a single input channel and a single sound event type 100 includes: emphasizing the directionally important frequencies by modeling frequency bias 104 ; separating sound events from background sounds by modeling accommodation 106 ; and detecting sound events by modeling beginning detection 108 .
  • the directionally important frequencies are emphasized by modeling the frequency bias of the human hearing mechanism 104 using frequency emphasis method.
  • the frequency emphasis method includes emphasizing the frequencies between about 500 Hz and about 4 kHz in each input channel of the input channel pair to produce a filtered signal in each input channel. These frequencies are emphasized because they have the most influence on the human hearing mechanism in terms of determining direction. Separating the sound events from the background sounds by modeling accommodation 106 generally includes examining the change in input power in the sound field and is shown in more detail in FIG. 2.
  • separating the sound events from the background sounds by modeling accommodation 106 includes: determining the power envelope of the input channel 201 ; determining the power in any steady-state sounds 202 , and subtracting the power in any steady-state sounds from the power envelope of the input channel.
  • Determining the power envelope of the input channel 201 generally includes squaring the voltage in the input channel.
  • a sound field, whether produced live or from a recording, is generally represented as a voltage signal in the time domain.
  • voltage signals do not have the additive property, which means that the magnitude of a voltage signal resulting from the combination of two voltage signals cannot be determined by simply adding the amplitudes of the two voltage signals.
  • power signals do have the additive property.
  • a power envelope results from which other power signals may be subtracted.
  • fluctuations in the power envelope that are not characteristic of the sound event type being detected may be removed. For example, if syllables are being detected, fluctuations with rise times faster than about 33 ms will be removed from the power envelope.
  • Determining the power in any steady-state signals 202 in the sound field includes determining the long-term average power of the sound field.
  • the long-term average power (referred to in this application as the “accommodation signal”) may be determined by integrating the power envelope over a time period equal to the rise time of the sound event type that is being detected. Once a sound event has reached its maximum value (at the end of the rise time of the sound event), the accommodation signal is gradually subtracted from the power envelope 204 at a rate equal to the accommodation rate of the human hearing mechanism (which is about 300 ms) to model the way in which the human hearing mechanism gradually ignores the effects of steady-state sounds in the presence of sound events.
  • the accommodation signal In order to control the accommodation, at all times the accommodation signal is forced to be less than or equal to the power envelope. Therefore, when the sound event ends, causing the power envelope to drop rapidly, the accommodation signal will be forced to equal the power envelope.
  • the result of subtracting the accommodation signal from the absolute signal power is a “differential signal.”
  • This difference signal includes positive pulses and other fluctuations. Each pulse in the difference signal indicates the occurrence of a sound event and the other fluctuations are caused by noise such as: reverberation, Gaussian noise, and other signals that may not currently be in the foreground stream.
  • FIG. 3 An example of how the accommodation method affects an input signal is shown in FIG. 3.
  • the time-domain plot A in FIG. 3 shows a sample segment of a sound field. This segment includes a collection of sound waves of varying frequencies that is characteristic of a syllable 300 . The magnitude of this syllable is the voltage in the signal “V.” The voltage V varies with time “t” and has a duration of “D.”
  • the time-domain plot B shows a pulse 302 , which is segment 300 after it has been converted into a power signal (step 201 in FIG. 2). Pulse 302 has an amplitude that represents the power in the signal (“V 2 ”) and also has duration D.
  • the time-domain plot C shows pulse 304 , which is pulse 302 after the fluctuations with rise times that are not characteristic of a syllable have been removed.
  • the pulse 304 also has an amplitude V 2 and has a rise time “t r ” where t r is equal to about 33 ms.
  • the time-domain plot D shows a pulse 306 , which is pulse 304 after the steady-state sounds have been accommodated to by subtracting the accommodation signal.
  • the rise-time of pulse 306 follows that of pulse 304 .
  • Modeling beginning detection includes distinguishing the sound events from the noise present in the differential signal.
  • the differential signal includes a series of rapidly-rising pulses and noise. While each pulse indicates the occurrence of a sound event, the noise includes fluctuations that may falsely indicate the occurrence of a sound event. Therefore, in order to detect the sound events, the sound events need to be distinguished from the noise.
  • Modeling beginning detection 108 is shown in more detail in FIG. 4 and includes deemphasizing the effects of volume; emphasizing the sound events 404 ; deemphasizing the noise 406 ; and detecting the sound events 408 .
  • steps 404 , 406 and 408 can be simultaneously accomplished is by scaling the differential signal by the short-term average power contained in the noise component of the differential signal (the “short-term average high frequency power”).
  • the short-term average high frequency power is isolated and used to divide the differential signal.
  • the short-term average high frequency power can be isolated by filtering the differential signal to obtain the component of the differential signal with frequencies higher than those characteristic of the sound event type being detected, and integrating the high-frequency portion over a short time period. This short time period may be different for different types of music or speech rates.
  • a value of about 160 ms was experimentally determined to work well for a wide variety of inputs.
  • Normalizing the differential signal by using an automatic gain methods has two effects, one obvious and two not-so-obvious.
  • the first effect is that it deemphasizes the effect volume has on the differential signal 402 by normalizing the differential signal with regard to the amplitude (which represents power) of the fluctuations in the differential signal. Because louder sound events and noise generally contain more power, the louder sound events and noise will be scaled more strongly that those that are not as loud. This is the normal function of an automatic gain control method.
  • a second effect is to de-emphasize noise 406 by de-emphasizing the fluctuations that occur more often than is characteristic of the sound event type being detected (for example, syllables generally occur or repeat once about every 200 ms).
  • the fluctuations in portions of the differential signal due to noise generally occur more frequently than is typical for a sound event.
  • the rapid repeat rate of these noise fluctuations increases the average high-frequency power with which that portion of the differential signal is divided. This de-emphasizes the portion of the differential signal that contains the noise.
  • the third effect of normalizing the differential signal is to emphasize the sound events 404 by emphasizing the fluctuations that occur or repeat no more often than is characteristic of the sound event type being detected.
  • the portions of the differential signal that contain fluctuations that occur less often than is typical for the sound event type being detected will have a lower average high frequency power (as compared to those containing noise) with which that portion of the differential signal is divided. This will provide a relative emphasis to the sound events.
  • De-emphasizing noise 406 may be further accomplished by removing many of the fluctuations in the differential signal due to noise. Some of the fluctuations due to noise can be identified and removed according to their rise time and other characteristics. For example, fluctuations that have rise-times not characteristic of the sound event type being detected are removed. For example, if the sound event type being detected is syllables, sounds with rise times faster than about 33 ms will be removed. In another example, if the sound event type being detected is impulsive sounds, sounds with rise times faster than about 3.3 ms will be removed. Additionally, low amplitude transients that occur more frequently than is characteristic of sound events such as, some noise and reverberation, can also be filtered out.
  • any transient simultaneously detected is noise or some other non-directional transient, transients occurring under such conditions can also be removed.
  • steps 402 , 404 and 406 many of the fluctuations in the differential signal are removed to create an improved differential signal that includes pulses each indicating the occurrence of a sound event. However, some noise remains.
  • the sound events need to be detected 408 from the noise. Because at this point, most of the fluctuations in the improved differential signal due to noise have a low amplitude as compared with the fluctuations caused by the sound events, the sound events are detected 408 by determining which fluctuations have an amplitude that exceeds a threshold using a threshold detection method. In the threshold detection method, the fluctuations that do not exceed the threshold are removed or ignored to produce the trigger signal.
  • a threshold detection method In the threshold detection method, the fluctuations that do not exceed the threshold are removed or ignored to produce the trigger signal.
  • One advantage of this method is that the triggering need not be ideal.
  • the sound event detection methods are not perfect in that in some sound events will not be detected, and some fluctuations due to noise will result in a pulse in the trigger signal falsely indicating the occurrence of a sound event. However, these occasional errors do not matter. Because the sound event detection methods are modeled after the human hearing mechanism, it produces the same types of errors as does the human hearing mechanism. Therefore, the result will be perceived as completely natural.
  • the threshold is chosen so that low amplitude fluctuations that are characteristic of noise or reverberation are not detected.
  • the threshold may be a fixed value which is determined experimentally. However, more accurate results are obtained if the threshold varies as a function of the sound field. For example, when the sound field includes many sound events, the threshold will generally be lower than when the sound field includes fewer sound events. This allows greater sensitivity when a greater number of sound events are present in the sound field.
  • the threshold can be manually selected from among two or more values experimentally determined according to the characteristics of the sound field. For example, if the sound field is that of modern or popular music, which typically includes many sound events, a lower threshold value can be selected, and alternatively, if the sound field is that of classical music, which typically includes few sound events and may be highly reverberant, a higher threshold value can be selected.
  • the threshold value can be chosen as a function of the number of sound events detected during a given time period. Therefore, during time periods where fewer sound events are detected, the threshold value is raised and during time periods where a greater number of sound events are detected, the threshold value is lowered. This provides an automatic and continuous adjustment of the threshold value as a function of the sound field and results in a more accurate trigger signal.
  • any of the sound event detection methods may also include performing only the beginning detection method, which is of particular use when detecting sound events of short duration (“short-duration sound events”), such as impulsive sounds.
  • Short-duration sound events such as impulsive sounds
  • the sound event detection methods can be simplified to include performing only the beginning detection method on the input signal (in decibels) directly.
  • the sound event detection methods that include performing only the beginning detection method may also include removing any fluctuations with rise-times slower than those characteristic of the short-duration sound event being detected before performing the beginning detection method. For example, if impulsive sounds are being detected, fluctuations in the input signal with rise times slower than about 3 ms will be removed.
  • Methods for detecting a single sound event type in the presence of a background signal may also be implemented when the sound field is generated in two or more input channels (“sound event detection methods for multiple input channels and a single sound event type”).
  • the sound event detection methods for multiple input channels and a single sound event type may include performing the sound event detection method for a single input channel and a single sound event type in parallel for each input channel to produce a trigger signal for each input channel indicating the occurrence of whatever sound event type is being detected.
  • the remaining power envelope pair is made up of power envelopes derived from the sum and the difference of the two input channels, the “left plus right power” and “left minus right power.” This pair is often also referred to as “center power” or the “surround power.”
  • Non-directional signals almost always will cause all four of these power envelopes to be equal in level.
  • a non-directional transient signal will cause all four of them to rise at the same time.
  • a non-directional transient will cause no rise in the difference signal, because it is equal in each input channel.
  • a directional signal for example a sound event in the left input channel only, will cause a large change in the “left power” envelope, and no corresponding increase in the “right power” envelope. Therefore, there will be a large increase in the difference signal created from the left and right accommodated signals, and this change can be used to generate triggers that preferentially distinguish directional signals from non-directional signals.
  • a difference signal is determined 601 be determining the difference between the differential signals in each input channel of an input channel pair, and rectifying this difference.
  • the difference between the input channels in the input channel pair includes both positive and negative pulses representing sound events that have occurred in either input channel of the input channel pair.
  • the pulses will be negative or positive depending upon which input channel in the input channel pair reflects the majority of the power in the particular sound event. Therefore, this difference is rectified to produce a difference signal with only positive pulses. While the difference signal still indicates sound events, as previously discussed, it does not indicate sound events common to both input channels. Therefore, the difference signal has a better signal to noise ratio then that of the differential signals of the individual input channels.
  • the sound event detection methods for multiple channels and a single sound event type may also include detecting a sound event type in multiple channel pairs (collectively, “sound event detection methods for multiple channel pairs and a single sound event type”). These methods generally include the sound event detection methods for multiple channels and a single sound event type that include an alternate beginning detection method implemented for more than one channel pair, or for a single channel pair from which four power envelopes can be derived.
  • the directionally important frequencies are emphasized 704 using a frequency bias method.
  • the sound events are separated from the background signal 706 using an accommodation method to produce a differential signal.
  • the accommodation method may also include removing fluctuations in the power envelope with rise times that not characteristic of syllables (those above about 33 ms).
  • syllables are detected 710 using a beginning detection method for single input channel (see FIG. 4) and impulsive sounds are detected using a sound event detection method for short-duration sound events.
  • a beginning detection method for single input channel see FIG. 4
  • impulsive sounds are detected using a sound event detection method for short-duration sound events.
  • detecting sound events includes deemphasizing the effects of volume, emphasizing sound events and deemphasizing noise using an automatic gain method and by removing certain fluctuations, and detecting sound events using a threshold detection method.
  • the short-term high frequency power used to normalize the differential signal is the power in the components of the differential signal above 30 Hz determined over about 167 ms.
  • the fluctuations removed are those with a rise time of less than 33 ms.
  • the sound event detection method for a single short-duration sound event includes: removing certain fluctuations; deemphasizing the effects of volume, emphasizing sound events and deemphasizing noise using an automatic gain method and removing certain fluctuations, and detecting sound events using a threshold detection method.
  • the automatic gain method uses the input signals (in decibels) instead of the differential signal.
  • the fluctuations that are removed are those with rise times less than about 3 ms.
  • Methods for detecting one or more sound events in the presence of a background signal may also be implemented so that multiple types of sound events may be detected for a sound field generated in multiple input channels (“sound event detection methods for multiple input channels and multiple sound events”).
  • the sound event detection method for multiple input channels and multiple sound events includes performing the sound event detection method for a single input channel and multiple sound event types implemented in parallel for each input channel of the sound field to produce a trigger signal for each sound event type in each input channel.
  • the sound event detection method for multiple input channels and multiple sound events 800 may detect each sound event only in each input channel pair, thus producing a trigger signal for each sound event type in each input channel pair.
  • FIG. 8 the example shown in FIG.
  • this method 800 is used to detect two sound event types (syllables and impulsive sounds) in two input channels (a right input channel and a left input channel) to produce two trigger signals for the input channel pair, the first indicating the occurrence of syllables and the second indicating the occurrence of impulsive sounds.
  • the method may be implemented to detect any number of sound events in any number of input channels to produce for each input channel pair a trigger signal for each sound event type.
  • the sound event detection method for multiple input channels and multiple sound event types 800 generally includes: emphasizing directionally important frequencies by modeling the frequency bias in the left input channel 802 ; emphasizing directionally important frequencies by modeling the frequency bias in the right input channel 804 ; separating sound events from the background sounds by modeling accommodation in the left input channel 806 ; separating sound events from the background sounds by modeling accommodation in the left input channel 808 ; detecting impulsive sounds in the left-right input channel pair by modeling beginning detection 812 ; and detecting syllables in the left-right input channel pair by modeling beginning detection 814 .
  • syllables may be detected using an alternate beginning detection method to produce a single trigger signal indicating the occurrence of syllables for the channel pair.
  • a sound event detection method for a single channel and a short-duration sound event is used to produce a trigger signal indicating the occurrence of an impulsive sound for each input channel.
  • the trigger signals may be combined to create fewer trigger signals.
  • the sound event detection methods for multiple channels and multiple sound event types may also include detecting more than one sound event type in multiple channel pairs (collectively, “sound event detection methods for multiple channel pairs and multiple sound event types”) These methods generally include the sound event detection methods for multiple channels and multiple sound event types that include an alternate beginning detection method implemented for more than one channel pair, or for a single channel pair from which four power envelopes can be derived.
  • the sound event detection and localization methods can be implemented to detect and localize any number and combination of sound events types in a sound field generated in any number of input channel pairs.
  • the sound event detection and localization methods are discussed in order of increasing complexity, with each subsequent sound event detection and localization method incorporating the steps of the prior methods, except as indicated.
  • FIG. 9 A sound event detection and localization method for a single input channel pair and a single sound event type is shown in FIG. 9.
  • This sound event detection and localization method for a single input channel pair and a single sound event type 900 detects a single sound event type and locates the sound events relative to a single input channel pair.
  • the location of a sound event as determined by this method 900 is given in terms of an angle between the input channels in the input channel pair (referred to in this application as a “true differential steering angle” or “true DSA”).
  • the sound event detection and localization method 900 generally includes: producing a differential signal and detecting sound events in an input channel pair for each input channel in the input channel pair 904 ; determining the initial sound event direction 906 ; and isolating the true DSA 908 .
  • 904 and 906 define a DSA determination method for a single input channel pair and a single sound event type 901 .
  • a differential signal is produced for each input channel of the input channel pair and sound events are detected in the input channel pair 904 using a sound event detection method for a single input channel pair and a single sound event type or a sound event detection method for a single input channel pair and a single short-duration sound event as appropriate for the sound event type being detected.
  • a trigger signal indicating the occurrence of a single sound event type is produced for the input channel pair and a differential signal is produced for each input channel in the input channel pair.
  • the initial direction determination method determines the direction of the sound event from the differential signal for each input channel in the input channel pair as these differential signals exist when the sound event occurs. More specifically, when the trigger signal indicates that a sound event is occurring, the differential signals for each input channel are converted into decibels to create a decibel differential signal for each input channel. The decibel differential signal for one input channel is then subtracted from the decibel differential signal from the other input channel to create a decibel ratio. This decibel ratio is converted into an equivalent angle using known methods to create a “differential steering angle” or “DSA” which indicates the location of the sound event between the input channels in the input channel pair.
  • DSA differential steering angle
  • the true direction of any syllables detected is obtained by isolating the DSA indicated in about the first 20 ms to about the first 30 ms of the syllable, which is then used (held) for about 200 ms.
  • the true direction of any impulsive sounds is obtained by isolating the DSA indicated in about the first 3.3 ms to about 5 ms of the impulsive sound, which is then used for about 50 ms.
  • the initial direction determination method may be implemented to obtain an even more accurate initial direction.
  • Sound event detection and localization methods may also be implemented to detect and localize a single sound event type in multiple input channel pairs (“sound event localization methods for multiple input channels pairs and a single sound event type”).
  • a typical surround sound system includes at least a left-right input channel pair and a center-surround input channel pair.
  • the true DSA for each channel pair is determined whenever a trigger signal in any channel pair indicates that a sound event is occurring, however, the DSA is used to indicate the direction of the sound event only if the DSA is accurate. If the DSA is not accurate, the direction indicated by a true OSA is used.
  • FIG. 10 An example of a sound event localization method for multiple input channel pairs and a single sound event type, is shown in FIG. 10, indicated by reference number 1000 and includes: performing a DSA determination method for a single input channel pair and a single sound event for each input channel pair (collectively a “DSA determination method for multiple input channel pairs a single sound event type”) 1051 ; determining the ordinary steering angle for each channel pair 1003 ; determining whether the DSA is accurate 1010 , where if the DSA is not accurate, isolating and using the true OSA for each input channel pair for the typical duration of the sound event type being detected 1014 ; where, if the DSA is accurate, isolating the true DSA and using the true DSA for the typical duration of the sound event type being detected 1012 .
  • the consistency check determines whether the sum of the absolute values of the true sound event directions in the left-right input channel pair and the center-surround input channel pair is less than about 45 degrees. Therefore, a sum of less than 45 degrees indicates that the steady-state signal did remain constant and that the direction indicated by the true sound event direction is accurate. However, a sum of greater than or equal to about 45 degrees indicates that the steady-state signal did not remain constant and that the direction indicated by the true sound event direction is not accurate. Based on this, the consistency check produces a signal whenever the true sound event is found to be inaccurate, thus indicating an error (an “error signal”).
  • Adj 1 - ⁇ 50 ( 2 )
  • the sound event detection and localization methods for multiple input channels pairs and a single sound event type may use the number of errors to adjust the threshold as part of error threshold adjustment methods.
  • An error threshold adjustment method determines the number of errors determined by the consistency check in a predetermined time period of about several seconds and uses this number to adjust the threshold. As the number of errors increases, the threshold is increased so that fewer sound events are detected. Therefore, a continuous adjustment to the threshold can be provided, which is in proportion to the number of errors detected by the consistency check.
  • the sound event detection and localization method for a single input channel pair and multiple sound event types 1100 generally includes: producing a differential signal for the input channel pair 1104 ; detecting syllables in a input channel pair 1106 ; detecting impulsive sounds in the input channel pair 1107 ; determining the initial direction 1108 ; and isolating the true DSA 1110 for the sound event detected.
  • steps 1104 , 1106 , 1107 , and 1108 define a DSA determination method for a single input channel pair and multiple sound event type 1101 .
  • Sound event detection and localization methods may also be implemented to detect and localize multiple sound event types in multiple input channel pairs (“sound event localization methods for multiple input channel pairs and multiple sound event types”).
  • the sound event localization methods for multiple input channel pairs and multiple sound event types generally involve performing a sound event detection and localization method for a single input channel pair and multiple sound event types in a parallel fashion for each input channel pair.
  • sound event localization methods for multiple input channel pairs and multiple sound event types may include detecting a sound event, determining the DSA for the sound event, determining whether the DSA is accurate and using the OSA if the DSA is not accurate.
  • FIG. 12 An example of a sound event localization methods for multiple input channel pairs and multiple sound event types that uses the OSA to indicate the direction of a sound event if the DSA is not accurate is shown in FIG. 12.
  • syllables and impulsive sounds are detected in two channel pairs.
  • this method may be implemented to detect any number of wound event types in any number of channels.
  • the true OSA is isolated 1214 to produce the true OSA.
  • the true OSA is then isolated from the OSA during the rise time of the impulsive sound and is used for the typical duration of an impulsive sound.
  • the sound event detection and localization methods for multiple channel pairs and a single sound event type may further include an accommodation adjustment method and/or an error threshold adjustment method.
  • the stereo/surround detection methods include performing a sound event detection and localization method for a single input channel pair for the center-surround input channel pair for each sound event type being detected, and determining the number of sound events with an associated true differential steering angle of about 0 degrees to about ⁇ 45 degrees (indicating a rear direction).
  • the surround detection method also determines whether the number of sound events detected for reproduction in the rear exceeds a predetermined value in a defined time period.
  • the properties of the human hearing mechanism may also be used to determine the location of all sounds within a sound field.
  • sound localization methods divide the sound field into sound events and non-sound events and localize the sound events in terms of a true differential steering angle or a true ordinary steering angle and the non-sound events in terms of a filtered ordinary steering angle. These methods essentially treat the non-sound events as a separate sound for which the direction is separately determined.
  • the sound localization methods can be implemented to specifically localize any number and combination of sound event types, in addition to localizing the remaining sounds, in a sound field generated in any number of input channel pairs.
  • the sound localization methods are discussed in order of increasing complexity, with each subsequent sound localization method incorporating the steps of the prior methods, except as indicated.
  • the filtered OSA is used to indicate the direction of the sound field unless a sound event is detected, in which case the true DSA is used.
  • the true DSA is used to indicate the direction of sound events only if the DSA is determined to be accurate. In these cases, if the DSA is not found to be accurate, the true OSA is used to indicate the direction of the sound events.
  • the DSA (which indicates the direction of the sound event) will be used (in some cases only if accurate) to define the direction of the entire sound field for the duration of the sound event.
  • the human hearing mechanism which generally perceives sound events as part of a foreground stream and noise as part of a background stream, will perceive only the sound event as having moved.
  • Determining the DSA and the trigger signal 1302 is generally accomplished by performing a DSA determination method for a single input channel pair and a single sound event that includes an alternate beginning detection method to produce a single trigger signal for the input channel pair. Even though sound events only occur whenever a pulse is present in the trigger signal, the DSA may be continuously determined. Alternatively, the trigger signal may be continuously determined and the DSA determined only when a pulse is present in the trigger signal.
  • the OSA is generally determined 1306 continuously using known methods. Determining which direction to use 1304 basically includes: determining when a sound event occurs, using the true DSA for the duration typical of the sound event type detected and decaying to the OSA at the end of the sound event. However, if at any time a subsequent sound event occurs (even during a sound event), the DSA for the subsequent sound event will be used for the duration typical of the subsequent sound event type.
  • a method for determining which direction to use (the “direction selection method for a single input channel pair and a single sound event type”) is shown in FIG. 14 and indicated by reference number 1304 . It generally includes, determining whether there is an input signal 1402 ; where if there is an input signal determining whether there is a sound event 1404 ; where if there is a sound event, generating and selecting the current true DSA for the typical duration of the sound event type being detected 1406 ; determining whether the typical duration has ended 1408 ; where if the typical duration has not ended, determining whether a subsequent sound event is detected 1410 , where if a subsequent sound event is not detected, repeating determining whether the typical duration has ended 1408 and whether a subsequent sound event has been detected 1410 until it is determined that either the typical duration has ended in 1408 or that a subsequent sound event has been detected in 1410 ; where if a subsequent sound event has been detected, defining the current DSA with that of the subsequent sound event 1414 and repeating 1406 .
  • a current true DSA is generated from the current DSA using a true direction isolation method and selected for the typical duration of the sound event type being detected 1408 .
  • the typical duration of the sound event will be about 50 ms to 200 ms (preferably about 150 ms) and if the sound event type being detected is an impulsive sound, than the typical duration will be about 50 ms.
  • the DSA is selected and used for the typical duration regardless of when the sound event being detected actually ends. However, during the typical duration, the input signal is monitored to determine if any subsequent sound events with an accurate DSA are detected ( 1408 and 1410 ).
  • the current DSA will be redefined by the DSA of the subsequent sound event 1414 , the redefined current DSA will be used to generate a current true DSA which will be selected for the typical duration of the sound event type detected 1406 , and 1408 , 1410 and 1414 will be repeated as appropriate. However, if no subsequent sound event is detected during the typical duration ( 1408 and 1410 ), then the entire method repeats, as appropriate, from 1402 .
  • a filtered OSA is selected, or continues to be selected 1418 and the process repeats, as appropriate from 1402 .
  • the filtered OSA is the OSA with the fluctuations having rise-times faster than a specified rise-time removed. For example, fluctuations with rise-times faster than approximately 300 ms may be removed. This prevents the OSA from reflecting rapid directional changes when no sound event is detected.
  • the OSA is selected and decayed to from the true DSA of the immediately preceding sound event 1420 .
  • the decay helps to provide a smooth transition from DSA of the immediately preceding sound event to the OSA.
  • the length of decay needed to provide this smooth transition depends on the sound event type of the immediately preceding sound event. If the immediately preceding sound event was a syllable, the decay will generally be about 300 ms seconds. The process then repeats, as appropriate, from 1402 until it is determined in 1402 that the input signal has ended.
  • the direction selection method includes determining whether the short-duration sound event has actually ended at the end of the typical duration, and immediately selecting the OSA without any decay if it is determined that the short-duration sound event has not actually ended.
  • Such a method (a “direction selection method for a single input channel pair and a short-duration sound event”) is shown in FIG. 15 and indicated by reference number 1500 .
  • the direction selection method for a single input channel pair and a short-duration sound event generally includes, determining whether there is an input signal 1502 ; where if there is an input signal, determining whether there is a short-duration sound event 1504 ; where if there is a short-duration sound event, generating and selecting the true DSA for the typical duration of the short-duration sound event detected 1506 ; determining whether the typical duration has ended 1508 ; where if the typical duration has not ended, determining whether a subsequent short-duration sound event is detected 1510 , where if a subsequent short-duration sound event is not detected, repeating determining whether the typical duration has ended 1508 and whether a subsequent short-duration sound event has been detected until it is determined that either the typical duration has ended in 1508 or that a subsequent short-duration sound event has been detected in step 1510 ; where if a subsequent short-duration sound event has been detected, defining the current DSA with that of the subsequent short-duration sound event 1512 and repeating
  • This direction selection method for a single input channel pair and a short-duration sound event is virtually the same as the previously discussed direction selection methods, except that the direction selection method for a single input channel pair and a short-duration sound event is implemented to detect short-duration sound events. Furthermore, at the end of a typical duration of a short-duration sound event (assuming no new sound event has or is occurring), a determination is made as to whether the sound event has actually ended 1514 ; and the filtered OSA is used either immediately 1520 or after a decay 1522 depending upon whether the short-duration sound event has actually ended.
  • the direction selection method for a single input channel pair and a short-duration sound event is implemented to detect short-duration sound events by using a sound event detection method for a single channel and a single short-duration sound event for each input channel of the channel pair to produce a trigger signal (or two trigger signals) that indicates whenever a short-duration sound event is detected.
  • Whether the immediately preceding short-duration sound event has actually ended may be determined 1518 by comparing the power envelope and the accommodation signal in each input channel of the channel pair. If the input power envelope I 2 is greater than the accommodation signal AccSig in any input channel of the input channel pair, it is determined that the short-duration sound event has not actually ended. Therefore, the filtered OSA is decayed to from the DSA of the immediately preceding sound event. If however, I 2 is about equal to or less than AccSig in each input channel of the input channel pair, it is determined that the short-duration sound event has actually ended and the filtered OSA is immediately selected 1522 . The length of decay needed to provide this smooth transition depends on the sound event type being detected. For example, if the immediately preceding sound event is an impulsive sound, the decay will take about 5 ms.
  • Sound localization methods may also be used to localize the sounds in a sound field when the sound field includes more than one input channel pair (“sound localization methods for multiple input channel pairs and a single sound event type”).
  • the sound localization methods for multiple input channel pairs and a single sound event include the same basic steps as the sound localization method for a single input channel pair and a single sound event type, which generally include: determining the OSA; determining the DSA and trigger signal; and determining which direction to use.
  • the step of determining the DSA and a trigger signal includes determining a DSA and a trigger signal for each input channel pair and is accomplished by performing a DSA determination method for multiple input channel pairs and a single sound event that includes an alternate beginning detection method.
  • the step of determining which direction to use includes responding to differential signals, trigger signals and DSAs from any of the input channel pairs; and generating and selecting the true differential steering angle for all the input channel pairs when a sound event with an accurate DSA is detected in any input channel pair.
  • the direction selection method for multiple channel pairs and a single sound event type is shown in more detail in FIG. 16 and is indicated by reference number 1600 .
  • this direction selection method includes determining whether, at the time any trigger signal indicates the occurrence of a sound event, the DSAs are accurate; and if the DSAs are not accurate using the true OSAs for the typical duration of the sound event type being detected instead of the DSAs.
  • this direction selection method 1600 includes: determining whether there is an input signal 1602 ; where if there is an input signal determining whether there is a sound event 1604 ; where if there is a sound event, determining whether the current true DSA is accurate 1606 ; where if the current true DSA is accurate, generating and selecting the current true DSA for the typical duration of the sound event type being detected 1608 ; where if the current true DSA is not accurate, generating and selecting the current true OSA for the typical duration of the sound event type being detected 1620 ; once either the true OSA or DSA is generated and selected, determining whether the typical duration has ended 1610 ; where if the typical duration has not ended, determining whether a subsequent sound event is detected 1612 , where if a subsequent sound event is not detected, repeating determining whether the typical duration has ended 1610 and whether a subsequent sound event has been detected 1612 until it is determined that either the typical duration has ended in 1610 or that a subsequent sound event has been detected in 1612 ; where
  • Determining whether there is an input signal 1602 includes determining whether there is an input signal in each input channel of each channel pair.
  • Determining whether there is a sound event 1604 includes performing a sound event detection method for a single channel pair and a single sound event type for each channel pair to produce a trigger signal that indicates the occurrence of any sound events for each pair.
  • Determining whether the current DSA is accurate 1606 includes determining whether the current DSA (the DSA determined at the moment any trigger signal indicates that a sound event is occurring) from every input channel pair is accurate using a method for determining DSA accuracy for multiple input channel pairs. A method for determining DSA accuracy for multiple input channel pairs is shown in more detail in FIG.
  • the current true DSAs are generated and selected for each channel pair 1608 using a true direction isolation method.
  • the current true OSAs (the OSA determined at the moment any of the trigger signals indicates that the current sound event is occurring) are generated and selected for each channel pair 1620 .
  • Either the current true OSA or the current true DSA is used for the typical duration of the sound event type being detected, unless a subsequent sound event is detected during the typical duration. As long as it is determined that the typical duration has not ended in 1610 , it is determined whether any subsequent sound events are detected 1612 .
  • a subsequent sound event it is determined whether the true DSAs for the subsequent sound event are accurate 1614 using a method for determining DSA accuracy for multiple input channel pairs. If the DSAs of the subsequent sound event are determined to be accurate, the DSAs of the subsequent sound event become the current DSAs 1616 to reflect the direction of the subsequent sound event. However, if the DSAs of the subsequent sound event are determined not to be accurate, the OSAs of the subsequent sound become the current OSAs 1618 to reflect the direction of the sound event. The process repeats from 1608 or 1620 , as appropriate, until the typical duration of any subsequent sound events ends.
  • any subsequent sound event ends, it is determined in 1622 whether there was an immediately preceding sound event. If there was an immediately preceding sound event, the filtered OSAs are selected and the direction of the sound field decays from that indicated by the DSAs to that indicated by the OSAs 1626 . However, if there was no immediately preceding sound event, the filtered OSA is used, or continues to be used 1624 .
  • any of the sound localization methods for multiple channel pairs and a single sound event type may further include using the consistency check to reduce the occurrence of inaccurate DSAs through use of an accommodation adjustment method and/or an error threshold adjustment method as previously described.
  • the direction selection method for multiple input channel pairs and a single sound event type includes determining whether the short-duration sound event has actually ended at the end of the typical duration, and immediately selecting the filtered OSA without any decay if it is determined that the short-duration sound event has not actually ended.
  • Such a method a “direction selection method for multiple input channel pairs and a short-duration sound event” is shown in FIG. 18 and indicated by reference number 1800 .
  • the direction selection method for a single input channel pair and a short-duration sound event generally includes: determining whether there is an input signal 1802 ; where if there is an input signal determining whether there is a short-duration sound event 1804 ; where if there is a short-duration sound event, determining whether the current true DSA is accurate 1806 ; where if the current true DSA is accurate, generating and selecting the current true DSA for the typical duration of the short-duration sound event being detected 1808 ; where if the current true DSA is not accurate, generating and selecting the current true OSA for the typical duration of the short-duration sound event being detected 1820 ; once either the true OSA or DSA is generated and selected, determining whether the typical duration has ended 1810 ; where if the typical duration has not ended, determining whether a subsequent short-duration sound event is detected 1812 , where if a subsequent sound event is not detected, repeating determining whether the typical duration has ended 1810 and whether a subsequent sound event has been detected 18
  • This direction selection method for multiple input channel pairs and a short-duration sound event is virtually the same as the previously discussed direction selection method for multiple input channels and a single sound event type, except that the direction selection method for multiple input channel pairs and a short-duration sound event is implemented to detect short-duration sound events. Furthermore, at the end of a typical duration of a short-duration sound event (assuming no new sound event has or is occurring), a determination is made as to whether the sound event has actually ended 1825 ; and the filtered OSA is used either immediately 1826 or after a decay 1828 depending upon whether the short-duration sound event has actually ended.
  • the direction selection method for multiple input channel pairs and a short-duration sound event is implemented to detect short-duration sound events by using a sound event detection method for multiple input channels and a single short-duration sound event for each input channel of the channel pair to produce a trigger signal (or two trigger signals) that indicates whenever a short-duration sound event is detected.
  • Whether the immediately preceding sound event has actually ended is determined 1825 by determining whether the input power envelope is greater than the accommodation signal for the short-duration sound event in any input channel. If the input power envelope is greater than the accommodation signal for the short-duration sound event in any input channel, it is determined that the short-duration sound event has not actually ended. Therefore, a decay is made from the DSAs of the immediately preceding short-duration sound event to the filtered OSA 1826 . If however, in each input channel, the power envelope is about equal to the accommodation signal, it is determined that the short-duration sound event has actually ended and the filtered OSA is immediately selected 1828 . The length of decay needed to provide this smooth transition depends on the sound event type being detected. For example, if the immediately preceding sound event is an impulsive sound, the decay will take about 5 ms.
  • Sound localization methods may also be used to localize the sounds in a sound field by distinguishing more than one sound event type (“sound localization methods for a single input channel pair and multiple sound event types”).
  • the sound localization methods for a single input channel pair and multiple sound event types include the same basic steps as the sound localization method for a single input channel pair and a single sound event type, which generally include: determining the OSA; determining the DSA and trigger signal; and determining which direction to use.
  • determining the DSA and the trigger signal includes determining the DSA and trigger signal for each sound event type by performing a DSA determination method for a single input channel pair and multiple sound event types that uses an alternate beginning detection method.
  • determining which direction to use includes performing a direction selection method for a single input channel pair (either for a single sound event type or a short-duration sound event) in parallel for each sound event type being detected. However, whenever a sound event of any type is detected in any channel pair, the DSA is used in each input channel pair for the typical duration of the sound event type that was actually detected.
  • Sound localization methods may also be used to localize the sounds in a sound field with more than one input channel pair by distinguishing more than one sound event type (“sound localization methods for multiple input channel pairs and multiple sound event types”).
  • the sound localization methods for multiple channel pairs and multiple sound event types include the same basic steps as the sound localization method for multiple input channel pairs and a single sound event type, which generally includes: determining the OSA; determining the DSA and trigger signal for each channel pair; and determining which direction to use according to a direction selection method for multiple input channel pairs and a single sound event type.
  • determining the DSA and trigger signal for each channel pair includes determining a DSA and trigger signal for each sound event type in each input channel pair.
  • Determining a DSA and trigger signal for each sound event type in each input channel pair is accomplished by performing a DSA determination method for multiple input channel pairs and multiple sound event types that includes an alternate beginning detection method.
  • performing the direction selection method for multiple input channel pairs and multiple sound event types in parallel for each sound event includes, responding to the trigger signals from any sound event type; generating and selecting the true DSA for all the input channel pairs if a sound event of any type with an accurate DSA is detected in any input channel pair; or selecting the true OSA for all the input channel pairs if a sound event of any type with an inaccurate DSA is detected in any channel.
  • the sound localization methods for multiple input channel pairs and multiple sound event types may also include using the accommodation adjustment methods and/or the error threshold adjustment methods as previously described.
  • the method will determine if the drum hit has actually ended, and if it has, the method will immediate revert to using the direction indicated by the filtered OSA and move the entire sound field back to the center input channel.
  • the human hearing mechanism will perceive the drum hit as originating from the rear and the music as continually originating from the front as if the music had never moved. If however, the DSA is used for longer than the typical duration of the drum hit or if at the end of the typical duration a decay is used to revert to the filtered OSA when the drum hit actually ends prior to the end of the typical duration, the entire sound field, including the music will be perceived as having moved to the rear.
  • the sound event detection methods, sound event detection and localization methods and sound localization methods and any methods included in any of these methods may be implemented in a sound event device as shown in FIG. 19 and indicated as reference number 1900 .
  • the optimization device 1900 generally includes a detection unit 1902 and may also include an interface unit 1904 .
  • the detection unit 1902 includes a processor 1908 coupled to a memory device 1906 .
  • the memory device 1908 may be any type of fixed or removable digital storage device and (if needed) a device for reading the digital storage device including, floppy disks and floppy drives, CD-ROM disks and drives, optical disks and drives, hard-drives, RAM, ROM and other such devices for storing digital information.
  • the processor 1908 may be any type of apparatus used to process digital information.
  • the memory device 1906 may store the sound field and at least one of the following methods: the sound event detection methods, sound event detection and localization methods and sound localization methods and any methods included in any of these methods (collectively, the “detection and/or localization methods”).
  • the memory Upon the relevant request from the processor 1908 via a processor signal 1910 , the memory communicates one of the detection and/or localization methods, and if necessary the sound field, via a memory signal 1912 to the processor 1908 .
  • the processor 1908 then performs the detection and/or localization method.
  • the interface unit 1904 generally includes an input device 1914 and an output device 1916 .
  • the output device 1916 is any type of visual, manual, audio, electronic or electromagnetic device capable of communicating information from a processor or memory to a person or other processor or memory. Examples of output devices include, but are not limited to, monitors, speakers, liquid crystal displays, networks, buses, and interfaces.
  • the input device 1914 is any type of visual, manual, mechanical, audio, electronic, or electromagnetic device capable of communicating information from a person or processor or memory to a processor or memory. Examples of input devices include keyboards, microphones, voice recognition systems, trackballs, mice, networks, buses, and interfaces.
  • the input and output devices 1914 and 1916 may be included in a single device such as a touch screen, computer, processor or memory coupled to the processor via a network.
  • the sound field may be communicated to the memory device 1918 from the input device 1914 through the processor 1920 .
  • the optimized model parameters may be communicated from the processor 1920 to the output device 1916 .
  • Sound event detectors may be used for a variety of applications for which the detection of specific types of sound events is helpful.
  • a sound event detector that detects syllables may be used as part of a phonic detector or as part of a speech recognition or speech coding system.
  • a sound event detector that detects syllables may be used in conjunction with an audio amplification device, such as a microphone.
  • Sound event detectors like the sound event detection methods, generally produce a trigger signal that includes some flag or marker, such as a pulse, that indicates the occurrence of a sound event. Also, like the sound event detection methods, the sound event detectors can be implemented to detect any number and combination of sound event types in a sound field that is generated in any number of input channels. In the following description, the sound event detectors are discussed in order of increasing complexity, with each subsequent sound detector incorporating the elements of the prior sound event detectors, except as indicated.
  • FIG. 20 One example of a sound event detector that detects a single type of sound event in a sound field generated in only one input channel (a “sound event detector for a single input channel and a single sound event type” is shown in FIG. 20 and indicated by reference number 2000 .
  • the sound event detector and localizer may be implemented to detect any number of sound event types in any number of input channel pairs.
  • the entire sound field is generated through a left input channel.
  • the term “left” as used in this example does not have any directional meaning because the entire sound field is contained in a single input channel and is used simply for the purposes of explanation.
  • a sound event detector for a single input channel and a single sound event type 1900 generally includes a frequency bias filter 2001 ; an accommodation circuit for a single input channel 2002 ; and a trigger generation circuit for a single input channel 2004 .
  • the accommodation circuit for a single input channel 2002 uses the sound field to produce a differential signal in the left input channel “Lo” for the sound event type being detected and the trigger generation circuit for a single input channel 2004 uses the accommodation signal Lo to produce a trigger signal “Tl” indicating whenever a sound event of the type being detected is detected.
  • the frequency bias filter 2001 models the frequency bias of the human hearing mechanism by emphasizing frequencies in the sound field from about 500 Hz to about 4000 Hz.
  • the accommodation circuit for a single input channel 1702 separates sound events from any background signals in the sound field by modeling accommodation.
  • This circuit 2102 is shown in more detail in FIG. 21 and generally includes: a multiplier 2102 ; a low-pass filter 2104 ; and an accommodation signal circuit 2006 .
  • the multiplier 2002 converts the sound field as generated in the left input channel (the “input signal” into a power signal “Lin 2 .”
  • the input signal is generally a voltage signal and can generally be converted into a power signal by squaring the input signal.
  • the resulting power signal Lin 2 includes many fluctuations, some of which indicate sound events, and some of which indicate noise.
  • the low-pass filter then removes the fluctuations with rise times faster than about 30 ms from the power signal Lin 2 to produce a filtered power signal L 2 .
  • This low-pass filter 2104 may be of any type, such as a filter with a roll-off of 12 dB/octave.
  • the accommodation circuit 2106 creates and subtracts the accommodation signal (which represents the long-term average power in the sound field as it is) to create the differential signal Lo.
  • the accommodation circuit 2106 generally includes an operational amplifier 2108 ; a resistor 2114 ; a diode 2112 and a capacitor 2110 .
  • the filtered power signal L 2 is coupled to the positive terminal of the operational amplifier 2108 or, alternatively any device capable of determining a difference between two signals.
  • capacitor 2110 will act as an open circuit resulting in an accommodation signal ⁇ overscore (L) ⁇ 2 that is about equal the filtered power signal L 2 to produce a differential signal Lo about equal to zero.
  • the filtered power signal L 2 when the filtered power signal L 2 does contain a sound event, the filtered power signal L 2 will rapidly increase according to the rise time of the sound event. This rapid increase in L 2 will cause a corresponding spike in the differential signal Lo.
  • capacitor 2110 After the rise-time of the sound event, capacitor 2110 will charge causing the accommodation signal ⁇ overscore (L) ⁇ 2 to gradually rise according to time constant defined by resistor 2114 and capacitor 2110 . This time constant is generally made equal to the accommodation rate of the human hearing mechanism, determined experimentally to be about 300 ms.
  • ⁇ overscore (L) ⁇ 2 will continue to rise until the voltage across capacitor 2110 (and thus ⁇ overscore (L) ⁇ 2 ) equals L 2 or until the sound event ends or starts to decay.
  • the differential signal Lo therefore, includes a series of fluctuations with on-times equal to or less than those characteristic of the sound event type being detected and with fall-times defined by the accommodation signal, and/or the end of the sound event.
  • the trigger generation circuit for a single input channel 1904 (shown in FIG. 19) then detects the sound events to produce a trigger signal Tl′ that includes a pulse whenever a sound event is detected.
  • a trigger generation circuit 1904 for a single input channel is shown in more detail in FIG. 22 and includes: a high-pass filter 2202 ; a normalization circuit 2206 and a low-pass filter 2208 .
  • the goal of the trigger generation circuit 1904 is to remove as many fluctuations caused by noise as possible and to deemphasize those that are not removed. As explained previously in connection with the sound event detection methods, this is accomplished by removing the fluctuations that have frequencies higher than those characteristic of the sound event being detected and by normalizing the differential signal Lo with the short-term high frequency power in the differential signal.
  • the normalization is accomplished using an automatic gain control circuit which includes the high-pass filter 2202 and the normalization circuit 2206 .
  • the high-pass filter includes a capacitor/resistor pair that defines the cutoff frequency as that which is characteristic of the sound event being detected. Additionally, a rectifier (not shown) may be included between the high-pass filter 2202 and the normalization circuit 2206 to rectify any negative pulses or fluctuations.
  • the normalized differential signal Nl is then filtered by a low-pass filter 2208 to remove fluctuations with frequencies higher than are characteristic of the sound event being detected to yield a filtered normalized differential signal Nl′.
  • additional noise can be removed from Nl′ by including circuitry that detects and removes fluctuations that occur more often than is characteristic of the sound event being detected and that remove any fluctuations that occur when a decrease in the sound field of at least 10 dB is detected.
  • Nl′ therefore, includes a series of pulses of varying amplitudes representing the occurrence of sound events and fluctuations due to noise.
  • a threshold detector 2218 In order to detect the sound events from the noise in the filtered normalized differential signal Nl′, a threshold detector 2218 detects only those pulses with an amplitude greater than a threshold. This helps to distinguish pulses indicating sound events from fluctuations due to noise.
  • the output of the threshold detector is a trigger signal “Tl” that indicates, generally by pulses, the occurrence of a sound event in the sole (left) input channel of the sound field.
  • the sound event detector for a single input channel and a single sound event type may also include a threshold adjustment circuit. The threshold adjustment circuit adjusts the threshold of the threshold detector in order to adjust the sensitivity of the sound event detector.
  • the comparator then produces a voltage which is inversely proportional to the number of sound events and communicates this voltage to the threshold detector which uses the voltage to define the threshold.
  • the threshold is increased when there are more sound events detected so that the sensitivity of the sound detector is increased.
  • any of the sound event detectors may not include an accommodation signal circuit, which is of particular use when detecting short-duration sound events, such as impulsive sounds.
  • An example of a sound event detector that does not include an accommodation signal circuit (the “sound event detectors for short-duration sound events” is shown in FIG. 23.
  • This sound event detector for short-duration sound events 2300 includes: a frequency bias filter 2301 ; a linear to dB converter 2302 ; a high-pass filter 2303 ; and a trigger generation circuit for a single channel pair 2304 .
  • the frequency bias filter 2301 emphasizes the frequencies in the input signal from about 500 Hz to about 4000 Hz to produce a filtered input signal.
  • the filtered input signal is then converted into decibels by the linear to dB converter 2302 to produce a decibel filtered input signal.
  • the filtered decibel input signal is again filtered, this time by the high-pass filter 2303 .
  • the high-pass filter 2303 removes any fluctuations with rise-times slower than are characteristic of the short-duration sound event being detected.
  • the trigger generation circuit for a single input channel pair 2304 then uses this twice filtered decibel input signal to create a trigger signal that indicates the occurrence of short-durations sound events.
  • the sound event detectors for short-duration sound events may also include a threshold adjustment circuit.
  • Sound event detectors may also be implemented when the sound field is generated in two or more input channels.
  • a sound event detector that detects a single sound event type in a sound field generated in two input channel may include a sound event detector for a single channel and a single channel pair for each input channel that produces a trigger signal for each input channel.
  • the trigger signals may be combined to form a single trigger signal that indicates the occurrence of a sound event in any input channel.
  • a sound event detector for multiple input channels and a single sound event type may produce only a single trigger signal for each channel pair from a difference signal.
  • FIG. 24 Such a sound event detector for multiple input channels and a single sound event type is shown in FIG. 24. In this example, the entire sound field is either generated through a single input channel pair including a left input channel and a right input channel. However, this method is applicable for any number of input channels or input channel pairs.
  • the sound event detector for multiple input channels and a single sound event type 2400 includes: a first accommodation circuit for a single input channel 2402 ; a second accommodation circuit for a single input channel 2404 ; and an alternate trigger generation circuit for a single input channel pair 2406 .
  • the first and second accommodation circuits 2402 and 2404 are generally identical.
  • the first accommodation circuit 2402 uses the left input channel signal (“Lin” to produce a differential signal for the left input channel (the “left differential signal” or “Lo”).
  • the second accommodation circuit 2404 uses the right input channel signal (“Rin” to produce a differential signal for the right input channel (the “right differential signal” or “Ro”).
  • the alternate trigger generation circuit for a single input channel pair 2406 uses both the right and left differential signals to produce a trigger signal that indicates the occurrence of sound events in either input channel (the “left-right trigger signal” or “Tlr”).
  • the trigger generation circuit for a single input channel pair 2406 is shown in more detail in FIG. 25 and includes: an operational amplifier 2501 ; a rectified high-pass filter 2502 ; a
  • the trigger generation circuit for a single input channel pair 2406 also includes an operational amplifier 2501 (or other device capable of determining a difference) that creates a signal equal to the difference between the left and right differential signal (the “left-right difference signal” or “Lo ⁇ Ro” and uses the left-right difference signal to create a trigger signal for the left-right input channel pair (the “left-right trigger signal” or Tlr”).
  • the left-right difference signal is obtained by subtracting Ro from Lo, it may alternatively be determined by subtracting Lo from Ro.
  • the left-right difference signal Lo ⁇ Ro includes a series of pulses and other fluctuations that indicated the occurrence of sound events and noise in either input channel of the input channel pair. However, because sound events that occur equally in both input channels are removed, many fluctuations due to noise are thus removed.
  • the pulses and fluctuations in the difference signal may have a positive or a negative amuplitude depending on whether the power in the transient is greater in the left input channel or the right input channel, respectively.
  • a first rectifier 2504 is included in the rectified high-pass filter 2502 and a second rectifier is included in the normalization circuit 2506 .
  • the rectified high-pass filter 2502 produces a rectified high-pass left-right difference signal (“HFlr”).
  • This rectified high-pass left-right difference signal is used by the normalization circuit 2506 to normalize the difference signal Lo ⁇ Ro, the result of which is rectified by the second rectifier 2507 to produce a normalized left-right signal (“Nlr”).
  • the low-pass filter 2508 removes fluctuations due to noise with rise-times faster than those characteristic of the sound event being detected to produce a filtered normalized left-right signal (“Nlr′”).
  • Nlr′ Just as in the trigger generation circuit for a single input channel, (not shown) additional noise can be removed from Nlr′ by including circuitry that detects and removes fluctuations that occur more often than is characteristic of the sound event being detected and that remove any fluctuations that occur when a decrease in the sound field of at least 10 dB is detected.
  • Nlr′ therefore, includes a series of pulses of varying amplitudes representing the occurrence of sound events and fluctuations due to the remaining noise.
  • the threshold detector 2510 detects the sound events as those pulses that have an amplitude greater than a threshold to create the left-right trigger signal Tlr. This sound event detector may be repeated in parallel for multiple input channel pairs to produce a trigger signal for each input channel pair.
  • this sound event detector for multiple input channels and a single sound event may also include a threshold adjustment circuit for each trigger generation circuit.
  • the sound event detectors for multiple input channels and a single sound event that include an alternate trigger generation circuit may also be referred to as “sound event detectors for a single channel pair and a single sound event type.” Additionally, any of the sound event detectors for multiple input channels (or a single input channel pair) and a single sound event may include a threshold adjustment circuit for each trigger generation circuit.
  • Sound event detectors may also be implemented so that more than one type of sound event is detected.
  • These “sound event detector for a single input channel and multiple sound event types” generally include a sound event detector for a single input channel and a single sound event type implemented in parallel for each sound event type being detected to produce a trigger signal for each sound event type being detected.
  • An example of such a sound event detector for a single input channel and multiple sound event types is shown in FIG. 26 and indicated by reference number 2600 .
  • the entire sound field is generated through a left input channel.
  • the term “left” as used in this example does not have any directional meaning because the entire sound field is contained in a single input channel and, in fact, the input channel can be given any designation.
  • the sound event detector for a single input channel and multiple sound event types 2600 is implemented to detect syllables and impulsive sounds. However, any number or combination of sound events may be detected.
  • the sound event detector for a single input channel and multiple sound event types 2600 generally includes: a accommodation circuit for a s ingle input channel 2602 ; a trigger generation circuit for a single input channel implemented for syllables 2604 ; and a sound event detector for a single input channel and a single short-duration sound event implemented for impulsive sounds 2606 .
  • the accommodation circuit for a single input channel 2602 uses the input signal Lin to produce a differential signal.
  • the trigger generation signal for a single input channel implemented to detect syllables 2604 uses the differential signal to produce a trigger signal that indicates the occurrence of syllables in the sole input channel (left) of the sound field (“Tl(s)”).
  • the trigger generation signal for a single input channel implemented to detect syllables 2604 includes filters (see FIG. 22) for which the cut-off frequency rise time is about 33 ms.
  • the trigger generation circuit for a single input channel implemented to detect impulsive sounds 2606 uses the input signal Lin to produce a trigger signal that indicates the occurrence of impulsive sounds in the sole input channel (left) of the sound field (“Tl(i)”).
  • This trigger generation circuit for a single input channel implemented to detect impulsive sounds 2606 includes a high-pass filter (see 2303 in FIG. 22) for which the cut-off rise-time is about 3 ms.
  • the sound event detectors for multiple input channels and a single sound event may include a threshold adjustment circuit for each trigger generation circuit.
  • Sound event detectors may also be implemented so that more than one type of sound event is detected in more than one input channel.
  • These “sound event detectors for multiple input channels and multiple sound event types” may produce a trigger signal for each sound event type in each input channel pair.
  • trigger signals in each channel pair may be combined in almost any manner to reduce the number of trigger signals.
  • An example of such a sound event detector for multiple input channels and multiple sound event types is shown in FIG. 27 and indicated by reference number 2700 . In this example, the entire sound field is either detected in or reproduced through a left and a right input channel.
  • the method may be implemented for any number and combination of input channels.
  • the sound event detector for multiple input channels and multiple sound event types is implemented to detect syllables and impulsive sounds. However, any number or combination of sound events may be detected.
  • the sound event detector for multiple input channels and multiple sound event types 2700 generally includes, a first accommodation circuit for a signal input channel 2702 ; a second accommodation circuit for a signal input channel 2706 ; a first sound event detector for a single channel and a short-duration sound event 2708 ; an alternate trigger generation circuit for a single input channel pair 2710 ; and a second sound event detector for a single channel and a short-duration sound event 2712 .
  • the first and second accommodation circuits for a single channel 2702 and 2703 are identical to each other.
  • the first accommodation circuit for a single channel 2702 produces a differential signal for the left input channel Lo.
  • the second accommodation circuit for a single channel 2703 produces a differential signal for right input channel Ro.
  • the alternative trigger generation circuit for a single channel pair 2718 uses Lo and Ro to produce a trigger signal that indicates the occurrence of syllables in the left-right channel pair Tlr(s).
  • the alternative trigger generation circuit for a single channel pair 2718 (shown in more detail in FIG. 25, indicated by reference number 2406 ) includes filters with a cut-off rise-time defined at about 33 ms.
  • the first and second sound event detector for a single channel and a single short-duration sound event 2708 and 2712 use Lin and Rin, respectively, to produce a trigger signal indicating the occurrence of impulsive sounds in the left and right input channels, respectively. These trigger signals may then be combined to produce a single trigger signal indicating the occurrence of trigger signals in the left-right input channel pair.
  • Both the first and second sound event detector for a single channel and a single short-duration sound event 2708 and 2712 respectively, include a high-pass filter (see 2303 in FIG. 23) with a cut-off rise-time of about 3 ms.
  • the sound event detector for multiple input channels and multiple sound event types includes a sound event detector for a single input channel and a single sound event implemented in parallel for each sound event type in each input channel. This sound event detector produces a trigger signal for each sound event in each input channel.
  • the sound event detector for multiple input channels and multiple sound event types may include a sound event detector for a single input channel a multiple sound event types implemented in parallel for each input channel. This sound event detector for multiple input channels and multiple sound event types also produces a trigger signal for each sound event in each input channel.
  • any of the sound event detectors for multiple input channels and multiple sound event types may also include a threshold adjustment circuit for each trigger generation circuit.
  • Sound event detectors and localizers The properties of the human hearing mechanism have been modeled to create electronic circuitry that detects and localizes sound events in the presence of a background signal (collectively, “sound event detectors and localizers”). Sound event detectors and localizers, like the sound event detection and localization methods, determine the direction of sound events in relation to one or more input channel pairs in terms of a differential steering angle or an ordinary steering angle and in some cases, verify the accuracy of the steering angles. Also, like the sound event detection and localization methods, the sound event detectors and localizers can be implemented to detect any number and combination of sound event types in a sound field generated in any number of input channels. In the following description, the sound event detectors and localizers are discussed in order of increasing complexity, with each subsequent sound detector and localizer incorporating the elements of the prior sound event detectors and localizers, except as indicated.
  • FIG. 28 One example of a sound event detector and localizer implemented to detect a single sound event type in a single input channel pair is shown in FIG. 28 (a “sound event detector and localizer for a single input channel pair and a single sound event type”).
  • sound events are detected and localized with respect to a right input channel and a left input channel.
  • the sound event detector and localizer for a single input channel pair and a single sound event type 2800 shown in FIG. 28 generally includes: a sound event detector for a single input channel and a single sound event type 2804 ; and a sound event localization circuit for a single input channel pair and a single sound event type 2806 .
  • the sound event detector for a single input channel pair and a single sound event type 2804 includes any of the sound event detectors for multiple input channels and a single sound event, which includes a trigger generation circuit for a single input channel pair and is implemented for whatever sound event is being detected.
  • the sound event detector for a single input channel pair and a single sound event type 2804 uses the left input signal Lin and a right input signal Rin to produce an differential signal for the left input channel Lo, a differential signal for the right input channel Ro, and a trigger signal indicating the occurrence in either input channel of whatever sound event is being detected Tlr.
  • the sound event localization circuit for a single input channel pair and a single sound event type 2806 then uses Lo, Ro and Tlr to produce a true differential steering angle indicating the direction of the detected sound events relative to the right and left input channel dlr′.
  • the sound event localization circuit for a single input channel pair and a single sound event type 2806 is shown in more detail in FIG. 29 and generally includes a DSA circuit 2904 , a switch 2518 ; a resistor 2906 ; a capacitor 2908 ; and a control circuit 2910 .
  • the DSA circuit 2904 uses the left accommodation signal Lo and the right accommodation signal Ro to determine the differential steering angle dlr.
  • the DSA circuit 2904 includes a first linear to decibel circuit 2912 ; a second linear to decibel circuit 2914 ; an operational amplifier 2916 and a decibel to equivalent angle circuit 2918 .
  • the first and second linear to decibel circuits 2912 and 2914 convert the left and right accommodation signals, respectively, from a power signal into a decibel signal.
  • the operational amplifier 2916 (or alternatively, any circuit that can determine a difference) determines the ratio between the left and right decibel signals by determining the difference between the two signals. This ratio is then converted into an equivalent angle by the decibel to equivalent angle circuit 2918 to produce the differential steering angle dlr.
  • the control circuit 2910 , the switch 2906 and the capacitor 2908 generally form a sample-and-hold circuit and can therefore be replaced with any device or circuit that performs a similar function.
  • the control circuit 2910 causes the switch 2906 to close and the capacitor 2908 to capture dlr during the rise time of the sound event to produce the true DSA (“dlr′”) for the typical duration of the sound event type being detected.
  • the control circuit 2910 receives the trigger signal Tlr and produces a control signal “con” that controls switch 2906 .
  • Switch 2906 is a two position switch and is generally in position C when no sound events are detected.
  • the switch 2906 When in position A, the switch 2906 is closed, when in position B, the switch is open, and when in position C, the switch 2906 is grounded.
  • the control circuit 2910 receives an indication from trigger signal Tlr that a sound event is occurring, it communicates to switch 2906 via “con” a command to close (go to position A). In response, the switch 2906 closes.
  • the control circuit 2910 communicates to switch 2906 via con a command to open (go to position B). After the typical duration of the sound event type being detected, the control circuit 2910 , communicates to switch 2906 a command via con to go to ground (go to position C).
  • the left-right differential steering angle dlr is captured by the capacitor 2908 to create the true DSA dlr′.
  • the true DSA is held until the end of the typical duration of the sound event, even if the sound event has not actually ended. For example, if the sound event being detected is a syllable, the true DSA will be held for about 50 ms to about 200 ms, preferably after about 150 ms. In another example, if the sound event being detected is an impulsive sound, the true DSA will be held for about 50 ms.
  • the sound event detector and localizer for a single input channel pair and a single sound event type is optimized for sound events with very short durations such as impulsive sounds.
  • this optimized sound event detector and localizer for a single input channel pair and a single sound event type further includes a circuit for determining the average DSA of all sound events detected in an about 3 ms time frame (the “DSA averaging circuit”).
  • the DSA averaging circuit is generally implemented in the sound event localization circuit 2806 between the DSA circuit 2804 and the switch 2906 .
  • the sound event detector and localizer may also be implemented to detect a single sound event type in a sound field generated in multiple input channel pairs (a “sound event detector and localizer for multiple input channel pairs and a single sound event type”).
  • a sound event detector and localizer for multiple input channel pairs and a single sound event type implemented to detect and localize a single sound event in both a right-left input channel pair (“LR input channel pair” and a center-surround input channel pair (“CS input channel pair” is shown in FIG. 30 and designated by reference number 3000 .
  • This detector and localizer may be implemented for any combination of input channel pairs with the LR input channel pair and the CS input channel pair used for only for the purposes of example.
  • the first and second sound event detectors for a single input channel pair and a single sound event type 3010 and 3012 are both implemented to detect the same sound event.
  • the first sound event detector for multiple input channels 3010 uses the input signals in the left and right input channels, Lin and Rin, respectively, to produce a left differential signal Lo, a right differential signal Ro, a left power envelope L 2 , a right power envelope R 2 , and a left-right trigger signal Tlr.
  • the second sound event detector for multiple input channels 3012 uses the input signals in the center and surround input channels, Cin and Rin, respectively, to produce a center differential signal Co, a surround differential signal So, a center power envelope C 2 , a surround power envelope S 2 , and a center-surround trigger signal Tcs.
  • the sound event localization circuit for multiple input channel pairs and a single sound event type 3014 uses the left differential signal Lo, the right differential signal Ro, and the left-right trigger signal to produce an angle indicating the direction of a detected sound event that equals either a true OSA or a true DSA for the left-right input channel pair (the “left-right sound event angle” or “d/lr′”), and the center differential signal Co, the surround differential signal So and the center-surround trigger signal Tcs to produce an angle indicating the direction of a detected sound event that equals either a true OSA or a true DSA for the center-surround input channel pair (the “center-surround sound event angle” or “d/cs′”).
  • circuit 3014 uses all the power envelopes, dlr and dcs to verify the accuracy of the DSA.
  • the sound event localization circuit for multiple input channel pairs and a single sound event type 2614 is shown in more detail in FIG. 31 and generally includes: a first OSA circuit 3102 ; a first DSA circuit 3104 ; a second DSA circuit 3106 ; a second OSA circuit 3108 ; a verification circuit 3116 ; a control circuit 3118 ; a first two-position switch 3110 ; a first three-position switch 3112 ; a first capacitor 3114 ; a second two-position switch 3120 ; a second three-position switch 3122 ; and a second capacitor 3124 .
  • d/lr′ is produced by the first OSA circuit 3102 , first DSA circuit 3104 , first two-position switch 3110 , first three-position switch 3112 and the first capacitor 3114 .
  • d/cs′ is produced by the second OSA circuit 3108 , second DSA circuit 3106 , second two-position switch 3120 , second three-position switch 3122 and the second capacitor 3124 .
  • the first and second OSA circuits 3102 and 3108 convert Lin and Rin and Cin and Sin, respectively, into ordinary steering angles, “lr” and “es” using known methods.
  • the control circuit 3118 produces a control signal “con” that is communicated to the three-way switches 3112 and 3122 .
  • the control signal causes the three-way switches 3112 and 3122 to move to (or stay in) position A whenever either trigger signal (Tlr or Tcs) indicates that a sound event is being detected in either input channel pair.
  • trigger signal Tlr or Tcs
  • con causes the three-way switches 3112 and 3122 to move to position B at the end of the typical duration of the rise-time of the sound event.
  • con causes the three-way switches 3112 and 3122 to move to position C at the end of the typical duration of the sound event type being detected.
  • the verification circuit determines whether at least two of the power envelopes (L 2 , R 2 , C 2 , S 2 ) have dropped by at least 3 dB or more. If at least two of the power envelopes have dropped by at least 3 dB or more from the last sound event, the verification circuit will communicate to both two-way switches 3110 and 3120 via vs causing them to move or stay in position E. However, if at least two of the power envelopes have not dropped by at least 3 dB or more, the verification circuit will communicate to both two-way switches 3110 and 3120 via vs causing the to move or stay in position D.
  • the verification circuit of the sound event localization circuit 3116 also includes a circuit for performing a consistency check (a “consistency check circuit”).
  • the consistency check circuit is coupled to both DSA circuits and uses the differential steering angles produced by each make a further determination of the accuracy of the differential steering angles.
  • the consistency check circuit uses a known circuit for determining the sum of the absolute values of dlr and dcs at the moment any trigger signal indicates that a sound event has been detected and then determines whether the sum is less than or equal to 45 degrees. If the sum is less than or equal to 45 degrees, and at least two of the power envelopes have not dropped by about 3 dB or more, the verification signal will communicate to both of the two-way switches 3110 and 3120 causing then to move to position D.
  • the error threshold circuit is also coupled to the threshold detector in the trigger generation circuits and further includes an error voltage source that is adjusted according to the number of errors counted by the counter. As the number of errors increases, the voltage produced by the error voltage source will increase to cause the threshold voltage to increase so that fewer sound events are detected.
  • the sound event detector and localizer for a single input channel pair and multiple sound event types may include a sound event detector for a single channel pair and a single sound event type implemented for each sound event type being detected and a sound event localization circuit for a single channel pair and a single sound event type that produces a differential steering angle indicating the direction of all types of sound events being detected.
  • a sound event detector and localizer for a single input channel pair and multiple sound event types that detects syllables and impulsive sounds is shown in FIG. 32.
  • the sound event detector and localizer for a single input channel pair and multiple sound event types includes: a sound event detector for a channel pair and a single sound event type 3202 ; a sound event detector for a single channel pair and a short-duration sound event type 3204 ; and a sound event localization circuit for a single channel pair and a single sound event type 3206 .
  • the sound event detector for a single channel pair and a single sound event type 3202 may include the sound event detector for multiple channels and a single sound event type shown in FIG.
  • the sound event detector for a single channel pair and a short-duration sound event type 3204 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG.
  • the sound event localization circuit for a single channel pair and a single sound event type 3206 may include the sound event localization circuit for a single channels pair and multiple sound event types as shown in FIG. 29 implemented to produce a left right differential steering angle indicating the direction of a detected syllable or impulsive sound whenever either trigger signal (Tlr(s) or Tlr(i)) indicates that a sound event is occurring.
  • this sound event detector and localizer is implemented to detect and localize syllables and impulsive sounds in a left-right input channel pair, it may be implemented to detect any number of sound event types in any channel pair simply by adding additional sound event detectors for a single channel pair implemented to detect the desired sound event type and having the sound event localization circuit for a single channel pair and a single sound event type respond to any of the trigger signals produced by the sound event detectors for a single channel pair.
  • the sound event detector and localizer may also be implemented to detect multiple sound event types in a sound field generated in multiple input channel pairs (a “sound event detector and localizer for multiple input channel pairs and multiple sound event types”).
  • a “sound event detector and localizer for multiple input channel pairs and multiple sound event types” An example of a sound event detector and localizer for multiple input channel pairs and multiple sound event types that detects syllables and impulsive sounds in both the left-right and center-surround channel pairs is shown in FIG. 33.
  • This sound event detector and localizer for multiple input channel pairs and multiple sound event types 3300 includes: a first and a second sound event detector for a single channel pair and a single short-duration sound event 3302 and 3308 , respectively; a first and a second sound event detector for a single channel pair and a single sound event type 3304 and 3306 , respectively; and a sound event localization circuit for multiple channel pairs and a single sound event type 3310 .
  • the first sound event detector for a single channel pair and a single sound event type 3304 may include the sound event detector for multiple channels and a single sound event type shown in FIG. 24 implemented to detect syllables and to produces a left differential signal Lo, a right differential signal Ro and a left-right trigger signal indicating the occurrence of syllables “Tlr(s).”
  • the second sound event detector for a single channel pair and a single sound event type 3306 may include the sound event detector for multiple channels and a single sound event type shown in FIG.
  • the first sound event detector for a single channel pair and a short-duration sound event type 3302 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG.
  • the second sound event detector for a single channel pair and a short-duration sound event type 3308 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG.
  • the sound event localization circuit for multiple channel pairs and a single sound event type 3310 may include the sound event localization circuit for multiple channel pairs and a single sound event type as shown in FIG.
  • this sound event detector and localizer is implemented to detect and localize syllables and impulsive sounds in left-right and center-surround input channel pairs, it may be implemented to detect any number of sound event types in number of channel pair simply by adding additional sound event detectors for a single channel pair implemented to detect the desired sound event types in the desired channel pairs and having the sound event localization circuit for multiple channel pairs and
  • any of the sound event detector and localizer for multiple input channel pairs and multiple sound event types may a dditionally include an accommodation adjustment circuit and/or an error threshold circuit.
  • Both the accommodation adjustment circuit and the error threshold circuit are coupled to the consistency check circuit and include a counter that counts the number of errors detected by the consistency check in a time period of about several seconds.
  • the accommodation adjustment circuit is also coupled to the accommodation signal in the transient detection circuits included in the DSA circuits.
  • the error threshold circuit is also coupled to the threshold detector in the trigger generation circuits and further includes an error voltage source that is adjusted according to the number of errors counted by the counter. As the number of errors increases, the voltage produced by the error voltage source will increase to cause the threshold voltage to increase so that fewer sound events are detected.
  • a stereo/surround detector determines whether a sound field is intended for reproduction in two input channels or in more than two input channels.
  • An example of a stereo/surround detector implemented to detect a single sound event type 3400 (a “stereo/surround detector for a single sound event type”) is shown in FIG. 34 and includes: a sound event detector and localizer for a single input channel pair and a single sound event type 3402 and a detector and counter 3404 .
  • the sound event detector and localizer for a single input channel pair and a single sound event type 3402 uses the signals in the center input channel Cin and the surround input channel Sin to produce a true differential steering angle dcs′ that reflects the direction of whatever sound event is being detected.
  • the threshold detector and counter 3404 determines the number of times dcs′ falls within the range of about 0 degrees to about ⁇ 45 degrees. If the number exceeds a predetermined value, the threshold detector and counter 3404 produces a signal sursig indicating that the sound field should be reproduced in surround. Conversely, if the number does not exceed a predetermined value, then the detector and counter 3404 produces a signal sursig indicating that the sound field should be reproduced in stereo.
  • the detector and counter 3404 will produce a signal sursig indicating that the sound field should be reproduced in surround. Additionally, the detector and counter may further determine the duration of the sound events and only count those with durations that exceed a predetermined value as sound events that are intended for reproduction in the rear. In one example, sound events with durations less than about 50 ms will not be counted as sound events that are intended to be reproduced in the rear. In another example, sound events with durations of about 200 ms to about 300 ms will be counted as sound events that are intended to be reproduced in the rear.
  • a stereo/surround detector implemented to detect a single sound event type (a “stereo/surround detector for multiple sound event types”) is shown in FIG. 35 and is indicated by reference number 3500 .
  • the stereo/surround detector 3500 is implemented to count the number of syllables and impulsive functions intended for reproduction in the rear. However, it may be implemented to count the number of any type and combination of types of sound events.
  • the detector 3500 includes: first and second sound event detector and localizers for a single input channel pair and a single sound event type 3502 and 3504 and a detector and counter 3506 .
  • the first sound event detector and localizer for a single input channel pair 3502 uses the signals in the center input channel Cin and the surround input channel Sin to produce a true differential steering angle that reflects the direction of the syllables dcs′ (s).
  • the second sound event detector and localizer for a single input channel pair 3504 uses the signals in the center input channel Cin and the surround input channel Sin to produce a true differential steering angle that reflects the direction of the impulsive sounds dcs′(i).
  • the detector and counter 3506 determines the number of then determines the number of times dcs′(s,i) falls within the range of about 0 degrees to about ⁇ 45 degrees to produce a signal sursig(s,i) that indicates whether the sound is to be reproduced in stereo or in surround.
  • the sound localizers can be implemented to detect any number and combination of sound event types in a sound field generated in any number of input channels.
  • the sound localizers are discussed in order of increasing complexity, with each subsequent sound localizer incorporating the elements of the prior sound localizers, except as indicated.
  • the sound localizer for a single input channel pair and a single sound event type 3600 includes: a sound event detector for a single channel pair and a single sound event type 3602 ; and a sound localization circuit for a single input channel pair and a single sound event 3604 .
  • the sound event detector for a single channel pair and a single sound event type 3602 may include the sound event detector for multiple channels and a single sound event type shown in FIG.
  • the sound localization circuit for a single input channel pair and a single sound event type 3604 is shown in more detail in FIG. 37 and includes: an ordinary steering angle (OSA) circuit 3702 ; a DSA circuit 3704 ; a control circuit 3706 ; a first switch 3708 ; a resistor 3710 ; a second switch 3712 and a capacitor 3714 .
  • the sound event localization circuit for a single input channel pair and a single sound event type 3604 uses the left and right differential signals Lo and Ro, respectively, and the left and right input signals Lin and Rin, respectively, to produce a left-right comprehensive steering angle clr′.
  • clr′ equals follows the OSA when no sound events are detected and follows the DSA whenever a sound event is detected.
  • the OSA circuit 3702 uses Lin and Rin to determine the ordinary steering angle lr.
  • the DSA circuit 3704 uses the differential signals Lo and Ro to produce the differential steering angle.
  • the control circuit 3706 , the first switch 3708 , the resistor 3710 ; the second switch 3712 , and the capacitor 3714 generally form a sample-and-hold circuit and can therefore be replaced with any device or circuit that performs a similar function.
  • the control circuit 3706 causes the first switch 3708 to close and the capacitor 3712 to capture dlr during the typical rise time of the sound event type being detected so that clr′ equals the true DSA for the typical duration of the sound event type being detected. More specifically, the control circuit 3716 receives the trigger signal Tlr and produces a control signal “con” that controls the first switch 3708 and the second switch 3712 .
  • the left-right differential steering angle dlr is captured by the capacitor 3714 to create the true DSA dlr′.
  • Clr′ is defined by the true DSA, which is held until the end of the typical duration of the sound event, even if the sound event has not actually ended. For example, if the sound event being detected is a syllable, the true DSA will be held for about 50 ms to about 200 ms, preferably after about 150 ms. In another example, if the sound event being detected is an impulsive sound, the true DSA will be held for about 50 ms.
  • the sound localization circuit can be specifically implemented for short-duration sound events.
  • a short-duration sound event such as an impulsive sound
  • An example of a sound event localization circuit that includes this functionality (referred to in this application as a “sound localization circuit for a single input channel pair and short-duration sound events”) is shown in FIG. 38.
  • the sound localization circuit for a single input channel pair and short-duration sound events 3800 includes: an OSA circuit 3802 ; a DSA circuit 3804 ; a control circuit 3808 ; a detector 3810 ; a first switch 3812 ; a second switch 3816 ; a first resistor 3814 ; a third switch 3813 ; a second resistor 3815 ; and a capacitor 3818 .
  • This sound localization circuit 3800 produces a left-right comprehensive steering angle clr′. When no sound events are detected, clr′ equals the filtered OSA (which is lr after it is filtered by resistor 3814 and capacitor 3816 ).
  • clr′ when a sound event is detected, clr′ either equals the true OSA or the true DSA.
  • the sound localization circuit for a single channel pair and a single short duration sound event generally behaves in the same way as the sound localization circuit shown in FIG. 37, except that at the end of a sound event, clr′ either decays to lr or goes to lr immediately depending on whether the sound event has actually ended.
  • the control circuit 3808 After the control circuit 3808 receives a trigger signal indicating that a sound event is occurring and has determined that the typical duration of the sound event being detected has ended, in addition to its other functions as previously described, it communicates with the detector 3810 to establish whether the sound event has actually ended.
  • the detector 3810 determines whether a sound event has actually ended by comparing the power envelope with the accommodation signal in each input channel of the channel pair. If the input power envelope is greater than the accommodation signal in any input channel of the input channel pair, the detector 3810 will determine that the short-duration sound event has not actually ended. Conversely, if the input power envelope is not greater than the accommodation signal in any input channel of the input channel pair, the detector 3810 will determine that the short-duration sound event has actually ended.
  • control circuit 3808 If the control circuit 3808 establishes that the sound event has not actually ended, it will then instruct the second switch 3816 to close via con. When the second switch 3816 closes, clr′ will drift to the filtered OSA at a specified rate. Therefore, capacitor 3818 and resistor 3814 are select so that their RC time constant is about equal to the specified rate (which is generally about 300 ms). For example, if impulsive sounds are being detected, the RC time constant of capacitor 3818 and resistor 3814 will be about 5 ms. However, if the control circuit 3808 establishes that the sound event has actually ended, it will then instruct the third switch 3813 to close via con. When the third switch 3813 closes, clr′ immediately goes to lr.
  • the first and second sound event detectors for a single input channel and a single sound event type 3902 and 3904 may include a sound event detector for multiple input channels and a single sound event type, such as that shown in FIG. 24, implemented to detect the same sound event.
  • the first sound event detector for a single input channel 3902 uses Lin′ and Rin′ to produce a left differential signal Lo, a right differential signal Ro, and a left-right trigger signal Tlr.
  • the second sound event detector for a single input channel 3904 uses Cin′ and Rin′ to produce a center differential signal Co, a surround differential signal So, and a center-surround trigger signal Tcs.
  • the sound localization circuit for multiple input channel pairs and a single sound event type 3906 uses the left differential signal Lo, the right differential signal Ro, and the left-right trigger signal to produce a left-right comprehensive steering angle clr′, and uses the center differential signal Co, the surround differential signal So and the center-surround trigger signal Tcs to produce a center-surround comprehensive steering angle ces′. Additionally, circuit 3906 uses all the power envelopes, dlr and dcs to verify the accuracy of the DSA. This circuit 3906 is shown in more detail in FIG.
  • a first OSA circuit 4002 and generally includes: a first OSA circuit 4002 ; a first DSA circuit 4004 ; a second DSA circuit 4006 ; a second OSA circuit 4008 ; a verification circuit 4020 ; a control circuit 4022 ; a first two-position switch 4010 ; a first switch 4012 ; a first resistor 4014 ; a first capacitor 4018 ; a second switch 4014 ; a second two-position switch 4030 ; a third switch 4032 ; a second capacitor 4038 ; a fourth switch 4036 ; and a second resistor 4034 .
  • clr′ is produced by the first OSA circuit 4002 , first DSA circuit 4004 , first two-position switch 4010 , first switch 4012 , the first resistor 4014 , the second switch 4016 , and the first capacitor 3114 .
  • d/cs′ is produced by the second OSA circuit 4008 , second DSA circuit 4006 , second two-position switch 4030 , the third switch 4032 , the second resistor 4034 and the second capacitor 4038 .
  • the first and second OSA circuits 4002 and 4008 convert Lin and Rin and Cin and Sin, respectively, into ordinary steering angles, lr and cs.
  • the control circuit 4022 , the first switch 4012 , the second switch 4016 , and the first capacitor 4018 form a first sample and hold circuit
  • the control circuit 4022 , the third switch 4032 , the fourth switch 4034 , the second resistor 4034 , and the second capacitor 4038 form a second sample and hold circuit.
  • Both the first and third switches 4012 and 4032 are normally open when no sound events are detected and are controlled by the control circuit so that each switch is closed when a sound event is detected and opened at the end of the typical rise time of the sound event type being detected.
  • Both the second and forth switches 4016 and 4036 are normally closed when no sound events are detected and opened after a sound event has been detected and the typical rise-time of the sound event type being detected has ended.
  • the control circuit 4022 produces a control signal “con” that is communicated to all the switches 4012 , 4016 , 4032 and 4036 .
  • the control signal causes the first and second switches 4012 and 4016 to close (or stay closed) whenever either trigger signal (Tlr or Tcs) indicates that a sound event is being detected in either input channel pair. Subsequently, con causes all the switches 4012 , 4016 , 4032 and 4036 to open at the end of the typical duration of the rise-time of the sound event.
  • Both two-way switches 4010 and 4030 include positions D and E.
  • the DSA for each channel pair are used to indicate the direction of the sound event.
  • the OSA for each channel pair is used to indicate the position of the sound event.
  • the verification circuit 4020 controls both two-way switches 4010 and 4020 via a verification signal “vs” according to whether the DSAs are correct when a sound event is detected.
  • the verification circuit determines whether at least two of the power envelopes (L 2 , R 2 , C 2 , S 2 ) have dropped by at least 3 dB or more. If at least two of the power envelopes have dropped by at least 3 dB or more from the last sound event, the verification circuit will communicate to both two-way switches 4010 and 4020 via vs causing them to move or stay in position E. However, if at least two of the power envelopes have not dropped by at least 3 dB or more, the verification circuit will communicate to both two-way switches 4010 and 4020 via vs causing the to move or stay in position D.
  • the verification circuit of the sound event localization circuit 4020 also includes a consistency check circuit.
  • the consistency check circuit is coupled to both DSA circuits and uses the differential steering angles produced by each to make a further determination of the accuracy of the differential steering angles as previously described.
  • the sound localizer for multiple sound event types and a single sound event type may further include an accommodation adjustment circuit and/or an error threshold circuit.
  • the sound localization circuit for multiple channel pairs and a single sound event type can be specifically implemented for short-duration sound events, such as impulsive sounds.
  • short-duration sound events such as impulsive sounds.
  • FIG. 41 An example of such a sound localization circuit for multiple input channel pairs and short-duration sound events is shown in FIG. 41 and indicated by reference number 4100 .
  • This sound localization circuit for multiple input channel pairs and short-duration sound events 4100 generally includes: a first OSA circuit 4102 ; a first DSA circuit 4104 ; a second DSA circuit 4106 ; a second OSA circuit 4108 ; a verification circuit 4120 ; a control circuit 4122 ; a detector circuit 4124 ; a first two-position switch 4110 ; a first switch 4112 ; a first resistor 4114 ; a first capacitor 4118 ; a second switch 4116 ; a second two-position switch 4130 ; a third switch 4132 ; a second capacitor 4138 ; a fourth switch 4136 ; a second resistor 4134 ; a fifth switch 4113 ; a third resistor 4115 ; a sixth switch 4133 ; and a fourth resistor 4135 .
  • This sound localization circuit 4100 produces a left-right comprehensive steering angle. When there are no sound events, clr′ will equal the filtered OSA (which is lr after being filtered by first resistor 4114 and first capacitor 4118 ). When there is a sound event detected, clr′ will equal either the true OSA or the true DSA of the left-right channel pair.
  • This sound localization circuit 4100 also produces a center-surround comprehensive steering ccs′. When no sound events are detected, clr′ equals the filtered OSA. When sound events are detected, clr′ will equal the true OSA or the true DSA of the center-surround channel pair.
  • the sound localization circuit for multiple channel pairs and a single short duration sound event generally behaves in the same way as the sound localization circuit for multiple channel pairs an a single sound event type shown in FIG. 40, except that at the end of a sound event, clr′ either decays to the filtered OSA or goes to the OSA immediately depending on whether the sound event has actually ended.
  • the control circuit 4122 After the control circuit 4122 receives a trigger signal indicating that a sound event is occurring and has determined that the typical duration of the sound event being detected has ended, in addition to its other functions as previously described, it communicates with the detector 4124 to establish whether the sound event has actually ended.
  • the detector 4122 determines whether a sound event has actually ended by comparing the power envelope with the accommodation signal in each input channel of the channel pair. If the input power envelope is greater than the accommodation signal in any input channel of the input channel pair, the detector 4122 will determine that the short-duration sound event has not actually ended. Conversely, if the input power envelope is not greater than the accommodation signal in any input channel of the input channel pair, the detector 4124 will determine that the short-duration sound event has actually ended.
  • control circuit 4122 If the control circuit 4122 establishes that the sound event has not actually ended, it will then instruct the second and fourth switches 4116 and 4136 , respectively, to close via con.
  • the second switch 4116 and the fourth switch 4136 close, clr′ will drift to the left-right filtered OSA (which is lr after it has been filtered by the first resistor 4114 and the first capacitor 4118 ) and ccs′ will drift to the center-surround filtered OSA (which is cs after it has been filtered by the second resistor 4136 and the second capacitor 4138 ) at a specified rate.
  • the first and second capacitors 4118 and 4138 , respectively, and the first and second resistors 4114 and 4134 , respectively, are select so that their RC time constant is about equal to the specified rate. For example, if impulsive sounds are being detected, the RC time constant of capacitor 4118 and resistor 4114 , as well as the RC time constant of capacitor 4138 and resistor 4134 will be about 5 ms.
  • control circuit 4122 establishes that the sound event has actually ended, it will then instruct the fifth and sixth switches 4113 and 4133 , respectively, to close via con.
  • fifth and sixth switches 4113 and 4133 respectively, close, clr′ and ccs′ both immediately goes to lr and cs, respectively. Therefore, the RC time constant of the third resistor 4115 and the first capacitor 4118 , and the RC time constant of the fourth resistor 4135 and the second capacitor 4138 will both be very low.
  • RC time constants may be about at least a factor of ten lower than the RC time constant of the first resistor 4114 and the first capacitor 4118 and the RC time constant of the second resistor 4134 and the second capacitor 4138 .
  • the fifth and sixth switches 4113 and 4133 will remain closed for a short time. This short time may be about 3 ms to about 10 ms.
  • the control circuit 4122 will instruct the fifth and sixth switches 4113 and 4135 , respectively, to open and the second and fourth switches 4166 and 4136 , respectively, to close so that clr′ and ccs′ will reflect the filtered OSAs.
  • the sound localizer for a multiple input channel pair and a short-duration sound event type may further include first and second DSA averaging circuits.
  • the first DSA averaging circuit may be implemented in the sound event localization circuit 4100 between the first DSA circuit 4104 and the first two-position switch 4110 .
  • the second DSA averaging circuit may be implemented in the sound event localization circuit 4100 between the second DSA circuit 4106 and the second two-position switch 4130 .
  • the verification circuit of the sound localization circuit for multiple channel pairs and a single short-duration sound event type 4100 also includes a consistency check circuit.
  • the consistency check circuit is coupled to both DSA circuits and uses the differential steering angles produced by each to make a further determination of the accuracy of the differential steering angles as previously described.
  • the sound localizer for multiple sound event types and a single sound event type may further include an accommodation adjustment circuit and/or an error threshold circuit.
  • the sound localizer may also be implemented to detect multiple sound event types in a sound field generated in a single input channel pair (a “sound localizer for a single input channel pair and multiple sound event types”).
  • a sound localizer for a single input channel pair and multiple sound event types An example of a sound event detector and localizer for a single input channel pair and a multiple sound event types implemented to detect and localize syllables and impulsive sounds in a left-right input channel pair is shown in FIG. 42 and indicated by reference number 4200 (however, this localizer may be implemented for any input channel pair to detect any combination of sound event types).
  • the sound localizer for a single input channel pair and multiple sound event types 4200 generally includes: a sound event detector for a single input channel and a single sound event type 4202 ; a sound event detector for a single-channel and a single short-duration sound event device 4206 ; and a sound localization circuit for a single input channel pair and multiple sound event types 4204 .
  • the sound event detector for a single channel pair and a single sound event type 4202 may include the sound event detector for multiple channels and a single sound event type shown in FIG. 24 implemented to detect syllables and to produces a left differential signal Lo, a right differential signal Ro and a left-right trigger signal indicating the occurrence of syllables Tlr(s).
  • the sound event detector for a single channel pair and a short-duration sound event type 4206 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG.
  • the sound localization circuit for a single channel pair and a single sound event type 4204 may include the sound localization circuit for a single channels pair and multiple sound event types as shown in FIG. 37 implemented to produce a left-right differential steering angle indicating the direction of a detected syllable or impulsive sound whenever either trigger signal (Tlr(s) or Tlr(i)) indicates that a sound event is occurring.
  • this sound localizer is implemented to detect and localize syllables and impulsive sounds in a left-right input channel pair, it may be implemented to detect any number of sound event types in any channel pair simply by adding additional sound event detectors for a single channel pair to implemented to detect the desired sound event type and having the sound localization circuit for a single channel pair and a single sound event type respond to any of the trigger signals produced by the sound event detectors for a single channel pair.
  • the sound localizer may also be implemented to detect multiple sound event types in a sound field generated in multiple input channel pairs (a “sound localizer for multiple input channel pairs and multiple sound event types”).
  • a sound localizer for multiple input channel pairs and multiple sound event types specifically localizes syllables and impulsive sounds in a left-right input channel pair and a center-surround input channel pair is shown in FIG. 43 (however, the sound localizer for multiple input channel pairs and multiple sound event types may be implemented to specifically localize any combination of sound events in any number of input channel pairs).
  • This sound localizer for multiple input channel pairs and multiple sound event types 4300 includes: a first and a second sound event detector for a single channel pair and a single short-duration sound event 4308 and 4306 , respectively; a first and a second sound event detector for a single channel pair and a single sound event type 4302 and 4304 , respectively; and a sound event localization circuit for multiple channel pairs and a single sound event type 4310 .
  • the first sound event detector for a single channel pair and a single sound event type 4302 may include the sound event detector for multiple channels and a single sound event type shown in FIG. 24 implemented to detect syllables and to produces a left differential signal Lo, a right differential signal Ro and a left-right trigger signal indicating the occurrence of syllables Tlr(s).
  • the second sound event detector for a single channel pair and a single sound event type 4304 may include the sound event detector for multiple channels and a single sound event type shown in FIG. 24 implemented to detect syllables and to produces a center differential signal Co, a surround differential signal So and a center-surround trigger signal indicating the occurrence of syllables Tcs(s).
  • the first sound event detector for a single channel pair and a short-duration sound event type 4308 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG. 23 implemented to detect impulsive sounds and to produce a left-right trigger signal (as a combination of a left trigger signal and a right trigger signal) indicating the occurrence of impulsive sounds Tlr(i).
  • the second sound event detector for a single channel pair and a short-duration sound event type 4306 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG.
  • the sound localization circuit for multiple channel pairs and a single sound event type 4310 may include the sound localization circuit for multiple channel pairs and a single sound event type as shown in FIG.
  • this sound localizer is implemented to detect and localize syllables and impulsive sounds in left-right and center-surround input channel pairs, it may be implemented to detect any number of sound event types in number of channel pair simply by adding additional sound event detectors for a single channel pair implemented to detect the desired sound event types in the desired channel pairs and having the sound localization circuit for multiple channel pairs and a single sound event type 4310
  • any of the sound localizers for multiple input channel pairs and multiple sound event types may additionally include an accommodation adjustment circuit and/or an error threshold circuit.
  • Both the accommodation adjustment circuit and the error threshold circuit are coupled to the consistency check circuit and include a counter that counts the number of errors detected by the consistency check in a time period of about several seconds.
  • the accommodation adjustment circuit is also coupled to the accommodation signal in the transient detection circuits included in the DSA circuits.
  • the error threshold circuit is also coupled to the threshold detector in the trigger generation circuits and further includes an error voltage source that is adjusted according to the number of errors counted by the counter. As the number of errors increases, the voltage produced by the error voltage source will increase to cause the threshold voltage to increase so that fewer sound events are detected.
  • Implementations of the sound event detection methods, sound event detection and localization methods and sound localization methods and any methods included in any of these methods include computer readable software code. These algorithms may be implemented together or independently. Such code may be stored on a processor, a memory device or on any other computer readable storage medium. Alternatively, the software code may be encoded in a computer readable electronic or optical signal. The code may be object code or any other code describing or controlling the functionality described in this application.
  • the computer readable storage medium may be a magnetic storage disk such as a floppy disk, an optical disk such as a CD-ROM, semiconductor memory or any other physical object storing program code or associated data.

Abstract

The beginning detection, accommodation and frequency bias properties of the human hearing mechanism have been modeled to create systems that can detect directional transients (“sound events”) in a sound field and separately localize them. These models break down a sound field into sound events and non-sound events and separately localize the sound events and non-sound events. Sound events are generally identified according to the frequency bias and beginning detection properties. Once detected, sound events are generally localized according to differential steering angles (steering angles to which the steady-state signals have been accommodated) or ordinary steering angles and which reflect the direction of a sound event indicated in its rise-time. When no sound events are detected, non-sound events are localized according to a steering angle that does not reflect rapid motion.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application is based on copending U.S. Provisional Application Serial No. 60/377,558, entitled “Adaptive Circuit for Determining the Direction and Timing of Sound Events Embedded in a Complex Sound Field.” The benefit of the filing date of the Provisional Application is claimed for this application.[0001]
  • BACKGROUND OF THE INVENTION
  • 1. Technical Field [0002]
  • The invention relates generally to detecting and localizing sound. More particularly, this invention relates to detecting and/or localizing sounds that include sound events in a complex sound field [0003]
  • 2. Related Art [0004]
  • Many sound-based applications, such as audio reproduction systems, audio and speech coding systems, speech recognition systems, and audio amplification systems require the ability to distinguish and detect certain types of sounds and to determine the directions from which sounds emanate or originate. The ability to detect certain types of sounds is important to applications such as sound amplification, while the ability to detect and localize sounds is crucial to applications involving sound reproduction. Unfortunately, the detection and localization of sounds can be very complicated because individual sounds rarely occur by themselves. [0005]
  • Whether live or reproduced, sound generally consists of a complicated combination of many different sounds occurring over time to form a complex sound field. The sounds in a sound field can overlap, occur one after the other, or in any combination. One way in which the individual sounds in a sound field are classified is according to whether a sound has emanated or originated from a particular location. Sounds that can be detected as emanating or originating from a particular direction are referred to as directional sounds, while sounds detected as emanating or originating from no particular direction at all are referred to as non-directional sounds. Another way of classifying sounds is according to whether a sound is a transient or a steady-state sound. Steady-state sounds are those that have a generally constant level of power over time, such as a sustained musical note. Steady-state sounds can be either directional or non-directional sounds. Transient sounds (or “transients”) are sounds that have an initial energy spike, such as a shout or a drum hit. Transients can also be either directional or non-directional sounds. An example of a non-directional transient sound is speech in a reverberant space where the direct sound is blocked by an object. In this case if the reverberation time is less than one second the time characteristics of the signal are preserved, but information about its direction is lost. [0006]
  • Directional transients are referred to in this application collectively as “sound events.” Two types of sound events are syllables and impulsive sounds. Syllables include phonemes and notes. Phonemes are transient sounds that are characteristic of phones in human speech and can be particularly useful in detecting and localizing syllables in human speech. Notes are the individual notes created by a musical instrument. Because notes and phonemes have common characteristics, they are collectively referred to in this application as “syllables.” Syllables, generally have the following characteristics: a finite duration of about at least 50 ms up to about 200 ms, but typically around 150 ms; rise times of about 33 ms; generally occur no more frequently than about once every 0.2 ms to about 0.5 ms; and may have low or high volume (amplitude). In contrast, impulsive sounds are transients of very short duration such as a drum hit or fricatives and explosives in speech. Impulsive sounds generally have the following characteristics: a short duration of about 5 ms to about 50 ms, rise times of about 1 ms to about 10 ms, and high volume. [0007]
  • To detect sounds in a sound field, whether generated live or as a reproduction, generally the sound field need only be generated in one input or “input channel.” However, to localize sounds, generally the sound field needs to be generated in at least two inputs or input channels. The archetype for sound localization is natural hearing, where the azimuth of the sound is detected primarily by the arrival time difference between the two input channels represented by the two ears. When localizing sounds, electronically, the azimuth of a sound source is determined primarily by the amplitude and phase relationships between the signals generated by two or more input channels. Generally, in order to describe the azimuth of the directional sounds at these input channels, the directions of the source of these sounds are described in terms of an angle between each corresponding pair of channels (each a “channel pair”). If sounds are generated in only two channels, the directions of the sounds are given in terms of an angle for that channel pair, generally a left/right angle “lr.” If the sounds are generated in four channels, the directions of the sounds are given in terms of an angle for each channel pair, generally, a left/right angle lr, and a front/back angle “cs.” For example, when the sound field is generated in two channels, the average direction of the sounds in the sound field is given in terms of lr only. In this case, the value for lr ranges from about −45 degrees to about 45 degrees, with −45 degrees indicating that the sound field originates from the left input channel, 45 degrees indicating that the sound field o riginates from the right input channel and 0 degrees indicating that the sound field originates from a position in the middle precisely between the right and left input channels (a position often referred to as “center”). [0008]
  • However, when the sound field is generated in two channel pairs, such as in a surround sound system, a second directional component is specified. Additionally, even if the sound field is generated in only one channel pair, a second directional component may also be specified because it is often possible to derive an additional channel pair from the one channel pair. In addition to specifying the direction in terms of lr, the direction is specified in terms of cs. The value for cs also ranges from about −45 degrees to about 45 degrees, with lr=0 and cs=45 degrees indicating that the sound field originates from the center input channel only, and lr=0 and cs=−45 indicating that the sound field originates from the rear input channel only. Similarly, lr=−45 degrees and cs=0 indicates a sound originating from the left and lr=45 and cs=0 indicating that the sound field originates from the right. Additionally, lr=−22.5 degrees and cs=−22.5 degrees indicates that the sound field originates from the left rear and lr=22.5 and cs=−22.5 indicates that the sound field originates from the right rear. [0009]
  • One known technique for determining these angles is used in reproducing recorded sound. In general, this known technique determines the intended direction of sounds by comparing the amplitudes of the signals in one input channel of a input channel pair with the signals in the corresponding input channel of the input channel pair (generally, the left with the right, and the center with the surround). More specifically, this ratio of amplitudes is used to determine what is generally referred to an “ordinary steering angle” or “OSA” for each input channel pair. To obtain the OSA, the voltage signals in each input channel of an input channel pair are rectified and the logarithms of the rectified voltages are taken. By subtracting the logarithm of the rectified voltage of one input channel from the log of the rectified voltage of the other input channel in the input channel pair, a signal is produced that equals the logarithm of the ratio of the voltages in the input channel pair which, when converted back into the magnitude domain, is the ordinary steering angle. In surround reproduction systems, this determination is often made by a device called a matrix decoder. [0010]
  • Unfortunately, this known technique treats the entire sound field as if it contained only a single sound because it determines the direction of the entire sound field according to the relative voltage strength in each input channel. Therefore, many directional individual sounds will not be properly localized. In order to treat the sound field as a complex combination of many sounds, attempts have been made to devise filters that will separate the directional transient sounds (sound events) so that their directions can be independently determined. However, a fundamental problem is encountered when designing such a filter. If the filter is made fast enough to distinguish the fluctuations of all directional transient signals, it will also distinguish fluctuations characteristic of non-directional transient signals such as reverberation and noise. As a result, the rapid fluctuations of reverberation and noise are reproduced as directional changes in the sound, which severely degrades the quality of the reproduced sound. On the other hand, if the filter is made slow enough not to distinguish the fluctuations characteristic of the non-directional signals, the filter is generally too slow to distinguish the fluctuations of certain sound events, particularly impulsive sounds. As a result, many sound events are not properly localized. No matter how these filters are designed, they generally work well on only one type of music but not on all. For example, the fast filter will work well on complex popular music, which is full of rapid changes, but will reflect false directional changes (steer too greatly) when a highly-reverberant classical piece is reproduced. [0011]
  • Additional problems arise when sounds are recorded in a given number of input channels and then reproduced over a different number of channels. For example, two common classes of sound recording and reproduction techniques are stereo and surround. Sounds recorded for reproduction in stereo (two channels) are intended to be perceived as originating only from the front. Sounds recorded for reproduction in surround (any number of input channels greater than two, but generally five or seven channels) are intended to be perceived as originating from all around, generally with one or two input channels used to reproduce sounds from the rear. The techniques used to record sounds intended for reproduction in stereo are generally different from those used to record sounds intended for reproduction in surround. However, because surround systems are not universally used, sounds recorded for reproduction in surround generally need to be capable of high-quality reproduction in stereo. For example, in a typical five channel surround system, the sounds in the center channel are encoded into the right and left input channels so that the sounds included in the center channel “c” equal the sum of the sounds included in the left and right input channels (c=1+r). Similarly, the sound in the surround channel are encoded into the left and right input channels so that the sounds included in the surround channel “s” equal the difference between the sounds included in the left and right input channels (s=1−r). In another example, the Dolby Surround® system, which records sounds for reproduction in surround, adds a negative phase to the sounds intended for reproduction from behind the listener (the rear). This negative phase is generally undetected by stereo reproduction systems and is transparent to the listener. However, the negative phase is detected by a surround reproduction system that then reproduces the associated sounds in the rear input channels. Unfortunately, many sounds naturally have negative phase, even when recorded in stereo format, and are therefore incorrectly reproduced in the rear input channels by a surround reproduction system. This can be distracting and unnatural. [0012]
  • SUMMARY
  • A need exists for a detector that can distinguish sound events from the remainder of the sound field, a localizer that can separately determine the directions of the detected sound events, and a sound localizer that can separately determine the directions of sound event as well as the remaining sounds in a sound field. Properties of the human hearing mechanism have been modeled and used to create systems with these features because many of the problems encountered by known sound detection and systems are not experienced by the human hearing mechanism. [0013]
  • The human hearing mechanism localizes sounds in a sound field by breaking down the sound field into individual sounds, determining the direction of the individual sounds, reassembling related sounds into streams (such as the melody line of a particular instrument or a line of dialogue from a particular speaker), and determining the direction of the stream based on the average direction of the individual sounds within the stream. In breaking down a sound field into individual sounds, the human hearing mechanism has the ability to distinguish transient sounds from other sounds and to distinguish sound events from non-directional transients. [0014]
  • Furthermore, the human hearing mechanism can distinguish transient from non-transient sounds and sound events from other transient sounds even in the presence of a background signal by relying on the properties of beginning detection and accommodation. Beginning detection results from human neurology which is highly evolved to determine the starting points and end points of sounds and makes the human hearing mechanism particularly sensitive to the rise and fall times of sounds. Accommodation is the property that allows the human hearing mechanism to distinguish transients from steady-state sounds by gradually ignoring the presence of the steady state sounds. Additionally, for purposes of localization, the human hearing mechanism relies more heavily on certain frequencies and the apparent direction of the beginnings of sounds. The human hearing mechanism relies most heavily on frequencies between about 500 Hz and about 4000 Hz for direction determination. This frequency bias is largely a result of the frequency response of the external ear (the pinnae, concha and ear canal) and by the frequency transfer function of the middle ear. The human hearing mechanism also relies on the direction indicated in the rise time of a sound more than the direction as indicated in other portions of the sound. This reliance is advantageous because the direction indicated in the rise time of a sound is less likely to be corrupted by sound reflections or reverberations even in a highly reflective environment. [0015]
  • After detecting the directions of the individual sounds, the human hearing mechanism reassembles related sounds into streams and separately determines the direction of each stream. The direction of each stream is generally the average direction of all sounds within the stream. In any case, the process of reassembling related sounds into streams and determining the direction of each stream is performed unconsciously and automatically by the human hearing mechanism. [0016]
  • The properties of the human hearing mechanism have been modeled to create sound event detection methods which produce a signal indicating the occurrence of sound events in a sound field. These sound event detection methods use the property of beginning detection to detect the occurrence of sound events. The beginning detection property is modeled to identify the onsets typical of sound events and uses these onsets to generate the trigger signal. Additionally, the sound event detection methods may use the accommodation property to improve the accuracy of the trigger signal. The accommodation property is modeled to reduce over time the effect any steady-states sounds may have on the sound field. These effects are reduced in the sound field to produce a differential signal. This differential signal may then be used by the beginning detection model to produce an improved trigger signal. The sound event detection methods may also include emphasizing the frequencies important to the human hearing mechanism. [0017]
  • The properties of the human hearing mechanism have also been used to create sound event detection and localization methods for determining the location of sound events that occur in the presence of a background signal. These sound event detection and localization methods produce at least one steering angle indicating the direction of the sound event whenever a trigger signal indicates that a sound event has occurred (“differential steering angles”). Sound event detection and localization methods use differential signals to determine the differential steering angles by comparing the relative power of the sound event in each input channel of each input channel pair after any steady-state sounds have been removed. By comparing the power in each input channel of an input channel pair instead of the voltage in each input channel of an input channel pair as done in known methods, individual sound events can be identified and separately localized. [0018]
  • The properties of the human hearing mechanism may also be used to determine the location of all sounds within a sound field. By combining the sound event detection and localization methods with known methods for determining the intended direction of sounds, sound localization methods have been developed that localize the individual sounds in the presence of background sounds for any combination of sounds more accurately. The sound localization methods divide the sound field into sound events and non-sound events and localize the sound events in terms of a true differential steering angle or a true ordinary steering angle and the non-sound events in terms of a filtered ordinary steering angle. These methods essentially treat the non-sound events as a separate sound for which the direction is separately determined. The sound localization methods, as well as the sound event detection and localization methods and the sound event detection methods, can be implemented for any number and combination of sound event types in a sound field generated in any number of input channel pairs. [0019]
  • The properties of the human hearing mechanism have been modeled to create electronic circuitry that detects sound events in the presence of a background signal. These sound event detectors may be used for a variety of applications for which the detection of specific types of sound events is helpful. For example, a sound event detector that detects syllables may be used as part of a phonic detector or as part of a speech recognition or speech coding system. In another example, a sound event detector that detects syllables may be used in conjunction with an audio amplification device, such as a microphone. This allows the microphone to remain off until a syllable is detected from a speaker, thus preventing the microphone from amplifying undesired sounds and feedback through the microphone itself when the speaker is silent. Sound event detectors, like the sound event detection methods, are based on models of the beginning detection property and the accommodation property and generally produce a trigger signal that includes some flag or marker, such as a pulse, that indicates the occurrence of a sound event. [0020]
  • The properties of the human hearing mechanism have been modeled to create electronic circuitry that detects and localizes sound events in the presence of a background signal. These sound event detectors and localizers, like the sound event detection and localization methods, determine the direction of sound events in relation to one or more input channel pairs in terms of a differential steering angle or an ordinary steering angle and in some cases, verify the accuracy of the steering angles. One application of a sound event detector and localizer is as a stereo/surround detector. A stereo/surround detector determines whether a sound field is intended for reproduction in two input channels or in more than two input channels. [0021]
  • The properties of the human hearing mechanism have been modeled to create electronic circuitry that detects sound events in the presence of a background signal (collectively, “sound event detectors”). Sound event detectors may be used for a variety of applications for which the detection of specific types of sound events is helpful. For example, a sound event detector that detects syllables may be used as part of a phonic detector or as part of a speech recognition or speech coding system. In another example, a sound event detector that detects syllables may be used in conjunction with an audio amplification device, such as a microphone. This allows the microphone to remain off until a syllable is detected from a speaker, thus preventing the microphone from amplifying undesired sounds and feedback through the microphone itself when the speaker is silent. Sound event detectors, like the sound event detection methods, generally produce a trigger signal that includes some flag or marker, such as a pulse, that indicates the occurrence of a sound event. Also, like the sound event detection methods, the sound event detectors can be implemented to detect any number and combination of sound event types in a sound field that is generated in any number of input channels. [0022]
  • The sound event detection methods, sound event detection and localization methods, sound localization methods, as well as any of the sound event detectors, sound event detectors and localizers and sound localizers, may be implemented in a sound event device and/or in computable readable software code. [0023]
  • Other systems, methods, features and advantages of the invention will be, or will become, apparent to one with skill in the art upon examination of the following figures and detailed description. It is intended that all such additional systems, methods, features and advantages be included within this description, be within the scope of the invention, and be protected by the following claims.[0024]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The invention can be better understood with reference to the following drawings and description. The components in the figures are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the invention. In the drawings: [0025]
  • FIG. 1 is a flow chart of a sound event detection method for a single input channel and single sound event type. [0026]
  • FIG. 2 is a flow chart of an accommodation method. [0027]
  • FIG. 3 is a series of time-domain plots of a sample segment of a sound field before, during and after the onsets of transients have been distinguished and the transients have been separated from background sounds. [0028]
  • FIG. 4 is a flow chart of a beginning detection method for a single input channel. [0029]
  • FIG. 5 is a flow chart of a sound event detection method for multiple input channels. [0030]
  • FIG. 6 is a flow chart of a beginning detection method for multiple input channels. [0031]
  • FIG. 7 is a flow chart of a sound event detection method for a single input channel and multiple sound event types. [0032]
  • FIG. 8 is a flow chart of a sound event detection method for multiple input channels and multiple sound event types. [0033]
  • FIG. 9 is a flow chart of a sound event detection and localization method for a single input channel pair and a single sound event type. [0034]
  • FIG. 10 is a flow chart of a sound event detection and localization method for multiple input channel pairs and a single sound event type. [0035]
  • FIG. 11 is a flow chart of a sound event detection and localization method for multiple input channel pairs and a single sound event type. [0036]
  • FIG. 12 is a flow chart of a sound event detection and localization method for a single input channel pair and multiple sound event types. [0037]
  • FIG. 13 is a flow chart of a sound localization method. [0038]
  • FIG. 14 is a flow chart of a direction selection method for a single input channel pair and a single sound event type. [0039]
  • FIG. 15 is a flow chart of a direction selection method for a single channel pair and a single short-duration sound event. [0040]
  • FIG. 16 is a flow chart of a direction selection method for multiple input channel pairs and a single sound event type. [0041]
  • FIG. 17 is a flow chart of a differential steering angle determination method. [0042]
  • FIG. 18 is a flow chart of a direction selection method for multiple input channel pairs and a single short-duration sound event. [0043]
  • FIG. 19 is block diagram of a sound event device. [0044]
  • FIG. 20 is a block diagram of a sound event detector for a single input channel and a single sound event type. [0045]
  • FIG. 21 is a circuit diagram of an accommodation circuit for a single input channel. [0046]
  • FIG. 22 is a circuit diagram of a trigger generation circuit for a single input channel. [0047]
  • FIG. 23 is a circuit diagram of a sound event detector for a single input channel and a short-duration sound event. [0048]
  • FIG. 24 is a block diagram of a sound event detector for multiple input channels and a single sound event type. [0049]
  • FIG. 25 is a circuit diagram of an alternate trigger generation circuit for multiple input channel pairs. [0050]
  • FIG. 26 is a block diagram of a sound event detector for a single input channel and multiple sound event types. [0051]
  • FIG. 27 is a block diagram of a sound event detector for multiple input channels and multiple sound event types. [0052]
  • FIG. 28 is a block diagram of a sound event detector and localizer for a single input channel pair and a single sound event type. [0053]
  • FIG. 29 is a circuit diagram of a sound event localization circuit for a single input channel pair and a single sound event type. [0054]
  • FIG. 30 is a block diagram of a sound event detector and localizer for multiple input channel pairs and a single sound event type. [0055]
  • FIG. 31 is a circuit diagram of a sound event localization circuit for multiple input channel pairs and a single sound event type. [0056]
  • FIG. 32 is a block diagram of a sound event detector and localizer for a single input channel pair and multiple sound event types. [0057]
  • FIG. 33 is a block diagram of a sound event detector and localizer for multiple input channel pairs and multiple sound event types. [0058]
  • FIG. 34 is a block diagram of a stereo/surround detector for a single sound event type. [0059]
  • FIG. 35 is a block diagram of a stereo/surround detector for multiple sound event types. [0060]
  • FIG. 36 is a block diagram of a sound localizer for a single input channel pair and a single sound event type. [0061]
  • FIG. 37 is a circuit diagram of a sound localization circuit for a single input channel pair and a single sound event type. [0062]
  • FIG. 38 is a circuit diagram of a sound localization circuit for a single input channel pair and a short-duration sound event. [0063]
  • FIG. 39 is a block diagram of a sound localizer for multiple input channel pairs and a single sound event type. [0064]
  • FIG. 40 is a circuit diagram of a sound localization circuit for multiple input channel pairs and a single sound event type. [0065]
  • FIG. 41 is a block diagram of a sound localization circuit for multiple input channel pairs and a short-duration sound event. [0066]
  • FIG. 42 is a block diagram of a sound localizer for a single input channel pair and multiple sound event types. [0067]
  • FIG. 43 is a block diagram of a sound localizer for multiple input channel pairs and multiple sound event types.[0068]
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • Sound detection and localization systems, including methods and devices, have been developed that detect and separately determine the direction of sound events in a complex sound field based on the properties of the human hearing mechanism. The human hearing mechanism, which includes the ears, auditory system and related portions of the brain, was chosen as a model for two basic reasons. The first is that the human hearing mechanism detects sounds and determines their direction very effectively and without the shortcomings of the known systems. The second reason is that many of the mistakes made by systems modeled after the human hearing mechanism will generally be the same mistakes made by the human hearing mechanism, thus making these mistakes inaudible. [0069]
  • The human hearing mechanism was studied and used as a model from which to create sound detection and localization systems. By modeling the frequency bias, beginning detection property, accommodation property and the onset reliance of the human hearing mechanism, extremely effective, high-quality methods have been created for the detection of sound events in the presence of steady-state sounds, for the detection and localization of sound events in the presence of steady-state sounds, and for the detection and localization of all sounds in a sound field. These methods can also be implemented in software and hardware to create programs, devices and even more complicated systems applicable to a wide range of applications. [0070]
  • Many of the problems encountered by known sound detection and localization systems are not experienced by the human hearing mechanism. For example, human listeners are not generally fooled by highly reverberant music into perceiving rapid motion of the music or its source. Furthermore, the human hearing mechanism localizes sounds in a sound field by breaking down the sound field into individual sounds, determining the direction of the individual sounds, reassembling related sounds into streams (such as the melody line of a particular instrument or a line of dialogue from a particular speaker), and determining the direction of the stream based on the average direction of the individual sounds within the stream. The robust nature of the human hearing mechanism enhances its ability to distinguish sounds and determine their direction, such that, even if the direction of the sounds in the stream are determined correctly only about 50% of the time, the direction of the entire stream will most likely be correctly determined. [0071]
  • In breaking down a sound field into individual sounds, the human hearing mechanism has the ability to distinguish transient sounds from other sounds and to distinguish sound events from non-directional transients. Furthermore, the human hearing mechanism can distinguish transient from non-transient sounds and sound events from other transient sounds even in the presence of a background signal by relying on the properties of beginning detection and accommodation. Beginning detection results from human neurology which is highly evolved to determine the starting points and end points of sounds and makes the human hearing mechanism particularly sensitive to the rise and fall times of sounds. Accommodation is the property that allows the human hearing mechanism to distinguish transients from steady-state sounds by gradually ignoring the presence of the steady state sounds. The properties of beginning detection and accommodation and their effect on localization were discovered and demonstrated experimentally. In experiments with headphones, it was discovered that the azimuth (as an indication of direction) of a tone can only be detected when the tone starts or stops. If the tone is continuous, it becomes very difficult to determine its direction. Furthermore, if new sounds are added to a steady state tone, the true direction of the new sounds is easy to determine. This demonstrates that some sounds have both a transient and a steady-state component. It also demonstrates that the human hearing mechanism is particularly sensitive to the transient components, which include the beginnings and ends of sounds and uses the beginnings and ends of sounds for localization (the beginning detection property). It also demonstrates that the human hearing mechanism ignores steady-state sounds over time and generally does not rely on them for localization (the accommodation property). It was further discovered that the rate at which the human hearing mechanism gradually ignores the steady-state sounds (the “accommodation rate”) is independent of the sound event type being detected. It was also found that the accommodation rate may not be constant and may adjust to acoustic conditions and to the rapidity of speech. However, a reasonable average value for the accommodation rate was found to be about 300 ms. [0072]
  • Additionally, for purposes of localization, the human hearing mechanism relies more heavily on certain frequencies and the apparent direction of the beginnings of sounds. The human hearing mechanism relies most heavily on frequencies between about 500 Hz and about 4000 Hz for direction determination. This frequency bias is largely a result of the frequency response of the external ear (the pinnae, concha and ear canal) and by the frequency transfer function of the middle ear. The human hearing mechanism also relies on the direction indicated in the rise time of a sound more than the direction as indicated in other portions of the sound. This reliance is advantageous because the direction indicated in the rise time of a sound is less likely to be corrupted by sound reflections or reverberations even in a highly reflective environment. [0073]
  • After detecting the directions of the individual sounds, the human hearing mechanism reassembles related sounds into streams and separately determines the direction of each stream. The direction of each stream is generally the average direction of all sounds within the stream. In many cases, sounds are reassembled into a foreground stream and a background stream. In one example, the foreground stream may consist of dialog and the background stream may consist of environmental sounds being produced around the source of the dialog. In any case, the process of reassembling related sounds into streams and determining the direction of each stream is performed unconsciously and automatically by the human hearing mechanism. [0074]
  • By applying the properties of the human hearing mechanism systems have been developed to (a) detect sound events within a sound field; (b) detect and localize sound events in a sound field; and (c) localize all sounds within a sound field. These methods, devices and systems use the properties of the human hearing mechanism to break the sound field down into individual sounds, distinguish the individual sound events, and determine the direction of the individual sound events and other sounds. When the directions of the individual sounds are used to reproduce the sound field, it is left to the human hearing mechanism to reassemble the related sounds into streams and to determine the direction of each stream. [0075]
  • 1. Methods for Detecting Sound Events [0076]
  • The properties of the human hearing mechanism have been modeled to create methods for detecting sound events in the presence of a background signal (collectively “sound event detection methods”). Sound event detection methods produce a signal (referred to in this application as a “trigger signal”), which includes a flag or marker, such as a pulse, that indicates the occurrence of a sound event. The sound event detection methods can be implemented to detect any number and combination of sound events in any number of input channels. In the following description, the sound event detection methods are discussed in order of increasing complexity, with each subsequent sound event detection method incorporating the steps of the prior methods, except as indicated. [0077]
  • An example of a method for detecting sound events in the presence of a background signal that only detects a single sound event type in a sound field that is generated in a single input channel (“sound event detection methods for a single input channel and a single sound event type”), is shown in FIG. 1 and indicated by [0078] reference number 100. The sound event detection method for a single input channel and a single sound event type 100 includes: emphasizing the directionally important frequencies by modeling frequency bias 104; separating sound events from background sounds by modeling accommodation 106; and detecting sound events by modeling beginning detection 108.
  • The directionally important frequencies are emphasized by modeling the frequency bias of the [0079] human hearing mechanism 104 using frequency emphasis method. The frequency emphasis method includes emphasizing the frequencies between about 500 Hz and about 4 kHz in each input channel of the input channel pair to produce a filtered signal in each input channel. These frequencies are emphasized because they have the most influence on the human hearing mechanism in terms of determining direction. Separating the sound events from the background sounds by modeling accommodation 106 generally includes examining the change in input power in the sound field and is shown in more detail in FIG. 2. More specifically, separating the sound events from the background sounds by modeling accommodation 106 (an “accommodation method”) includes: determining the power envelope of the input channel 201; determining the power in any steady-state sounds 202, and subtracting the power in any steady-state sounds from the power envelope of the input channel. Determining the power envelope of the input channel 201 generally includes squaring the voltage in the input channel. A sound field, whether produced live or from a recording, is generally represented as a voltage signal in the time domain. Unfortunately, voltage signals do not have the additive property, which means that the magnitude of a voltage signal resulting from the combination of two voltage signals cannot be determined by simply adding the amplitudes of the two voltage signals. However, power signals do have the additive property. Therefore, by converting the voltage signal into a power signal, (generally by squaring the voltage signal) a power envelope results from which other power signals may be subtracted. In order to obtain a cleaner power envelope (one containing fewer fluctuations), fluctuations in the power envelope that are not characteristic of the sound event type being detected may be removed. For example, if syllables are being detected, fluctuations with rise times faster than about 33 ms will be removed from the power envelope.
  • Determining the power in any steady-[0080] state signals 202 in the sound field includes determining the long-term average power of the sound field. The long-term average power (referred to in this application as the “accommodation signal”) may be determined by integrating the power envelope over a time period equal to the rise time of the sound event type that is being detected. Once a sound event has reached its maximum value (at the end of the rise time of the sound event), the accommodation signal is gradually subtracted from the power envelope 204 at a rate equal to the accommodation rate of the human hearing mechanism (which is about 300 ms) to model the way in which the human hearing mechanism gradually ignores the effects of steady-state sounds in the presence of sound events. In order to control the accommodation, at all times the accommodation signal is forced to be less than or equal to the power envelope. Therefore, when the sound event ends, causing the power envelope to drop rapidly, the accommodation signal will be forced to equal the power envelope. The result of subtracting the accommodation signal from the absolute signal power is a “differential signal.” This difference signal includes positive pulses and other fluctuations. Each pulse in the difference signal indicates the occurrence of a sound event and the other fluctuations are caused by noise such as: reverberation, Gaussian noise, and other signals that may not currently be in the foreground stream.
  • An example of how the accommodation method affects an input signal is shown in FIG. 3. The time-domain plot A in FIG. 3 shows a sample segment of a sound field. This segment includes a collection of sound waves of varying frequencies that is characteristic of a [0081] syllable 300. The magnitude of this syllable is the voltage in the signal “V.” The voltage V varies with time “t” and has a duration of “D.” The time-domain plot B shows a pulse 302, which is segment 300 after it has been converted into a power signal (step 201 in FIG. 2). Pulse 302 has an amplitude that represents the power in the signal (“V2”) and also has duration D. The time-domain plot C shows pulse 304, which is pulse 302 after the fluctuations with rise times that are not characteristic of a syllable have been removed. The pulse 304 also has an amplitude V2 and has a rise time “tr” where tr is equal to about 33 ms. The time-domain plot D shows a pulse 306, which is pulse 304 after the steady-state sounds have been accommodated to by subtracting the accommodation signal. The rise-time of pulse 306 follows that of pulse 304. However, after the syllable has reached its maximum value (after tr) the effects of any simultaneously-occurring steady-state signals are removed from pulse 306 at a rate that is equal to the accommodation rate of the human hearing mechanism (about 300 ms). When the sound event ends, the power envelope V2 (pulse 304) decreases rapidly. At the point where V2 is about to become less than the accommodation signal “AccSig,” AccSig is forced to about equal V2 until the end of the sound event. This causes the differential signal to go to zero.
  • Referring again to FIG. 1, once the sound events have been separated from the background sounds by modeling [0082] accommodation 106, the sound events are detected by modeling beginning detection 108. Modeling beginning detection includes distinguishing the sound events from the noise present in the differential signal. As previously discussed, the differential signal includes a series of rapidly-rising pulses and noise. While each pulse indicates the occurrence of a sound event, the noise includes fluctuations that may falsely indicate the occurrence of a sound event. Therefore, in order to detect the sound events, the sound events need to be distinguished from the noise. Modeling beginning detection 108 is shown in more detail in FIG. 4 and includes deemphasizing the effects of volume; emphasizing the sound events 404; deemphasizing the noise 406; and detecting the sound events 408.
  • One way in which steps [0083] 404, 406 and 408 can be simultaneously accomplished is by scaling the differential signal by the short-term average power contained in the noise component of the differential signal (the “short-term average high frequency power”). Using the techniques generally involved in automatic gain control (an “automatic gain method”), the short-term average high frequency power is isolated and used to divide the differential signal. The short-term average high frequency power can be isolated by filtering the differential signal to obtain the component of the differential signal with frequencies higher than those characteristic of the sound event type being detected, and integrating the high-frequency portion over a short time period. This short time period may be different for different types of music or speech rates. However, a value of about 160 ms was experimentally determined to work well for a wide variety of inputs.
  • Normalizing the differential signal by using an automatic gain methods has two effects, one obvious and two not-so-obvious. The first effect is that it deemphasizes the effect volume has on the [0084] differential signal 402 by normalizing the differential signal with regard to the amplitude (which represents power) of the fluctuations in the differential signal. Because louder sound events and noise generally contain more power, the louder sound events and noise will be scaled more strongly that those that are not as loud. This is the normal function of an automatic gain control method.
  • A second effect is to de-emphasize [0085] noise 406 by de-emphasizing the fluctuations that occur more often than is characteristic of the sound event type being detected (for example, syllables generally occur or repeat once about every 200 ms). The fluctuations in portions of the differential signal due to noise generally occur more frequently than is typical for a sound event. The rapid repeat rate of these noise fluctuations increases the average high-frequency power with which that portion of the differential signal is divided. This de-emphasizes the portion of the differential signal that contains the noise. The third effect of normalizing the differential signal is to emphasize the sound events 404 by emphasizing the fluctuations that occur or repeat no more often than is characteristic of the sound event type being detected. The portions of the differential signal that contain fluctuations that occur less often than is typical for the sound event type being detected will have a lower average high frequency power (as compared to those containing noise) with which that portion of the differential signal is divided. This will provide a relative emphasis to the sound events.
  • [0086] De-emphasizing noise 406 may be further accomplished by removing many of the fluctuations in the differential signal due to noise. Some of the fluctuations due to noise can be identified and removed according to their rise time and other characteristics. For example, fluctuations that have rise-times not characteristic of the sound event type being detected are removed. For example, if the sound event type being detected is syllables, sounds with rise times faster than about 33 ms will be removed. In another example, if the sound event type being detected is impulsive sounds, sounds with rise times faster than about 3.3 ms will be removed. Additionally, low amplitude transients that occur more frequently than is characteristic of sound events such as, some noise and reverberation, can also be filtered out. Furthermore, because it can be assumed that whenever the power level of the sound field has dropped more than about 10 dB from that of a previous sound event, any transient simultaneously detected is noise or some other non-directional transient, transients occurring under such conditions can also be removed. As a result of steps 402, 404 and 406, many of the fluctuations in the differential signal are removed to create an improved differential signal that includes pulses each indicating the occurrence of a sound event. However, some noise remains.
  • Therefore, to produce a trigger signal that only includes pulses indicating the occurrence of sound events, the sound events need to be detected [0087] 408 from the noise. Because at this point, most of the fluctuations in the improved differential signal due to noise have a low amplitude as compared with the fluctuations caused by the sound events, the sound events are detected 408 by determining which fluctuations have an amplitude that exceeds a threshold using a threshold detection method. In the threshold detection method, the fluctuations that do not exceed the threshold are removed or ignored to produce the trigger signal. One advantage of this method is that the triggering need not be ideal. The sound event detection methods are not perfect in that in some sound events will not be detected, and some fluctuations due to noise will result in a pulse in the trigger signal falsely indicating the occurrence of a sound event. However, these occasional errors do not matter. Because the sound event detection methods are modeled after the human hearing mechanism, it produces the same types of errors as does the human hearing mechanism. Therefore, the result will be perceived as completely natural.
  • The threshold is chosen so that low amplitude fluctuations that are characteristic of noise or reverberation are not detected. The threshold may be a fixed value which is determined experimentally. However, more accurate results are obtained if the threshold varies as a function of the sound field. For example, when the sound field includes many sound events, the threshold will generally be lower than when the sound field includes fewer sound events. This allows greater sensitivity when a greater number of sound events are present in the sound field. [0088]
  • Alternatively, the threshold can be manually selected from among two or more values experimentally determined according to the characteristics of the sound field. For example, if the sound field is that of modern or popular music, which typically includes many sound events, a lower threshold value can be selected, and alternatively, if the sound field is that of classical music, which typically includes few sound events and may be highly reverberant, a higher threshold value can be selected. Alternatively, the threshold value can be chosen as a function of the number of sound events detected during a given time period. Therefore, during time periods where fewer sound events are detected, the threshold value is raised and during time periods where a greater number of sound events are detected, the threshold value is lowered. This provides an automatic and continuous adjustment of the threshold value as a function of the sound field and results in a more accurate trigger signal. [0089]
  • Any of the sound event detection methods may also include performing only the beginning detection method, which is of particular use when detecting sound events of short duration (“short-duration sound events”), such as impulsive sounds. Short-duration sound events, such as impulsive sounds, tend to end before the effects of accommodation can be sensed. Therefore, the sound event detection methods can be simplified to include performing only the beginning detection method on the input signal (in decibels) directly. The sound event detection methods that include performing only the beginning detection method (the “sound event detection methods for short-duration sound events”) may also include removing any fluctuations with rise-times slower than those characteristic of the short-duration sound event being detected before performing the beginning detection method. For example, if impulsive sounds are being detected, fluctuations in the input signal with rise times slower than about 3 ms will be removed. [0090]
  • Methods for detecting a single sound event type in the presence of a background signal may also be implemented when the sound field is generated in two or more input channels (“sound event detection methods for multiple input channels and a single sound event type”). The sound event detection methods for multiple input channels and a single sound event type may include performing the sound event detection method for a single input channel and a single sound event type in parallel for each input channel to produce a trigger signal for each input channel indicating the occurrence of whatever sound event type is being detected. [0091]
  • Alternatively, a sound event detection method for multiple input channels and a single sound event may include detecting sound events in pairs of channels, instead of in each channel separately. Generally, this includes subtracting the accommodated signal in one channel from the accommodated signal in the other channel to create a “difference signal,” which is then used to create the trigger signal. Often the difference signal is created by subtracting the accommodation signal of one input channel in an input channel pair from that of the other input channel in the input channel pair. For example, the input channel pairs may include a left-right channel pair and/or a center-surround input channel pair. Throughout this application, the term “input channel pair” includes any combination of two input channels and the channels that can be derived from the input channels. The The purpose of using more than one input channel to derive trigger signals is to provide a focus on directional signals. For example, in electronically recorded music and speech, directional signals can be differentiated from non-directional signals by the phase and amplitude relationships between the input channels. For example, sound fields are often generated in two input channels (ordinary stereo). From these two input channels it is useful to derive four power envelopes and to organize the four power envelopes into two power envelope pairs. For example, if the original two input channels are designated with the conventional “left” and “right” names, the resulting power envelopes can be designated “left power” and “right power,” respectively and together make up a power envelope pair. The remaining power envelope pair is made up of power envelopes derived from the sum and the difference of the two input channels, the “left plus right power” and “left minus right power.” This pair is often also referred to as “center power” or the “surround power.” Non-directional signals almost always will cause all four of these power envelopes to be equal in level. A non-directional transient signal will cause all four of them to rise at the same time. When the difference signal is used to create the trigger signal, a non-directional transient will cause no rise in the difference signal, because it is equal in each input channel. By contrast, a directional signal, for example a sound event in the left input channel only, will cause a large change in the “left power” envelope, and no corresponding increase in the “right power” envelope. Therefore, there will be a large increase in the difference signal created from the left and right accommodated signals, and this change can be used to generate triggers that preferentially distinguish directional signals from non-directional signals. [0092]
  • An example of a sound event detection method for multiple input channels and a [0093] single sound event 500 that includes detecting sound events in pairs of channels is shown in FIG. 5 and includes: emphasizing the directionally important frequencies by modeling frequency bias 504; separating the sound events from the background sounds by modeling accommodation in each input channel 506; and detecting sound events in each input channel pair by modeling beginning detection in each input channel pair. The directionally important frequencies are emphasized 504 using a frequency emphasis method. Then, the sound events are separated from the background signal in each input channel 506 using an accommodation method to produce a differential signal for each input channel. Subsequently, the sound events are detected in each input channel pair 508 using an alternate beginning detection method in parallel for each input channel pair in and the differential signals for each channel pair.
  • An alternate beginning detection method is shown in more detail in FIG. 6 and may include, determining the difference signal for each [0094] input channel pair 601; deemphasizing the effects of volume in each input channel pair 602; emphasizing sound events in each input channel pair 604; deemphasizing noise in each input channel pair 606; and detecting sound events in each input channel pair 608. Therefore, the sound event distinction method for multiple input channels includes the same steps as the sound event distinction method but performed on the difference signal of each input channel pair.
  • A difference signal is determined [0095] 601 be determining the difference between the differential signals in each input channel of an input channel pair, and rectifying this difference. The difference between the input channels in the input channel pair includes both positive and negative pulses representing sound events that have occurred in either input channel of the input channel pair. The pulses will be negative or positive depending upon which input channel in the input channel pair reflects the majority of the power in the particular sound event. Therefore, this difference is rectified to produce a difference signal with only positive pulses. While the difference signal still indicates sound events, as previously discussed, it does not indicate sound events common to both input channels. Therefore, the difference signal has a better signal to noise ratio then that of the differential signals of the individual input channels. Furthermore, the difference signal does not contain any signals that are equal in both input channels of the input channel pair, which helps to elimination certain non-directional signals, such as noise, which are generally equal in all input channels. Once the difference signal has been determined for each input channel pair 601, the effects of volume on the difference signal are de-emphasized 602, sound events are emphasized 604, and noise is de-emphasized in the difference signal 606 using an automatic gain method to produce a series of pulses indicating the presence of any remaining transients. Subsequently, the sound events detected in each input channel 608 by determining which of the pulses are greater than a given threshold. Generally, steps 602, 604 and 608 include a rectification step so that the pulses indicating the remaining transients are all positive. The result of step 608 is a trigger signal with positive pulses indicating the occurrence of a sound event for every input channel pair. This sound event detection method may be implemented for multiple input channel pairs by performing the method in parallel for each input channel pair to create a separate trigger signal for each input channel pair.
  • The sound event detection methods for multiple channels and a single sound event type may also include detecting a sound event type in multiple channel pairs (collectively, “sound event detection methods for multiple channel pairs and a single sound event type”). These methods generally include the sound event detection methods for multiple channels and a single sound event type that include an alternate beginning detection method implemented for more than one channel pair, or for a single channel pair from which four power envelopes can be derived. [0096]
  • Methods for detecting one or more sound events in the presence of a background signal may also be implemented so that multiple sound event types, such as syllables and impulsive sounds, are detected in a single input channel (“sound event detection methods for a single input channel and multiple sound event types”). One example of a sound event detection method for a single input channel and multiple sound event types that detects both syllables and impulsive sounds is shown in FIG. 7 (although this method may be implemented to detect any number and combination of sound event types). The example of a sound event detection method for a single input channel and multiple [0097] sound event 700 shown in FIG. 7, basically includes: emphasizing the directionally important frequencies by modeling beginning detection 704; separating sound events from the background sounds by modeling accommodation 706; detecting syllables by modeling beginning detection 710; and detecting impulsive sounds by modeling beginning detection 712. It is important to understand that syllables and impulsive sounds may often occur simultaneously.
  • The directionally important frequencies are emphasized [0098] 704 using a frequency bias method. Then, the sound events are separated from the background signal 706 using an accommodation method to produce a differential signal. To improve the results, the accommodation method may also include removing fluctuations in the power envelope with rise times that not characteristic of syllables (those above about 33 ms). Subsequently, syllables are detected 710 using a beginning detection method for single input channel (see FIG. 4) and impulsive sounds are detected using a sound event detection method for short-duration sound events. As describe in connection with FIG. 4, detecting sound events includes deemphasizing the effects of volume, emphasizing sound events and deemphasizing noise using an automatic gain method and by removing certain fluctuations, and detecting sound events using a threshold detection method. However, in this case, when detecting syllables, the short-term high frequency power used to normalize the differential signal is the power in the components of the differential signal above 30 Hz determined over about 167 ms. Additionally, the fluctuations removed are those with a rise time of less than 33 ms. Conversely, in these cases, when detecting impulsive sounds, the sound event detection method for a single short-duration sound event includes: removing certain fluctuations; deemphasizing the effects of volume, emphasizing sound events and deemphasizing noise using an automatic gain method and removing certain fluctuations, and detecting sound events using a threshold detection method. However, when detecting impulsive sounds, the automatic gain method uses the input signals (in decibels) instead of the differential signal. Further, the fluctuations that are removed are those with rise times less than about 3 ms.
  • Methods for detecting one or more sound events in the presence of a background signal may also be implemented so that multiple types of sound events may be detected for a sound field generated in multiple input channels (“sound event detection methods for multiple input channels and multiple sound events”). The sound event detection method for multiple input channels and multiple sound events includes performing the sound event detection method for a single input channel and multiple sound event types implemented in parallel for each input channel of the sound field to produce a trigger signal for each sound event type in each input channel. Alternatively, as shown in FIG. 8, the sound event detection method for multiple input channels and multiple [0099] sound events 800 may detect each sound event only in each input channel pair, thus producing a trigger signal for each sound event type in each input channel pair. In the example shown in FIG. 8, this method 800 is used to detect two sound event types (syllables and impulsive sounds) in two input channels (a right input channel and a left input channel) to produce two trigger signals for the input channel pair, the first indicating the occurrence of syllables and the second indicating the occurrence of impulsive sounds. However, the method may be implemented to detect any number of sound events in any number of input channels to produce for each input channel pair a trigger signal for each sound event type. The sound event detection method for multiple input channels and multiple sound event types 800 generally includes: emphasizing directionally important frequencies by modeling the frequency bias in the left input channel 802; emphasizing directionally important frequencies by modeling the frequency bias in the right input channel 804; separating sound events from the background sounds by modeling accommodation in the left input channel 806; separating sound events from the background sounds by modeling accommodation in the left input channel 808; detecting impulsive sounds in the left-right input channel pair by modeling beginning detection 812; and detecting syllables in the left-right input channel pair by modeling beginning detection 814.
  • Directionally important frequencies are emphasized separately in both the left and [0100] right input channels 802 and 804, respectively, using a frequency bias method. Subsequently, sound events are separated from the background sounds using an accommodation method in both the left and right input channels 806 and 810, respectively, to produce a differential signal for the left and right input channels, respectively. The differential signals from both channels are then used to detect syllables by modeling beginning detection 814. Syllables may be detected using a beginning detection method to separately detect syllables in each input channel to produce a trigger signal indicating the occurrence of syllables in each channel. Alternatively, syllables may be detected using an alternate beginning detection method to produce a single trigger signal indicating the occurrence of syllables for the channel pair. To detect impulsive sounds by modeling beginning detection 812, a sound event detection method for a single channel and a short-duration sound event is used to produce a trigger signal indicating the occurrence of an impulsive sound for each input channel. In this, as in all the sound event detection methods, the trigger signals may be combined to create fewer trigger signals.
  • The sound event detection methods for multiple channels and multiple sound event types may also include detecting more than one sound event type in multiple channel pairs (collectively, “sound event detection methods for multiple channel pairs and multiple sound event types”) These methods generally include the sound event detection methods for multiple channels and multiple sound event types that include an alternate beginning detection method implemented for more than one channel pair, or for a single channel pair from which four power envelopes can be derived. [0101]
  • 2. Methods for Detecting and Localizing Sound Events [0102]
  • The properties of the human hearing mechanism have also been used to create methods for determining the location of sound events that occur in the presence of a background signal (collectively “sound event detection and localization methods”). As previously discussed, the sound event detection methods for multiple input channels not only produce a trigger signal indicating whenever a sound event occurs, but also produce differential signals from which the direction of a sound event can be determined. Sound event detection and localization methods use these differential signals to determine the direction of a sound event by comparing the relative power of the sound event in each input channel of each input channel pair after any steady-state sounds have been removed. By comparing the power in each input channel of an input channel pair instead of the voltage in each input channel of a input channel pair as done in known methods, individual sound events can be identified and separately localized. The sound event detection and localization methods can be implemented to detect and localize any number and combination of sound events types in a sound field generated in any number of input channel pairs. In the following description, the sound event detection and localization methods are discussed in order of increasing complexity, with each subsequent sound event detection and localization method incorporating the steps of the prior methods, except as indicated. [0103]
  • A sound event detection and localization method for a single input channel pair and a single sound event type is shown in FIG. 9. This sound event detection and localization method for a single input channel pair and a single [0104] sound event type 900 detects a single sound event type and locates the sound events relative to a single input channel pair. The location of a sound event as determined by this method 900 is given in terms of an angle between the input channels in the input channel pair (referred to in this application as a “true differential steering angle” or “true DSA”). The sound event detection and localization method 900 generally includes: producing a differential signal and detecting sound events in an input channel pair for each input channel in the input channel pair 904; determining the initial sound event direction 906; and isolating the true DSA 908. Collectively, 904 and 906 define a DSA determination method for a single input channel pair and a single sound event type 901.
  • A differential signal is produced for each input channel of the input channel pair and sound events are detected in the [0105] input channel pair 904 using a sound event detection method for a single input channel pair and a single sound event type or a sound event detection method for a single input channel pair and a single short-duration sound event as appropriate for the sound event type being detected. As a result, a trigger signal indicating the occurrence of a single sound event type is produced for the input channel pair and a differential signal is produced for each input channel in the input channel pair. When a sound event is detected 904 in the input channel pair, as indicated by the trigger signal, the initial direction of that sound event is determined 906 using an initial direction determination method. The initial direction determination method determines the direction of the sound event from the differential signal for each input channel in the input channel pair as these differential signals exist when the sound event occurs. More specifically, when the trigger signal indicates that a sound event is occurring, the differential signals for each input channel are converted into decibels to create a decibel differential signal for each input channel. The decibel differential signal for one input channel is then subtracted from the decibel differential signal from the other input channel to create a decibel ratio. This decibel ratio is converted into an equivalent angle using known methods to create a “differential steering angle” or “DSA” which indicates the location of the sound event between the input channels in the input channel pair.
  • For example, if syllables are being detected, the true direction of any syllables detected is obtained by isolating the DSA indicated in about the first 20 ms to about the first 30 ms of the syllable, which is then used (held) for about 200 ms. In another example, if impulsive sounds are being detected, the true direction of any impulsive sounds is obtained by isolating the DSA indicated in about the first 3.3 ms to about 5 ms of the impulsive sound, which is then used for about 50 ms. However, in short-duration sound events, such as impulsive sounds, the initial direction determination method may be implemented to obtain an even more accurate initial direction. This initial direction determination method detects all triggers that occur in a very short segment of the trigger signal (generally, about 3 ms) determines the DSAs associated with those pulses, determines the short-time average of these DSAs and uses the short-term average DSA to indicate the direction of all the short-term sound events detected in the very short segment for the typical duration of the short-duration sound event type being detected. For example, if impulsive sounds are being detected, the short-time average is then used for the next about 50 ms. The idea of using a time average over the rise-time of a sound event as a measure of its direction also works with sounds that have longer durations, but it is useful to use a longer time period for the averaging, as the rise time of these signals is frequently longer. [0106]
  • Sound event detection and localization methods may also be implemented to detect and localize a single sound event type in multiple input channel pairs (“sound event localization methods for multiple input channels pairs and a single sound event type”). For example, a typical surround sound system includes at least a left-right input channel pair and a center-surround input channel pair. In the sound event localization methods for multiple input channel pairs and a single sound event type, the true DSA for each channel pair is determined whenever a trigger signal in any channel pair indicates that a sound event is occurring, however, the DSA is used to indicate the direction of the sound event only if the DSA is accurate. If the DSA is not accurate, the direction indicated by a true OSA is used. [0107]
  • An example of a sound event localization method for multiple input channel pairs and a single sound event type, is shown in FIG. 10, indicated by [0108] reference number 1000 and includes: performing a DSA determination method for a single input channel pair and a single sound event for each input channel pair (collectively a “DSA determination method for multiple input channel pairs a single sound event type”) 1051; determining the ordinary steering angle for each channel pair 1003; determining whether the DSA is accurate 1010, where if the DSA is not accurate, isolating and using the true OSA for each input channel pair for the typical duration of the sound event type being detected 1014; where, if the DSA is accurate, isolating the true DSA and using the true DSA for the typical duration of the sound event type being detected 1012.
  • The DSA determination method for multiple channel pairs and a single sound event type [0109] 1051 includes: detecting a sound event in each input channel pair and producing a differential signal for each channel in the channel pair 1004; and determining the initial DSA for each input channel in the channel pair 1006. Detecting a sound event in each input channel pair and producing a differential signal for each channel in the channel pair 1004 includes performing a sound event detection method for a single channel pair and a single sound event type in parallel for each channel pair. Determining the initial DSA for each input channel in the channel pair 1006 includes performing an initial direction determination method in parallel for each channel pair. Determining the OSA for each channel pair 1003 is generally accomplished using known methods.
  • Determining whether the DSA is accurate includes determining, at the moment the trigger signal indicates that a sound event is occurring, whether a sound event begins in one channel just as a sound event ends in another channel using a “background signal check” and determining whether the values obtained for the DSA are consistent with each other a “consistency check.” In general, when one input channel pair is derived from another input channel pair (for example, the center-surround input channel pair is derived from the left-right input channel pair), the sound event detection and localization methods do not yield accurate results when a sound event occurs precisely at the moment that another sound event in a different direction ends. This is generally characterized by an about simultaneous drop of about 3 dB in the power envelopes of at least two of the input channels. Although signals of this type are not common in natural environments, they are common in many popular test signals used for sound equipment and are sometimes used in popular music. Therefore, the background signal check examines the power envelopes in each input channel and when two of them drop about simultaneously by about 3 dB, the DSA is considered to be inaccurate. [0110]
  • The consistency check determines whether the sum of the absolute values of the true sound event directions in the left-right input channel pair and the center-surround input channel pair is less than about 45 degrees. Therefore, a sum of less than 45 degrees indicates that the steady-state signal did remain constant and that the direction indicated by the true sound event direction is accurate. However, a sum of greater than or equal to about 45 degrees indicates that the steady-state signal did not remain constant and that the direction indicated by the true sound event direction is not accurate. Based on this, the consistency check produces a signal whenever the true sound event is found to be inaccurate, thus indicating an error (an “error signal”). [0111]
  • If the DSA is found to be accurate, the true DSA is isolated and used for the typical duration of the sound event type being detected [0112] 1012. For example, if syllables are being detected the DSA is used for the typical duration of a sound event. In another example, if impulsive sounds are being detected, the DSA is used for the typical duration of an impulsive sound. If however, the DSA is not accurate, the true OSA is isolated and used for the typical duration of the sound event type being detected 1014. The true OSA is the OSA occurring during the rise-time of the sound event type being detected.
  • The sound event detection and localization methods for multiple channel pairs and a single sound event type may also include methods that reduce the occurrence of errors in the true direction of the sound events using an accommodation adjustment method. The accommodation adjustment method uses the consistency check to adjust the degree of accommodation in the accommodation signal. As previously discussed, the accommodation signal represents the steady-state signals present in the sound field as they are gradually ignored by the human hearing mechanism. The rate at which these steady-state signals are ignored is referred to as the degree of accommodation. The degree of accommodation, which is defined by the accommodation signal (“AccSig”). The accommodation adjustment method multiplies the accommodation signal AccSig by an adjustment factor Adj, so that AccSig is defined by the following equation:[0113]
  • AccSig=AccSig(Adj)  (1)
  • where Adj is defined by the following equation: [0114] Adj = 1 - α 50 ( 2 )
    Figure US20040005065A1-20040108-M00001
  • where α is the number of errors determined by the consistency check in a time period of about several seconds. The precise time period over which the number of errors is determined is relatively unimportant. The purpose of Adj is to reduce the degree of accommodation so that the accommodation signal will be reduced therefore, causing the steady-state signals to be ignored at a slower rate. This results in fewer transients being detected, thus reducing the number of transients incorrectly located. It has been determined experimentally that if Adj is about greater or equal to 0.998, the number of errors indicated by the consistency check are reduced and usable results are still obtained by the sound event detection and localization methods for multiple input channel pairs. However, if Adj is about equal to or lower than 0.995, too many transients fail to be detected so that the sound event detection and localization methods for multiple input channel pairs do not yield usable results. [0115]
  • Alternatively, the sound event detection and localization methods for multiple input channels pairs and a single sound event type may use the number of errors to adjust the threshold as part of error threshold adjustment methods. An error threshold adjustment method determines the number of errors determined by the consistency check in a predetermined time period of about several seconds and uses this number to adjust the threshold. As the number of errors increases, the threshold is increased so that fewer sound events are detected. Therefore, a continuous adjustment to the threshold can be provided, which is in proportion to the number of errors detected by the consistency check. [0116]
  • Sound event detection and localization methods may also be implemented to detect and localize multiple sound event types in a single input channel pair (“sound event localization methods for a single input channel pair and multiple sound event types”). FIG. 11 shows a sound event detection and [0117] localization method 1100 implemented to detect and localize two types of sound events (syllables, and impulsive sounds) to produce a DSA that indicates their whenever they occur. However, this method may be implemented in a similar manner to detect and localize any number and types of sound events. The sound event detection and localization method for a single input channel pair and multiple sound event types 1100 generally includes: producing a differential signal for the input channel pair 1104; detecting syllables in a input channel pair 1106; detecting impulsive sounds in the input channel pair 1107; determining the initial direction 1108; and isolating the true DSA 1110 for the sound event detected. Collectively, steps 1104, 1106, 1107, and 1108 define a DSA determination method for a single input channel pair and multiple sound event type 1101.
  • Producing a differential signal for the [0118] channel pair 1104 includes performing an accommodation method for each input channel. Detecting syllables in the input channel pair 1106 includes performing a beginning detection method for a single channel pair and a single sound event type to produce a trigger signal indicating the occurrence of syllables in the channel pair. Detecting impulsive sounds in the input channel pair 1107 includes performing a sound event detection method for a single input channel and a single short-duration sound event in parallel for each channel in the channel pair (the resulting two trigger signals may be combined to form a single trigger signal indicating the occurrence of impulsive sounds in the channel pair).
  • The initial direction is then determined [0119] 1108 at the moment any trigger signal indicates that a syllable or an impulsive sound is detected. At this moment, an initial direction determination is used to determine the initial direction of the sound event detected using the differential signals of each input channel in the input channel pair. At the moment any trigger signal indicates that a sound event of either type is occurring, the true DSA is isolated 1110 according to the true direction isolation method which uses the differential signal to produce the true DSA. The true DSA is then used for the typical duration of the sound event type that was detected. For example, if a syllable was detected, the DSA will be used for the typical duration of a syllable. Conversely, if an impulsive sound was detected, the DSA will be used for the typical duration of an impulsive sound.
  • Sound event detection and localization methods may also be implemented to detect and localize multiple sound event types in multiple input channel pairs (“sound event localization methods for multiple input channel pairs and multiple sound event types”). The sound event localization methods for multiple input channel pairs and multiple sound event types generally involve performing a sound event detection and localization method for a single input channel pair and multiple sound event types in a parallel fashion for each input channel pair. Alternately, sound event localization methods for multiple input channel pairs and multiple sound event types may include detecting a sound event, determining the DSA for the sound event, determining whether the DSA is accurate and using the OSA if the DSA is not accurate. An example of a sound event localization methods for multiple input channel pairs and multiple sound event types that uses the OSA to indicate the direction of a sound event if the DSA is not accurate is shown in FIG. 12. In this example, syllables and impulsive sounds are detected in two channel pairs. However, this method may be implemented to detect any number of wound event types in any number of channels. [0120]
  • In FIG. 12, the sound event localization methods for multiple input channel pairs and multiple [0121] sound event types 1200 includes: producing a differential signal for the input channel pair 1204; detecting syllables in each input channel pair 1206; detecting impulsive sounds in each input channel pair 1207; determining the initial direction for each channel pair 1208; determining whether the DSA is accurate 1210, where if the DSA is accurate, the true DSA is isolated and used for the typical duration of the sound event type detected 1212; where if the DSA is not accurate, the true OSA is isolated and used for the typical duration of the sound event type detected 1214. Collectively, steps 1204, 1206, 1207, and 1208 define a DSA determination method for multiple input channel pairs and multiple sound event types 1201.
  • Producing a differential signal for each [0122] channel pair 1204 includes performing an accommodation method in parallel for each input channel. Detecting syllables in each input channel pair 1206 includes performing a beginning detection method in parallel for a single channel pair and a single sound event type to produce a trigger signal for each channel pair indicating the occurrence of syllables in any channel pair. Detecting impulsive sounds in each input channel pair 1207 includes performing a sound event detection method for a single input channel and a single short-duration sound event in parallel for each channel in each channel pair (for each channel pair the resulting two trigger signals may be combined to form a single trigger signal indicating the occurrence of impulsive sounds in the channel pair). Simultaneously, the OSA is determined in each channel pair 1203 using known methods.
  • The initial direction is then determined in each [0123] channel 1208 at the moment any trigger signal indicates that a syllable or an impulsive sound is detected. At this moment, an initial direction determination is used to determine the initial direction of the sound event detected using the differential signals of each input channel in the input channel pair. It is then determined whether the DSAs are accurate 1210 using a beginning detection method and/or a consistency check. If the DSAs are found to be accurate, the true DSA is isolated 1210 according to the true direction isolation method which uses the differential signal to produce the true DSA. The true DSA is then used for the typical duration of a syllable. Conversely, if an impulsive sound was detected, the DSA will be used for the typical duration of an impulsive sound. If however, the DSA is not found to be accurate, the true OSA is isolated 1214 to produce the true OSA. The true OSA is then isolated from the OSA during the rise time of the impulsive sound and is used for the typical duration of an impulsive sound. Additionally, the sound event detection and localization methods for multiple channel pairs and a single sound event type may further include an accommodation adjustment method and/or an error threshold adjustment method.
  • In any of the sound event detection and localization methods, the direction of any sound events is used to indicate the direction of the entire sound field for the typical duration of the sound event type that was detected. However, if a subsequent sound event occurs during the typical duration, the direction of the subsequent sound event will be used for all input channels as soon as the subsequent sound event occurs and will continue to be used for the typical duration of the sound event type of the subsequent sound event. In generally means that when multiple sound events overlap, the direction of the most recent sound event will be used. [0124]
  • One application of the sound event detection and localization methods is in methods that determine whether recorded sounds are intended for reproduction in stereo (two channels) or in surround (more than two channels) (“stereo/surround detection methods”). Stereo/surround detection methods generally determine the number of sound events intended to be reproduced behind the listener and whether this number exceeds a predetermined value. These methods have an advantage over known methods that use negative phase as an indicator of sounds that are to be reproduced behind the listener because transients that naturally have negative phase will not falsely indicate a rear direction. [0125]
  • As previously discussed, sound fields intended for reproduction in surround include sound events that are intended to be reproduced behind the listener. Therefore, the stereo/surround detection methods include performing a sound event detection and localization method for a single input channel pair for the center-surround input channel pair for each sound event type being detected, and determining the number of sound events with an associated true differential steering angle of about 0 degrees to about −45 degrees (indicating a rear direction). In order to avoid a false designation of surround caused by inaccurate differential steering angles, the surround detection method also determines whether the number of sound events detected for reproduction in the rear exceeds a predetermined value in a defined time period. Generally, if just a few sound events intended for reproduction in the rear are detected (on the order of 2 or 3) during a time period of about 10 s to about 15 s, it is determined that the sound field or that particular portion of the sound field is intended for reproduction in surround. Additionally, if the duration of the sound events may be determined and only those sound events with durations that exceed a predetermined duration will be counted as sound events that are intended for reproduction in the rear. For example, sound events of very short duration, on the order of less than about 50 ms, often provide a poor indication of the intended direction of the sound field. Therefore, only sound events with durations that exceed about 50 ms will be counted as sound events intended for reproduction in the rear. In another example, only sound events with durations of about 200 ms to about 300 ms will be counted as sound events intended for reproduction in the rear. [0126]
  • 3. Methods for Localizing Sounds [0127]
  • The properties of the human hearing mechanism may also be used to determine the location of all sounds within a sound field. By combining the sound event detection and localization methods with known methods for determining the intended direction of sounds, methods have been developed that localize the individual sounds in the presence of background sounds more accurately for any combination of sounds (“sound localization methods”). The sound localization methods divide the sound field into sound events and non-sound events and localize the sound events in terms of a true differential steering angle or a true ordinary steering angle and the non-sound events in terms of a filtered ordinary steering angle. These methods essentially treat the non-sound events as a separate sound for which the direction is separately determined. The sound localization methods can be implemented to specifically localize any number and combination of sound event types, in addition to localizing the remaining sounds, in a sound field generated in any number of input channel pairs. In the following description, the sound localization methods are discussed in order of increasing complexity, with each subsequent sound localization method incorporating the steps of the prior methods, except as indicated. [0128]
  • Generally, in all the sound localization methods, the filtered OSA is used to indicate the direction of the sound field unless a sound event is detected, in which case the true DSA is used. In sound localization methods implemented for a sound field generated in multiple channel pairs (or in a single channel pair from which another channel pair can be generated), the true DSA is used to indicate the direction of sound events only if the DSA is determined to be accurate. In these cases, if the DSA is not found to be accurate, the true OSA is used to indicate the direction of the sound events. Additionally, no matter which type of steering angle (OSA or DSA) is selected, that type of steering angle is used for all input channels of the sound field. For example, if a sound event in the presence of noise is detected where the sound event is detected in only one input channel and the noise is present in every input channel, the DSA (which indicates the direction of the sound event) will be used (in some cases only if accurate) to define the direction of the entire sound field for the duration of the sound event. Fortunately, the human hearing mechanism, which generally perceives sound events as part of a foreground stream and noise as part of a background stream, will perceive only the sound event as having moved. [0129]
  • A sound localization method for detecting a single sound event type in a single input channel pair (“sound localization method for a single input channel pair and a single sound event type”) is shown in FIG. 13 and generally includes: determining the DSA and the [0130] trigger signal 1302; determining the OSA 1306; and determining which direction to use 1304. 1302, 1304 and 1306 are generally performed simultaneously and concurrently, for as long as a sound field is sensed.
  • Determining the DSA and the [0131] trigger signal 1302 is generally accomplished by performing a DSA determination method for a single input channel pair and a single sound event that includes an alternate beginning detection method to produce a single trigger signal for the input channel pair. Even though sound events only occur whenever a pulse is present in the trigger signal, the DSA may be continuously determined. Alternatively, the trigger signal may be continuously determined and the DSA determined only when a pulse is present in the trigger signal. The OSA is generally determined 1306 continuously using known methods. Determining which direction to use 1304 basically includes: determining when a sound event occurs, using the true DSA for the duration typical of the sound event type detected and decaying to the OSA at the end of the sound event. However, if at any time a subsequent sound event occurs (even during a sound event), the DSA for the subsequent sound event will be used for the duration typical of the subsequent sound event type.
  • A method for determining which direction to use (the “direction selection method for a single input channel pair and a single sound event type”) is shown in FIG. 14 and indicated by [0132] reference number 1304. It generally includes, determining whether there is an input signal 1402; where if there is an input signal determining whether there is a sound event 1404; where if there is a sound event, generating and selecting the current true DSA for the typical duration of the sound event type being detected 1406; determining whether the typical duration has ended 1408; where if the typical duration has not ended, determining whether a subsequent sound event is detected 1410, where if a subsequent sound event is not detected, repeating determining whether the typical duration has ended 1408 and whether a subsequent sound event has been detected 1410 until it is determined that either the typical duration has ended in 1408 or that a subsequent sound event has been detected in 1410; where if a subsequent sound event has been detected, defining the current DSA with that of the subsequent sound event 1414 and repeating 1406, 1408, 1410, and 1414 as appropriate; where if it is determined in 1408 that the typical duration has ended, repeating the entire method as appropriate from 1402; where if there is no sound event detected in 1404, determining whether there was an immediately preceding sound event 1416; where if there was no immediately preceding sound event, selecting the filtered OSA 1418; and repeating the entire method as appropriate from 1402; and if there was an immediately preceding sound event, selecting and decaying to the filtered OSA from the true DSA 1420; and repeating the entire method as appropriate from 1402; where the entire method is repeated as appropriate until there is no input signal detected in 1402, where if there is no input signal, stopping the method.
  • Determining whether there is an [0133] input signal 1402 includes determining whether the input power of the sound field in all input channels (“I2”) is greater than about zero. Additionally, it may also include determining whether the input power has dropped in all input channels by more than about 30 dB from a prior sound event. If it has, it can generally be assumed that the input signal has stopped. If there is an input signal, it is then determined whether there is a sound event 1404 by examining the trigger signal. Whenever the trigger signal contains a pulse or other indication of the occurrence of a sound event, a sound event exists. Conversely, whenever the trigger signal does not contain a pulse or other indication of the occurrence of a sound event no sound event exists.
  • If there is a sound event, then a current true DSA is generated from the current DSA using a true direction isolation method and selected for the typical duration of the sound event type being detected [0134] 1408. As previously discussed, if the sound event type being detected is a syllable, than the typical duration of the sound event will be about 50 ms to 200 ms (preferably about 150 ms) and if the sound event type being detected is an impulsive sound, than the typical duration will be about 50 ms. The DSA is selected and used for the typical duration regardless of when the sound event being detected actually ends. However, during the typical duration, the input signal is monitored to determine if any subsequent sound events with an accurate DSA are detected (1408 and 1410). If during the typical duration a subsequent sound event is detected, then the current DSA will be redefined by the DSA of the subsequent sound event 1414, the redefined current DSA will be used to generate a current true DSA which will be selected for the typical duration of the sound event type detected 1406, and 1408, 1410 and 1414 will be repeated as appropriate. However, if no subsequent sound event is detected during the typical duration (1408 and 1410), then the entire method repeats, as appropriate, from 1402.
  • In contrast, if it is determined in [0135] 1404 that there is no sound event, it is then determined whether there was an immediately preceding sound event 1416. If there was no immediately preceding sound event (a sound event for which the typical duration had just ended), a filtered OSA is selected, or continues to be selected 1418 and the process repeats, as appropriate from 1402. The filtered OSA is the OSA with the fluctuations having rise-times faster than a specified rise-time removed. For example, fluctuations with rise-times faster than approximately 300 ms may be removed. This prevents the OSA from reflecting rapid directional changes when no sound event is detected. If however, there was an immediately preceding sound event (a sound event for which the typical duration had just ended), the OSA is selected and decayed to from the true DSA of the immediately preceding sound event 1420. The decay helps to provide a smooth transition from DSA of the immediately preceding sound event to the OSA. The length of decay needed to provide this smooth transition depends on the sound event type of the immediately preceding sound event. If the immediately preceding sound event was a syllable, the decay will generally be about 300 ms seconds. The process then repeats, as appropriate, from 1402 until it is determined in 1402 that the input signal has ended.
  • In certain cases, particularly when sound events with a very short-duration are being localized, more accurate results are obtained if the direction selection method includes determining whether the short-duration sound event has actually ended at the end of the typical duration, and immediately selecting the OSA without any decay if it is determined that the short-duration sound event has not actually ended. Such a method (a “direction selection method for a single input channel pair and a short-duration sound event”) is shown in FIG. 15 and indicated by [0136] reference number 1500. The direction selection method for a single input channel pair and a short-duration sound event generally includes, determining whether there is an input signal 1502; where if there is an input signal, determining whether there is a short-duration sound event 1504; where if there is a short-duration sound event, generating and selecting the true DSA for the typical duration of the short-duration sound event detected 1506; determining whether the typical duration has ended 1508; where if the typical duration has not ended, determining whether a subsequent short-duration sound event is detected 1510, where if a subsequent short-duration sound event is not detected, repeating determining whether the typical duration has ended 1508 and whether a subsequent short-duration sound event has been detected until it is determined that either the typical duration has ended in 1508 or that a subsequent short-duration sound event has been detected in step 1510; where if a subsequent short-duration sound event has been detected, defining the current DSA with that of the subsequent short-duration sound event 1512 and repeating steps 1506, 1508, 1510 and 1512 as appropriate; where if it is determined in step 1508 that the typical duration has ended, repeating the entire method as appropriate from step 1502; where if there is no short-duration sound event detected in 1504, determining whether there was an immediately preceding short-duration sound event 1514; where if there was not an immediately preceding short-duration sound event, selecting or continuing to select the filtered OSA 1516; and repeating the entire method as appropriate from step 1502; and if there was an immediately preceding short-duration sound event, determining whether the immediately preceding short-duration sound event actually ended 1518; where if the immediately preceding short-duration sound event has actually ended, immediately going to the filtered OSA 1520; and repeating the entire method as appropriate from 1502; where if the immediately preceding short-duration sound event has not actually ended, selecting and decaying to the filtered OSA 1522 and repeating the entire method as appropriate from 1522; where the entire method is repeated as appropriate until there is no input signal detected in 1502, where if there is no input signal, stopping the method.
  • This direction selection method for a single input channel pair and a short-duration sound event is virtually the same as the previously discussed direction selection methods, except that the direction selection method for a single input channel pair and a short-duration sound event is implemented to detect short-duration sound events. Furthermore, at the end of a typical duration of a short-duration sound event (assuming no new sound event has or is occurring), a determination is made as to whether the sound event has actually ended [0137] 1514; and the filtered OSA is used either immediately 1520 or after a decay 1522 depending upon whether the short-duration sound event has actually ended. The direction selection method for a single input channel pair and a short-duration sound event is implemented to detect short-duration sound events by using a sound event detection method for a single channel and a single short-duration sound event for each input channel of the channel pair to produce a trigger signal (or two trigger signals) that indicates whenever a short-duration sound event is detected.
  • Whether the immediately preceding short-duration sound event has actually ended may be determined [0138] 1518 by comparing the power envelope and the accommodation signal in each input channel of the channel pair. If the input power envelope I2 is greater than the accommodation signal AccSig in any input channel of the input channel pair, it is determined that the short-duration sound event has not actually ended. Therefore, the filtered OSA is decayed to from the DSA of the immediately preceding sound event. If however, I2 is about equal to or less than AccSig in each input channel of the input channel pair, it is determined that the short-duration sound event has actually ended and the filtered OSA is immediately selected 1522. The length of decay needed to provide this smooth transition depends on the sound event type being detected. For example, if the immediately preceding sound event is an impulsive sound, the decay will take about 5 ms.
  • Sound localization methods may also be used to localize the sounds in a sound field when the sound field includes more than one input channel pair (“sound localization methods for multiple input channel pairs and a single sound event type”). The sound localization methods for multiple input channel pairs and a single sound event include the same basic steps as the sound localization method for a single input channel pair and a single sound event type, which generally include: determining the OSA; determining the DSA and trigger signal; and determining which direction to use. However, the step of determining the DSA and a trigger signal includes determining a DSA and a trigger signal for each input channel pair and is accomplished by performing a DSA determination method for multiple input channel pairs and a single sound event that includes an alternate beginning detection method. In addition, the step of determining which direction to use (the “direction selection method for multiple channel pairs and a single sound event type”) includes responding to differential signals, trigger signals and DSAs from any of the input channel pairs; and generating and selecting the true differential steering angle for all the input channel pairs when a sound event with an accurate DSA is detected in any input channel pair. [0139]
  • The direction selection method for multiple channel pairs and a single sound event type is shown in more detail in FIG. 16 and is indicated by [0140] reference number 1600. Generally, this direction selection method includes determining whether, at the time any trigger signal indicates the occurrence of a sound event, the DSAs are accurate; and if the DSAs are not accurate using the true OSAs for the typical duration of the sound event type being detected instead of the DSAs. More specifically, this direction selection method 1600 includes: determining whether there is an input signal 1602; where if there is an input signal determining whether there is a sound event 1604; where if there is a sound event, determining whether the current true DSA is accurate 1606; where if the current true DSA is accurate, generating and selecting the current true DSA for the typical duration of the sound event type being detected 1608; where if the current true DSA is not accurate, generating and selecting the current true OSA for the typical duration of the sound event type being detected 1620; once either the true OSA or DSA is generated and selected, determining whether the typical duration has ended 1610; where if the typical duration has not ended, determining whether a subsequent sound event is detected 1612, where if a subsequent sound event is not detected, repeating determining whether the typical duration has ended 1610 and whether a subsequent sound event has been detected 1612 until it is determined that either the typical duration has ended in 1610 or that a subsequent sound event has been detected in 1612; where if a subsequent sound event has been detected, determining whether the true DSA of subsequent sound event is accurate 1614; where if the true DSA of the subsequent sound event is accurate, defining the current DSA with that of the subsequent sound event 1616 and repeating 1610, 1612, 1614, 1616 and 1618 as appropriate; where if it is determined in 1614 that the true DSA of the subsequent sound event is not accurate, defining the current filtered OSA with that of the subsequent sound event 1618 and repeating 1610, 1612, 1614, 1616 and 1618 as appropriate; where if it is determined in 1610 that the typical duration has ended, repeating the entire method as appropriate from 1602; where if there is no sound event detected in 1604, determining whether there was an immediately preceding sound event 1622; where if there was no immediately preceding sound event, selecting or continuing to select the filtered OSA 1624; and repeating the entire method as appropriate from 1602; and if there was an immediately preceding sound event, selecting and decaying to the filtered OSA from the true DSA 1626; and repeating the entire method as appropriate from 1602; where the entire method is repeated as appropriate until there is no input signal detected in 1602, where if there is no input signal, stopping the method.
  • Determining whether there is an [0141] input signal 1602 includes determining whether there is an input signal in each input channel of each channel pair. Determining whether there is a sound event 1604 includes performing a sound event detection method for a single channel pair and a single sound event type for each channel pair to produce a trigger signal that indicates the occurrence of any sound events for each pair. Determining whether the current DSA is accurate 1606 includes determining whether the current DSA (the DSA determined at the moment any trigger signal indicates that a sound event is occurring) from every input channel pair is accurate using a method for determining DSA accuracy for multiple input channel pairs. A method for determining DSA accuracy for multiple input channel pairs is shown in more detail in FIG. 17, is indicated by reference number 1700, and includes: determining whether the background signal has dropped by about 3 dB or more in at least two input channels 1702; and where if the background signal has not dropped by about 3 dB or more in at least two input channels, determining whether all the current DSAs pass a consistency check 1704. Determining whether there the background signal has dropped by about 3 dB or more in at least two input channels is accomplished by performing a background signal check on all input channels. If the background signal has dropped by about 3 dB or more in at least two input channels, the DSA is not considered accurate. However, if background signal has not dropped by about 3 dB or more in at least two input channels, the consistency check is performed as previously described. If the DSAs pass the consistency check, they are considered accurate.
  • As shown in FIG. 16, if the DSAs are considered accurate, the current true DSAs are generated and selected for each [0142] channel pair 1608 using a true direction isolation method. However, if the DSAs are not considered accurate, the current true OSAs (the OSA determined at the moment any of the trigger signals indicates that the current sound event is occurring) are generated and selected for each channel pair 1620. Either the current true OSA or the current true DSA is used for the typical duration of the sound event type being detected, unless a subsequent sound event is detected during the typical duration. As long as it is determined that the typical duration has not ended in 1610, it is determined whether any subsequent sound events are detected 1612. If a subsequent sound event is detected, it is determined whether the true DSAs for the subsequent sound event are accurate 1614 using a method for determining DSA accuracy for multiple input channel pairs. If the DSAs of the subsequent sound event are determined to be accurate, the DSAs of the subsequent sound event become the current DSAs 1616 to reflect the direction of the subsequent sound event. However, if the DSAs of the subsequent sound event are determined not to be accurate, the OSAs of the subsequent sound become the current OSAs 1618 to reflect the direction of the sound event. The process repeats from 1608 or 1620, as appropriate, until the typical duration of any subsequent sound events ends.
  • Once any subsequent sound event ends, it is determined in [0143] 1622 whether there was an immediately preceding sound event. If there was an immediately preceding sound event, the filtered OSAs are selected and the direction of the sound field decays from that indicated by the DSAs to that indicated by the OSAs 1626. However, if there was no immediately preceding sound event, the filtered OSA is used, or continues to be used 1624. In addition, any of the sound localization methods for multiple channel pairs and a single sound event type may further include using the consistency check to reduce the occurrence of inaccurate DSAs through use of an accommodation adjustment method and/or an error threshold adjustment method as previously described.
  • In certain cases, particularly when sound events with a very short-duration are being localized, more accurate results are obtained if the direction selection method for multiple input channel pairs and a single sound event type includes determining whether the short-duration sound event has actually ended at the end of the typical duration, and immediately selecting the filtered OSA without any decay if it is determined that the short-duration sound event has not actually ended. Such a method (a “direction selection method for multiple input channel pairs and a short-duration sound event” is shown in FIG. 18 and indicated by [0144] reference number 1800. The direction selection method for a single input channel pair and a short-duration sound event generally includes: determining whether there is an input signal 1802; where if there is an input signal determining whether there is a short-duration sound event 1804; where if there is a short-duration sound event, determining whether the current true DSA is accurate 1806; where if the current true DSA is accurate, generating and selecting the current true DSA for the typical duration of the short-duration sound event being detected 1808; where if the current true DSA is not accurate, generating and selecting the current true OSA for the typical duration of the short-duration sound event being detected 1820; once either the true OSA or DSA is generated and selected, determining whether the typical duration has ended 1810; where if the typical duration has not ended, determining whether a subsequent short-duration sound event is detected 1812, where if a subsequent sound event is not detected, repeating determining whether the typical duration has ended 1810 and whether a subsequent sound event has been detected 1812 until it is determined that either the typical duration has ended in 1810 or that a subsequent sound event has been detected in 1812; where if a subsequent sound event has been detected, determining whether the true DSA of subsequent sound event is accurate 1814; where if the true DSA of the subsequent sound event is accurate, defining the current DSA with that of the subsequent sound event 1816 and repeating 1810, 1812, 1814, 1816 and 1818 as appropriate; where if it is determined in 1814 that the true DSA of the subsequent sound event is not accurate, defining the current OSA with that of the subsequent sound event 1818 and repeating 1810, 1812, 1814, 1816 and 1818 as appropriate; where if it is determined in 1810 that the typical duration has ended, repeating the entire method as appropriate from 1802; where if there is no sound event detected in 1804, determining whether there was an immediately preceding sound event 1622; where if there was no immediately preceding sound event, selecting or continuing to select the filtered OSA 1824; and repeating the entire method as appropriate from 1802; and if there was an immediately preceding sound event, determining if the immediately preceding sound event has actually ended 1825; where if the immediately preceding sound event has actually ended, selecting and immediately going to the filtered OSA from the true DSA 1826; and repeating the entire method as appropriate from 1802; where if the immediately preceding sound event has not actually ended, selecting and decaying to the filtered OSA 1828 and repeating the entire method as appropriate from 1802; where the entire method is repeated as appropriate until there is no input signal detected in 1802, where if there is no input signal, stopping the method.
  • This direction selection method for multiple input channel pairs and a short-duration sound event is virtually the same as the previously discussed direction selection method for multiple input channels and a single sound event type, except that the direction selection method for multiple input channel pairs and a short-duration sound event is implemented to detect short-duration sound events. Furthermore, at the end of a typical duration of a short-duration sound event (assuming no new sound event has or is occurring), a determination is made as to whether the sound event has actually ended [0145] 1825; and the filtered OSA is used either immediately 1826 or after a decay 1828 depending upon whether the short-duration sound event has actually ended. The direction selection method for multiple input channel pairs and a short-duration sound event is implemented to detect short-duration sound events by using a sound event detection method for multiple input channels and a single short-duration sound event for each input channel of the channel pair to produce a trigger signal (or two trigger signals) that indicates whenever a short-duration sound event is detected.
  • Whether the immediately preceding sound event has actually ended is determined [0146] 1825 by determining whether the input power envelope is greater than the accommodation signal for the short-duration sound event in any input channel. If the input power envelope is greater than the accommodation signal for the short-duration sound event in any input channel, it is determined that the short-duration sound event has not actually ended. Therefore, a decay is made from the DSAs of the immediately preceding short-duration sound event to the filtered OSA 1826. If however, in each input channel, the power envelope is about equal to the accommodation signal, it is determined that the short-duration sound event has actually ended and the filtered OSA is immediately selected 1828. The length of decay needed to provide this smooth transition depends on the sound event type being detected. For example, if the immediately preceding sound event is an impulsive sound, the decay will take about 5 ms.
  • Sound localization methods may also be used to localize the sounds in a sound field by distinguishing more than one sound event type (“sound localization methods for a single input channel pair and multiple sound event types”). The sound localization methods for a single input channel pair and multiple sound event types include the same basic steps as the sound localization method for a single input channel pair and a single sound event type, which generally include: determining the OSA; determining the DSA and trigger signal; and determining which direction to use. However, determining the DSA and the trigger signal includes determining the DSA and trigger signal for each sound event type by performing a DSA determination method for a single input channel pair and multiple sound event types that uses an alternate beginning detection method. In addition, determining which direction to use includes performing a direction selection method for a single input channel pair (either for a single sound event type or a short-duration sound event) in parallel for each sound event type being detected. However, whenever a sound event of any type is detected in any channel pair, the DSA is used in each input channel pair for the typical duration of the sound event type that was actually detected. [0147]
  • Sound localization methods may also be used to localize the sounds in a sound field with more than one input channel pair by distinguishing more than one sound event type (“sound localization methods for multiple input channel pairs and multiple sound event types”). The sound localization methods for multiple channel pairs and multiple sound event types include the same basic steps as the sound localization method for multiple input channel pairs and a single sound event type, which generally includes: determining the OSA; determining the DSA and trigger signal for each channel pair; and determining which direction to use according to a direction selection method for multiple input channel pairs and a single sound event type. However, in the method for detecting multiple sound events, determining the DSA and trigger signal for each channel pair includes determining a DSA and trigger signal for each sound event type in each input channel pair. Determining a DSA and trigger signal for each sound event type in each input channel pair is accomplished by performing a DSA determination method for multiple input channel pairs and multiple sound event types that includes an alternate beginning detection method. In addition, performing the direction selection method for multiple input channel pairs and multiple sound event types in parallel for each sound event includes, responding to the trigger signals from any sound event type; generating and selecting the true DSA for all the input channel pairs if a sound event of any type with an accurate DSA is detected in any input channel pair; or selecting the true OSA for all the input channel pairs if a sound event of any type with an inaccurate DSA is detected in any channel. Alternatively, the sound localization methods for multiple input channel pairs and multiple sound event types may also include using the accommodation adjustment methods and/or the error threshold adjustment methods as previously described. [0148]
  • Determining which direction to use in this manner works well for simple and more complicated sound events. For example, consider a sound field that contains popular music which includes prominent drum hits, where the bulk of the music is originates from the front (the center input channel) and the drum hits originate from the rear (the surround input channel). The direction of the music will be defined by the filtered OSA as localized in the center input channel (assuming the music contains no other sound events). However, when a drum hit occurs, the method will detect an impulsive sound and (assuming the DSA is accurate) will use the direction indicated in the true DSA to move the entire sound field to the rear. At the end of the typical duration of an impulsive sound, the method will determine if the drum hit has actually ended, and if it has, the method will immediate revert to using the direction indicated by the filtered OSA and move the entire sound field back to the center input channel. When the direction is determined in this manner, the human hearing mechanism will perceive the drum hit as originating from the rear and the music as continually originating from the front as if the music had never moved. If however, the DSA is used for longer than the typical duration of the drum hit or if at the end of the typical duration a decay is used to revert to the filtered OSA when the drum hit actually ends prior to the end of the typical duration, the entire sound field, including the music will be perceived as having moved to the rear. [0149]
  • In a more complicated example, consider a sound field that contains a speech sound that has a very sharp attack, such as a shout, in the presence of ambient noise. This speech sound is complex because it contains both an impulsive sound (the attack) and a syllable (the remainder of the shout). The beginning of the shout will be detected as an impulse, and assuming the current impulsive DSA is correct, the impulsive DSA will be selected for the typical duration of an impulsive sound, however, either during the typical duration or immediately after, the syllable portion of the shout will be detected and assuming it is correct, the syllable DSA will be selected and used for the typical duration of a syllable. Because at the moment the syllable is detected, the DSA of the previously detected impulsive sound will equal that of the syllable, no change in direction will occur. Therefore, the direction indicated in the sharp onset of the shout will be quickly captured according to its impulsive nature and the direction will be used for time characteristic of its syllabic nature. [0150]
  • 4. Sound Event Device [0151]
  • The sound event detection methods, sound event detection and localization methods and sound localization methods and any methods included in any of these methods may be implemented in a sound event device as shown in FIG. 19 and indicated as [0152] reference number 1900. The optimization device 1900 generally includes a detection unit 1902 and may also include an interface unit 1904. The detection unit 1902 includes a processor 1908 coupled to a memory device 1906. The memory device 1908 may be any type of fixed or removable digital storage device and (if needed) a device for reading the digital storage device including, floppy disks and floppy drives, CD-ROM disks and drives, optical disks and drives, hard-drives, RAM, ROM and other such devices for storing digital information. The processor 1908 may be any type of apparatus used to process digital information. The memory device 1906 may store the sound field and at least one of the following methods: the sound event detection methods, sound event detection and localization methods and sound localization methods and any methods included in any of these methods (collectively, the “detection and/or localization methods”). Upon the relevant request from the processor 1908 via a processor signal 1910, the memory communicates one of the detection and/or localization methods, and if necessary the sound field, via a memory signal 1912 to the processor 1908. The processor 1908 then performs the detection and/or localization method.
  • The [0153] interface unit 1904 generally includes an input device 1914 and an output device 1916. The output device 1916 is any type of visual, manual, audio, electronic or electromagnetic device capable of communicating information from a processor or memory to a person or other processor or memory. Examples of output devices include, but are not limited to, monitors, speakers, liquid crystal displays, networks, buses, and interfaces. The input device 1914 is any type of visual, manual, mechanical, audio, electronic, or electromagnetic device capable of communicating information from a person or processor or memory to a processor or memory. Examples of input devices include keyboards, microphones, voice recognition systems, trackballs, mice, networks, buses, and interfaces. Alternatively, the input and output devices 1914 and 1916, respectively, may be included in a single device such as a touch screen, computer, processor or memory coupled to the processor via a network. The sound field may be communicated to the memory device 1918 from the input device 1914 through the processor 1920. Additionally, the optimized model parameters may be communicated from the processor 1920 to the output device 1916.
  • 6. Sound Event Detector [0154]
  • The properties of the human hearing mechanism have been modeled to create electronic circuitry that detects sound events in the presence of a background signal (collectively, “sound event detectors”). Sound event detectors may be used for a variety of applications for which the detection of specific types of sound events is helpful. For example, a sound event detector that detects syllables may be used as part of a phonic detector or as part of a speech recognition or speech coding system. In another example, a sound event detector that detects syllables may be used in conjunction with an audio amplification device, such as a microphone. This allows the microphone to remain off until a syllable is detected from a speaker, thus preventing the microphone from amplifying undesired sounds and feedback through the microphone itself when the speaker is silent. Sound event detectors, like the sound event detection methods, generally produce a trigger signal that includes some flag or marker, such as a pulse, that indicates the occurrence of a sound event. Also, like the sound event detection methods, the sound event detectors can be implemented to detect any number and combination of sound event types in a sound field that is generated in any number of input channels. In the following description, the sound event detectors are discussed in order of increasing complexity, with each subsequent sound detector incorporating the elements of the prior sound event detectors, except as indicated. [0155]
  • One example of a sound event detector that detects a single type of sound event in a sound field generated in only one input channel (a “sound event detector for a single input channel and a single sound event type” is shown in FIG. 20 and indicated by [0156] reference number 2000. However, the sound event detector and localizer may be implemented to detect any number of sound event types in any number of input channel pairs. In the present example, the entire sound field is generated through a left input channel. The term “left” as used in this example, does not have any directional meaning because the entire sound field is contained in a single input channel and is used simply for the purposes of explanation. A sound event detector for a single input channel and a single sound event type 1900 generally includes a frequency bias filter 2001; an accommodation circuit for a single input channel 2002; and a trigger generation circuit for a single input channel 2004. In general, the accommodation circuit for a single input channel 2002 uses the sound field to produce a differential signal in the left input channel “Lo” for the sound event type being detected and the trigger generation circuit for a single input channel 2004 uses the accommodation signal Lo to produce a trigger signal “Tl” indicating whenever a sound event of the type being detected is detected.
  • The [0157] frequency bias filter 2001 models the frequency bias of the human hearing mechanism by emphasizing frequencies in the sound field from about 500 Hz to about 4000 Hz. The accommodation circuit for a single input channel 1702 separates sound events from any background signals in the sound field by modeling accommodation. This circuit 2102 is shown in more detail in FIG. 21 and generally includes: a multiplier 2102; a low-pass filter 2104; and an accommodation signal circuit 2006. The multiplier 2002 converts the sound field as generated in the left input channel (the “input signal” into a power signal “Lin2.” The input signal is generally a voltage signal and can generally be converted into a power signal by squaring the input signal. The resulting power signal Lin2 includes many fluctuations, some of which indicate sound events, and some of which indicate noise. In order to reduce the fluctuations due to noise, the low-pass filter then removes the fluctuations with rise times faster than about 30 ms from the power signal Lin2 to produce a filtered power signal L2. This low-pass filter 2104 may be of any type, such as a filter with a roll-off of 12 dB/octave.
  • It is from the filtered power signal L[0158] 2 that the accommodation signal circuit 2106 creates and subtracts the accommodation signal (which represents the long-term average power in the sound field as it is) to create the differential signal Lo. The accommodation circuit 2106 generally includes an operational amplifier 2108; a resistor 2114; a diode 2112 and a capacitor 2110. The filtered power signal L2 is coupled to the positive terminal of the operational amplifier 2108 or, alternatively any device capable of determining a difference between two signals. When the filtered power signal L2 does not include a sound event, capacitor 2110 will act as an open circuit resulting in an accommodation signal {overscore (L)}2 that is about equal the filtered power signal L2 to produce a differential signal Lo about equal to zero.
  • However, when the filtered power signal L[0159] 2 does contain a sound event, the filtered power signal L2 will rapidly increase according to the rise time of the sound event. This rapid increase in L2 will cause a corresponding spike in the differential signal Lo. After the rise-time of the sound event, capacitor 2110 will charge causing the accommodation signal {overscore (L)}2 to gradually rise according to time constant defined by resistor 2114 and capacitor 2110. This time constant is generally made equal to the accommodation rate of the human hearing mechanism, determined experimentally to be about 300 ms. {overscore (L)}2 will continue to rise until the voltage across capacitor 2110 (and thus {overscore (L)}2) equals L2 or until the sound event ends or starts to decay. This increasing {overscore (L)}2 is subtracted from L2 resulting in a gradual decay in Lo that models the accommodation property of the human hearing mechanism. When the sound event decays, the power envelope L2 will quickly drop according to the fall-time of the sound event. As L2 approaches the value of {overscore (L)}2 , capacitor 2110 will discharge through diode 2112 preventing the accommodation signal {overscore (L)}2 from exceeding the filtered power signal L2. This prevents the production of a negative pulse in Lo at the end of the sound event.
  • The differential signal Lo, therefore, includes a series of fluctuations with on-times equal to or less than those characteristic of the sound event type being detected and with fall-times defined by the accommodation signal, and/or the end of the sound event. The trigger generation circuit for a single input channel [0160] 1904 (shown in FIG. 19) then detects the sound events to produce a trigger signal Tl′ that includes a pulse whenever a sound event is detected. A trigger generation circuit 1904 for a single input channel is shown in more detail in FIG. 22 and includes: a high-pass filter 2202; a normalization circuit 2206 and a low-pass filter 2208.
  • The goal of the [0161] trigger generation circuit 1904 is to remove as many fluctuations caused by noise as possible and to deemphasize those that are not removed. As explained previously in connection with the sound event detection methods, this is accomplished by removing the fluctuations that have frequencies higher than those characteristic of the sound event being detected and by normalizing the differential signal Lo with the short-term high frequency power in the differential signal. The normalization is accomplished using an automatic gain control circuit which includes the high-pass filter 2202 and the normalization circuit 2206. The high-pass filter includes a capacitor/resistor pair that defines the cutoff frequency as that which is characteristic of the sound event being detected. Additionally, a rectifier (not shown) may be included between the high-pass filter 2202 and the normalization circuit 2206 to rectify any negative pulses or fluctuations. The normalization circuit 2206, which includes an integrator 2210 and a divide by circuit 2212, then averages the high-frequency component of Lo over a short time period defined by the integrator 2210. The short time period defined by the integrator may equal about 160 ms, however, this time period may be adjusted as a function of the type of sound field. The divide-by circuit 2212 then divides Lo by the averaged HFl to yield the normalized differential signal Nl. Additionally, a rectifier (not shown) may be included between the normalization circuit 2206 and the low-pass filter 2208 to rectify any negative pulses or fluctuations.
  • The normalized differential signal Nl is then filtered by a low-[0162] pass filter 2208 to remove fluctuations with frequencies higher than are characteristic of the sound event being detected to yield a filtered normalized differential signal Nl′. Although not shown, additional noise can be removed from Nl′ by including circuitry that detects and removes fluctuations that occur more often than is characteristic of the sound event being detected and that remove any fluctuations that occur when a decrease in the sound field of at least 10 dB is detected. Nl′, therefore, includes a series of pulses of varying amplitudes representing the occurrence of sound events and fluctuations due to noise.
  • In order to detect the sound events from the noise in the filtered normalized differential signal Nl′, a [0163] threshold detector 2218 detects only those pulses with an amplitude greater than a threshold. This helps to distinguish pulses indicating sound events from fluctuations due to noise. The output of the threshold detector is a trigger signal “Tl” that indicates, generally by pulses, the occurrence of a sound event in the sole (left) input channel of the sound field. Alternatively, the sound event detector for a single input channel and a single sound event type may also include a threshold adjustment circuit. The threshold adjustment circuit adjusts the threshold of the threshold detector in order to adjust the sensitivity of the sound event detector. The threshold detector may allow manual adjustment of the threshold and may include a voltage source and a variable resistor coupled to the threshold detector in the trigger generation circuit. The resistance of the resistor may be manually controlled by a knob or switch or other such device to control the voltage supplied by the voltage s ource t o the threshold d etector which is used by the threshold detector to define the threshold. Alternatively, the threshold detector provides automatic adjustment of the threshold and includes a counter coupled to the output of the trigger generation circuit and a comparator coupled to the counter and the threshold detector in the trigger generation circuit. The counter counts the number of sound events that occur in a specified time period and communicates this number to the comparator. This specified time period is generally on the order of about a few seconds. The comparator then produces a voltage which is inversely proportional to the number of sound events and communicates this voltage to the threshold detector which uses the voltage to define the threshold. Generally, the threshold is increased when there are more sound events detected so that the sensitivity of the sound detector is increased.
  • Any of the sound event detectors may not include an accommodation signal circuit, which is of particular use when detecting short-duration sound events, such as impulsive sounds. An example of a sound event detector that does not include an accommodation signal circuit (the “sound event detectors for short-duration sound events” is shown in FIG. 23. This sound event detector for short-[0164] duration sound events 2300 includes: a frequency bias filter 2301; a linear to dB converter 2302; a high-pass filter 2303; and a trigger generation circuit for a single channel pair 2304. The frequency bias filter 2301 emphasizes the frequencies in the input signal from about 500 Hz to about 4000 Hz to produce a filtered input signal. The filtered input signal is then converted into decibels by the linear to dB converter 2302 to produce a decibel filtered input signal. The filtered decibel input signal is again filtered, this time by the high-pass filter 2303. The high-pass filter 2303 removes any fluctuations with rise-times slower than are characteristic of the short-duration sound event being detected. The trigger generation circuit for a single input channel pair 2304 then uses this twice filtered decibel input signal to create a trigger signal that indicates the occurrence of short-durations sound events. Alternately, the sound event detectors for short-duration sound events may also include a threshold adjustment circuit.
  • Sound event detectors may also be implemented when the sound field is generated in two or more input channels. A sound event detector that detects a single sound event type in a sound field generated in two input channel (a “sound event detector for multiple input channels and a single sound event type” may include a sound event detector for a single channel and a single channel pair for each input channel that produces a trigger signal for each input channel. Alternatively, the trigger signals may be combined to form a single trigger signal that indicates the occurrence of a sound event in any input channel. Alternatively, a sound event detector for multiple input channels and a single sound event type may produce only a single trigger signal for each channel pair from a difference signal. Such a sound event detector for multiple input channels and a single sound event type is shown in FIG. 24. In this example, the entire sound field is either generated through a single input channel pair including a left input channel and a right input channel. However, this method is applicable for any number of input channels or input channel pairs. [0165]
  • The sound event detector for multiple input channels and a single [0166] sound event type 2400 includes: a first accommodation circuit for a single input channel 2402; a second accommodation circuit for a single input channel 2404; and an alternate trigger generation circuit for a single input channel pair 2406. The first and second accommodation circuits 2402 and 2404, respectively, are generally identical. The first accommodation circuit 2402 uses the left input channel signal (“Lin” to produce a differential signal for the left input channel (the “left differential signal” or “Lo”). The second accommodation circuit 2404 uses the right input channel signal (“Rin” to produce a differential signal for the right input channel (the “right differential signal” or “Ro”).
  • The alternate trigger generation circuit for a single [0167] input channel pair 2406 uses both the right and left differential signals to produce a trigger signal that indicates the occurrence of sound events in either input channel (the “left-right trigger signal” or “Tlr”). The trigger generation circuit for a single input channel pair 2406 is shown in more detail in FIG. 25 and includes: an operational amplifier 2501; a rectified high-pass filter 2502; a
  • rectified [0168] normalization circuit 2506; a low-pass filter 2508 and a threshold detector 2510. Although similar to the trigger generation circuit for a single input channel (as shown in FIG. 22), the trigger generation circuit for a single input channel pair 2406 also includes an operational amplifier 2501 (or other device capable of determining a difference) that creates a signal equal to the difference between the left and right differential signal (the “left-right difference signal” or “Lo−Ro” and uses the left-right difference signal to create a trigger signal for the left-right input channel pair (the “left-right trigger signal” or Tlr”). Although in this example, the left-right difference signal is obtained by subtracting Ro from Lo, it may alternatively be determined by subtracting Lo from Ro. The left-right difference signal Lo−Ro includes a series of pulses and other fluctuations that indicated the occurrence of sound events and noise in either input channel of the input channel pair. However, because sound events that occur equally in both input channels are removed, many fluctuations due to noise are thus removed. The pulses and fluctuations in the difference signal may have a positive or a negative amuplitude depending on whether the power in the transient is greater in the left input channel or the right input channel, respectively.
  • In order to produce a trigger signal that includes only positive pulses, a [0169] first rectifier 2504 is included in the rectified high-pass filter 2502 and a second rectifier is included in the normalization circuit 2506. The rectified high-pass filter 2502 produces a rectified high-pass left-right difference signal (“HFlr”). This rectified high-pass left-right difference signal is used by the normalization circuit 2506 to normalize the difference signal Lo−Ro, the result of which is rectified by the second rectifier 2507 to produce a normalized left-right signal (“Nlr”). The low-pass filter 2508 removes fluctuations due to noise with rise-times faster than those characteristic of the sound event being detected to produce a filtered normalized left-right signal (“Nlr′”). Just as in the trigger generation circuit for a single input channel, (not shown) additional noise can be removed from Nlr′ by including circuitry that detects and removes fluctuations that occur more often than is characteristic of the sound event being detected and that remove any fluctuations that occur when a decrease in the sound field of at least 10 dB is detected. Nlr′, therefore, includes a series of pulses of varying amplitudes representing the occurrence of sound events and fluctuations due to the remaining noise. The threshold detector 2510 then detects the sound events as those pulses that have an amplitude greater than a threshold to create the left-right trigger signal Tlr. This sound event detector may be repeated in parallel for multiple input channel pairs to produce a trigger signal for each input channel pair. Alternatively, this sound event detector for multiple input channels and a single sound event may also include a threshold adjustment circuit for each trigger generation circuit. The sound event detectors for multiple input channels and a single sound event that include an alternate trigger generation circuit may also be referred to as “sound event detectors for a single channel pair and a single sound event type.” Additionally, any of the sound event detectors for multiple input channels (or a single input channel pair) and a single sound event may include a threshold adjustment circuit for each trigger generation circuit.
  • Sound event detectors may also be implemented so that more than one type of sound event is detected. These “sound event detector for a single input channel and multiple sound event types” generally include a sound event detector for a single input channel and a single sound event type implemented in parallel for each sound event type being detected to produce a trigger signal for each sound event type being detected. An example of such a sound event detector for a single input channel and multiple sound event types is shown in FIG. 26 and indicated by [0170] reference number 2600. In the example, the entire sound field is generated through a left input channel. The term “left” as used in this example, does not have any directional meaning because the entire sound field is contained in a single input channel and, in fact, the input channel can be given any designation. Also, in this example, the sound event detector for a single input channel and multiple sound event types 2600 is implemented to detect syllables and impulsive sounds. However, any number or combination of sound events may be detected.
  • In this example, the sound event detector for a single input channel and multiple [0171] sound event types 2600 generally includes: a accommodation circuit for a s ingle input channel 2602; a trigger generation circuit for a single input channel implemented for syllables 2604; and a sound event detector for a single input channel and a single short-duration sound event implemented for impulsive sounds 2606. The accommodation circuit for a single input channel 2602 uses the input signal Lin to produce a differential signal. The trigger generation signal for a single input channel implemented to detect syllables 2604 uses the differential signal to produce a trigger signal that indicates the occurrence of syllables in the sole input channel (left) of the sound field (“Tl(s)”). The trigger generation signal for a single input channel implemented to detect syllables 2604 includes filters (see FIG. 22) for which the cut-off frequency rise time is about 33 ms. The trigger generation circuit for a single input channel implemented to detect impulsive sounds 2606 uses the input signal Lin to produce a trigger signal that indicates the occurrence of impulsive sounds in the sole input channel (left) of the sound field (“Tl(i)”). This trigger generation circuit for a single input channel implemented to detect impulsive sounds 2606 includes a high-pass filter (see 2303 in FIG. 22) for which the cut-off rise-time is about 3 ms. Alternatively, the sound event detectors for multiple input channels and a single sound event may include a threshold adjustment circuit for each trigger generation circuit.
  • Sound event detectors may also be implemented so that more than one type of sound event is detected in more than one input channel. These “sound event detectors for multiple input channels and multiple sound event types” may produce a trigger signal for each sound event type in each input channel pair. Alternatively, trigger signals in each channel pair may be combined in almost any manner to reduce the number of trigger signals. An example of such a sound event detector for multiple input channels and multiple sound event types is shown in FIG. 27 and indicated by [0172] reference number 2700. In this example, the entire sound field is either detected in or reproduced through a left and a right input channel. However, the method may be implemented for any number and combination of input channels. Additionally, in this example, the sound event detector for multiple input channels and multiple sound event types is implemented to detect syllables and impulsive sounds. However, any number or combination of sound events may be detected.
  • The sound event detector for multiple input channels and multiple [0173] sound event types 2700 generally includes, a first accommodation circuit for a signal input channel 2702; a second accommodation circuit for a signal input channel 2706; a first sound event detector for a single channel and a short-duration sound event 2708; an alternate trigger generation circuit for a single input channel pair 2710; and a second sound event detector for a single channel and a short-duration sound event 2712.
  • The first and second accommodation circuits for a [0174] single channel 2702 and 2703, respectively, are identical to each other. The first accommodation circuit for a single channel 2702 produces a differential signal for the left input channel Lo. The second accommodation circuit for a single channel 2703 produces a differential signal for right input channel Ro. The alternative trigger generation circuit for a single channel pair 2718 uses Lo and Ro to produce a trigger signal that indicates the occurrence of syllables in the left-right channel pair Tlr(s). The alternative trigger generation circuit for a single channel pair 2718 (shown in more detail in FIG. 25, indicated by reference number 2406) includes filters with a cut-off rise-time defined at about 33 ms. The first and second sound event detector for a single channel and a single short- duration sound event 2708 and 2712, respectively, use Lin and Rin, respectively, to produce a trigger signal indicating the occurrence of impulsive sounds in the left and right input channels, respectively. These trigger signals may then be combined to produce a single trigger signal indicating the occurrence of trigger signals in the left-right input channel pair. Both the first and second sound event detector for a single channel and a single short- duration sound event 2708 and 2712, respectively, include a high-pass filter (see 2303 in FIG. 23) with a cut-off rise-time of about 3 ms.
  • Alternatively, the sound event detector for multiple input channels and multiple sound event types includes a sound event detector for a single input channel and a single sound event implemented in parallel for each sound event type in each input channel. This sound event detector produces a trigger signal for each sound event in each input channel. Alternatively, the sound event detector for multiple input channels and multiple sound event types may include a sound event detector for a single input channel a multiple sound event types implemented in parallel for each input channel. This sound event detector for multiple input channels and multiple sound event types also produces a trigger signal for each sound event in each input channel. Alternatively, any of the sound event detectors for multiple input channels and multiple sound event types may also include a threshold adjustment circuit for each trigger generation circuit. [0175]
  • 7. Sound Event Detectors and Localizers [0176]
  • The properties of the human hearing mechanism have been modeled to create electronic circuitry that detects and localizes sound events in the presence of a background signal (collectively, “sound event detectors and localizers”). Sound event detectors and localizers, like the sound event detection and localization methods, determine the direction of sound events in relation to one or more input channel pairs in terms of a differential steering angle or an ordinary steering angle and in some cases, verify the accuracy of the steering angles. Also, like the sound event detection and localization methods, the sound event detectors and localizers can be implemented to detect any number and combination of sound event types in a sound field generated in any number of input channels. In the following description, the sound event detectors and localizers are discussed in order of increasing complexity, with each subsequent sound detector and localizer incorporating the elements of the prior sound event detectors and localizers, except as indicated. [0177]
  • One example of a sound event detector and localizer implemented to detect a single sound event type in a single input channel pair is shown in FIG. 28 (a “sound event detector and localizer for a single input channel pair and a single sound event type”). In this example, sound events are detected and localized with respect to a right input channel and a left input channel. However this method is applicable to any combination of input channels with left and right being used here solely for the purposes of explanation. The sound event detector and localizer for a single input channel pair and a single [0178] sound event type 2800, shown in FIG. 28 generally includes: a sound event detector for a single input channel and a single sound event type 2804; and a sound event localization circuit for a single input channel pair and a single sound event type 2806.
  • The sound event detector for a single input channel pair and a single [0179] sound event type 2804 includes any of the sound event detectors for multiple input channels and a single sound event, which includes a trigger generation circuit for a single input channel pair and is implemented for whatever sound event is being detected. The sound event detector for a single input channel pair and a single sound event type 2804 uses the left input signal Lin and a right input signal Rin to produce an differential signal for the left input channel Lo, a differential signal for the right input channel Ro, and a trigger signal indicating the occurrence in either input channel of whatever sound event is being detected Tlr. The sound event localization circuit for a single input channel pair and a single sound event type 2806 then uses Lo, Ro and Tlr to produce a true differential steering angle indicating the direction of the detected sound events relative to the right and left input channel dlr′.
  • The sound event localization circuit for a single input channel pair and a single [0180] sound event type 2806 is shown in more detail in FIG. 29 and generally includes a DSA circuit 2904, a switch 2518; a resistor 2906; a capacitor 2908; and a control circuit 2910. The DSA circuit 2904 uses the left accommodation signal Lo and the right accommodation signal Ro to determine the differential steering angle dlr. The DSA circuit 2904 includes a first linear to decibel circuit 2912; a second linear to decibel circuit 2914; an operational amplifier 2916 and a decibel to equivalent angle circuit 2918. The first and second linear to decibel circuits 2912 and 2914, respectively, convert the left and right accommodation signals, respectively, from a power signal into a decibel signal. The operational amplifier 2916 (or alternatively, any circuit that can determine a difference) determines the ratio between the left and right decibel signals by determining the difference between the two signals. This ratio is then converted into an equivalent angle by the decibel to equivalent angle circuit 2918 to produce the differential steering angle dlr.
  • The [0181] control circuit 2910, the switch 2906 and the capacitor 2908 generally form a sample-and-hold circuit and can therefore be replaced with any device or circuit that performs a similar function. In general, when a sound event is detected, the control circuit 2910 causes the switch 2906 to close and the capacitor 2908 to capture dlr during the rise time of the sound event to produce the true DSA (“dlr′”) for the typical duration of the sound event type being detected. More specifically, the control circuit 2910 receives the trigger signal Tlr and produces a control signal “con” that controls switch 2906. Switch 2906 is a two position switch and is generally in position C when no sound events are detected. When in position A, the switch 2906 is closed, when in position B, the switch is open, and when in position C, the switch 2906 is grounded. When the control circuit 2910 receives an indication from trigger signal Tlr that a sound event is occurring, it communicates to switch 2906 via “con” a command to close (go to position A). In response, the switch 2906 closes. At the end of the rise-time typical for the sound event type being detected, the control circuit 2910 communicates to switch 2906 via con a command to open (go to position B). After the typical duration of the sound event type being detected, the control circuit 2910, communicates to switch 2906 a command via con to go to ground (go to position C). During the time that switch 2906 is closed (the rise-time of the sound event), the left-right differential steering angle dlr is captured by the capacitor 2908 to create the true DSA dlr′. The true DSA is held until the end of the typical duration of the sound event, even if the sound event has not actually ended. For example, if the sound event being detected is a syllable, the true DSA will be held for about 50 ms to about 200 ms, preferably after about 150 ms. In another example, if the sound event being detected is an impulsive sound, the true DSA will be held for about 50 ms. At the end of the typical duration of the sound event type being detected, the capacitor is ground through switch C causing the voltage held by capacitor 2908 and thus, dlr′ to go to zero. The capacitor 2908 is chosen so that it can sufficiently capture dlr during the rise time of the sound event being detected. For example, if the sound event being detected is a syllable, the capacitor must be able to capture dlr in 20 ms to about 30 ms. In another example, if the sound event being detected is an impulsive sound, the capacitor must be able to capture dlr in about 5 ms.
  • Alternatively, the sound event detector and localizer for a single input channel pair and a single sound event type is optimized for sound events with very short durations such as impulsive sounds. In some cases, it is very difficult to obtain an accurate DSA for short-duration sound events therefore it is beneficial to use an average of the directions of all the short-duration sound events that occur in a given time period (generally, about 3 ms) as the direction for all the short-duration sound events in that time period. Therefore, this optimized sound event detector and localizer for a single input channel pair and a single sound event type further includes a circuit for determining the average DSA of all sound events detected in an about 3 ms time frame (the “DSA averaging circuit”). The DSA averaging circuit is generally implemented in the sound [0182] event localization circuit 2806 between the DSA circuit 2804 and the switch 2906.
  • The sound event detector and localizer may also be implemented to detect a single sound event type in a sound field generated in multiple input channel pairs (a “sound event detector and localizer for multiple input channel pairs and a single sound event type”). An example of a sound event detector and localizer for multiple input channel pairs and a single sound event type implemented to detect and localize a single sound event in both a right-left input channel pair (“LR input channel pair” and a center-surround input channel pair (“CS input channel pair” is shown in FIG. 30 and designated by [0183] reference number 3000. This detector and localizer may be implemented for any combination of input channel pairs with the LR input channel pair and the CS input channel pair used for only for the purposes of example. The sound event detector and localizer for multiple input channel pairs and a single sound event type 3000 produces a true differential steering angle for the LR input channel pair (dlr′) and the CS input channel pair (dcs′) and generally includes: a first and second sound event detector for a single input channel pair and a single sound event 3010 and 3012, respectively; and a sound event localization circuit for multiple input channel pairs and a single sound event type 3014.
  • The first and second sound event detectors for a single input channel pair and a single [0184] sound event type 3010 and 3012 are both implemented to detect the same sound event. The first sound event detector for multiple input channels 3010 uses the input signals in the left and right input channels, Lin and Rin, respectively, to produce a left differential signal Lo, a right differential signal Ro, a left power envelope L2, a right power envelope R2, and a left-right trigger signal Tlr. Similarly, the second sound event detector for multiple input channels 3012 uses the input signals in the center and surround input channels, Cin and Rin, respectively, to produce a center differential signal Co, a surround differential signal So, a center power envelope C2, a surround power envelope S2, and a center-surround trigger signal Tcs.
  • The sound event localization circuit for multiple input channel pairs and a single [0185] sound event type 3014 uses the left differential signal Lo, the right differential signal Ro, and the left-right trigger signal to produce an angle indicating the direction of a detected sound event that equals either a true OSA or a true DSA for the left-right input channel pair (the “left-right sound event angle” or “d/lr′”), and the center differential signal Co, the surround differential signal So and the center-surround trigger signal Tcs to produce an angle indicating the direction of a detected sound event that equals either a true OSA or a true DSA for the center-surround input channel pair (the “center-surround sound event angle” or “d/cs′”). Additionally, circuit 3014 uses all the power envelopes, dlr and dcs to verify the accuracy of the DSA. The sound event localization circuit for multiple input channel pairs and a single sound event type 2614 is shown in more detail in FIG. 31 and generally includes: a first OSA circuit 3102; a first DSA circuit 3104; a second DSA circuit 3106; a second OSA circuit 3108; a verification circuit 3116; a control circuit 3118; a first two-position switch 3110; a first three-position switch 3112; a first capacitor 3114; a second two-position switch 3120; a second three-position switch 3122; and a second capacitor 3124.
  • d/lr′ is produced by the [0186] first OSA circuit 3102, first DSA circuit 3104, first two-position switch 3110, first three-position switch 3112 and the first capacitor 3114. Similarly, d/cs′ is produced by the second OSA circuit 3108, second DSA circuit 3106, second two-position switch 3120, second three-position switch 3122 and the second capacitor 3124. The first and second OSA circuits 3102 and 3108 convert Lin and Rin and Cin and Sin, respectively, into ordinary steering angles, “lr” and “es” using known methods. The control circuit 3118, the first three-way switch 3112, and the first capacitor 3114 form a first sample and hold circuit, while the control circuit 3118, the second three-way switch 3122, the second capacitor 3124 form a second sample and hold circuit. Both the first and second three way switches 3112 and 3122, respectively, are normally in position C when no sound events are detected and controlled by the control circuit so that each switch is closed when a sound event is detected (moved to position A), opened at the end of the typical rise time of the sound event type being detected (moved to position B) and grounded at the end of the typical duration of the sound event type being detected (move to position C). The control circuit 3118 produces a control signal “con” that is communicated to the three- way switches 3112 and 3122. The control signal causes the three- way switches 3112 and 3122 to move to (or stay in) position A whenever either trigger signal (Tlr or Tcs) indicates that a sound event is being detected in either input channel pair. Subsequently, con causes the three- way switches 3112 and 3122 to move to position B at the end of the typical duration of the rise-time of the sound event. Then con causes the three- way switches 3112 and 3122 to move to position C at the end of the typical duration of the sound event type being detected.
  • Both two-[0187] way switches 3110 and 3120 include positions D and E. When a sound event is detected and both two-way switches 3110 and 3120 are in position D, the DSA for each channel pair are used to indicate the direction of the sound event. However, when a sound event is detected and both two-way switches 3110 and 3120 are in position E, the OSA for each channel pair is used to indicate the position of the sound event. The verification circuit 3116 controls both two-way switches 3110 and 3120 via a verification signal “vs” according to whether the DSAs are correct when a sound event is detected. When either trigger signal (Tlr or Tcs) indicates that a sound event is being detected, the verification circuit determines whether at least two of the power envelopes (L2, R2, C2, S2) have dropped by at least 3 dB or more. If at least two of the power envelopes have dropped by at least 3 dB or more from the last sound event, the verification circuit will communicate to both two-way switches 3110 and 3120 via vs causing them to move or stay in position E. However, if at least two of the power envelopes have not dropped by at least 3 dB or more, the verification circuit will communicate to both two-way switches 3110 and 3120 via vs causing the to move or stay in position D.
  • Alternatively, the verification circuit of the sound [0188] event localization circuit 3116 also includes a circuit for performing a consistency check (a “consistency check circuit”). The consistency check circuit is coupled to both DSA circuits and uses the differential steering angles produced by each make a further determination of the accuracy of the differential steering angles. The consistency check circuit uses a known circuit for determining the sum of the absolute values of dlr and dcs at the moment any trigger signal indicates that a sound event has been detected and then determines whether the sum is less than or equal to 45 degrees. If the sum is less than or equal to 45 degrees, and at least two of the power envelopes have not dropped by about 3 dB or more, the verification signal will communicate to both of the two-way switches 3110 and 3120 causing then to move to position D.
  • Furthermore, additional circuitry may be added to adjust the degree of accommodation (an “accommodation adjustment circuit”) and the threshold as a function of the consistency check (an “error threshold circuit”). Both the accommodation adjustment circuit and the error threshold circuit (not shown) are coupled to the consistency check circuit and include a counter that counts the number of errors detected by the consistency check in a time period of about several seconds. The accommodation adjustment circuit is also coupled to the accommodation signal in the transient detection circuits included in the [0189] DSA circuits 3104 and 3102 and further includes an accommodation voltage source that is adjusted according to the number of errors counted by the counter. As the number of errors increases, the voltage produced by the accommodation voltage source will increase to reduce the degree of accommodation. In contrast, the error threshold circuit is also coupled to the threshold detector in the trigger generation circuits and further includes an error voltage source that is adjusted according to the number of errors counted by the counter. As the number of errors increases, the voltage produced by the error voltage source will increase to cause the threshold voltage to increase so that fewer sound events are detected.
  • The sound event detector and localizer may also be implemented to detect multiple sound event types in a sound field generated in a single input channel pair (a “sound event detector and localizer for a single input channel pair and multiple sound event types”). A sound event detector and localizer for a single input channel pair and a multiple sound event types generally includes a sound event detector and localizer for a single input channel and a single sound even type implemented in parallel for each sound event type being detected to produce a differential steering angle for each sound event type in the input channel pair. Alternatively, the sound event detector and localizer for a single input channel pair and multiple sound event types may include a sound event detector for a single channel pair and a single sound event type implemented for each sound event type being detected and a sound event localization circuit for a single channel pair and a single sound event type that produces a differential steering angle indicating the direction of all types of sound events being detected. An example of such a sound event detector and localizer for a single input channel pair and multiple sound event types that detects syllables and impulsive sounds is shown in FIG. 32. [0190]
  • In FIG. 32, the sound event detector and localizer for a single input channel pair and multiple sound event types includes: a sound event detector for a channel pair and a single [0191] sound event type 3202; a sound event detector for a single channel pair and a short-duration sound event type 3204; and a sound event localization circuit for a single channel pair and a single sound event type 3206. The sound event detector for a single channel pair and a single sound event type 3202 may include the sound event detector for multiple channels and a single sound event type shown in FIG. 24 implemented to detect syllables and to produces a left differential signal Lo, a right differential signal Ro and a left-right trigger signal indicating the occurrence of syllables “Tlr(s).” The sound event detector for a single channel pair and a short-duration sound event type 3204 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG. 23 implemented to detect impulsive sounds and to produce a left-right trigger signal (as a combination of a left trigger signal and a right trigger signal) indicating the occurrence of impulsive sounds “Tlr(i).” The sound event localization circuit for a single channel pair and a single sound event type 3206 may include the sound event localization circuit for a single channels pair and multiple sound event types as shown in FIG. 29 implemented to produce a left right differential steering angle indicating the direction of a detected syllable or impulsive sound whenever either trigger signal (Tlr(s) or Tlr(i)) indicates that a sound event is occurring. Although this sound event detector and localizer is implemented to detect and localize syllables and impulsive sounds in a left-right input channel pair, it may be implemented to detect any number of sound event types in any channel pair simply by adding additional sound event detectors for a single channel pair implemented to detect the desired sound event type and having the sound event localization circuit for a single channel pair and a single sound event type respond to any of the trigger signals produced by the sound event detectors for a single channel pair.
  • The sound event detector and localizer may also be implemented to detect multiple sound event types in a sound field generated in multiple input channel pairs (a “sound event detector and localizer for multiple input channel pairs and multiple sound event types”). An example of a sound event detector and localizer for multiple input channel pairs and multiple sound event types that detects syllables and impulsive sounds in both the left-right and center-surround channel pairs is shown in FIG. 33. This sound event detector and localizer for multiple input channel pairs and multiple [0192] sound event types 3300 includes: a first and a second sound event detector for a single channel pair and a single short- duration sound event 3302 and 3308, respectively; a first and a second sound event detector for a single channel pair and a single sound event type 3304 and 3306, respectively; and a sound event localization circuit for multiple channel pairs and a single sound event type 3310.
  • The first sound event detector for a single channel pair and a single [0193] sound event type 3304 may include the sound event detector for multiple channels and a single sound event type shown in FIG. 24 implemented to detect syllables and to produces a left differential signal Lo, a right differential signal Ro and a left-right trigger signal indicating the occurrence of syllables “Tlr(s).” Similarly, the second sound event detector for a single channel pair and a single sound event type 3306 may include the sound event detector for multiple channels and a single sound event type shown in FIG. 24 implemented to detect syllables and to produces a center differential signal Co, a surround differential signal So and a center-surround trigger signal indicating the occurrence of syllables “Tcs(s).” The first sound event detector for a single channel pair and a short-duration sound event type 3302 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG. 23 implemented to detect impulsive sounds and to produce a left-right trigger signal (as a combination of a left trigger signal and a right trigger signal) indicating the occurrence of impulsive sounds “Tlr(i).” Similarly, the second sound event detector for a single channel pair and a short-duration sound event type 3308 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG. 23 implemented to detect impulsive sounds and to produce a center-surround trigger signal (as a combination of a center trigger signal and a surround trigger signal) indicating the occurrence of impulsive sounds “Tcs(i).” The sound event localization circuit for multiple channel pairs and a single sound event type 3310 may include the sound event localization circuit for multiple channel pairs and a single sound event type as shown in FIG. 31 implemented to produce left-right sound event angle indicating the direction of a detected syllable or impulsive sound whenever any trigger signal (Tlr(s), Tlr(i), Tcs(s), or Tcs(i)) indicates that a sound event is occurring “d/lr′(s,i)” and a center-surround sound event angle indicating the direction of a detected syllable or impulsive sound whenever any trigger signal (Tlr(s), Tlr(i), Tcs(s), or Tcs(i)) indicates that a sound event is occurring “d/cs′(s,i).” Although this sound event detector and localizer is implemented to detect and localize syllables and impulsive sounds in left-right and center-surround input channel pairs, it may be implemented to detect any number of sound event types in number of channel pair simply by adding additional sound event detectors for a single channel pair implemented to detect the desired sound event types in the desired channel pairs and having the sound event localization circuit for multiple channel pairs and a single sound event type respond to any of the trigger signals produced by the sound event detectors for a single channel pair.
  • Furthermore, any of the sound event detector and localizer for multiple input channel pairs and multiple sound event types may a dditionally include an accommodation adjustment circuit and/or an error threshold circuit. Both the accommodation adjustment circuit and the error threshold circuit (not shown) are coupled to the consistency check circuit and include a counter that counts the number of errors detected by the consistency check in a time period of about several seconds. The accommodation adjustment circuit is also coupled to the accommodation signal in the transient detection circuits included in the DSA circuits. In contrast, the error threshold circuit is also coupled to the threshold detector in the trigger generation circuits and further includes an error voltage source that is adjusted according to the number of errors counted by the counter. As the number of errors increases, the voltage produced by the error voltage source will increase to cause the threshold voltage to increase so that fewer sound events are detected. [0194]
  • 8. Surround Detector [0195]
  • One application of a sound event detector and localizer is as a stereo/surround detector. A stereo/surround detector determines whether a sound field is intended for reproduction in two input channels or in more than two input channels. An example of a stereo/surround detector implemented to detect a single sound event type [0196] 3400 (a “stereo/surround detector for a single sound event type”) is shown in FIG. 34 and includes: a sound event detector and localizer for a single input channel pair and a single sound event type 3402 and a detector and counter 3404. The sound event detector and localizer for a single input channel pair and a single sound event type 3402 uses the signals in the center input channel Cin and the surround input channel Sin to produce a true differential steering angle dcs′ that reflects the direction of whatever sound event is being detected. The threshold detector and counter 3404 then determines the number of times dcs′ falls within the range of about 0 degrees to about −45 degrees. If the number exceeds a predetermined value, the threshold detector and counter 3404 produces a signal sursig indicating that the sound field should be reproduced in surround. Conversely, if the number does not exceed a predetermined value, then the detector and counter 3404 produces a signal sursig indicating that the sound field should be reproduced in stereo. In general, if the number of sound events detected during a relatively long time period on the order of about 10 s to about 15 s is on the order of about 2 or 3, the detector and counter 3404 will produce a signal sursig indicating that the sound field should be reproduced in surround. Additionally, the detector and counter may further determine the duration of the sound events and only count those with durations that exceed a predetermined value as sound events that are intended for reproduction in the rear. In one example, sound events with durations less than about 50 ms will not be counted as sound events that are intended to be reproduced in the rear. In another example, sound events with durations of about 200 ms to about 300 ms will be counted as sound events that are intended to be reproduced in the rear.
  • Alternatively, a stereo/surround detector implemented to detect a single sound event type (a “stereo/surround detector for multiple sound event types”) is shown in FIG. 35 and is indicated by [0197] reference number 3500. In the example of FIG. 35, the stereo/surround detector 3500 is implemented to count the number of syllables and impulsive functions intended for reproduction in the rear. However, it may be implemented to count the number of any type and combination of types of sound events. The detector 3500 includes: first and second sound event detector and localizers for a single input channel pair and a single sound event type 3502 and 3504 and a detector and counter 3506. The first sound event detector and localizer for a single input channel pair 3502 uses the signals in the center input channel Cin and the surround input channel Sin to produce a true differential steering angle that reflects the direction of the syllables dcs′ (s). Similarly, the second sound event detector and localizer for a single input channel pair 3504 uses the signals in the center input channel Cin and the surround input channel Sin to produce a true differential steering angle that reflects the direction of the impulsive sounds dcs′(i). The detector and counter 3506 then determines the number of then determines the number of times dcs′(s,i) falls within the range of about 0 degrees to about −45 degrees to produce a signal sursig(s,i) that indicates whether the sound is to be reproduced in stereo or in surround.
  • 9. Sound Localizer [0198]
  • The properties of the human hearing mechanism have been modeled to create electronic circuitry that localizes sound events in the presence of a background signal, as well as all remaining sounds (collectively, “sound localizers”). Sound localizers separately detect and localize sound events and non-sound events in a sound field to produce a continuous indication of the direction of the sound field. These sound localizers may be used in a variety of applications, such as the reproduction of recorded sounds, particularly if the sounds are part of a complex sound field that includes sound events occurring simultaneously with steady-state sounds. When a sound field is to be reproduced in surround but was stored in stereo format, the sound localizers can be used as part of a matrix decoder to derive the true directions of the sounds from a two input channel mix. Also, the sound localizers can be implemented to detect any number and combination of sound event types in a sound field generated in any number of input channels. In the following description, the sound localizers are discussed in order of increasing complexity, with each subsequent sound localizer incorporating the elements of the prior sound localizers, except as indicated. [0199]
  • An example of a sound localizer that separately localizes a single sound event type in a single input channel pair (a “sound localizer for a single input channel pair and a single sound event type”) is shown in FIG. 36. The sound localizer for a single input channel pair and a single [0200] sound event type 3600 includes: a sound event detector for a single channel pair and a single sound event type 3602; and a sound localization circuit for a single input channel pair and a single sound event 3604. The sound event detector for a single channel pair and a single sound event type 3602 may include the sound event detector for multiple channels and a single sound event type shown in FIG. 24 implemented to detect and localize whatever sound event type is being detected and produces left and right differential signals Lo and Ro, respectively, and a left-right trigger signal Tlr. The sound localization circuit 3604 uses Tlr, Lo, Ro, and the signals in the left and right input channels Lin and Rin, respectively, to produce a steering angle that indicates the direction of the sound field with respect to the left-right input channel pair in terms of an ordinary steering angle and a differential steering angle (generally referred to in this application as a “comprehensive steering angle” and the comprehensive steering angle with respect to the left-right input channel pair is referred to as “clr′”).
  • The sound localization circuit for a single input channel pair and a single [0201] sound event type 3604 is shown in more detail in FIG. 37 and includes: an ordinary steering angle (OSA) circuit 3702; a DSA circuit 3704; a control circuit 3706; a first switch 3708; a resistor 3710; a second switch 3712 and a capacitor 3714. The sound event localization circuit for a single input channel pair and a single sound event type 3604 uses the left and right differential signals Lo and Ro, respectively, and the left and right input signals Lin and Rin, respectively, to produce a left-right comprehensive steering angle clr′. In general, clr′ equals follows the OSA when no sound events are detected and follows the DSA whenever a sound event is detected. The OSA circuit 3702 uses Lin and Rin to determine the ordinary steering angle lr. The DSA circuit 3704 uses the differential signals Lo and Ro to produce the differential steering angle. The control circuit 3706, the first switch 3708, the resistor 3710; the second switch 3712, and the capacitor 3714 generally form a sample-and-hold circuit and can therefore be replaced with any device or circuit that performs a similar function.
  • In general, when no sound events are detected, the [0202] first switch 3708 will be open and the second switch 3712 will be closed. In this state, clr′ will follow lr at a rate defined by the resistor 3710 and the capacitor 3714. However, when a sound event is detected, the control circuit 3706 causes the first switch 3708 to close and the capacitor 3712 to capture dlr during the typical rise time of the sound event type being detected so that clr′ equals the true DSA for the typical duration of the sound event type being detected. More specifically, the control circuit 3716 receives the trigger signal Tlr and produces a control signal “con” that controls the first switch 3708 and the second switch 3712. When Tlr indicates that a sound event is occurring, the control circuit 3706 communicates to the first switch 3708 via “con” a command to close causing the first switch 3708 to close. At the end of the rise-time typical for the sound event type being detected, the control circuit 3706 communicates to the first and second switches 3708 and 3712, respectively, via con a command to open causing the first and second switches 3708 and 3712 to open. After the typical duration of the sound event type being detected, the control circuit 3706, communicates to the second switch 3712 a command via con to go close. During the time that the first switch 3706 is closed (the rise-time of the sound event), the left-right differential steering angle dlr is captured by the capacitor 3714 to create the true DSA dlr′. Clr′ is defined by the true DSA, which is held until the end of the typical duration of the sound event, even if the sound event has not actually ended. For example, if the sound event being detected is a syllable, the true DSA will be held for about 50 ms to about 200 ms, preferably after about 150 ms. In another example, if the sound event being detected is an impulsive sound, the true DSA will be held for about 50 ms. At the end of the typical duration of the sound event type being detected, the capacitor 3714 will charge or discharge until it reflects lr at a specified rate. The capacitor 3714 and resistor 3710 are chosen so that they define an RC time constant that will cause the specified rate of decay. For example, the RC time constant be equal to about 300 ms.
  • Alternatively, the sound localization circuit can be specifically implemented for short-duration sound events. As previously discussed, if a short-duration sound event, such as an impulsive sound, is localized according to a DSA, it is often advantageous at the end of the typical duration of the short-duration sound event to determine if the sound event has actually ended and switch immediately to the OSA if the sound event has ended. An example of a sound event localization circuit that includes this functionality (referred to in this application as a “sound localization circuit for a single input channel pair and short-duration sound events”) is shown in FIG. 38. The sound localization circuit for a single input channel pair and short-[0203] duration sound events 3800 includes: an OSA circuit 3802; a DSA circuit 3804; a control circuit 3808; a detector 3810; a first switch 3812; a second switch 3816; a first resistor 3814; a third switch 3813; a second resistor 3815; and a capacitor 3818. This sound localization circuit 3800 produces a left-right comprehensive steering angle clr′. When no sound events are detected, clr′ equals the filtered OSA (which is lr after it is filtered by resistor 3814 and capacitor 3816). However, when a sound event is detected, clr′ either equals the true OSA or the true DSA. The sound localization circuit for a single channel pair and a single short duration sound event generally behaves in the same way as the sound localization circuit shown in FIG. 37, except that at the end of a sound event, clr′ either decays to lr or goes to lr immediately depending on whether the sound event has actually ended.
  • After the [0204] control circuit 3808 receives a trigger signal indicating that a sound event is occurring and has determined that the typical duration of the sound event being detected has ended, in addition to its other functions as previously described, it communicates with the detector 3810 to establish whether the sound event has actually ended. The detector 3810 determines whether a sound event has actually ended by comparing the power envelope with the accommodation signal in each input channel of the channel pair. If the input power envelope is greater than the accommodation signal in any input channel of the input channel pair, the detector 3810 will determine that the short-duration sound event has not actually ended. Conversely, if the input power envelope is not greater than the accommodation signal in any input channel of the input channel pair, the detector 3810 will determine that the short-duration sound event has actually ended. If the control circuit 3808 establishes that the sound event has not actually ended, it will then instruct the second switch 3816 to close via con. When the second switch 3816 closes, clr′ will drift to the filtered OSA at a specified rate. Therefore, capacitor 3818 and resistor 3814 are select so that their RC time constant is about equal to the specified rate (which is generally about 300 ms). For example, if impulsive sounds are being detected, the RC time constant of capacitor 3818 and resistor 3814 will be about 5 ms. However, if the control circuit 3808 establishes that the sound event has actually ended, it will then instruct the third switch 3813 to close via con. When the third switch 3813 closes, clr′ immediately goes to lr. Therefore, the second resistor 3815 is chosen so that the RC time constant of the second resistor 3815 and the capacitor 3818 is much lower than that of the first resistor 3814 and the capacitor 3818 (generally about a factor of ten lower). Generally, the third switch will remain closed for a very short time (generally about 3 ms to about 10 ms). After this very short time has ended, the control circuit 3803 will instruct the third switch 3813 to open and the second switch 3816 to close so that clr′ will go to the filtered OSA. Additionally, the sound localizer for a single input channel pair and a short-duration sound event type may further include a DSA averaging circuit. The DSA averaging circuit may be implemented in the sound event localization circuit 3800 between the DSA circuit 3804 and the switch 3812.
  • The sound localizer may also be implemented to detect a single sound event type in a sound field generated in multiple input channel pairs (a “sound localizer for multiple input channel pairs and a single sound event type”). An example of a sound localizer for multiple input channel pairs and a single sound event type implemented to detect and localize a single sound event in both an LR input channel pair and a CS input channel pair is shown in FIG. 39 and designated by [0205] reference number 3900. This sound localizer may be implemented for any combination of input channel pairs with the LR input channel pair and the CS input channel pair used in this instance for explanation purposes only. The sound localizer for multiple input channel pairs and a single sound event type 3900 produces a comprehensive steering angle for the LR input channel pair (“clr′”) and the CS input channel pair (“ccs′”) and generally includes: a first and second sound event detector for a single input channel pair and a single sound event 3902 and 3904, respectively; and a sound localization circuit for multiple input channel pairs and a single sound event type 3906.
  • The first and second sound event detectors for a single input channel and a single [0206] sound event type 3902 and 3904 may include a sound event detector for multiple input channels and a single sound event type, such as that shown in FIG. 24, implemented to detect the same sound event. The first sound event detector for a single input channel 3902 uses Lin′ and Rin′ to produce a left differential signal Lo, a right differential signal Ro, and a left-right trigger signal Tlr. Similarly, the second sound event detector for a single input channel 3904 uses Cin′ and Rin′ to produce a center differential signal Co, a surround differential signal So, and a center-surround trigger signal Tcs.
  • The sound localization circuit for multiple input channel pairs and a single [0207] sound event type 3906 uses the left differential signal Lo, the right differential signal Ro, and the left-right trigger signal to produce a left-right comprehensive steering angle clr′, and uses the center differential signal Co, the surround differential signal So and the center-surround trigger signal Tcs to produce a center-surround comprehensive steering angle ces′. Additionally, circuit 3906 uses all the power envelopes, dlr and dcs to verify the accuracy of the DSA. This circuit 3906 is shown in more detail in FIG. 40 and generally includes: a first OSA circuit 4002; a first DSA circuit 4004; a second DSA circuit 4006; a second OSA circuit 4008; a verification circuit 4020; a control circuit 4022; a first two-position switch 4010; a first switch 4012; a first resistor 4014; a first capacitor 4018; a second switch 4014; a second two-position switch 4030; a third switch 4032; a second capacitor 4038; a fourth switch 4036; and a second resistor 4034.
  • clr′ is produced by the [0208] first OSA circuit 4002, first DSA circuit 4004, first two-position switch 4010, first switch 4012, the first resistor 4014, the second switch 4016, and the first capacitor 3114. Similarly, d/cs′ is produced by the second OSA circuit 4008, second DSA circuit 4006, second two-position switch 4030, the third switch 4032, the second resistor 4034 and the second capacitor 4038. The first and second OSA circuits 4002 and 4008 convert Lin and Rin and Cin and Sin, respectively, into ordinary steering angles, lr and cs. The control circuit 4022, the first switch 4012, the second switch 4016, and the first capacitor 4018 form a first sample and hold circuit, while the control circuit 4022, the third switch 4032, the fourth switch 4034, the second resistor 4034, and the second capacitor 4038 form a second sample and hold circuit. Both the first and third switches 4012 and 4032, respectively, are normally open when no sound events are detected and are controlled by the control circuit so that each switch is closed when a sound event is detected and opened at the end of the typical rise time of the sound event type being detected. Both the second and forth switches 4016 and 4036 are normally closed when no sound events are detected and opened after a sound event has been detected and the typical rise-time of the sound event type being detected has ended. The control circuit 4022 produces a control signal “con” that is communicated to all the switches 4012, 4016, 4032 and 4036. The control signal causes the first and second switches 4012 and 4016 to close (or stay closed) whenever either trigger signal (Tlr or Tcs) indicates that a sound event is being detected in either input channel pair. Subsequently, con causes all the switches 4012, 4016, 4032 and 4036 to open at the end of the typical duration of the rise-time of the sound event.
  • Both two-[0209] way switches 4010 and 4030 include positions D and E. When a sound event is detected and both two- way switches 4010 and 4030 are in position D, the DSA for each channel pair are used to indicate the direction of the sound event. However, when a sound event is detected and both two- way switches 4010 and 4020 are in position E, the OSA for each channel pair is used to indicate the position of the sound event. The verification circuit 4020 controls both two- way switches 4010 and 4020 via a verification signal “vs” according to whether the DSAs are correct when a sound event is detected. When either trigger signal (Tlr or Tcs) indicates that a sound event is being detected, the verification circuit determines whether at least two of the power envelopes (L2, R2, C2, S2) have dropped by at least 3 dB or more. If at least two of the power envelopes have dropped by at least 3 dB or more from the last sound event, the verification circuit will communicate to both two- way switches 4010 and 4020 via vs causing them to move or stay in position E. However, if at least two of the power envelopes have not dropped by at least 3 dB or more, the verification circuit will communicate to both two- way switches 4010 and 4020 via vs causing the to move or stay in position D.
  • Alternatively, the verification circuit of the sound [0210] event localization circuit 4020 also includes a consistency check circuit. The consistency check circuit is coupled to both DSA circuits and uses the differential steering angles produced by each to make a further determination of the accuracy of the differential steering angles as previously described. Furthermore, the sound localizer for multiple sound event types and a single sound event type may further include an accommodation adjustment circuit and/or an error threshold circuit.
  • Alternatively, the sound localization circuit for multiple channel pairs and a single sound event type can be specifically implemented for short-duration sound events, such as impulsive sounds. As previously discussed, if a short-duration sound event is localized, it is often advantageous at the end of the typical duration of the sound event type being detected to determine if the sound event has actually ended and switch immediately to the OSA if the sound event has not actually ended. An example of such a sound localization circuit for multiple input channel pairs and short-duration sound events is shown in FIG. 41 and indicated by [0211] reference number 4100. This sound localization circuit for multiple input channel pairs and short-duration sound events 4100 generally includes: a first OSA circuit 4102; a first DSA circuit 4104; a second DSA circuit 4106; a second OSA circuit 4108; a verification circuit 4120; a control circuit 4122; a detector circuit 4124; a first two-position switch 4110; a first switch 4112; a first resistor 4114; a first capacitor 4118; a second switch 4116; a second two-position switch 4130; a third switch 4132; a second capacitor 4138; a fourth switch 4136; a second resistor 4134; a fifth switch 4113; a third resistor 4115; a sixth switch 4133; and a fourth resistor 4135. This sound localization circuit 4100 produces a left-right comprehensive steering angle. When there are no sound events, clr′ will equal the filtered OSA (which is lr after being filtered by first resistor 4114 and first capacitor 4118). When there is a sound event detected, clr′ will equal either the true OSA or the true DSA of the left-right channel pair. This sound localization circuit 4100 also produces a center-surround comprehensive steering ccs′. When no sound events are detected, clr′ equals the filtered OSA. When sound events are detected, clr′ will equal the true OSA or the true DSA of the center-surround channel pair. The sound localization circuit for multiple channel pairs and a single short duration sound event generally behaves in the same way as the sound localization circuit for multiple channel pairs an a single sound event type shown in FIG. 40, except that at the end of a sound event, clr′ either decays to the filtered OSA or goes to the OSA immediately depending on whether the sound event has actually ended.
  • After the [0212] control circuit 4122 receives a trigger signal indicating that a sound event is occurring and has determined that the typical duration of the sound event being detected has ended, in addition to its other functions as previously described, it communicates with the detector 4124 to establish whether the sound event has actually ended. The detector 4122 determines whether a sound event has actually ended by comparing the power envelope with the accommodation signal in each input channel of the channel pair. If the input power envelope is greater than the accommodation signal in any input channel of the input channel pair, the detector 4122 will determine that the short-duration sound event has not actually ended. Conversely, if the input power envelope is not greater than the accommodation signal in any input channel of the input channel pair, the detector 4124 will determine that the short-duration sound event has actually ended. If the control circuit 4122 establishes that the sound event has not actually ended, it will then instruct the second and fourth switches 4116 and 4136, respectively, to close via con. When the second switch 4116 and the fourth switch 4136 close, clr′ will drift to the left-right filtered OSA (which is lr after it has been filtered by the first resistor 4114 and the first capacitor 4118) and ccs′ will drift to the center-surround filtered OSA (which is cs after it has been filtered by the second resistor 4136 and the second capacitor 4138) at a specified rate. Therefore, the first and second capacitors 4118 and 4138, respectively, and the first and second resistors 4114 and 4134, respectively, are select so that their RC time constant is about equal to the specified rate. For example, if impulsive sounds are being detected, the RC time constant of capacitor 4118 and resistor 4114, as well as the RC time constant of capacitor 4138 and resistor 4134 will be about 5 ms.
  • However, if the [0213] control circuit 4122 establishes that the sound event has actually ended, it will then instruct the fifth and sixth switches 4113 and 4133, respectively, to close via con. When fifth and sixth switches 4113 and 4133, respectively, close, clr′ and ccs′ both immediately goes to lr and cs, respectively. Therefore, the RC time constant of the third resistor 4115 and the first capacitor 4118, and the RC time constant of the fourth resistor 4135 and the second capacitor 4138 will both be very low. These RC time constants may be about at least a factor of ten lower than the RC time constant of the first resistor 4114 and the first capacitor 4118 and the RC time constant of the second resistor 4134 and the second capacitor 4138. The fifth and sixth switches 4113 and 4133, respectively, will remain closed for a short time. This short time may be about 3 ms to about 10 ms. After this short time, the control circuit 4122 will instruct the fifth and sixth switches 4113 and 4135, respectively, to open and the second and fourth switches 4166 and 4136, respectively, to close so that clr′ and ccs′ will reflect the filtered OSAs. Additionally, the sound localizer for a multiple input channel pair and a short-duration sound event type may further include first and second DSA averaging circuits. The first DSA averaging circuit may be implemented in the sound event localization circuit 4100 between the first DSA circuit 4104 and the first two-position switch 4110. Likewise, the second DSA averaging circuit may be implemented in the sound event localization circuit 4100 between the second DSA circuit 4106 and the second two-position switch 4130.
  • Alternatively, the verification circuit of the sound localization circuit for multiple channel pairs and a single short-duration [0214] sound event type 4100 also includes a consistency check circuit. The consistency check circuit is coupled to both DSA circuits and uses the differential steering angles produced by each to make a further determination of the accuracy of the differential steering angles as previously described. Furthermore, the sound localizer for multiple sound event types and a single sound event type may further include an accommodation adjustment circuit and/or an error threshold circuit.
  • The sound localizer may also be implemented to detect multiple sound event types in a sound field generated in a single input channel pair (a “sound localizer for a single input channel pair and multiple sound event types”). An example of a sound event detector and localizer for a single input channel pair and a multiple sound event types implemented to detect and localize syllables and impulsive sounds in a left-right input channel pair is shown in FIG. 42 and indicated by reference number [0215] 4200 (however, this localizer may be implemented for any input channel pair to detect any combination of sound event types). The sound localizer for a single input channel pair and multiple sound event types 4200 generally includes: a sound event detector for a single input channel and a single sound event type 4202; a sound event detector for a single-channel and a single short-duration sound event device 4206; and a sound localization circuit for a single input channel pair and multiple sound event types 4204.
  • The sound event detector for a single channel pair and a single [0216] sound event type 4202 may include the sound event detector for multiple channels and a single sound event type shown in FIG. 24 implemented to detect syllables and to produces a left differential signal Lo, a right differential signal Ro and a left-right trigger signal indicating the occurrence of syllables Tlr(s). The sound event detector for a single channel pair and a short-duration sound event type 4206 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG. 23 implemented to detect impulsive sounds and to produce a left-right trigger signal (as a combination of a left trigger signal and a right trigger signal) indicating the occurrence of impulsive sounds Tlr(i). The sound localization circuit for a single channel pair and a single sound event type 4204 may include the sound localization circuit for a single channels pair and multiple sound event types as shown in FIG. 37 implemented to produce a left-right differential steering angle indicating the direction of a detected syllable or impulsive sound whenever either trigger signal (Tlr(s) or Tlr(i)) indicates that a sound event is occurring. Although this sound localizer is implemented to detect and localize syllables and impulsive sounds in a left-right input channel pair, it may be implemented to detect any number of sound event types in any channel pair simply by adding additional sound event detectors for a single channel pair to implemented to detect the desired sound event type and having the sound localization circuit for a single channel pair and a single sound event type respond to any of the trigger signals produced by the sound event detectors for a single channel pair.
  • The sound localizer may also be implemented to detect multiple sound event types in a sound field generated in multiple input channel pairs (a “sound localizer for multiple input channel pairs and multiple sound event types”). An example of a sound localizer for multiple input channel pairs and multiple sound event types specifically localizes syllables and impulsive sounds in a left-right input channel pair and a center-surround input channel pair is shown in FIG. 43 (however, the sound localizer for multiple input channel pairs and multiple sound event types may be implemented to specifically localize any combination of sound events in any number of input channel pairs). This sound localizer for multiple input channel pairs and multiple [0217] sound event types 4300 includes: a first and a second sound event detector for a single channel pair and a single short- duration sound event 4308 and 4306, respectively; a first and a second sound event detector for a single channel pair and a single sound event type 4302 and 4304, respectively; and a sound event localization circuit for multiple channel pairs and a single sound event type 4310.
  • The first sound event detector for a single channel pair and a single [0218] sound event type 4302 may include the sound event detector for multiple channels and a single sound event type shown in FIG. 24 implemented to detect syllables and to produces a left differential signal Lo, a right differential signal Ro and a left-right trigger signal indicating the occurrence of syllables Tlr(s). Similarly, the second sound event detector for a single channel pair and a single sound event type 4304 may include the sound event detector for multiple channels and a single sound event type shown in FIG. 24 implemented to detect syllables and to produces a center differential signal Co, a surround differential signal So and a center-surround trigger signal indicating the occurrence of syllables Tcs(s). The first sound event detector for a single channel pair and a short-duration sound event type 4308 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG. 23 implemented to detect impulsive sounds and to produce a left-right trigger signal (as a combination of a left trigger signal and a right trigger signal) indicating the occurrence of impulsive sounds Tlr(i). Similarly, the second sound event detector for a single channel pair and a short-duration sound event type 4306 may include the sound event detector for a single channel pair and a single short-duration sound event shown in FIG. 23 implemented to detect impulsive sounds and to produce a center-surround trigger signal (as a combination of a center trigger signal and a surround trigger signal) indicating the occurrence of impulsive sounds Tcs(i). The sound localization circuit for multiple channel pairs and a single sound event type 4310 may include the sound localization circuit for multiple channel pairs and a single sound event type as shown in FIG. 40 implemented to produce left-right sound event angle indicating the direction of a detected syllable or impulsive sound whenever any trigger signal (Tlr(s), Tlr(i), Tcs(s), or Tcs(i)) indicates that a sound event is occurring “clr′(s,i)” and a center-surround sound event angle indicating the direction of a detected syllable or impulsive sound whenever any trigger signal (Tlr(s), Tlr(i), Tcs(s), or Tcs(i)) indicates that a sound event is occurring “ccs(s,i).” Although this sound localizer is implemented to detect and localize syllables and impulsive sounds in left-right and center-surround input channel pairs, it may be implemented to detect any number of sound event types in number of channel pair simply by adding additional sound event detectors for a single channel pair implemented to detect the desired sound event types in the desired channel pairs and having the sound localization circuit for multiple channel pairs and a single sound event type 4310 respond to any of the trigger signals produced by the sound event detectors for a single channel pair.
  • Furthermore, any of the sound localizers for multiple input channel pairs and multiple sound event types may additionally include an accommodation adjustment circuit and/or an error threshold circuit. Both the accommodation adjustment circuit and the error threshold circuit (not shown) are coupled to the consistency check circuit and include a counter that counts the number of errors detected by the consistency check in a time period of about several seconds. The accommodation adjustment circuit is also coupled to the accommodation signal in the transient detection circuits included in the DSA circuits. In contrast, the error threshold circuit is also coupled to the threshold detector in the trigger generation circuits and further includes an error voltage source that is adjusted according to the number of errors counted by the counter. As the number of errors increases, the voltage produced by the error voltage source will increase to cause the threshold voltage to increase so that fewer sound events are detected. [0219]
  • 10. Software [0220]
  • Implementations of the sound event detection methods, sound event detection and localization methods and sound localization methods and any methods included in any of these methods, include computer readable software code. These algorithms may be implemented together or independently. Such code may be stored on a processor, a memory device or on any other computer readable storage medium. Alternatively, the software code may be encoded in a computer readable electronic or optical signal. The code may be object code or any other code describing or controlling the functionality described in this application. The computer readable storage medium may be a magnetic storage disk such as a floppy disk, an optical disk such as a CD-ROM, semiconductor memory or any other physical object storing program code or associated data. [0221]
  • While various embodiments of the invention have been described, it will be apparent to those of ordinary skill in the art that many more embodiments and implementations are possible within the scope of the invention. Accordingly, the invention is not to be restricted except in light of the attached claims and their equivalents. [0222]

Claims (7)

What is claimed is:
1. A method for detecting sound events in a sound field, comprising:
modeling an accommodation property of a human hearing mechanism;
separating the sound events from any background sounds in response to the accommodation property;
modeling a beginning detection property of the human hearing mechanism; and
detecting the sound events in response to the beginning detection property.
2. The method, as claimed in claim 1, further comprising:
modeling a frequency bias property of the human hearing mechanism; and
emphasizing directionally important frequencies in the sound field in response to the frequency bias property.
3. A computer readable storage medium storing computer readable program code, comprising:
data encoding a sound field; and
a computer code for implementing a sound event detection method for detecting sound events in the sound field.
4. The computer readable storage medium storing computer readable program code, as claimed in claim 3, where the sound field is generated in a single channel; and where the sound event detection method includes a sound event detection method for a single channel and a single sound event type, and detects sound events of the single sound event type.
5. A sound event detection device, comprising:
an interface unit for receiving a sound field; and
a detection unit, where the detector device includes,
a processor, where the processor is coupled to the interface unit and receives the sound field; and
a memory, where the memory is coupled to the processor and stores a sound event detection method; where in response to a request from the processor, the memory communicates the sound event detection method to the processor; where the processor uses the sound event detection method to detect sound events in the sound field to produce a trigger signal, and where the processor communicates the trigger signal to the interface unit.
6. A sound event detector, comprising:
a linear to decibel converter, where the linear to decibel converter receives the sound field and produces a decibel voltage signal; and
a trigger generation circuit, where the trigger generation circuit receives the decibel voltage signal and produces a trigger signal indicating the occurrence of one or more sound events in a sound field.
7. A sound event detector for detecting sound events in a sound field, comprising:
a beginning detection modeling means, where the beginning detection modeling means receives the sound field and produces a filtered power signal;
an accommodation modeling means, where the accommodation modeling means receives the filtered power signal and produces an accommodated signal;
a means for distinguishing the sound events from any noise present in the accommodated signal, where the means for distinguishing the sound events is couple to the accommodation modeling means, receives the accommodated signal and produces a trigger signal.
US10/428,451 2002-05-03 2003-05-02 Sound event detection system Abandoned US20040005065A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US10/428,451 US20040005065A1 (en) 2002-05-03 2003-05-02 Sound event detection system

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US37755802P 2002-05-03 2002-05-03
US10/428,451 US20040005065A1 (en) 2002-05-03 2003-05-02 Sound event detection system

Publications (1)

Publication Number Publication Date
US20040005065A1 true US20040005065A1 (en) 2004-01-08

Family

ID=29401529

Family Applications (4)

Application Number Title Priority Date Filing Date
US10/428,366 Active 2025-12-14 US7492908B2 (en) 2002-05-03 2003-05-02 Sound localization system based on analysis of the sound field
US10/428,405 Active 2025-07-08 US7567676B2 (en) 2002-05-03 2003-05-02 Sound event detection and localization system using power analysis
US10/428,451 Abandoned US20040005065A1 (en) 2002-05-03 2003-05-02 Sound event detection system
US10/810,989 Expired - Lifetime US7499553B2 (en) 2002-05-03 2004-03-26 Sound event detector system

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US10/428,366 Active 2025-12-14 US7492908B2 (en) 2002-05-03 2003-05-02 Sound localization system based on analysis of the sound field
US10/428,405 Active 2025-07-08 US7567676B2 (en) 2002-05-03 2003-05-02 Sound event detection and localization system using power analysis

Family Applications After (1)

Application Number Title Priority Date Filing Date
US10/810,989 Expired - Lifetime US7499553B2 (en) 2002-05-03 2004-03-26 Sound event detector system

Country Status (8)

Country Link
US (4) US7492908B2 (en)
EP (1) EP1585947B1 (en)
JP (2) JP4744874B2 (en)
KR (1) KR101047194B1 (en)
CN (1) CN1830009B (en)
AU (1) AU2003265935A1 (en)
CA (2) CA2773294C (en)
WO (1) WO2003093775A2 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020196755A1 (en) * 2001-05-11 2002-12-26 Hiroyuki Hidaka Portable communication terminal and wireless communication system therefor
US20030040822A1 (en) * 2001-05-07 2003-02-27 Eid Bradley F. Sound processing system using distortion limiting techniques
US20040005064A1 (en) * 2002-05-03 2004-01-08 Griesinger David H. Sound event detection and localization system
US20060088175A1 (en) * 2001-05-07 2006-04-27 Harman International Industries, Incorporated Sound processing system using spatial imaging techniques
US20070140513A1 (en) * 2005-12-15 2007-06-21 Harman International Industries, Incorporated Distortion compensation
EP2316942A1 (en) 2004-12-22 2011-05-04 Alnylam Pharmaceuticals, Inc. Conserved HBV and HCV sequences useful for gene silencing
TWI728632B (en) * 2019-12-31 2021-05-21 財團法人工業技術研究院 Positioning method for specific sound source
US11158330B2 (en) * 2016-11-17 2021-10-26 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for decomposing an audio signal using a variable threshold
US11183199B2 (en) 2016-11-17 2021-11-23 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for decomposing an audio signal using a ratio as a separation characteristic
US11568867B2 (en) 2013-06-27 2023-01-31 Amazon Technologies, Inc. Detecting self-generated wake expressions
US20230080485A1 (en) * 2013-11-05 2023-03-16 Alarm.Com Incorporated Handling duress input

Families Citing this family (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7908077B2 (en) 2003-06-10 2011-03-15 Itt Manufacturing Enterprises, Inc. Land use compatibility planning software
US7782256B2 (en) 1999-03-05 2010-08-24 Era Systems Corporation Enhanced passive coherent location techniques to track and identify UAVs, UCAVs, MAVs, and other objects
US7889133B2 (en) 1999-03-05 2011-02-15 Itt Manufacturing Enterprises, Inc. Multilateration enhancements for noise and operations management
US8446321B2 (en) 1999-03-05 2013-05-21 Omnipol A.S. Deployable intelligence and tracking system for homeland security and search and rescue
US7667647B2 (en) 1999-03-05 2010-02-23 Era Systems Corporation Extension of aircraft tracking and positive identification from movement areas into non-movement areas
US7570214B2 (en) 1999-03-05 2009-08-04 Era Systems, Inc. Method and apparatus for ADS-B validation, active and passive multilateration, and elliptical surviellance
US7739167B2 (en) 1999-03-05 2010-06-15 Era Systems Corporation Automated management of airport revenues
US8203486B1 (en) 1999-03-05 2012-06-19 Omnipol A.S. Transmitter independent techniques to extend the performance of passive coherent location
US7777675B2 (en) 1999-03-05 2010-08-17 Era Systems Corporation Deployable passive broadband aircraft tracking
US7443987B2 (en) * 2002-05-03 2008-10-28 Harman International Industries, Incorporated Discrete surround audio system for home and automotive listening
US20050108024A1 (en) * 2003-11-13 2005-05-19 Fawcett John Jr. Systems and methods for retrieving data
DK1989853T3 (en) * 2006-02-23 2017-03-20 Togewa Holding Ag SWITCHING SYSTEM AND SIMILAR PROCEDURE FOR UNICAST OR MULTICAST TRANSMISSIONS FROM START TO END OF DATA AND / OR MULTIMEDIA FLOW BETWEEN NETWORK Nodes
JP4786384B2 (en) * 2006-03-27 2011-10-05 株式会社東芝 Audio processing apparatus, audio processing method, and audio processing program
US7965227B2 (en) 2006-05-08 2011-06-21 Era Systems, Inc. Aircraft tracking using low cost tagging as a discriminator
JP4867516B2 (en) * 2006-08-01 2012-02-01 ヤマハ株式会社 Audio conference system
EP1947471B1 (en) * 2007-01-16 2010-10-13 Harman Becker Automotive Systems GmbH System and method for tracking surround headphones using audio signals below the masked threshold of hearing
US8908873B2 (en) * 2007-03-21 2014-12-09 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Method and apparatus for conversion between multi-channel audio formats
US9015051B2 (en) * 2007-03-21 2015-04-21 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Reconstruction of audio channels with direction parameters indicating direction of origin
CN100505837C (en) 2007-05-10 2009-06-24 华为技术有限公司 System and method for controlling image collector for target positioning
US8050414B2 (en) * 2008-10-16 2011-11-01 Gas Technology Institute Robust pipe-strike pulse detector
US8045738B2 (en) * 2008-10-31 2011-10-25 Zounds Hearing, Inc. System for managing feedback
WO2011139772A1 (en) * 2010-04-27 2011-11-10 James Fairey Sound wave modification
TWI403304B (en) 2010-08-27 2013-08-01 Ind Tech Res Inst Method and mobile device for awareness of linguistic ability
US9111526B2 (en) 2010-10-25 2015-08-18 Qualcomm Incorporated Systems, method, apparatus, and computer-readable media for decomposition of a multichannel music signal
CN102592597B (en) * 2011-01-17 2014-08-13 鸿富锦精密工业(深圳)有限公司 Electronic device and audio data copyright protection method
US9143571B2 (en) * 2011-03-04 2015-09-22 Qualcomm Incorporated Method and apparatus for identifying mobile devices in similar sound environment
EP2495581B1 (en) 2011-03-04 2017-03-22 BlackBerry Limited Human audible localization for sound emitting devices
JP5994470B2 (en) * 2012-08-08 2016-09-21 株式会社Jvcケンウッド Sound source direction detecting device, sound source direction detecting method, sound source direction detecting program
DE102013207149A1 (en) * 2013-04-19 2014-11-06 Siemens Medical Instruments Pte. Ltd. Controlling the effect size of a binaural directional microphone
KR102195897B1 (en) 2013-06-05 2020-12-28 삼성전자주식회사 Apparatus for dectecting aucoustic event, operating method thereof, and computer-readable recording medium having embodied thereon a program which when executed by a computer perorms the method
CN103531202B (en) * 2013-10-14 2015-10-28 无锡儒安科技有限公司 Distributed Detection sound event also chooses the method for similar events point
US9478229B2 (en) 2013-12-10 2016-10-25 Massachusetts Institute Of Technology Methods and apparatus for recording impulsive sounds
US10134416B2 (en) * 2015-05-11 2018-11-20 Microsoft Technology Licensing, Llc Privacy-preserving energy-efficient speakers for personal sound
US9977645B2 (en) * 2015-10-01 2018-05-22 Moodelizer Ab Dynamic modification of audio content
US10134422B2 (en) 2015-12-01 2018-11-20 Qualcomm Incorporated Determining audio event based on location information
WO2017184149A1 (en) * 2016-04-21 2017-10-26 Hewlett-Packard Development Company, L.P. Electronic device microphone listening modes
WO2017192200A1 (en) * 2016-05-05 2017-11-09 The Research Foundation For The State Unversity Of New York Compositions for treating periodontitis and dental calculus accumulation
US10264999B2 (en) 2016-09-07 2019-04-23 Massachusetts Institute Of Technology High fidelity systems, apparatus, and methods for collecting noise exposure data
CN106814670A (en) * 2017-03-22 2017-06-09 重庆高略联信智能技术有限公司 A kind of river sand mining intelligent supervision method and system
CN108806711A (en) * 2018-08-07 2018-11-13 吴思 A kind of extracting method and device
US10811032B2 (en) * 2018-12-19 2020-10-20 Cirrus Logic, Inc. Data aided method for robust direction of arrival (DOA) estimation in the presence of spatially-coherent noise interferers

Citations (93)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4251688A (en) * 1979-01-15 1981-02-17 Ana Maria Furner Audio-digital processing system for demultiplexing stereophonic/quadriphonic input audio signals into 4-to-72 output audio signals
US4382158A (en) * 1980-03-22 1983-05-03 Sharp Kabushiki Kaisha Tone control of the operational type
US4759066A (en) * 1987-05-27 1988-07-19 Polk Investment Corporation Sound system with isolation of dimensional sub-speakers
US4761814A (en) * 1985-06-20 1988-08-02 Pioneer Electronic Corporation Variable bandwidth multivoice demodulating circuit
US4799260A (en) * 1985-03-07 1989-01-17 Dolby Laboratories Licensing Corporation Variable matrix decoder
US4862502A (en) * 1988-01-06 1989-08-29 Lexicon, Inc. Sound reproduction
US4866776A (en) * 1983-11-16 1989-09-12 Nissan Motor Company Limited Audio speaker system for automotive vehicle
US4891839A (en) * 1984-12-31 1990-01-02 Peter Scheiber Signal re-distribution, decoding and processing in accordance with amplitude, phase and other characteristics
US4932059A (en) * 1988-01-11 1990-06-05 Fosgate Inc. Variable matrix decoder for periphonic reproduction of sound
US4940977A (en) * 1987-09-25 1990-07-10 Dolby Laboratories Licensing Corporation Adaptive-filter single-bit digital encoder and decoder and adaptation control circuit responsive to bit-stream loading
US4941177A (en) * 1985-03-07 1990-07-10 Dolby Laboratories Licensing Corporation Variable matrix decoder
US4953213A (en) * 1989-01-24 1990-08-28 Pioneer Electronic Corporation Surround mode stereophonic reproducing equipment
US5046098A (en) * 1985-03-07 1991-09-03 Dolby Laboratories Licensing Corporation Variable matrix decoder with three output channels
US5109419A (en) * 1990-05-18 1992-04-28 Lexicon, Inc. Electroacoustic system
US5119422A (en) * 1990-10-01 1992-06-02 Price David A Optimal sonic separator and multi-channel forward imaging system
US5136650A (en) * 1991-01-09 1992-08-04 Lexicon, Inc. Sound reproduction
US5146507A (en) * 1989-02-23 1992-09-08 Yamaha Corporation Audio reproduction characteristics control device
US5189703A (en) * 1988-01-06 1993-02-23 Lucasarts Entertainment Company Timbre correction units for use in sound systems
US5199075A (en) * 1991-11-14 1993-03-30 Fosgate James W Surround sound loudspeakers and processor
US5222143A (en) * 1990-08-14 1993-06-22 Samsung Electronics Co., Ltd. Compatible multivoice broadcasting receiver
US5280528A (en) * 1990-06-08 1994-01-18 Fosgate James W Band pass filter circuit for rear channel filtering in a surround processor
US5295189A (en) * 1990-06-08 1994-03-15 Fosgate James W Control voltage generator for surround sound processor
US5319713A (en) * 1992-11-12 1994-06-07 Rocktron Corporation Multi dimensional sound circuit
US5333201A (en) * 1992-11-12 1994-07-26 Rocktron Corporation Multi dimensional sound circuit
US5337196A (en) * 1991-01-31 1994-08-09 Samsung Electronics Co., Ltd. Stereo/multivoice recording and reproducing video tape recorder including a decoder developing a switch control signal
US5339363A (en) * 1990-06-08 1994-08-16 Fosgate James W Apparatus for enhancing monophonic audio signals using phase shifters
US5386473A (en) * 1994-01-21 1995-01-31 Harrison; Robert W. Passive surround sound circuit
US5400433A (en) * 1991-01-08 1995-03-21 Dolby Laboratories Licensing Corporation Decoder for variable-number of channel presentation of multidimensional sound fields
US5412732A (en) * 1992-01-16 1995-05-02 Pioneer Electronic Corporation Stereo surround system
US5428687A (en) * 1990-06-08 1995-06-27 James W. Fosgate Control voltage generator multiplier and one-shot for integrated surround sound processor
US5463424A (en) * 1993-08-03 1995-10-31 Dolby Laboratories Licensing Corporation Multi-channel transmitter/receiver system providing matrix-decoding compatible signals
US5497425A (en) * 1994-03-07 1996-03-05 Rapoport; Robert J. Multi channel surround sound simulation device
US5504819A (en) * 1990-06-08 1996-04-02 Harman International Industries, Inc. Surround sound processor with improved control voltage generator
US5524054A (en) * 1993-06-22 1996-06-04 Deutsche Thomson-Brandt Gmbh Method for generating a multi-channel audio decoder matrix
US5594800A (en) * 1991-02-15 1997-01-14 Trifield Productions Limited Sound reproduction system having a matrix converter
US5610985A (en) * 1993-01-22 1997-03-11 U.S. Philips Corporation Digital 3-channel transmission of left and right stereo signals and a center signal
US5617480A (en) * 1993-02-25 1997-04-01 Ford Motor Company DSP-based vehicle equalization design system
US5625696A (en) * 1990-06-08 1997-04-29 Harman International Industries, Inc. Six-axis surround sound processor with improved matrix and cancellation control
US5638452A (en) * 1995-04-21 1997-06-10 Rocktron Corporation Expandable multi-dimensional sound circuit
US5642423A (en) * 1995-11-22 1997-06-24 Sony Corporation Digital surround sound processor
US5666424A (en) * 1990-06-08 1997-09-09 Harman International Industries, Inc. Six-axis surround sound processor with automatic balancing and calibration
US5708719A (en) * 1995-09-07 1998-01-13 Rep Investment Limited Liability Company In-home theater surround sound speaker system
US5727067A (en) * 1995-08-28 1998-03-10 Yamaha Corporation Sound field control device
US5727068A (en) * 1996-03-01 1998-03-10 Cinema Group, Ltd. Matrix decoding method and apparatus
US5748749A (en) * 1993-03-24 1998-05-05 Noise Cancellation Technologies, Inc. Active noise cancelling muffler
US5761313A (en) * 1995-06-30 1998-06-02 Philips Electronics North America Corp. Circuit for improving the stereo image separation of a stereo signal
US5768394A (en) * 1995-08-18 1998-06-16 Samsung Electronics Co., Ltd. Surround audio signal reproducing apparatus having a sub-woofer signal mixing function
US5771295A (en) * 1995-12-26 1998-06-23 Rocktron Corporation 5-2-5 matrix system
US5796844A (en) * 1996-07-19 1998-08-18 Lexicon Multichannel active matrix sound reproduction with maximum lateral separation
US5798818A (en) * 1995-10-17 1998-08-25 Sony Corporation Configurable cinema sound system
US5802181A (en) * 1994-03-07 1998-09-01 Sony Corporation Theater sound system with upper surround channels
US5862228A (en) * 1997-02-21 1999-01-19 Dolby Laboratories Licensing Corporation Audio matrix encoding
US5870480A (en) * 1996-07-19 1999-02-09 Lexicon Multichannel active matrix encoder and decoder with maximum lateral separation
US5930370A (en) * 1995-09-07 1999-07-27 Rep Investment Limited Liability In-home theater surround sound speaker system
US5974380A (en) * 1995-12-01 1999-10-26 Digital Theater Systems, Inc. Multi-channel audio decoder
US6032081A (en) * 1995-09-25 2000-02-29 Korea Telecommunication Authority Dematrixing processor for MPEG-2 multichannel audio decoder
US6038324A (en) * 1997-02-21 2000-03-14 Ambourn; Paul R. Automotive surround sound circuit background of the invention
US6108584A (en) * 1997-07-09 2000-08-22 Sony Corporation Multichannel digital audio decoding method and apparatus
US6118876A (en) * 1995-09-07 2000-09-12 Rep Investment Limited Liability Company Surround sound speaker system for improved spatial effects
US6122381A (en) * 1996-05-17 2000-09-19 Micronas Interuetall Gmbh Stereophonic sound system
US6141597A (en) * 1997-09-08 2000-10-31 Picturetel Corporation Audio processor
US6198826B1 (en) * 1997-05-19 2001-03-06 Qsound Labs, Inc. Qsound surround synthesis from stereo
US20020097882A1 (en) * 2000-11-29 2002-07-25 Greenberg Jeffry Allen Method and implementation for detecting and characterizing audible transients in noise
US6442278B1 (en) * 1999-06-15 2002-08-27 Hearing Enhancement Company, Llc Voice-to-remaining audio (VRA) interactive center channel downmix
US6442277B1 (en) * 1998-12-22 2002-08-27 Texas Instruments Incorporated Method and apparatus for loudspeaker presentation for positional 3D sound
US6453047B1 (en) * 1998-09-28 2002-09-17 Creative Technology Ltd Matrix encoding system with improved behavior frequency
US6466913B1 (en) * 1998-07-01 2002-10-15 Ricoh Company, Ltd. Method of determining a sound localization filter and a sound localization control system incorporating the filter
US20030040822A1 (en) * 2001-05-07 2003-02-27 Eid Bradley F. Sound processing system using distortion limiting techniques
US6539357B1 (en) * 1999-04-29 2003-03-25 Agere Systems Inc. Technique for parametric coding of a signal containing information
US6556685B1 (en) * 1998-11-06 2003-04-29 Harman Music Group Companding noise reduction system with simultaneous encode and decode
US6577736B1 (en) * 1998-10-15 2003-06-10 Central Research Laboratories Limited Method of synthesizing a three dimensional sound-field
US6587565B1 (en) * 1997-03-13 2003-07-01 3S-Tech Co., Ltd. System for improving a spatial effect of stereo sound or encoded sound
US6590983B1 (en) * 1998-10-13 2003-07-08 Srs Labs, Inc. Apparatus and method for synthesizing pseudo-stereophonic outputs from a monophonic input
US6611212B1 (en) * 1999-04-07 2003-08-26 Dolby Laboratories Licensing Corp. Matrix improvements to lossless encoding and decoding
US6624873B1 (en) * 1998-05-05 2003-09-23 Dolby Laboratories Licensing Corporation Matrix-encoded surround-sound channels in a discrete digital sound format
US6636608B1 (en) * 1997-11-04 2003-10-21 Tatsuya Kishii Pseudo-stereo circuit
US20040005064A1 (en) * 2002-05-03 2004-01-08 Griesinger David H. Sound event detection and localization system
US6683964B1 (en) * 1994-06-30 2004-01-27 Lucent Technologies Inc. Direction finder
US6683962B1 (en) * 1997-12-23 2004-01-27 Harman International Industries, Incorporated Method and system for driving speakers with a 90 degree phase shift
US6694027B1 (en) * 1999-03-09 2004-02-17 Smart Devices, Inc. Discrete multi-channel/5-2-5 matrix system
US6697491B1 (en) * 1996-07-19 2004-02-24 Harman International Industries, Incorporated 5-2-5 matrix encoder and decoder system
US6711266B1 (en) * 1997-02-07 2004-03-23 Bose Corporation Surround sound channel encoding and decoding
US20040086130A1 (en) * 2002-05-03 2004-05-06 Eid Bradley F. Multi-channel sound processing systems
US20050018860A1 (en) * 2001-05-07 2005-01-27 Harman International Industries, Incorporated: Sound processing system for configuration of audio signals in a vehicle
US6850622B2 (en) * 1997-05-29 2005-02-01 Sony Corporation Sound field correction circuit
US20050031128A1 (en) * 2003-06-02 2005-02-10 Yuji Tomita Apparatus for generating surround signal from two-channel stereo signal
US20050063551A1 (en) * 2003-09-18 2005-03-24 Yiou-Wen Cheng Multi-channel surround sound expansion method
US20060008175A1 (en) * 1997-04-22 2006-01-12 Koichiro Tanaka Image processing apparatus, image processing method, and storage medium
US6996239B2 (en) * 2001-05-03 2006-02-07 Harman International Industries, Inc. System for transitioning from stereo to simulated surround sound
US7065217B2 (en) * 2001-03-05 2006-06-20 Harman/Becker Automotive Systems (Becker Division) Gmbh Apparatus and method for multichannel sound reproduction system
US7177432B2 (en) * 2001-05-07 2007-02-13 Harman International Industries, Incorporated Sound processing system with degraded signal optimization
US7206413B2 (en) * 2001-05-07 2007-04-17 Harman International Industries, Incorporated Sound processing system using spatial imaging techniques
US7236838B2 (en) * 2000-08-29 2007-06-26 Matsushita Electric Industrial Co., Ltd. Signal processing apparatus, signal processing method, program and recording medium

Family Cites Families (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3845572A (en) * 1972-08-02 1974-11-05 Singer Co Modular vehicle trainer sound system having a plurality of separately controllable sound generators and a polyphonic speaker array
US4972482A (en) 1987-09-18 1990-11-20 Sanyo Electric Co., Ltd. Fm stereo demodulator
JPH0256600A (en) * 1988-08-23 1990-02-26 Ricoh Co Ltd Speech dialing system
JP3118023B2 (en) * 1990-08-15 2000-12-18 株式会社リコー Voice section detection method and voice recognition device
DE69210689T2 (en) 1991-01-08 1996-11-21 Dolby Lab Licensing Corp ENCODER / DECODER FOR MULTI-DIMENSIONAL SOUND FIELDS
JPH06276599A (en) * 1991-07-26 1994-09-30 Gijutsu Kenkyu Kumiai Iryo Fukushi Kiki Kenkyusho Impulsive sound suppressing device
US5161197A (en) 1991-11-04 1992-11-03 Lexicon, Inc. Acoustic analysis
WO1994010816A1 (en) * 1992-10-29 1994-05-11 Wisconsin Alumni Research Foundation Methods and apparatus for producing directional sound
US5357574A (en) 1992-12-14 1994-10-18 Ford Motor Company Coherent signal generation in digital radio receiver
DE69428119T2 (en) 1993-07-07 2002-03-21 Picturetel Corp REDUCING BACKGROUND NOISE FOR LANGUAGE ENHANCEMENT
US6144747A (en) 1997-04-02 2000-11-07 Sonics Associates, Inc. Head mounted surround sound system
US5561737A (en) 1994-05-09 1996-10-01 Lucent Technologies Inc. Voice actuated switching system
US5626643A (en) * 1994-09-26 1997-05-06 Owens-Corning Fiberglas Technology Inc. Contact drying of fibers to form composite strands
US5841993A (en) 1996-01-02 1998-11-24 Ho; Lawrence Surround sound system for personal computer for interfacing surround sound with personal computer
US5850455A (en) 1996-06-18 1998-12-15 Extreme Audio Reality, Inc. Discrete dynamic positioning of audio signals in a 360° environment
FI105522B (en) 1996-08-06 2000-08-31 Sample Rate Systems Oy Arrangement for home theater or other audio equipment
US6144474A (en) 1996-10-21 2000-11-07 Fujitsu Limited Optical transmission system including optical repeaters with selectively enabled gain equalizers contained therein and including an add/drop apparatus with a plurality of individually selectable filters
DE19651308C2 (en) 1996-12-10 1998-10-22 Becker Gmbh Audio sound system for a motor vehicle
US5983087A (en) 1997-06-26 1999-11-09 Delco Electronics Corporation Distributed digital signal processing for vehicle audio systems
JP4151110B2 (en) 1998-05-14 2008-09-17 ソニー株式会社 Audio signal processing apparatus and audio signal reproduction apparatus
EP0980064A1 (en) * 1998-06-26 2000-02-16 Ascom AG Method for carrying an automatic judgement of the transmission quality of audio signals
JP2000032434A (en) * 1998-07-08 2000-01-28 Victor Co Of Japan Ltd Image-pickup device
JP3484988B2 (en) 1998-09-22 2004-01-06 ヤマハ株式会社 Performance information editing method and recording medium storing performance information editing program
FI113935B (en) 1998-09-25 2004-06-30 Nokia Corp Method for Calibrating the Sound Level in a Multichannel Audio System and a Multichannel Audio System
US6711536B2 (en) * 1998-10-20 2004-03-23 Canon Kabushiki Kaisha Speech processing apparatus and method
JP2001028799A (en) 1999-05-10 2001-01-30 Sony Corp Onboard sound reproduction device
CN100429960C (en) 2000-07-19 2008-10-29 皇家菲利浦电子有限公司 Multi-channel stereo converter for deriving a stereo surround and/or audio centre signal
JP4264686B2 (en) 2000-09-14 2009-05-20 ソニー株式会社 In-vehicle sound reproduction device

Patent Citations (99)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4251688A (en) * 1979-01-15 1981-02-17 Ana Maria Furner Audio-digital processing system for demultiplexing stereophonic/quadriphonic input audio signals into 4-to-72 output audio signals
US4382158A (en) * 1980-03-22 1983-05-03 Sharp Kabushiki Kaisha Tone control of the operational type
US4866776A (en) * 1983-11-16 1989-09-12 Nissan Motor Company Limited Audio speaker system for automotive vehicle
US4891839A (en) * 1984-12-31 1990-01-02 Peter Scheiber Signal re-distribution, decoding and processing in accordance with amplitude, phase and other characteristics
US5046098A (en) * 1985-03-07 1991-09-03 Dolby Laboratories Licensing Corporation Variable matrix decoder with three output channels
US4799260A (en) * 1985-03-07 1989-01-17 Dolby Laboratories Licensing Corporation Variable matrix decoder
US4941177A (en) * 1985-03-07 1990-07-10 Dolby Laboratories Licensing Corporation Variable matrix decoder
US4761814A (en) * 1985-06-20 1988-08-02 Pioneer Electronic Corporation Variable bandwidth multivoice demodulating circuit
US4759066A (en) * 1987-05-27 1988-07-19 Polk Investment Corporation Sound system with isolation of dimensional sub-speakers
US4940977A (en) * 1987-09-25 1990-07-10 Dolby Laboratories Licensing Corporation Adaptive-filter single-bit digital encoder and decoder and adaptation control circuit responsive to bit-stream loading
US4862502A (en) * 1988-01-06 1989-08-29 Lexicon, Inc. Sound reproduction
US5189703A (en) * 1988-01-06 1993-02-23 Lucasarts Entertainment Company Timbre correction units for use in sound systems
US4932059A (en) * 1988-01-11 1990-06-05 Fosgate Inc. Variable matrix decoder for periphonic reproduction of sound
US4953213A (en) * 1989-01-24 1990-08-28 Pioneer Electronic Corporation Surround mode stereophonic reproducing equipment
US5146507A (en) * 1989-02-23 1992-09-08 Yamaha Corporation Audio reproduction characteristics control device
US5109419A (en) * 1990-05-18 1992-04-28 Lexicon, Inc. Electroacoustic system
US5295189A (en) * 1990-06-08 1994-03-15 Fosgate James W Control voltage generator for surround sound processor
US5307415A (en) * 1990-06-08 1994-04-26 Fosgate James W Surround processor with antiphase blending and panorama control circuitry
US5504819A (en) * 1990-06-08 1996-04-02 Harman International Industries, Inc. Surround sound processor with improved control voltage generator
US5339363A (en) * 1990-06-08 1994-08-16 Fosgate James W Apparatus for enhancing monophonic audio signals using phase shifters
US5280528A (en) * 1990-06-08 1994-01-18 Fosgate James W Band pass filter circuit for rear channel filtering in a surround processor
US5428687A (en) * 1990-06-08 1995-06-27 James W. Fosgate Control voltage generator multiplier and one-shot for integrated surround sound processor
US5666424A (en) * 1990-06-08 1997-09-09 Harman International Industries, Inc. Six-axis surround sound processor with automatic balancing and calibration
US5625696A (en) * 1990-06-08 1997-04-29 Harman International Industries, Inc. Six-axis surround sound processor with improved matrix and cancellation control
US5644640A (en) * 1990-06-08 1997-07-01 Harman International Industries, Inc. Surround sound processor with improved control voltage generator
US5222143A (en) * 1990-08-14 1993-06-22 Samsung Electronics Co., Ltd. Compatible multivoice broadcasting receiver
US5119422A (en) * 1990-10-01 1992-06-02 Price David A Optimal sonic separator and multi-channel forward imaging system
US5400433A (en) * 1991-01-08 1995-03-21 Dolby Laboratories Licensing Corporation Decoder for variable-number of channel presentation of multidimensional sound fields
US5136650A (en) * 1991-01-09 1992-08-04 Lexicon, Inc. Sound reproduction
US5337196A (en) * 1991-01-31 1994-08-09 Samsung Electronics Co., Ltd. Stereo/multivoice recording and reproducing video tape recorder including a decoder developing a switch control signal
US5594800A (en) * 1991-02-15 1997-01-14 Trifield Productions Limited Sound reproduction system having a matrix converter
US5301237A (en) * 1991-11-14 1994-04-05 Fosgate James W Surround sound loudspeakers
US5199075A (en) * 1991-11-14 1993-03-30 Fosgate James W Surround sound loudspeakers and processor
US5412732A (en) * 1992-01-16 1995-05-02 Pioneer Electronic Corporation Stereo surround system
US5333201A (en) * 1992-11-12 1994-07-26 Rocktron Corporation Multi dimensional sound circuit
US5319713A (en) * 1992-11-12 1994-06-07 Rocktron Corporation Multi dimensional sound circuit
US5610985A (en) * 1993-01-22 1997-03-11 U.S. Philips Corporation Digital 3-channel transmission of left and right stereo signals and a center signal
US5617480A (en) * 1993-02-25 1997-04-01 Ford Motor Company DSP-based vehicle equalization design system
US5748749A (en) * 1993-03-24 1998-05-05 Noise Cancellation Technologies, Inc. Active noise cancelling muffler
US5524054A (en) * 1993-06-22 1996-06-04 Deutsche Thomson-Brandt Gmbh Method for generating a multi-channel audio decoder matrix
US5463424A (en) * 1993-08-03 1995-10-31 Dolby Laboratories Licensing Corporation Multi-channel transmitter/receiver system providing matrix-decoding compatible signals
US5386473A (en) * 1994-01-21 1995-01-31 Harrison; Robert W. Passive surround sound circuit
US5802181A (en) * 1994-03-07 1998-09-01 Sony Corporation Theater sound system with upper surround channels
US5497425A (en) * 1994-03-07 1996-03-05 Rapoport; Robert J. Multi channel surround sound simulation device
US6683964B1 (en) * 1994-06-30 2004-01-27 Lucent Technologies Inc. Direction finder
US5638452A (en) * 1995-04-21 1997-06-10 Rocktron Corporation Expandable multi-dimensional sound circuit
US5761313A (en) * 1995-06-30 1998-06-02 Philips Electronics North America Corp. Circuit for improving the stereo image separation of a stereo signal
US5768394A (en) * 1995-08-18 1998-06-16 Samsung Electronics Co., Ltd. Surround audio signal reproducing apparatus having a sub-woofer signal mixing function
US5727067A (en) * 1995-08-28 1998-03-10 Yamaha Corporation Sound field control device
US5708719A (en) * 1995-09-07 1998-01-13 Rep Investment Limited Liability Company In-home theater surround sound speaker system
US6118876A (en) * 1995-09-07 2000-09-12 Rep Investment Limited Liability Company Surround sound speaker system for improved spatial effects
US5930370A (en) * 1995-09-07 1999-07-27 Rep Investment Limited Liability In-home theater surround sound speaker system
US6032081A (en) * 1995-09-25 2000-02-29 Korea Telecommunication Authority Dematrixing processor for MPEG-2 multichannel audio decoder
US5798818A (en) * 1995-10-17 1998-08-25 Sony Corporation Configurable cinema sound system
US5642423A (en) * 1995-11-22 1997-06-24 Sony Corporation Digital surround sound processor
US5974380A (en) * 1995-12-01 1999-10-26 Digital Theater Systems, Inc. Multi-channel audio decoder
US5771295A (en) * 1995-12-26 1998-06-23 Rocktron Corporation 5-2-5 matrix system
US5727068A (en) * 1996-03-01 1998-03-10 Cinema Group, Ltd. Matrix decoding method and apparatus
US6122381A (en) * 1996-05-17 2000-09-19 Micronas Interuetall Gmbh Stereophonic sound system
US5796844A (en) * 1996-07-19 1998-08-18 Lexicon Multichannel active matrix sound reproduction with maximum lateral separation
US6697491B1 (en) * 1996-07-19 2004-02-24 Harman International Industries, Incorporated 5-2-5 matrix encoder and decoder system
US5870480A (en) * 1996-07-19 1999-02-09 Lexicon Multichannel active matrix encoder and decoder with maximum lateral separation
US7107211B2 (en) * 1996-07-19 2006-09-12 Harman International Industries, Incorporated 5-2-5 matrix encoder and decoder system
US6711266B1 (en) * 1997-02-07 2004-03-23 Bose Corporation Surround sound channel encoding and decoding
US5862228A (en) * 1997-02-21 1999-01-19 Dolby Laboratories Licensing Corporation Audio matrix encoding
US6038324A (en) * 1997-02-21 2000-03-14 Ambourn; Paul R. Automotive surround sound circuit background of the invention
US6587565B1 (en) * 1997-03-13 2003-07-01 3S-Tech Co., Ltd. System for improving a spatial effect of stereo sound or encoded sound
US20060008175A1 (en) * 1997-04-22 2006-01-12 Koichiro Tanaka Image processing apparatus, image processing method, and storage medium
US6198826B1 (en) * 1997-05-19 2001-03-06 Qsound Labs, Inc. Qsound surround synthesis from stereo
US6850622B2 (en) * 1997-05-29 2005-02-01 Sony Corporation Sound field correction circuit
US6108584A (en) * 1997-07-09 2000-08-22 Sony Corporation Multichannel digital audio decoding method and apparatus
US6141597A (en) * 1997-09-08 2000-10-31 Picturetel Corporation Audio processor
US6636608B1 (en) * 1997-11-04 2003-10-21 Tatsuya Kishii Pseudo-stereo circuit
US6683962B1 (en) * 1997-12-23 2004-01-27 Harman International Industries, Incorporated Method and system for driving speakers with a 90 degree phase shift
US6624873B1 (en) * 1998-05-05 2003-09-23 Dolby Laboratories Licensing Corporation Matrix-encoded surround-sound channels in a discrete digital sound format
US6466913B1 (en) * 1998-07-01 2002-10-15 Ricoh Company, Ltd. Method of determining a sound localization filter and a sound localization control system incorporating the filter
US6453047B1 (en) * 1998-09-28 2002-09-17 Creative Technology Ltd Matrix encoding system with improved behavior frequency
US6590983B1 (en) * 1998-10-13 2003-07-08 Srs Labs, Inc. Apparatus and method for synthesizing pseudo-stereophonic outputs from a monophonic input
US6577736B1 (en) * 1998-10-15 2003-06-10 Central Research Laboratories Limited Method of synthesizing a three dimensional sound-field
US6556685B1 (en) * 1998-11-06 2003-04-29 Harman Music Group Companding noise reduction system with simultaneous encode and decode
US6442277B1 (en) * 1998-12-22 2002-08-27 Texas Instruments Incorporated Method and apparatus for loudspeaker presentation for positional 3D sound
US6694027B1 (en) * 1999-03-09 2004-02-17 Smart Devices, Inc. Discrete multi-channel/5-2-5 matrix system
US6611212B1 (en) * 1999-04-07 2003-08-26 Dolby Laboratories Licensing Corp. Matrix improvements to lossless encoding and decoding
US6539357B1 (en) * 1999-04-29 2003-03-25 Agere Systems Inc. Technique for parametric coding of a signal containing information
US6442278B1 (en) * 1999-06-15 2002-08-27 Hearing Enhancement Company, Llc Voice-to-remaining audio (VRA) interactive center channel downmix
US7236838B2 (en) * 2000-08-29 2007-06-26 Matsushita Electric Industrial Co., Ltd. Signal processing apparatus, signal processing method, program and recording medium
US20020097882A1 (en) * 2000-11-29 2002-07-25 Greenberg Jeffry Allen Method and implementation for detecting and characterizing audible transients in noise
US7065217B2 (en) * 2001-03-05 2006-06-20 Harman/Becker Automotive Systems (Becker Division) Gmbh Apparatus and method for multichannel sound reproduction system
US6996239B2 (en) * 2001-05-03 2006-02-07 Harman International Industries, Inc. System for transitioning from stereo to simulated surround sound
US20030040822A1 (en) * 2001-05-07 2003-02-27 Eid Bradley F. Sound processing system using distortion limiting techniques
US7206413B2 (en) * 2001-05-07 2007-04-17 Harman International Industries, Incorporated Sound processing system using spatial imaging techniques
US20050018860A1 (en) * 2001-05-07 2005-01-27 Harman International Industries, Incorporated: Sound processing system for configuration of audio signals in a vehicle
US7177432B2 (en) * 2001-05-07 2007-02-13 Harman International Industries, Incorporated Sound processing system with degraded signal optimization
US20040179697A1 (en) * 2002-05-03 2004-09-16 Harman International Industries, Incorporated Surround detection system
US20040022392A1 (en) * 2002-05-03 2004-02-05 Griesinger David H. Sound detection and localization system
US20040005064A1 (en) * 2002-05-03 2004-01-08 Griesinger David H. Sound event detection and localization system
US20040086130A1 (en) * 2002-05-03 2004-05-06 Eid Bradley F. Multi-channel sound processing systems
US20050031128A1 (en) * 2003-06-02 2005-02-10 Yuji Tomita Apparatus for generating surround signal from two-channel stereo signal
US20050063551A1 (en) * 2003-09-18 2005-03-24 Yiou-Wen Cheng Multi-channel surround sound expansion method

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8031879B2 (en) 2001-05-07 2011-10-04 Harman International Industries, Incorporated Sound processing system using spatial imaging techniques
US20030040822A1 (en) * 2001-05-07 2003-02-27 Eid Bradley F. Sound processing system using distortion limiting techniques
US20060088175A1 (en) * 2001-05-07 2006-04-27 Harman International Industries, Incorporated Sound processing system using spatial imaging techniques
US8472638B2 (en) 2001-05-07 2013-06-25 Harman International Industries, Incorporated Sound processing system for configuration of audio signals in a vehicle
US20080319564A1 (en) * 2001-05-07 2008-12-25 Harman International Industries, Incorporated Sound processing system for configuration of audio signals in a vehicle
US20080317257A1 (en) * 2001-05-07 2008-12-25 Harman International Industries, Incorporated Sound processing system for configuration of audio signals in a vehicle
US7760890B2 (en) 2001-05-07 2010-07-20 Harman International Industries, Incorporated Sound processing system for configuration of audio signals in a vehicle
US20020196755A1 (en) * 2001-05-11 2002-12-26 Hiroyuki Hidaka Portable communication terminal and wireless communication system therefor
US20040005064A1 (en) * 2002-05-03 2004-01-08 Griesinger David H. Sound event detection and localization system
US20040022392A1 (en) * 2002-05-03 2004-02-05 Griesinger David H. Sound detection and localization system
US20040179697A1 (en) * 2002-05-03 2004-09-16 Harman International Industries, Incorporated Surround detection system
EP2316942A1 (en) 2004-12-22 2011-05-04 Alnylam Pharmaceuticals, Inc. Conserved HBV and HCV sequences useful for gene silencing
US8036402B2 (en) * 2005-12-15 2011-10-11 Harman International Industries, Incorporated Distortion compensation
US20070140513A1 (en) * 2005-12-15 2007-06-21 Harman International Industries, Incorporated Distortion compensation
US8942391B2 (en) 2005-12-15 2015-01-27 Harman International Industries, Incorporated Distortion compensation
US11568867B2 (en) 2013-06-27 2023-01-31 Amazon Technologies, Inc. Detecting self-generated wake expressions
US11600271B2 (en) * 2013-06-27 2023-03-07 Amazon Technologies, Inc. Detecting self-generated wake expressions
US20230080485A1 (en) * 2013-11-05 2023-03-16 Alarm.Com Incorporated Handling duress input
US11158330B2 (en) * 2016-11-17 2021-10-26 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for decomposing an audio signal using a variable threshold
US11183199B2 (en) 2016-11-17 2021-11-23 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for decomposing an audio signal using a ratio as a separation characteristic
US11869519B2 (en) 2016-11-17 2024-01-09 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for decomposing an audio signal using a variable threshold
TWI728632B (en) * 2019-12-31 2021-05-21 財團法人工業技術研究院 Positioning method for specific sound source

Also Published As

Publication number Publication date
WO2003093775A2 (en) 2003-11-13
KR101047194B1 (en) 2011-07-06
JP2011022602A (en) 2011-02-03
CA2483609C (en) 2012-09-18
EP1585947A4 (en) 2011-04-27
CA2483609A1 (en) 2003-11-13
US7499553B2 (en) 2009-03-03
US7492908B2 (en) 2009-02-17
US20040022392A1 (en) 2004-02-05
EP1585947A2 (en) 2005-10-19
AU2003265935A1 (en) 2003-11-17
CN1830009B (en) 2010-05-05
US20040179697A1 (en) 2004-09-16
AU2003265935A8 (en) 2003-11-17
US20040005064A1 (en) 2004-01-08
KR20040105252A (en) 2004-12-14
US7567676B2 (en) 2009-07-28
CA2773294A1 (en) 2003-11-13
JP4744874B2 (en) 2011-08-10
WO2003093775A3 (en) 2006-03-30
JP2005539413A (en) 2005-12-22
EP1585947B1 (en) 2020-01-01
CN1830009A (en) 2006-09-06
CA2773294C (en) 2013-03-12

Similar Documents

Publication Publication Date Title
US7567676B2 (en) Sound event detection and localization system using power analysis
US10685638B2 (en) Audio scene apparatus
JP4952698B2 (en) Audio processing apparatus, audio processing method and program
Ratnam et al. Blind estimation of reverberation time
EP0637011B1 (en) Speech signal discrimination arrangement and audio device including such an arrangement
US20100185308A1 (en) Sound Signal Processing Device And Playback Device
JP4826625B2 (en) Volume correction device, volume correction method, volume correction program, and electronic device
JPS61500329A (en) Signal processing device for determining the angular position of the acoustic source
US5572593A (en) Method and apparatus for detecting and extending temporal gaps in speech signal and appliances using the same
CN108235181B (en) Method for noise reduction in an audio processing apparatus
EP2949133B1 (en) Automatic loudspeaker polarity detection
EP3757993A1 (en) Pre-processing for automatic speech recognition
KR101073632B1 (en) A zero-crossing-based multiple source localization apparatus in reverberant environments
CN113287169A (en) Apparatus, method and computer program for blind source separation and remixing
US20050004792A1 (en) Speech characteristic extraction method speech charateristic extraction device speech recognition method and speech recognition device
Kolotzek et al. Fast processing explains the effect of sound reflection on binaural unmasking
EP4303874A1 (en) Providing a measure of intelligibility of an audio signal
JPH08298698A (en) Environmental sound analyzer
Vesa Estimation of reverberation time from binaural signals without using controlled excitation
Morgan et al. Automated evaluation of acoustic talker direction finder algorithms in the varechoic chamber
Henderson Estimating azimuth from speech in a natural auditory environment
Vesa Jälkikaiunta-ajan estimointi binauraalisesta signaalista ilman tunnettua herätettä
JPS6264197A (en) Houling detector
JPH04340598A (en) Voice recognition device

Legal Events

Date Code Title Description
AS Assignment

Owner name: HARMAN INTERNATIONAL INDUSTRIES, INCORPORATED, CAL

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GRIESINGER, DAVID H.;REEL/FRAME:014442/0888

Effective date: 20030821

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION