|Publication number||US5212733 A|
|Application number||US 07/486,151|
|Publication date||May 18, 1993|
|Filing date||Feb 28, 1990|
|Priority date||Feb 28, 1990|
|Also published as||DE69130528D1, DE69130528T2, EP0517848A1, EP0517848A4, EP0517848B1, WO1991013497A1|
|Publication number||07486151, 486151, US 5212733 A, US 5212733A, US-A-5212733, US5212733 A, US5212733A|
|Inventors||Douglas J. Devitt, Kristoph D. Krug|
|Original Assignee||Voyager Sound, Inc.|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (11), Non-Patent Citations (27), Referenced by (147), Classifications (34), Legal Events (8)|
|External Links: USPTO, USPTO Assignment, Espacenet|
A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent disclosure, as it appears in the Patent and Trademark Office patent files or records, but otherwise reserves all copyright rights whatsoever.
The present invention relates to interactively controlling multiple parameters affecting an audio output.
It is known to mix sound input signals into sound output signals using a sound mixing board which includes separate manually activated controls for each sound input signal. The sound input signals may include signals from a plurality of musical instruments and voices as well as sound effects. The controls may include pan (i.e., position) and gain (i.e., amplitude) controls. The sound output signals may be provided to speakers or any recording device.
Sound input signals are mixed to control the position of the signal relative to a listener as well as to control the relative level of an input signal with respect to other input signals. The position of an input signal is controlled by adjusting the amount of the input signal which is provided to each output channel. E.g., for a stereo mix environment (i.e., an environment which includes a left output channel and a right output channel) an input signal which is located directly to the right of the listener is provided to the right output channel while the left output channel is attenuated. The relative level/position of the input signal is controlled by adjusting the amplitude of the input signal which is provided to each output channel.
It is also known to mix sound input signals using a computer system which imitates the controls of a sound mixing board. In such a computer system, the controls are displayed on the screen of the computer as they might appear on the mixing board. Steinberg/Jones, Northridge, Calif. produces such a computer system under the trade designation MIMIX.
It is also known to use external processors such as compressors, limiters, equalizers, reverb units, delay units, to achieve certain effects, e.g., in synthesizing sound or in editing of audio and audiovisual works.
Our invention features in general a system for controlling multiple parameters affecting an audio output. The system employs a controller with a visual display (e.g., a cathode ray tube) for displaying an icon that is a visual representation of an input sound signal in a multidimensional space. An input device (e.g., a mouse) is used to control the location of the icon on the display. The controller generates a multiple parameter control signal that is based upon the location of the icon and is used by a sound signal processing circuit to control multiple parameters affecting an audio output. With the invention, the states of multiple parameters can be simultaneously controlled and usefully displayed, permitting precision and flexibility in real time dynamic sound control, and permitting the sound engineer to achieve complex mixes with a large number of parameters.
In preferred embodiments a plurality of icons can be used to control a plurality of input sound signals. The parameters could include gains for respective outputs in a multiple output channel environment (e.g., a stereo mix or Dolby system), one or more reverb parameters, equalization parameters, filtering parameters, and compression parameters. The multidimensional space could represent the physical location of a listener in a stereo environment relative to the sources of sound, and the display could include hard left and hard right regions and a nonlinear fadeout region. The units of gain associated with the display could be rescalable. The multidimensional space could alternatively represent a scene of an audiovisual work; locations in the space associated with locations in the scene would have associated audio parameter sets, e.g., stored in a look-up table. State information as to positions of the icons at any instantaneous time and changes in position with time can be stored for later playback, editing and synthesis.
In another aspect, our invention features, in general, a system employing icons that are images of the sources of input sound signals on a visual display for interactively controlling multiple parameters affecting an audio output. This facilitates identification of the source of the input sound signal by the sound engineer, something that is particularly advantageous when there are a large number of input sound signals to control and monitor.
In another aspect, our invention features, in general, a system for interactively controlling the modification of M input sound signals to provide N output sound signals. The system includes a sound signal processing circuit having M times N controllable amplifiers arranged in an M by N matrix in which each of the M inputs is distributed to N controllable amplifiers and the outputs of M controllable amplifiers are combined to provide each of the N outputs. Each controllable amplifier has a unique set of an input signal and an output signal and receives a unique, continuously variable gain control signal. Such a system provides great flexibility in achieving complex sound mixes and complex real time changes in sound mixes.
In preferred embodiments, each controllable amplifier includes a digital-to-analog converter (DAC) that receives a multibit gain control signal and a voltage controlled amplifier that receives the control voltage output of the DAC; the DACs have slew limiting of the control voltage to eliminate control voltage artifacts in the output signals; and a decoder circuit receives gain control signals for all controllable amplifiers with addresses identifying which gain control signals apply to which amplifiers.
Other advantages and features of the invention will be apparent from the following description in a preferred embodiment thereof and from the claims.
The drawings are first briefly described.
FIG. 1 shows a diagrammatic-block diagram of a sound mixing system according to the invention.
FIG. 2 shows a schematic-block diagram of a decoder circuit of the FIG. 1 sound mixing system.
FIG. 3 shows a schematic-block diagram of an arrangement of controllable amplifiers of the FIG. 1 sound mixing system.
FIG. 4 shows a schematic-block diagram of line level trim circuits of the FIG. 1 sound mixing system.
FIGS. 5A and 5B show examples of displays for the FIG. 1 mixing system.
FIGS. 6A to 6C show a flow chart of a partial control module of the FIG. 1 sound mixing system, which is responsible for the real time mixing.
FIG. 7 is a block diagram of an alternative sound signal processing matrix configuration for the FIG. 1 system.
Referring to FIG. 1, in real time, interactive sound mixing system 10, computer system 12 communicates with mixing circuit 14 via 16-bit data bus 16 (only 12 bits are used in this embodiment) and 8-bit address and control bus 17 (6 bits for address and 2 bits for control), as well as with display 18 (e.g., a color cathode ray tube (CRT)) via display interconnect 20. Computer system 12 includes computer 30 (e.g., an Atari 1040ST), which may be controlled via either keyboard 32 or mouse 34, as well as storage media 36 (e.g., a disk drive), RGB (red, green, blue) interface 38 and control module 40, which is software loaded on computer 30. (A computer listing for the software is incorporated herein as the Appendix.) Computer 30 utilizes a GEM operating system. The software for control module 40 is implemented in an interpretative and compilable GFA Basic (3.0 Version) available from Antic Software, San Francisco, Calif. Mouse 34 includes two activation buttons, right button 35a and left button 35b. Computer system 12 acts as an interactive controller that generates multiple parameter control signals provided to mixing circuit 14, which acts as a sound signal processing circuit that modifies input sound signals to provide output sound signals based upon the multiple parameter control signal.
Mixing circuit 14 includes voltage-controlled amplifier (VCA) cross-point matrix 44, a programmable-gain matrix, which receives sound input signals (e.g., from analog sound sources such as tapes, compact disks, synthesizers, musical instruments and voices) and provides sound output signals to right and left line level trim circuits 46, 48. Trim circuits 46, 48 provide sound output signals to audio amplifier 50, which provides audio output signals to left and right speakers 52, 54 or recording devices.
Cross-point matrix 44 includes address decoder circuit 60, which receives address and control information over address and control bus 17, as well as a plurality of controllable amplifiers 62, which receive amplitude control information over data bus 16 as well as sound input signals via input channels 64, 66, 68, 70. Controllable amplifiers 62 provide sound output signals via output channels 72, 74. Each input channel 64, 66, 68, 70 provides its respective input signal to a pair of controllable amplifiers 62, which each provide an output signal on a respective output channel 72, 74. Accordingly, the number of controllable amplifiers 62 in the VCA cross-point matrix is defined by the number of input channels times the number of output channels. (I.e., 2 output channels×4 input channels=8 amplifiers.)
Referring to FIG. 2, decoder circuit 60 includes an arrangement of three 3:8 decoders. More specifically, enable decoder 80 (available under the trade designation 74LS138) receives three enable bits via address and control bus 17 and provides enable signals to decoders 82, 84 (available under the trade designation 74LS138), which also receive three address bits and DATA CLK (one of the control bits) via bus 17. Decoders 82, 84 provide a plurality of active low enable signals LATCH DAC 0/ to LATCH DAC 9/.
Referring to FIG. 3, each controllable amplifier 62 includes register circuit 90, which receives twelve bits of data via data bus 16, digital-to-analog converter (DAC) 92 (available from Analog Devices under the trade designation AD565), which receives registered data from register circuit 90, and voltage controlled amplifier (VCA) 94 (available from P.M.I., Inc. under the trade designation 2120), which receives a V DAC control signal from DAC 92 as well as a sound input signal via an input channel. VCAs 94 provide a logarithmic output gain for a linear voltage control input signal. A remote ground signal (VGND) is also provided to the gain adjust input terminal of each VCA 94; thus, the gain of all the VCAs is adjusted equally.
Still referring to FIG. 3, the output terminals of four controllable amplifiers 62 are tied together to provide a summation of output currents and provide a signal to operational amplifier OPI which provides a first sound output signal (e.g., the right sound output signal). The output terminals of four other controllable amplifiers 62 are tied together to provide a second sound output signal (e.g. the left sound output signal).
Referring to FIG. 4, line level trim circuits 46, 48 each include a DAC (not shown), which receives a digital signal from data bus 16 and provides an input control signal (VDAC 8 or VDAC 9) to a respective VCA 194. Each VCA also receives a sound output signal from one of the chains of controllable amplifiers. Each level trim circuit VCA 194 provides a sound output to the respective left or right channel of audio amplifier 50. Each channel of audio amplifier 50 includes a pair of cascaded operational amplifiers.
Referring to FIGS. 5A and 5B, sound mixing system 10 uses display 18 to provide video display 100 which displays sound input signals in a two-dimensional sound output channel space. FIG. 5A shows an empty video display (i.e., the sound output channel space), while FIG. 5B shows a video display 100 which displays the mix of four input signals. The video display 100 of an output signal space for a stereo mix environment includes a mix portion 102 (the output channel space) which displays angle of displacement from center as θ and radial distance, which is related to relative sound amplitude, in decibels (dB).
Mix portion 102 (where icons 106 are moved to actively mix an input signal) includes elliptical boundary 103 which delineates main mix portion 104 (where the majority of the mixing is accomplished) and nonlinear fadeout portion 107 (where icons are moved to gradually fade an input sound signal to mute, i.e., -90 dB). At the bottom of main mix portion 104 are hard left portion 104a and hard right portion 104b. Main mix portion 102 also includes amplitude indicators 109, which generally indicate attenuation relative to listener origin 111, and position indicators 112, which generally indicate angular position relative to listener origin 111. Amplitude indicators 109 could be rescaled (e.g, to show 3 dB increments instead of 5 dB increments), if desired. The position of mouse 34 is displayed by cursor 105, and input channels are displayed as icons 106, which are visual representations of the input sound signal sources. Icons 106 may have the appearance of the element providing the respective sound input signal. (E.g., a guitar which provides a sound input signal would have a guitar shaped icon.)
Video display 100 also includes information portion 99, which provides information about input channels which are displayed on video display 100. Information portion 99 includes active icon portion 108, which displays the "active" icon as well as mix information (e.g., left and right gain and angular displacement relating to that icon), output channel portion 110, which provides information regarding the status of the output channels, events portion 114, which displays different stored states (i.e., a particular configuration of icons) of a particular mix, groups (GRPS) portion 115, which displays whether the active icon is part of a group of icons which are configured to move together, and SMPTE (Society of Motion Picture and Television Engineers) time-code readout 117, which provides information relating to time reference coding of a particular mix.
Referring to FIGS. 1, 5A and 5B, sound mixing system 10 is used to mix one or more sound input signals into a plurality of sound output signals by displaying the sound input signals in a multidimensional sound output signal space. (FIGS. 5A and 5B show the output signal space for a stereo mix environment.) Mouse 34 and keyboard 32 are used to move icons representing sound input signals around the screen display representing the space of the sound output signals. As the icons are moved, the mix of the input signals changes.
More specifically, when cursor 105 is positioned over an icon 106, right button 35a of mouse 34 is depressed, and that icon 106 becomes the active icon (i.e., the icon 106 which represents the input signal which is being mixed). Left button 35b is depressed, and the active icon is moved around display 100 by moving mouse 34 according to known techniques. If both right and left buttons are depressed over the active icon, then the icon may be moved along or around a constant radius. For example, in the stereo mix environment, as an icon moves around the central radius, the relative position of the sound input signal (i.e., the amount of the sound input signal being supplied to the right output channel and the amount of the sound input signal being supplied to the left output channel) changes. Also, as the icon moves along the radius, the strength (i.e., the gain) of the sound input signal changes. Thus, by simply moving an icon to a desired mix position, the mix of the input sound signal is accomplished. After the mix of the input signal has been completed, the position of the icon on the display may be stored for future use with disk drive 36.
Referring to FIGS. 6A-6C, a flow chart of control module 40 of real time mix system 10 is shown. Line references in the flow chart are to the lines of the listing in the Appendix hereto. More specifically, at initialization step 120, normalized coefficients are precomputed and ranges of variables are initialized. The left and right gain coefficients are computed using the following equations:
left-- dB-- coef=20×LOG10 (MAX (SIN (90-i)/2),0.001)
right-- dB-- coef=20×LOG10 (MAX (COS (90-i)/2),0.001)
MAX=a function which provides the greater of the two numbers in the parenthetical, and
i=an angular position relative to a listener.
The coefficients of dB v. angular position (relative to listener) are stored in a two-dimensional look-up table, thus, allowing module 40 to function faster because these coefficients need not be generated for every mix action. At real time mix step 122, control module 40 has finished its initialization and starts a mix of input sound signals to stereo output signals.
At setup graphic buffers step 124, graphic buffers are provided so that a previous mix (such as a mix which is retrieved from disk drive 36) may be stored while a present mix is being generated. Also, in step 124, the resolution of display 18 is configured to allow computer 30 to function in an environment having 320 locations in the x direction by 200 locations in the y direction. Also, computer 30 causes 320 by 200 pixels to be displayed on display 18 and mouse 34 to be initialized for this configuration. By initializing mouse 34 for this configuration, a location on display 18 corresponds to a functional location of computer 30. Thus, computer 30 may be controlled simply by referring to cursor 105 which is controlled by mouse 34 and displayed on display 18.
At icon step 126, control module 40 determines if any icons are present on display 100. If icons are present, then there is a preexisting mix of information; if no icons are present, then there is no mix present and, thus, the mix need not be updated at this time.
If there is an icon present, then at step 128, the graphics are updated. More specifically, the x and y coordinates of the icon are used to update all the values displayed on screen 100 as well as to update each controllable amplifier 62.
At step 130, mouse 34 and keyboard 32 are sampled to determine if any action has been taken. At step 132, the results of the sample are analyzed to determine if any special function keys have been pressed. If so, then step 134, which includes subroutines for the special functions, is used to instruct computer 30 how to react to a particular function key. Functions include: increase gain, decrease gain, pan left, pan right, etc.
At step 136, video display 100 is scanned to determine if cursor 105, which is controlled by mouse 34, is in mix portion 102. If cursor 105 is not in the mix area, then at step 138, the mouse may be used to control special functions of control module 40 (e.g., the mouse may be used to control a pull-down menu). The pull-down menu is an alternative to the special function keys used in step 132.
At step 140, control module 40 checks to see if right button 35a of mouse 34 has been activated. If so then, at step 142, control module 40 checks to see if any icon is targeted by mouse 34 (i.e., to see if the cursor, which is controlled by mouse 34, is positioned over an icon). If so then, at step 144, control module 40 initializes a new icon and updates the graphics of active icon portion 108 of display 100 to indicate the new active icon. If no new icon is activated, then control module 40 continues to operate with a prior icon.
At step 146, a moving average is set up for the moving mouse. More specifically, seven previous locations of the mouse are stored and weighted together in a conventional manner. Because module 40 is functioning so quickly, the seven previous locations are generated at a speed which is virtually transparent to the user and listener. Each time the mouse is updated (i.e., moved) then the standard traveling average is automatically updated. Thus, movements of the mouse provide a substantially continuously flowing output signal. The weighing function may be adjusted in a conventional manner to provide a faster or slower ramping movement of mouse 34, and thus of the mix.
At step 148, the mouse position is analyzed to determine if DACs 92 need to be updated (e.g., due to movement of mouse 34 or use of special functions). If so then, at step 150, control module 40 updates the coordinates displayed by video display 100 and provided to controllable amplifiers 62; mouse cursor 105 is also hidden so that mouse 34 moves an icon 106 rather than cursor 105.
At step 152, control module 40 determines whether mouse 34 is activated by toggling one or both of its buttons 35a, 35b. If mouse 34 is activated, then at step 154, the x, y coordinates of mouse 34 are sampled, a weighted path for mouse 34 is computed, and a limit is imposed to prevent the icon from moving beyond the mix zone.
At step 156, gain information (r) and pan information (θ) are computed relative to listener origin 111 by converting standard x, y location values to polar coordinates relative to the listener origin. The listener origin may be different from the computer screen origin. E.g., in the stereo environment shown in FIGS. 5A and 5B, the listener origin is at the center, bottom screen.
At step 158, left dB and right dB values are computed based upon the r and θ values. In general, the r value is first used to calculate a value for a "bels" variable; if the icon is within region 104, r is simply multiplied times a scaling factor "d-scale", which has a value of 0.1875; if the icon is outside of region 104 in region 107, bels is calculated by first multiplying r times d-scale and then rescaling the value so obtained in a manner that provides realistic fadeout of the signal, the rescaling being a function of the value of bels and θ. Following the computation of bels, it is determined if the icon is in an angle of acceptance 159 (see FIG. 5B). If not, left dB and right dB values are now computed using the coefficients stored in the look-up table which was generated at step 120. More specifically,
dB-- left=MAX(bels+dB-- coef(a+90,0),-90)
dB-- right=MAX(bels+dB.sub. coef(a+90,1),-90)
dB-- left=the variable left dB component for the active icon,
dB-- right=the variable right dB component for the active icon,
MAX=a function which provides the greater of the two numbers in parenthetical,
If the icon is within angle of acceptance 159, then mono output signals (left and right gain equal to -3 dB below the computed gain) are provided. More specifically, if the icon is located in the angle of acceptance, then:
dB-- left=MAX(bels-3,-90), and
MAX=a function which provides the greater of the two numbers in parenthetical.
Thus, when an icon is in the angle of acceptance, dB left and dB right are equal, each being 3 dB less than the bels radius value.
At step 160, the left and right dB values, which were computed in step 158, are displayed.
At step 166, mouse 34 is checked to see if it is moving by checking the x, y location values. If it is, then, at step 163, a final check is performed to determine if cross-point matrix 44 should be enabled. If matrix 44 is enabled then, at step 164, a six-bit address is generated; the address represents controllable amplifiers 62 which are assigned to active channels. This address is provided to decoder circuit 60. Decoder circuit 60 then provides enable signals to the DACs associated with the appropriate controllable amplifiers 62.
At step 166, control module 40 checks to confirm that the addressed controllable amplifiers 62 correspond to active channels. If the channels are inactive then at step 168, control module 40 instructs computer 30 to provide a mute signal to the addressed controllable amplifiers 62.
If the channels are active, then, at step 170, a 12-bit data word, which represents right gain, is generated for the right output VCA 194, and a 12-bit data word, which represents left gain, is generated for the left output 194. A unity offset is included in these computations. The unity offset compensates for the offset of individual controllable amplifiers 62. More specifically,
right-- gain=(offset1-dB-- right)×20,
left-- gain=(offset2-dB-- left)×20,
offset1=the offset of the right controllable amplifier, and
offset2=the offset of the left controllable amplifier.
Because VCAs 94 of controllable amplifiers 62 provide a logarithmic output level for a linear input, the left and right input gain signals may be linear.
At step 172, these data words are provided to controllable amplifiers 62. More specifically, DAC 92 of each activated controllable amplifier 62 receives this data word via register circuit 90 and provides a voltage (V DAC) to a respective VCA 94. VCA 94 receives the voltage which is linear and provides an output signal which is the appropriate dB level. DACs 92 provide slew limiting of the control voltage to eliminate control voltage artifacts in the signal output.
At step 174, the mouse coordinates of the active icon are updated, and, at step 176, the entire mix is updated. At step 178, the x and y coordinates are updated and the r and θ values are stored for the active icon. At step 180, the keyboard is sampled to determine if there was any keyboard activity and at step 182 the mouse is scanned to determine if a constant R or θ mode is chosen. If so then, at step 184, the data is computed to provide the icon along a constant R path or along a constant θ radius.
Control module then returns to step 152, and if mouse 34 is activated, repeats the mix for another icon. If mouse 34 is not activated then, at step 186, R and θ are computed for the mouse location and, at step 188, the keyboard is scanned to determine if any special commands were generated. Special commands may include whether to store the mix and prepare to end the present mix session. If not, then control module 40 returns to sample mouse 34 at step 130. If so, then the special commands are processed at step 190. Control module 40 then checks to see if the ESC key was pressed at step 192. If not, then control module 40 again determines whether any icons are present at step 126. If so, then control module 40 terminates the mix at step 194.
Other embodiments of the invention are within the scope of the following claims.
For example, system 10 may be used to mix input signals into greater than two output channels, e.g., a quadraphonic mix environment or a Dolby surround sound mix environment (having left front, center front, right front, and rear outputs), and the visual display could be in three or more dimensions.
Also, the icons could change in appearance (e.g., color, size, brightness) to indicate a change in parameter value (e.g., compression, equalization, grouping).
Additionally, regions may be defined in the multidimensional space, such that movement of an icon to that location will cause automatic generation of a control signal affecting any number of parameters.
This mapping ability of the invention can, e.g., be used in mixing voice and music for film or television. When mixing for film or television, often music is mixed with the narrative to enhance the presentation of the work. The music is often mixed somewhat lower than the narrative when narrative is present and is brought to near full level when there is no narrative, to fill in the gap. It is often necessary to limit the dynamic ranges of the narrative and the music to achieve better intelligibility. The narrative often has compression and limiting employed to contain the dynamic peaks of the announcer while the music is heavily compressed, so that no offending peaks interfere with the voice over and to keep the music from "dropping out" or disappearing in the mix. This compression of the music is made even more necessary if the music has a great dynamic range. When the music level is brought up to full level, in a case where there is no narrative, the excessive compression used for mixing under will be heard, often resulting in music without dynamics and sometimes audible "pumping" from the compressor's inability to deal with the complex waveform. The engineer typically selects a compression ratio that achieves punch when the music is full and contains the music sufficiently when it is mixed under. Unless the engineer has special equipment at his disposal, the compression ratio parameter is not changed during a typical mix because it is too complicated to change that parameter or many parameters dynamically. Tracking a change in the mix of voice and music with a change in a parameter, such as compression ratio, is very difficult, especially if many changes are required.
With the interactive control apparatus according to the invention, the change in compression slope can be mapped into the output space so that simply moving the icon to change gain also outputs control data to an external, dynamically controlled compressor that has its compression slope altered to fit the requirements for the music being full level or being mixed under to accommodate the narrative. In such a case, the compression slope can be adjusted to precisely provide the optimum compression for the music under and music full periods.
Another example for film mixing is the use of mapping templates for each scene of the movie. A template would consist of a 2-D or 3-D "floor plan" of the scene, setup by the engineer. Different areas would be outputted to dynamically control external audio processing devices, e.g., compressors, limiters, equalizers, reverb and delay units. If an icon is placed into any of these regions so mapped, the coordinates of the icon position are used to address data in a look-up table or derived from a function, and the data are outputted dynamically to the external device processing the audio channel represented by the icon. E.g., if the actor were to walk behind an object, the parameters for reverb, early reflection, and equalization could be altered to achieve the effect of someone talking from behind the object. These data outputted are in addition to the gain change to one or more output channels, e.g., based upon the position of the actor. The multiple parameter control data would also be layered, so that an icon placed in a region so mapped, would effect the change of one or more external devices associated with that channel or any other channel. Movement of one sound source could also be programmed to affect another sound source; e.g., if one sound source (associated with the active icon) moves in front of another sound source, the second source could be automatically attenuated even though its icon is not active. Different layers could also be selected or deactivated at will. An extra output bus of the gain matrix could be the source of the signal to the external processing, allowing one or more channels to contribute to the signal sent to an external device, e.g., a reverb unit.
The overall mapping concept could involve primary gain settings to the mix, secondary gain settings (dependent on or independent of the primary gain setting) to feed external processing, and control information for the dynamic alteration of external processing parameters (reverb decay, delay time, compression slope, limiter attack time, limiter release time, etc.).
The return from the external processing device could re-enter the gain matrix on another previously unused or additional input channel and also be represented by an icon. As shown in FIG. 7, two output busses 202, 204 from matrix 206 feed signals into two independent audio effects processors 208, 210 that are dynamically reprogrammable via MIDI control line 212.
Output busses 224a, 224b are for the main mix and are used to feed the recording device or listening apparatus. Busses 226a, 226b are the effect busses used to provide primary feeding back into processors 208, 210. Module 220 allows monitoring the output of mix busses and effect busses to confirm proper operation.
Due to the universal nature of the variable gain matrix, all input channels can contribute to a main stereo mix while one or more dependent or independent submixes can be outputted to the external effects (reverb, delay, pitch . . . ), as well as monitoring busses (222a, 222b). Dependent means software simulation of postfade, and independent means software simulation of prefade. Prefade and postfade are terms used on conventional consoles to indicate if a submix is after the main channel fader (dependent) or before the main channel fader (independent). The stereo returns from the external effect devices are re-entered into the variable gain matrix, and can therefore contribute to the main mix or one or more submixes. The software will disallow or mute the various nodes that might contribute to feedback.
With an M×N variable gain matrix employed, main channel signals as well as effect returns can be uniquely controlled from the interactive controller. The outputs of the matrix could feed recording devices, effect processors, monitoring busses. For instance two busses 222a, 222b could be dedicated as a monitoring buss. Because the buss is easily reconfigurable via software, this monitoring buss could be dynamically reconfigured for stereo "in-place" solo, normal stereo monitoring, group solo, mute in monitor but not in main mix, prefade (independent) listen, etc.
The variable gain matrix in a large enough configuration functions as a universal, dynamically reprogrammable "console" with a graphics front end. Under software control, configurations that must be "hardwired" into a current console, can be simulated via manipulation of the various gains at various matrix nodes.
The system may also be used for theater playback to play back special effects, in recording studios for the main mix or for special mix effects, in television post production for the main mix or for easy repeatable automated stereo mixing to timecode, in live performances to allow the mixing to be remote from the amplifiers (e.g., by using an optical link for busses 16, 17), in pavilions to allow live mixing to the audience from one or more tracks, in radio production rooms where space and versatility are a factor, in mixing sound effects for simulators (e.g., aircraft, space, submarine) or in satellite mix-minus setups to allow easy recall of complicated mixes.
Besides interactive input devices that permit direct interaction with the sound engineer, other sources of controlling information can be used to change the position of icons and appearance, e.g., a completely independent computer drive console, a source of MIDI data, a MIDI sequencer, a source of computer generated data, and an external audio or visual editing system. ##SPC1##
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US3757046 *||Jul 23, 1970||Sep 4, 1973||T Williams||Control signal generating device moving sound speaker systems including a plurality of speakers and a|
|US3982071 *||Aug 20, 1974||Sep 21, 1976||Weiss Edward A||Multichannel sound signal processing system employing voltage controlled amplifiers|
|US4399328 *||Feb 23, 1981||Aug 16, 1983||U.S. Philips Corporation||Direction and frequency independent column of electro-acoustic transducers|
|US4683589 *||Nov 27, 1985||Jul 28, 1987||Scholz Research & Development||Electronic audio system|
|US4731848 *||Oct 22, 1984||Mar 15, 1988||Northwestern University||Spatial reverberator|
|US4792974 *||Aug 26, 1987||Dec 20, 1988||Chace Frederic I||Automated stereo synthesizer for audiovisual programs|
|US4817149 *||Jan 22, 1987||Mar 28, 1989||American Natural Sound Company||Three-dimensional auditory display apparatus and method utilizing enhanced bionic emulation of human binaural sound localization|
|US4864625 *||Sep 2, 1986||Sep 5, 1989||Casio Computer Co., Ltd.||Effector for electronic musical instrument|
|US4885792 *||Oct 27, 1988||Dec 5, 1989||The Grass Valley Group, Inc.||Audio mixer architecture using virtual gain control and switching|
|US4933768 *||Jul 19, 1989||Jun 12, 1990||Sanyo Electric Co., Ltd.||Sound reproducer|
|US5027689 *||Aug 31, 1989||Jul 2, 1991||Yamaha Corporation||Musical tone generating apparatus|
|1||"Andrew Von Gamm takes a look at AKG's Creative Audio Processor", Studio Sound, Oct. 1989, pp. 38, 40.|
|2||"Ballade", sequencer for MT-32, advertisement (Pynaware Corp., 1988) in Electronic Musician, Dec., 1988, p. 19.|
|3||*||A brochure by AKG Acoustics, Oct. 1988, printed in Austria.|
|4||*||Andrew Von Gamm takes a look at AKG s Creative Audio Processor , Studio Sound, Oct. 1989, pp. 38, 40.|
|5||*||Ballade , sequencer for MT 32, advertisement (Pynaware Corp., 1988) in Electronic Musician, Dec., 1988, p. 19.|
|6||*||Chamberlin, Musical Applications of Microprocessors, 1980, p. 268.|
|7||Croft, Simon, "JMS C-Mix Fader Automation" Sound (Oct. 1987) pp. 52, 54.|
|8||*||Croft, Simon, JMS C Mix Fader Automation Sound (Oct. 1987) pp. 52, 54.|
|9||Digital Music Services brochure entitled "DMP7 PRO" and dated 1987.|
|10||*||Digital Music Services brochure entitled DMP7 PRO and dated 1987.|
|11||JL Cooper Electronics brochure entitled "Affordable Automation" and distributed in 1987.|
|12||*||JL Cooper Electronics brochure entitled Affordable Automation and distributed in 1987.|
|13||Keller, Helmut, "Universal Switchpoint Matrix and Automation Fader/Mixer System", An Audio Engineering Society Preprint, presented Mar. 1-4, 1988, Preprint 2611 (G-8).|
|14||*||Keller, Helmut, Universal Switchpoint Matrix and Automation Fader/Mixer System , An Audio Engineering Society Preprint, presented Mar. 1 4, 1988, Preprint 2611 (G 8).|
|15||Lehrman, Paul D., "The Future of MIDI Time Code", Recording Engineer/Producer, Oct. 1987, pp. 106-107, 110-113.|
|16||*||Lehrman, Paul D., The Future of MIDI Time Code , Recording Engineer/Producer, Oct. 1987, pp. 106 107, 110 113.|
|17||*||MegaMix advertisement in Post, Sep. 1987, p. 34.|
|18||*||Nadler, Wilhelm, Dipl Ing., Delta Stereo Compact Processor (DSP 610) to Utilize a New Directional Sound Reinforcement System , An Audio Engineering Society Preprint, presented Nov. 12 16, 1986, Preprint 2391 (D 17).|
|19||*||Nadler, Wilhelm, Dipl Ing., The DSP 610 A Compact to Utilise a New Directional Sound Reinforcement System , An Audio Engineering Society Preprint, presented Mar. 10 13, 1987, Preprint 2472 (K 2).|
|20||Nadler, Wilhelm, Dipl-Ing., "Delta Stereo Compact Processor (DSP 610) to Utilize a New Directional Sound Reinforcement System", An Audio Engineering Society Preprint, presented Nov. 12-16, 1986, Preprint 2391 (D-17).|
|21||Nadler, Wilhelm, Dipl-Ing., "The DSP 610--A Compact to Utilise a New Directional Sound Reinforcement System", An Audio Engineering Society Preprint, presented Mar. 10-13, 1987, Preprint 2472 (K-2).|
|22||Richter, F. and Persterer, A. "Design and Applications of a Creative Audio Processor" An Audio Engineering Society Preprint, presented Mar. 7-10, 1989, Preprint 2782 (U-4).|
|23||*||Richter, F. and Persterer, A. Design and Applications of a Creative Audio Processor An Audio Engineering Society Preprint, presented Mar. 7 10, 1989, Preprint 2782 (U 4).|
|24||Ryle, G., "Multi-Tasking for Musicians" Start, vol. 3, No. 12 Jul. 1989 pp. 24-25.|
|25||*||Ryle, G., Multi Tasking for Musicians Start, vol. 3, No. 12 Jul. 1989 pp. 24 25.|
|26||Sippl, Fritz, D. Ing., F.A.E.S., "New Trends in Professional Audio", Issue 1990.|
|27||*||Sippl, Fritz, D. Ing., F.A.E.S., New Trends in Professional Audio , Issue 1990.|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US5390138 *||Sep 13, 1993||Feb 14, 1995||Taligent, Inc.||Object-oriented audio system|
|US5402501 *||Jul 27, 1993||Mar 28, 1995||Euphonix, Inc.||Automated audio mixer|
|US5420933 *||Oct 13, 1993||May 30, 1995||Sony Electronics Inc.||Up and down-loadable VTR configuration for an audio follow video mixer|
|US5487067 *||Feb 1, 1994||Jan 23, 1996||Sony Corporation||Audio data communications|
|US5524060 *||Feb 14, 1994||Jun 4, 1996||Euphonix, Inc.||Visuasl dynamics management for audio instrument|
|US5555310 *||Feb 14, 1994||Sep 10, 1996||Kabushiki Kaisha Toshiba||Stereo voice transmission apparatus, stereo signal coding/decoding apparatus, echo canceler, and voice input/output apparatus to which this echo canceler is applied|
|US5576507 *||Dec 27, 1994||Nov 19, 1996||Lamarra; Frank||Wireless remote channel-MIDI switching device|
|US5608807 *||Apr 23, 1996||Mar 4, 1997||Brunelle; Thoedore M.||Audio mixer sound instrument I.D. panel|
|US5617480 *||Jan 6, 1995||Apr 1, 1997||Ford Motor Company||DSP-based vehicle equalization design system|
|US5636283 *||Apr 15, 1994||Jun 3, 1997||Solid State Logic Limited||Processing audio signals|
|US5675557 *||Jul 27, 1994||Oct 7, 1997||Carlos Lores Borras||Integrated mixing system for synchronizing video and audio signals|
|US5700966 *||Jul 13, 1995||Dec 23, 1997||Lamarra; Frank||Wireless remote channel-MIDI switching device|
|US5715318 *||Nov 2, 1995||Feb 3, 1998||Hill; Philip Nicholas Cuthbertson||Audio signal processing|
|US5721851 *||Jul 31, 1995||Feb 24, 1998||International Business Machines Corporation||Transient link indicators in image maps|
|US5734731 *||Nov 29, 1994||Mar 31, 1998||Marx; Elliot S.||Real time audio mixer|
|US5740260 *||May 22, 1995||Apr 14, 1998||Presonus L.L.P.||Midi to analog sound processor interface|
|US5754660 *||Sep 20, 1996||May 19, 1998||Nintendo Co., Ltd.||Sound generator synchronized with image display|
|US5774566 *||Dec 11, 1995||Jun 30, 1998||Studer Professional Audio Ag||Device for treating signals|
|US5812674 *||Aug 20, 1996||Sep 22, 1998||France Telecom||Method to simulate the acoustical quality of a room and associated audio-digital processor|
|US5812688 *||Apr 18, 1995||Sep 22, 1998||Gibson; David A.||Method and apparatus for using visual images to mix sound|
|US5852800 *||Oct 20, 1995||Dec 22, 1998||Liquid Audio, Inc.||Method and apparatus for user controlled modulation and mixing of digitally stored compressed data|
|US5862229 *||Oct 9, 1997||Jan 19, 1999||Nintendo Co., Ltd.||Sound generator synchronized with image display|
|US5864868 *||Feb 13, 1996||Jan 26, 1999||Contois; David C.||Computer control system and user interface for media playing devices|
|US5910996 *||Mar 21, 1997||Jun 8, 1999||Eggers; Philip E.||Dual audio program system|
|US5949887 *||May 23, 1997||Sep 7, 1999||International Business Machines Corporation||Method and graphical user interface for configuring an audio system|
|US5959610 *||Jan 11, 1995||Sep 28, 1999||Euphonix||Computer-mirrored panel input device|
|US5969719 *||Jun 17, 1997||Oct 19, 1999||Matsushita Electric Industrial Co., Ltd.||Computer generating a time-variable icon for an audio signal|
|US6011851 *||Jun 23, 1997||Jan 4, 2000||Cisco Technology, Inc.||Spatial audio processing method and apparatus for context switching between telephony applications|
|US6057829 *||Aug 18, 1997||May 2, 2000||Euphonix, Inc.||Computer-mirrored panel input device|
|US6271841||Apr 2, 1999||Aug 7, 2001||Matsushita Electric Industrial Co., Ltd.||Information processor for changing a display in response to an input audio signal|
|US6359632 *||Oct 23, 1998||Mar 19, 2002||Sony United Kingdom Limited||Audio processing system having user-operable controls|
|US6390919 *||Aug 31, 2000||May 21, 2002||Sony Computer Entertainment Inc.||Entertainment system, entertainment apparatus, recording medium, and program|
|US6421692||Nov 20, 1998||Jul 16, 2002||Object Technology Licensing Corporation||Object-oriented multimedia [data routing system] presentation control system|
|US6459797 *||Apr 1, 1998||Oct 1, 2002||International Business Machines Corporation||Audio mixer|
|US6490359||Jun 17, 1998||Dec 3, 2002||David A. Gibson||Method and apparatus for using visual images to mix sound|
|US6599244 *||Dec 23, 1999||Jul 29, 2003||Siemens Medical Solutions, Usa, Inc.||Ultrasound system and method for direct manipulation interface|
|US6686522||Jun 22, 2001||Feb 3, 2004||Shinko Corporation||Musical instrument with a body made of polyurethane foam|
|US6702677 *||Oct 13, 2000||Mar 9, 2004||Sony Computer Entertainment Inc.||Entertainment system, entertainment apparatus, recording medium, and program|
|US6728382 *||Aug 10, 1999||Apr 27, 2004||Euphonix, Inc.||Functional panel for audio mixer|
|US6829017||Feb 1, 2001||Dec 7, 2004||Avid Technology, Inc.||Specifying a point of origin of a sound for audio effects using displayed visual information from a motion picture|
|US6839441||Jan 20, 1998||Jan 4, 2005||Showco, Inc.||Sound mixing console with master control section|
|US6850496||Jun 9, 2000||Feb 1, 2005||Cisco Technology, Inc.||Virtual conference room for voice conferencing|
|US6898291 *||Jun 30, 2004||May 24, 2005||David A. Gibson||Method and apparatus for using visual images to mix sound|
|US6901149 *||Jan 10, 2001||May 31, 2005||Korg Incorporated||Audio mixer|
|US6977653 *||Mar 8, 2000||Dec 20, 2005||Tektronix, Inc.||Surround sound display|
|US6981208||Jun 12, 2002||Dec 27, 2005||Object Technology Licensing Corporation||Multimedia data routing system and method|
|US7019205||Oct 13, 2000||Mar 28, 2006||Sony Computer Entertainment Inc.||Entertainment system, entertainment apparatus, recording medium, and program|
|US7054449 *||Sep 25, 2001||May 30, 2006||Bernafon Ag||Method for adjusting a transmission characteristic of an electronic circuit|
|US7123728 *||Aug 15, 2001||Oct 17, 2006||Apple Computer, Inc.||Speaker equalization tool|
|US7158844 *||Jul 21, 2003||Jan 2, 2007||Paul Cancilla||Configurable surround sound system|
|US7162045 *||Jun 16, 2000||Jan 9, 2007||Yamaha Corporation||Sound processing method and apparatus|
|US7224811 *||Nov 10, 1998||May 29, 2007||Yamaha Corporation||Audio system utilizing personal computer|
|US7305097 *||Feb 14, 2003||Dec 4, 2007||Bose Corporation||Controlling fading and surround signal level|
|US7343210 *||Jul 2, 2004||Mar 11, 2008||James Devito||Interactive digital medium and system|
|US7549123 *||Jun 15, 2005||Jun 16, 2009||Apple Inc.||Mixing input channel signals to generate output channel signals|
|US7572971||Nov 3, 2006||Aug 11, 2009||Verax Technologies Inc.||Sound system and method for creating a sound event based on a modeled sound field|
|US7636448||Oct 28, 2005||Dec 22, 2009||Verax Technologies, Inc.||System and method for generating sound events|
|US7643640 *||Oct 13, 2004||Jan 5, 2010||Bose Corporation||System and method for designing sound systems|
|US7698009||Oct 27, 2005||Apr 13, 2010||Avid Technology, Inc.||Control surface with a touchscreen for editing surround sound|
|US7729673||Mar 3, 2005||Jun 1, 2010||Sony Ericsson Mobile Communications Ab||Method and apparatus for multichannel signal limiting|
|US7742609||Apr 3, 2003||Jun 22, 2010||Gibson Guitar Corp.||Live performance audio mixing system with simplified user interface|
|US7774707||Apr 22, 2005||Aug 10, 2010||Creative Technology Ltd||Method and apparatus for enabling a user to amend an audio file|
|US7859533 *||Apr 4, 2006||Dec 28, 2010||Yamaha Corporation||Data processing apparatus and parameter generating apparatus applied to surround system|
|US7885396||Jun 23, 2005||Feb 8, 2011||Cisco Technology, Inc.||Multiple simultaneously active telephone calls|
|US7907738||Aug 30, 2006||Mar 15, 2011||Apple Inc.||Speaker equalization tool|
|US7994412||May 18, 2005||Aug 9, 2011||Verax Technologies Inc.||Sound system and method for creating a sound event based on a modeled sound field|
|US8027483 *||Sep 10, 2009||Sep 27, 2011||Bose Corporation||System and method for designing sound systems|
|US8046689 *||Mar 6, 2006||Oct 25, 2011||Apple Inc.||Media presentation with supplementary media|
|US8068105 *||Jul 18, 2008||Nov 29, 2011||Adobe Systems Incorporated||Visualizing audio properties|
|US8073160||Jul 18, 2008||Dec 6, 2011||Adobe Systems Incorporated||Adjusting audio properties and controls of an audio mixer|
|US8073169||Oct 31, 2007||Dec 6, 2011||Bose Corporation||Controlling fading and surround signal level|
|US8085269 *||Jul 18, 2008||Dec 27, 2011||Adobe Systems Incorporated||Representing and editing audio properties|
|US8193437 *||Mar 18, 2011||Jun 5, 2012||Yamaha Corporation||Electronic music apparatus and tone control method|
|US8255069 *||Aug 6, 2007||Aug 28, 2012||Apple Inc.||Digital audio processor|
|US8331575||Nov 22, 2010||Dec 11, 2012||Yamaha Corporation||Data processing apparatus and parameter generating apparatus applied to surround system|
|US8483410||Dec 3, 2007||Jul 9, 2013||Lg Electronics Inc.||Apparatus and method for inputting a command, method for displaying user interface of media signal, and apparatus for implementing the same, apparatus for processing mix signal and method thereof|
|US8520858||Apr 21, 2006||Aug 27, 2013||Verax Technologies, Inc.||Sound system and method for capturing and reproducing sounds originating from a plurality of sound sources|
|US8577676||Apr 17, 2009||Nov 5, 2013||Dolby Laboratories Licensing Corporation||Method and apparatus for maintaining speech audibility in multi-channel audio with minimal impact on surround experience|
|US8634491 *||Mar 10, 2010||Jan 21, 2014||Rockstar Consortium USLP||Method and apparatus for reducing the contribution of noise to digitally sampled signals|
|US8645864 *||Nov 5, 2007||Feb 4, 2014||Nvidia Corporation||Multidimensional data input interface|
|US9059690||Dec 17, 2013||Jun 16, 2015||Rpx Clearinghouse Llc||Method and apparatus for reducing the contribution of noise to digitally sampled signals|
|US9208821||Aug 6, 2007||Dec 8, 2015||Apple Inc.||Method and system to process digital audio data|
|US9544705||Aug 23, 2013||Jan 10, 2017||Verax Technologies, Inc.||Sound system and method for capturing and reproducing sounds originating from a plurality of sound sources|
|US9606620 *||Oct 14, 2015||Mar 28, 2017||Spotify Ab||Multi-track playback of media content during repetitive motion activities|
|US20010009584 *||Jan 10, 2001||Jul 26, 2001||Michio Suruga||Audio mixer|
|US20020044148 *||Sep 25, 2001||Apr 18, 2002||Bernafon Ag||Method for adjusting a transmission characteristic of an electronic circuit|
|US20030035559 *||Aug 16, 2001||Feb 20, 2003||Laurent Cohen||Trackball controller for built-in effects|
|US20040030425 *||Apr 3, 2003||Feb 12, 2004||Nathan Yeakel||Live performance audio mixing system with simplified user interface|
|US20040161126 *||Feb 14, 2003||Aug 19, 2004||Rosen Michael D.||Controlling fading and surround signal level|
|US20040240686 *||Jun 30, 2004||Dec 2, 2004||Gibson David A.||Method and apparatus for using visual images to mix sound|
|US20050036628 *||Jul 2, 2004||Feb 17, 2005||James Devito||Interactive digital medium and system|
|US20050129256 *||Feb 3, 2005||Jun 16, 2005||Metcalf Randall B.||Sound system and method for capturing and reproducing sounds originating from a plurality of sound sources|
|US20050185806 *||Mar 4, 2005||Aug 25, 2005||Salvador Eduardo T.||Controlling fading and surround signal level|
|US20050223877 *||May 18, 2005||Oct 13, 2005||Metcalf Randall B||Sound system and method for creating a sound event based on a modeled sound field|
|US20060078130 *||Oct 13, 2004||Apr 13, 2006||Morten Jorgensen||System and method for designing sound systems|
|US20060109988 *||Oct 28, 2005||May 25, 2006||Metcalf Randall B||System and method for generating sound events|
|US20060117261 *||Apr 22, 2005||Jun 1, 2006||Creative Technology Ltd.||Method and Apparatus for Enabling a User to Amend an Audio FIle|
|US20060148435 *||Mar 3, 2005||Jul 6, 2006||Sony Ericsson Mobile Communications Ab||Method and apparatus for multichannel signal limiting|
|US20060159286 *||Feb 21, 2006||Jul 20, 2006||Stiles Enrique M||Bessel array with non-empty null positions|
|US20060159287 *||Feb 21, 2006||Jul 20, 2006||Stiles Enrique M||MTM of bessels loudspeaker|
|US20060159288 *||Feb 21, 2006||Jul 20, 2006||Stiles Enrique M||Bessel dipole loudspeaker|
|US20060159289 *||Feb 21, 2006||Jul 20, 2006||Stiles Enrique M||Bessel array with full amplitude signal to half amplitude position transducers|
|US20060168150 *||Mar 6, 2006||Jul 27, 2006||Apple Computer, Inc.||Media presentation with supplementary media|
|US20060182298 *||Apr 11, 2006||Aug 17, 2006||Stiles Enrique M||Bessel soundbar|
|US20060206221 *||Feb 22, 2006||Sep 14, 2006||Metcalf Randall B||System and method for formatting multimode sound content and metadata|
|US20060251260 *||Apr 4, 2006||Nov 9, 2006||Yamaha Corporation||Data processing apparatus and parameter generating apparatus applied to surround system|
|US20060262948 *||Apr 21, 2006||Nov 23, 2006||Metcalf Randall B|
|US20060291670 *||Aug 30, 2006||Dec 28, 2006||Nick King||Speaker equalization tool|
|US20070003044 *||Jun 23, 2005||Jan 4, 2007||Cisco Technology, Inc.||Multiple simultaneously active telephone calls|
|US20070056434 *||Nov 3, 2006||Mar 15, 2007||Verax Technologies Inc.||Sound system and method for creating a sound event based on a modeled sound field|
|US20070136695 *||Jan 24, 2007||Jun 14, 2007||Chris Adam||Graphical user interface (GUI), a synthesiser and a computer system including a GUI|
|US20080025530 *||Jul 26, 2006||Jan 31, 2008||Sony Ericsson Mobile Communications Ab||Method and apparatus for normalizing sound playback loudness|
|US20080107293 *||Oct 31, 2007||May 8, 2008||Bose Corporation||Controlling Fading And Surround Signal Level|
|US20090043410 *||Aug 6, 2007||Feb 12, 2009||Matt Evans||Digital audio processor|
|US20090044122 *||Aug 6, 2007||Feb 12, 2009||Matt Evans||Method and system to process digital audio data|
|US20090222731 *||May 8, 2009||Sep 3, 2009||William George Stewart||Mixing input channel signals to generate output channel signals|
|US20100002889 *||Sep 10, 2009||Jan 7, 2010||Bose Corporation||System and method for designing sound systems|
|US20100014693 *||Dec 3, 2007||Jan 21, 2010||Lg Electronics Inc.||Apparatus and method for inputting a command, method for displaying user interface of media signal, and apparatus for implementing the same, apparatus for processing mix signal and method thereof|
|US20100223552 *||Mar 2, 2009||Sep 2, 2010||Metcalf Randall B||Playback Device For Generating Sound Events|
|US20100287476 *||May 10, 2010||Nov 11, 2010||Sony Corporation, A Japanese Corporation||System and interface for mixing media content|
|US20110054887 *||Apr 17, 2009||Mar 3, 2011||Dolby Laboratories Licensing Corporation||Method and Apparatus for Maintaining Speech Audibility in Multi-Channel Audio with Minimal Impact on Surround Experience|
|US20110064228 *||Nov 22, 2010||Mar 17, 2011||Yamaha Corporation||Data processing apparatus and parameter generating apparatus applied to surround system|
|US20110162513 *||Mar 18, 2011||Jul 7, 2011||Yamaha Corporation||Electronic music apparatus and tone control method|
|US20110222637 *||Mar 10, 2010||Sep 15, 2011||Nortel Networks Limited||Method and Apparatus for Reducing the Contribution of Noise to Digitally Sampled Signals|
|US20110230990 *||Nov 24, 2009||Sep 22, 2011||Creative Technology Ltd||Method and device for modifying playback of digital musical content|
|USRE44611||Oct 30, 2009||Nov 26, 2013||Verax Technologies Inc.||System and method for integral transference of acoustical events|
|CN101133552B||Aug 25, 2005||Jan 12, 2011||索尼爱立信移动通讯股份有限公司||Method and apparatus for multichannel signal limiting|
|DE19634155B4 *||Aug 23, 1996||Nov 18, 2010||France Telecom||Verfahren zur Simulation der akustischen Qualität eines Raumes und damit verbundener Audio-Digitaler Prozessor|
|DE102007034723A1 *||Jul 23, 2007||Mar 12, 2009||Hennings, Detlef, Dr.||Spatial sound pattern producing method, involves implementing automatic production of spatial acoustic pattern by modulation of sound channels of sound sources, and implementing automatic modulation of sound channels|
|EP0743767A2 *||Apr 17, 1996||Nov 20, 1996||Sony United Kingdom Limited||Audio mixing console|
|EP0743767A3 *||Apr 17, 1996||Sep 15, 1999||Sony United Kingdom Limited||Audio mixing console|
|EP1385146A1 *||Jul 10, 2003||Jan 28, 2004||Texas Instruments Incorporated||System and method for digitally processing one or more audio signals|
|EP1851656A2 *||Feb 22, 2006||Nov 7, 2007||Verax Technologies Inc.||System and method for formatting multimode sound content and metadata|
|EP1851656A4 *||Feb 22, 2006||Sep 23, 2009||Verax Technologies Inc||System and method for formatting multimode sound content and metadata|
|EP2092409A1 *||Dec 3, 2007||Aug 26, 2009||LG Electronics Inc.||Apparatus and method for inputting a command, method for displaying user interface of media signal, and apparatus for implementing the same, apparatus for processing mix signal and method thereof|
|EP2092409A4 *||Dec 3, 2007||Jan 13, 2010||Lg Electronics Inc|
|WO1998032223A2 *||Jan 15, 1998||Jul 23, 1998||Mackie Designs Inc.||Digital signal mixing architecture|
|WO1998032223A3 *||Jan 15, 1998||Nov 19, 1998||Mackie Designs Inc||Digital signal mixing architecture|
|WO1999008180A1 *||Aug 11, 1998||Feb 18, 1999||Hewlett-Packard Company||Multi-media display system|
|WO2001063593A1 *||Feb 19, 2001||Aug 30, 2001||Lucian Gontko||A mode for band imitation, of a symphonic orchestra in particular, and the equipment for imitation utilising this mode|
|WO2003087980A2 *||Apr 4, 2003||Oct 23, 2003||Gibson Guitar Corp.||Live performance audio mixing system with simplified user interface|
|WO2003087980A3 *||Apr 4, 2003||Feb 12, 2004||Gibson Guitar Corp||Live performance audio mixing system with simplified user interface|
|WO2005003927A2 *||Jul 2, 2004||Jan 13, 2005||James Devito||Interactive digital medium and system|
|WO2005003927A3 *||Jul 2, 2004||Jun 23, 2005||James Devito||Interactive digital medium and system|
|WO2006073515A1||Aug 25, 2005||Jul 13, 2006||Sony Ericsson Mobile Communications Ab||Method and apparatus for multichannel signal limiting|
|WO2006091540A3 *||Feb 22, 2006||Apr 16, 2009||Randall B Metcalf||System and method for formatting multimode sound content and metadata|
|WO2008066364A1||Dec 3, 2007||Jun 5, 2008||Lg Electronics Inc.|
|U.S. Classification||381/119, 715/727, G9B/27.051, G9B/27.012, 715/839, G9B/27.008, 715/978|
|International Classification||G10H7/00, H04S3/00, G11B27/028, G10H1/00, H04S7/00, G11B27/034, G11B27/34, H04H60/04|
|Cooperative Classification||Y10S715/978, G10H1/0008, G10H2210/295, G11B27/034, H04S3/00, H04S7/40, G11B27/34, G10H2220/111, G11B27/028, G10H7/002, H04H60/04|
|European Classification||H04S7/40, G11B27/028, G10H7/00C, H04H60/04, H04S3/00, G11B27/034, G11B27/34, G10H1/00M|
|Feb 28, 1990||AS||Assignment|
Owner name: VOYAGER SOUND, INC., MASSACHUSETTS
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST.;ASSIGNORS:DE VITT, DOUGLAS J.;KRUG, KRISTOPH D.;REEL/FRAME:005245/0867
Effective date: 19900228
|Dec 26, 1996||REMI||Maintenance fee reminder mailed|
|Apr 11, 1997||SULP||Surcharge for late payment|
|Apr 11, 1997||FPAY||Fee payment|
Year of fee payment: 4
|Dec 5, 2000||FPAY||Fee payment|
Year of fee payment: 8
|Dec 5, 2000||SULP||Surcharge for late payment|
Year of fee payment: 7
|Dec 12, 2000||REMI||Maintenance fee reminder mailed|
|Nov 18, 2004||FPAY||Fee payment|
Year of fee payment: 12