|Publication number||US20060013266 A1|
|Application number||US 10/990,339|
|Publication date||Jan 19, 2006|
|Filing date||Nov 16, 2004|
|Priority date||Sep 21, 1998|
|Also published as||US6839734|
|Publication number||10990339, 990339, US 2006/0013266 A1, US 2006/013266 A1, US 20060013266 A1, US 20060013266A1, US 2006013266 A1, US 2006013266A1, US-A1-20060013266, US-A1-2006013266, US2006/0013266A1, US2006/013266A1, US20060013266 A1, US20060013266A1, US2006013266 A1, US2006013266A1|
|Inventors||Andres Vega-Garcia, Mu Han, Donald Ryan, Thomas Pfenning, Rajeev Byrisetty, Stefan Solomon, Michael Clark|
|Original Assignee||Microsoft Corporation|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (3), Referenced by (18), Classifications (9), Legal Events (1)|
|External Links: USPTO, USPTO Assignment, Espacenet|
The present invention concerns computer systems, particularly systems, methods, and software which support real-time multimedia communications across a computer network.
Computers, particularly personal computers, have enjoyed, in recent years, an enormous growth in utility. Early computers allowed users to perform tasks such as word-processing and bookkeeping. Today, however, computers are being used also to manage, display, and manipulate multimedia data, such as digital video and audio. Additionally, computers have become everyday communications devices, nearly as common as telephones and televisions.
Much of this utility growth, especially for communications, stems from the fantastic, compounded growth of computer networks, such as the much heralded Internet. The Internet, a worldwide network of computers interconnected through private wiring systems and public telephone systems, functions as a planetary communications system for sending and receiving information from one computer to one or more other computers. The information can take almost any form, including text, audio, and video information.
Communicating audio and video information between computers can be problematic in at least two ways. First, many conventional computers include software for handling audio and video information which can be inconvenient to use. For example, many conventional computers include Microsoft's DirectShow software—a system of interconnectable software modules (or filters)—which allows computers to capture multimedia data into data-storage devices, such as hard drives, and to playback, or render, the captured data through then audio-video equipment. Thus, a user having the DirectShow software can link her computer to a website or other computer featuring an audio-video clip, download (or copy) the clip into her local bard drive, and then play back the downloaded copy on her computer. Unfortunately, audio-video clips often include a great amount of data which, in some cases, requires several minutes to download, ultimately inconveniencing the user with a long wait. Accordingly, there is a need to extend software, such as Microsoft's DirectShow software, with features that allow concurrent download and playback, or streaming, of multimedia data.
Second, many conventional computers also include conferencing software, such as Microsoft's NetMeeting (version 2.1) software, which allows two or more computer users to communicate interactively across a computer network via audio-only or audio-video transmissions. A network conference typically requires that each party to the conference communicate using a common protocol not only for organizing and transmitting the audio and video data, but especially for encoding and decoding it. The common protocol ensures that each party to the conference ultimately understands what the other parties are communicating. Unfortunately, not all conferencing software uses the same protocol, frequently preventing users from network conferencing with users having different software. Accordingly, there is a need for network conferencing software that operates with more than one communications protocol.
In sum, there remains a need for a systems, methods and software which support streaming of multimedia data and multimedia conferencing via more than one communications protocol.
To address these and other needs, the inventors devised several new “pluggable,” or interconnectable, software modules for addition to Microsoft's DirectShow software. Some of the modules allow the DirectShow software to handle real-time multimedia data exchange on the network (full-duplex communications), which ultimately allows streaming and network conferencing using more than one communications protocol. One embodiment of the invention includes a module that supports the Real-time Transport Protocol (RIP) for streaming and another includes a demultiplexer module for routing the data, based on its type, to type-specific data handlers and decoders. In another embodiment, the demultiplexer module handles large multiparty conferences by monitoring incoming data streams and dynamically routing a subset of them to appropriate data handlers and decoders.
Another aspect of the invention concerns its modular, dynamically connectable architecture. The architecture includes a core set of modules which may be connected to emulate or form any type of conferencing software. In the exemplary embodiment, this conference software architecture includes an RTP source and rendering module, and RTP demultiplexer, send and receive payload handlers, and a set of decoder and encoder modules to handle various data types.
In its various embodiments, the invention encompasses systems, methods, computers, and computer-readable media of varying scope. In addition to the embodiments and advantages mentioned here, other embodiments and advantages will become apparent from the drawings and the following description.
The following detailed description, which references and incorporates
The exemplary embodiment of the invention concerns extensions of Microsoft's Direct Show software, specifically version 2.0, to support streaming and network conferencing using more than one communications protocols.
Exemplary Computer System Embodying the Invention
Moreover, those skilled in the art will appreciate that the invention may be practiced with other computer system configurations, including hand-held devices, multiprocessor systems, microprocessor-based or programmable consumer electronics, network PCS, minicomputers, mainframe computers, and the like. The invention may also be practiced in distributed computing environments where tasks are performed by remote processing devices linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote memory storage devices.
More particularly, computer system 10 includes a general purpose computing device in the form of a computer 20, including a processing unit 21, a system memory 22, and a system bus 23 that operatively couples various system, components including the system memory to processing unit 21. There may be only one or there may be more than one processing unit 21, such that the processor of computer 20 comprises a single central-processing unit (CPU), or a plurality of processing units, commonly referred to as a parallel processing environment. Computer 20 may be a conventional computer, a distributed computer, or any other type of computer. Thus, the invention is not limited to a particular computer.
System bus 23 may be any of several types of bus structures or memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures. The system memory may also be referred to as simply the memory, and includes read only memory (ROM) 24 and random access memory (RAM) 25. ROM 24 stores a basic input/output system (BIOS) 26, containing the basic routines that help to transfer information between elements within the computer 20, such as during start-up. Computer 20 further includes a hard disk drive 27 for reading from and writing to a hard disk, not shown, a magnetic disk drive 28 for reading from or writing to a removable magnetic disk 29, and an optical disk drive 30 for reading from or writing to a removable optical disk 31, such as a CD ROM or other optical media.
Hard disk drive 27, magnetic disk drive 28, and optical disk drive 30 are connected to system bus 23 by a hard disk drive interface 32, a magnetic disk drive interface 33, and an optical disk drive interface 34, respectively. The drives and their associated computer-readable media provide nonvolatile storage of computer-readable instructions, data structures, program modules and other data for computer 20. Any type of computer-readable media which can store data accessible by a computer, such as magnetic cassettes, flash memory cards, digital video disks, Bernoulli cartridges, random access memories (RAMs), read only memories (ROMs), and the like, can be used in the exemplary operating environment.
A number of program modules are stored on the hard disk, magnetic disk 29, optical disk 31, ROM 24, or RAM 25, including an operating system 35, one or more application programs 36, other program modules 37, and program data 38. Operating system 35 provides numerous functions and services to application programs 36 stored by system memory 22, hard-disk drive 27, and/or hard-disk drive 50. Examples of suitable operating systems include past, present, and future versions of the Microsoft Windows operating system, the Microsoft Windows NT operating system, the IBM OS/2 operating system, and the Apple Computer AppleTalk operating system.
Of particular relevance to the present invention is an augmented DirectShow multimedia software 62 included as one of application programs 36.
System 10, as
Computer 20 can operate in a networked environment using logical connections to one or more remote computers, such as remote computer 49. These logical connections are achieved by a communication device coupled to or a part of computer 20. However, the invention is not limited to a particular type of communications device. Remote computer 49, which may be another computer, a server, a router, a network personal computer (PC), a client, a peer device or other common network node, typically includes many or all of the elements of computer 20, although
When used in a LAN-networking environment, computer 20 is connected to the local network 51 through a network interface or adapter 53, which is one type of communications device. When used in a WAN-networking environment, computer 20 typically includes a modem 54, a type of communications device or any other type of communications device for establishing communications over wide area network 52, such as the Internet. Modem 54, which maybe internal or external, is connected to system bus 23 via serial port interface 46. In a networked environment, program modules depicted relative to personal computer 20, or portions thereof, may be stored in the remote memory storage device. It is appreciated that the network connections shown are exemplary and other means of and communications devices for establishing a communications link between the computers may be used.
The exemplary computer may be a conventional computer, a distributed computer, or any other type of computer, since the invention is not limited to any particular computer. A distributed computer typically includes one or more processing units as its processor, and a computer-readable medium such as a memory. The computer may also include a communications device such as a network adapter or a modem, so that it is able to communicatively couple with other computers to form a computer network.
Operation of the Exemplary Computer System
The invention primarily concerns operation of augmented DirectShow software 62, which as already noted is one of application programs 36 and has the exemplary structure shown in
Receiver portion 64, which includes a network source 640, demultiplexer module 641, one or more receiver payload handlers 642 a and 642 b for specific payload types, and one or more corresponding decoder modules 643 a and 643 b, audio mixer module 644, and media rendering (or playback) module 645. Network Source receives and validates data packets based on their headers, each of which identifies the format, or payload type, of its packet. Validation also includes discarding duplicates and rejecting “old” packets. In some embodiments, validation also includes matching SSRCs to IP addresses. Examples of data types include audio G.711, audio G.723.1, video H.261, and video H.263. The validated data then passes to demux module 641, which separates and routes the data packets based on their payload type to specific parts of the receive chain for further type-specific processing.
More precisely, demux module 641, in the exemplary embodiment, passes the data to either receiver payload handler 642 a or 642 b, whichever is appropriate to the specific payload type. In the exemplary embodiment, demux module 641 routes data based on SSRCs and payload type. Although the exemplary embodiment shows only two payload handlers, other embodiments provide as many as possible to accommodate a wide, if not exhaustive, variety of data types.
In addition, the exemplary demux module 641 includes unique features for handling large, multiparty conferences, for example, those having tens, hundreds, or even thousands of parties. In this context, demux module 641 singles out a limited number of the incoming data streams and dynamically maps, or routes, this subset of streams to the output modules, or filters, for further processing and eventual rendering. This dynamic mapping mechanism limits the number of concurrent speakers that a conference client can actively hear during the multiparty conference to a manageable number, both from the perspective of a listening party and system processing capability.
To this end, demux module 641 monitors incoming data stream activity for each conference party, or speaker, and decides to map or not to map the data stream to downstream modules for further processing based on the active or inactive status of the stream. Demux module 641 also monitors for new speakers. If a new speaker is detected, for example, based on detection of a new SSRC field in an incoming data packet, and one of the currently mapped streams has been inactive for a certain timeout period, demux module 641 replaces the inactive stream with the new stream. And if necessary, demux module 641 maps it to a different type-specific payload handler and decoder.
The payload handlers perform functions necessary to ensure proper decoding of the data. Example of such functions, all of which are well known, include reassembling or combining several data packets into a larger data package, reordering a received sequence of data packets, detecting and rejecting duplicated packets, or computing and compensating delay jitter.
Receiver payload handler 642 a and 642 b then pass on the data to respective decoder modules 643 a and 642 b, which independently decode the data to forms appropriate to rendering, or playback, by appropriate components of computer system 10. In the exemplary embodiment, this entails restoring the packetized data to a form which, barring any transmission or reception errors, resembles its form prior to being transmitted across network 51 or 52 by remote computer 49. After decoding, video data passes directly to media rendering module 645 for playback. Audio data, on the other hand, passes first to mixer 644, which combines acoustic information coming from several sources into a single audio stream compatible with media rendering module 645. Media rendering module 645 is generally associated with a video window, a data file, a sound card, or other audio or video transducer. (The exemplary embodiment associates media rendering module 645 with a call-control module (not shown) and processes H.323 media formats using the H.245 protocol.)
A user of system 10 can also rely on sender portion 66 of augmented DirectShow software 62 to transmit audio-only or audio-video data (as well as other types of data) back to remote computer 49. In the exemplary embodiment, system 10 responds with the same format of data as it received; however, other embodiments respond with data having respond with a different format. In general, data passes from module to module of sender portion 66, which includes media source module 661, silence supressor 662, encoder 663, a sender payload handler 664, and a network rendering module 665, before being transmitted across network 51 or 52 to computer 49.
More specifically, the data, such as audio samples or a video frame, intended for transmission originate from media source module 661 and then proceed to encoder module 663. Media source 661 is associated with a camera, video-capture board, a sound card coupled to a microphone, or other audio source such as a radio, cassette player, compact-disc player, or an audio, video, or audio-video file. In the case of audio data, the data passes first through silence suppressor 663 to reject data, for example, silences or low-level background sounds, which provide little or no useful information and would otherwise waste available bandwidth. However if the data is not audio, it passes directly to encoder 663, which converts the data, according to conventional techniques, into a more compact form to conserve bandwidth.
Encoder 663 passes the compacted data to sender payload handler 664, and handler 664 converts it to a form suitable for transmission over a packet switched network, network 50 or 51 for example. For video data, this conversion includes splitting a video frame into smaller pieces or subframes according to conventional techniques. Network rendering module 664 then sends the subdivided data onto the network, adding headers according to a specific transport protocol. Examples of suitable transport protocols include UDP, TCP, HTTP, and ATM.
In the exemplary embodiment, network source and network rendering module 664 support conventional streaming, bandwidth reservation, communications quality, telephony, and encryption protocols. Examples of protocols or application program interfaces supported by these modules include RTP (real-time transport protocol), RTCP (real-time transport control protocol), RSVP (resource reservation protocol), QOS (quality of service), GQOS (generic quality of service), TAPI 3.0 (telephony interface) and CryptoAPI (cryptographic application program interface). Further information on these aspects of the exemplary embodiment are publicly available in numerous RFCs (request for comments), SDKs (software development kits), and DDKs (device driver kits) as well as other sources. For instance, RFC 1889 entitled RTP: A Transport Protocol for Real-Time Applications (January 1996) and RFC 1890 entitled RTP Profile for Audio and Video Conferences with Minimal Control (January 1996) describe relevant aspects of the type of streaming supported in the exemplary embodiment.
Many conditions that occur in RTP, RTCP, QOS, and local computer-system are signaled up through DirectShow events, that is, short messages that the filters send to a user of the filter graph. These include, for example, detection of a new participant, detection of departing session participant, presence of QOS receivers or senders, no memory or network (socket) errors. The RTCP protocol allows the exchange of network information through the RTCP reports. This information can serve a variety of purposes, including, for example, modifying the rate media source module 661 produces data, modifying bit output rate of encoder 663, or even selecting an alternative encoder (or encoding protocol) for sender portion 66. This kind of feedback response ultimately adapts the characteristics of sender portion 66 toward an optimal quality level for a given set of network resources.
In furtherance of the art, the inventors devised several new software modules, or filters, for addition to Microsoft's DirectShow software. Some of the modules allow the DirectShow software to support streaming and network conferencing using more than one communications protocol. For example, one module supports the Real-time Transport Protocol for streaming and another routes incoming network data, based on its type, to type-specific data handlers and decoders.
The embodiments described above are intended only to illustrate and teach one or more ways of practicing or implementing the present invention, not to restrict its breadth or scope. The actual scope of the invention, which embraces all ways of practicing or implementing the invention, is defined only by the following claims and their equivalents.
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US5990933 *||Mar 22, 1999||Nov 23, 1999||Videoserver, Inc.||Apparatus and method for storage and playback of video images and audio messages in multipoint videoconferencing|
|US6122259 *||Feb 26, 1997||Sep 19, 2000||Hitachi, Ltd.||Video conference equipment and multipoint video conference system using the same|
|US6839734 *||Sep 21, 1998||Jan 4, 2005||Microsoft Corporation||Multimedia communications software with network streaming and multi-format conferencing|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US7260108 *||Sep 27, 2002||Aug 21, 2007||Alpine Electronics Inc.||Multimedia information providing method and apparatus|
|US7975283||Dec 14, 2005||Jul 5, 2011||At&T Intellectual Property I, L.P.||Presence detection in a bandwidth management system|
|US8024438||Dec 14, 2005||Sep 20, 2011||At&T Intellectual Property, I, L.P.||Methods, systems, and computer program products for implementing bandwidth management services|
|US8098582||Dec 14, 2005||Jan 17, 2012||At&T Intellectual Property I, L.P.||Methods, systems, and computer program products for implementing bandwidth control services|
|US8104054||Dec 9, 2005||Jan 24, 2012||At&T Intellectual Property I, L.P.||Methods, systems, and devices for bandwidth conservation|
|US8306033 *||Dec 14, 2005||Nov 6, 2012||At&T Intellectual Property I, L.P.||Methods, systems, and computer program products for providing traffic control services|
|US8335239||Dec 14, 2005||Dec 18, 2012||At&T Intellectual Property I, L.P.||Methods, systems, and devices for bandwidth conservation|
|US8605755||Nov 14, 2012||Dec 10, 2013||At&T Intellectual Property I, L.P.||Methods, systems, and devices for bandwidth conservation|
|US8621500||Dec 2, 2011||Dec 31, 2013||At&T Intellectual Property I, L.P.||Methods, systems, and devices for bandwidth conservation|
|US8670448 *||Sep 14, 2012||Mar 11, 2014||At&T Intellectual Property I, L.P.||Methods, systems, and computer program products for providing traffic control services|
|US8701148||Dec 9, 2005||Apr 15, 2014||At&T Intellectual Property I, L.P.||Methods, systems, and devices for bandwidth conservation|
|US9137385||Sep 17, 2013||Sep 15, 2015||Digifonica (International) Limited||Determining a time to permit a communications session to be conducted|
|US9143608||Apr 15, 2013||Sep 22, 2015||Digifonica (International) Limited||Intercepting voice over IP communications and other data communications|
|US20060221826 *||Dec 14, 2005||Oct 5, 2006||Bedingfield James C Sr||Methods, systems, and computer program products for providing traffic control services|
|US20060251116 *||Dec 14, 2005||Nov 9, 2006||Bedingfield James C Sr||Methods, systems, and computer program products for implementing bandwidth management services|
|US20060288011 *||Jun 21, 2005||Dec 21, 2006||Microsoft Corporation||Finding and consuming web subscriptions in a web browser|
|US20130010596 *||Jan 10, 2013||Bellsouth Intellectual Property Corporation||Methods, systems, and computer program products for providing traffic control services|
|US20140153477 *||Nov 27, 2013||Jun 5, 2014||Digifonica (International) Limited||Uninterrupted transmission of internet protocol transmissions during endpoint changes|
|International Classification||H04L29/06, H04J3/04|
|Cooperative Classification||H04L65/4046, H04L65/608, H04L29/06027|
|European Classification||H04L29/06C2, H04L29/06M6P, H04L29/06M4C4|
|Jan 15, 2015||AS||Assignment|
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034766/0001
Effective date: 20141014