Search Images Maps Play YouTube News Gmail Drive More »
Sign in
Screen reader users: click this link for accessible mode. Accessible mode has the same essential features but works better with your reader.

Patents

  1. Advanced Patent Search
Publication numberUS20020059170 A1
Publication typeApplication
Application numberUS 09/835,838
Publication dateMay 16, 2002
Filing dateApr 16, 2001
Priority dateApr 17, 2000
Also published asUS6990531, US7020783, US7043563, US7111006, US7120662, US7127518, US7143195, US7155539, US7962654, US7975066, US8386641, US8417770, US8463935, US20020002602, US20020002603, US20020002611, US20020002618, US20020002622, US20020002625, US20020002636, US20020002686, US20020004796, US20020004816, US20020007404, US20020019853, US20020023159, US20020056006, US20060129697, US20100157998, US20110238860, US20110246665, US20110302321, US20120054308, US20120059907, US20140074981, WO2001080002A1, WO2001080003A2, WO2001080003A3, WO2001080004A2, WO2001080004A3, WO2001080014A2, WO2001080014A3, WO2001080024A2, WO2001080024A3, WO2001080033A2, WO2001080033A3, WO2001080062A2, WO2001080062A3, WO2001080063A2, WO2001080063A3, WO2001080064A2, WO2001080064A3, WO2001080093A2, WO2001080093A3, WO2001080515A2, WO2001080515A3, WO2001080516A2, WO2001080516A3, WO2001080517A2, WO2001080517A3, WO2001080524A2, WO2001080524A3
Publication number09835838, 835838, US 2002/0059170 A1, US 2002/059170 A1, US 20020059170 A1, US 20020059170A1, US 2002059170 A1, US 2002059170A1, US-A1-20020059170, US-A1-2002059170, US2002/0059170A1, US2002/059170A1, US20020059170 A1, US20020059170A1, US2002059170 A1, US2002059170A1
InventorsMark Vange
Original AssigneeMark Vange
Export CitationBiBTeX, EndNote, RefMan
External Links: USPTO, USPTO Assignment, Espacenet
Load balancing between multiple web servers
US 20020059170 A1
Abstract
A system for load balancing in a network environment including a plurality of network resources coupled to a network wherein at least some of the network resources provide redundant services. A client is coupled to the network and a gateway machine is coupled to the network in communication with the client. The gateway machine is configured to receive requests from the client, establish communication channel through the network with a network resource specified by the client, and access the specified network resources to service the received client requests. The gateway machine includes or is coupled to mechanisms for selecting amongst providers of redundant services a particular provider for a received request so as to balance load amongst the plurality of resources providing redundant services.
Images(8)
Previous page
Next page
Claims(19)
We claim:
1. A system for load balancing in a network environment comprising:
a plurality of servers coupled to a network;
a set of network resources associated with each of the servers, wherein at least some of the network resources are redundant;
a client coupled to the network and generating a request specifying some of the redundant resources;
a gateway machine coupled to the network in communication with the client, the gateway machine configured to receive the request from the client, select from amongst the servers that are associated with the request-specified redundant services, establish a communication channel with the selected server, and access the specified server to service the received client request; and
means coupled to the gateway machine for selecting amongst servers of redundant resources a particular server for a received request so as to balance load amongst the servers providing redundant resources.
2. The system of claim 1 further comprising means within the gateway machine for selecting amongst a plurality of channels between the gateway machine and the servers associated with the network resources specified by the client request.
3. The system of claim 1 wherein at least some of the plurality of servers comprise world wide web servers.
4. The system of claim 3 wherein the client comprises a web browser.
5. The system of claim 1 wherein each of the plurality of servers is associated with a network address and the gateway machine comprises a web server having an network address distinct from the plurality of servers.
6. The system of claim 1 wherein the network comprises the Internet.
7. The system of claim 1 further comprising means for monitoring quality of service between the gateway and each of the servers.
8. The system of claim 7 wherein the means for monitoring quality of service is implemented within the gateway machine.
9. The system of claim 7 further comprising means for selecting from amongst the servers providing redundant services using the relative quality of service between the servers.
10. The system of claim 7 further comprising:
means for allocating an additional server with redundant services in response to the quality of service falling below a preselected level, wherein the gateway machine is configured to establish a new communication channel through the network with the additional server.
11. The system of claim 1 wherein the gateway machine further comprises:
a means for generating a response to the client request using services provided to the gateway machine by the servers.
12. A method for load balancing in a network environment comprising:
providing a communication network;
providing a plurality of servers coupled to the network wherein at least some of the servers provide redundant services;
generating a request for a redundant service in a network-coupled client machine;
directing the request to a network-coupled gateway machine;
causing the gateway machine to select amongst servers providing the redundant services a particular server for the received request so as to balance load amongst the plurality of servers providing the redundant services; and
after selecting a particular server, causing the gateway machine to generate a request to the selected server for the specified resources.
13. The method of claim 12 selecting amongst a plurality of channels between the gateway machine and the servers of redundant services specified by the client request.
14. The method of claim 12 wherein at least some of the plurality of servers comprise world wide web servers and the client comprises a web browser.
15. The method of claim 12 wherein the network comprises the Internet.
16. The method of claim 12 further comprising monitoring quality of service between the gateway and each of the servers.
17. The method of claim 16 further comprising selecting from amongst the servers providing redundant services using the relative quality of service between the servers.
18. The method of claim 12 further comprising using the gateway machine to generate a response to the client request using services provided to the gateway machine by the servers.
19. A system for improving performance of a network connected server comprising:
an intermediary server receiving requests for server access from a plurality of sources;
means for monitoring at least one variable affecting server performance; and
a queue data structure within the intermediary server, wherein the queue data structure is responsive to the means for monitoring to release requests from the queue data structure in a manner determined to improve server performance.
Description
RELATED APPLICATIONS

[0001] The present invention claims priority from U.S. Provisional Patent Application No. 60/197,490 entitled CONDUCTOR GATEWAY filed on Apr. 17, 2000.

BACKGROUND OF THE INVENTION

[0002] 1. Field of the Invention

[0003] The present invention relates, in general, to network information access and, more particularly, to software, systems and methods for serving web pages in a coordinated fashion from multiple cooperating web servers.

[0004] 2. Relevant Background

[0005] Increasingly, business data processing systems, entertainment systems, and personal communications systems are implemented by computers across networks that are interconnected by internetworks (e.g., the Internet). The Internet is rapidly emerging as the preferred system for distributing and exchanging data. Data exchanges support applications including electronic commerce, broadcast and multicast messaging, videoconferencing, gaming, and the like.

[0006] The Internet is a collection of disparate computers and networks coupled together by a web of interconnections using standardized communication protocols. The Internet is characterized by its vast reach as a result of its wide and increasing availability and easy access protocols. Unfortunately, the heterogenous nature of the Internet results in variable bandwidth and quality of service between points. The latency and reliability of data transport is largely determined by the total amount of traffic on the Internet and so varies wildly seasonally and throughout the day. Other factors that affect quality of service include equipment outages and line degradation that force packets to be rerouted, damaged and/or dropped. Also, routing software and hardware limitations within the Internet infrastructure may create bandwidth bottlenecks, even when the mechanisms are operating within specifications.

[0007] Internet transport protocols do not discriminate between users. Data packets are passed between routers and switches that make up the Internet fabric based on the hardware's instantaneous view of the best path between source and destination nodes specified in the packet. Because each packet may take a different path, the latency of a packet cannot be guaranteed and, in practice, varies significantly. Likewise, data packets are routed through the Internet without any prioritization based on content.

[0008] Prioritization has not been an issue with conventional networks such as local area networks (LANs) and wide area networks (WANs) because the average latency of such networks has been sufficiently low and sufficiently uniform to provide acceptable performance. However, there is an increasing demand for network applications that cannot tolerate high and variable latency. This situation is complicated when the application is to be run over the Internet where latency and variability in latency are many times greater than in LAN and WAN environments.

[0009] A particular need exists in environments that involve multiple users accessing a network resource such as a web server. Examples include broadcast, multicast and videoconferencing as well as most electronic commerce (e-commerce) applications. In these applications, it is important to maintain a reliable connection so that the server and clients remain synchronized and information is not lost.

[0010] In electronic commerce (e-commerce) applications, it is important to provide a satisfying buying experience that leads to a purchase transaction. To provide this high level of service, a web site operator must ensure that data is delivered to the customer in the most usable and efficient fashion. Also, the web site operator must ensure that critical data received from the customer is handled with priority.

[0011] Until now, however, the e-commerce site owner has had little or no control over the transport mechanisms through the Internet that affect the latency and quality of service. This is akin to a retailer being forced to deal with a customer by shouting across the street, never certain how often what was said must be repeated, and knowing that during rush hour communication would be nearly impossible. While efforts are continually being made to increase the capacity and quality of service afforded by the Internet, it is contemplated that congestion will always impact the ability to predictably and reliably offer a specified level of service. Moreover, the change in the demand for bandwidth increases at a greater rate than does the change in bandwidth supply, ensuring that congestion will continue to be an issue into the foreseeable future. A need exists for a system to exchange data over the Internet that provides a high quality of service even during periods of congestion.

[0012] Many e-commerce transactions are abandoned by the user because system performance degradations frustrate the purchaser before the transaction is consummated. While a transaction that is abandoned while a customer is merely browsing through a catalog may be tolerable, abandonment when the customer is just a few clicks away from a purchase is highly undesirable. However, existing Internet transport mechanisms and systems do not allow the e-commerce site owner any ability to distinguish between the “just browsing” and the “about-to-buy” customers. In fact, the vagaries of the Internet may lead to the casual browser receiving a higher quality of service while the about-to-buy customer becomes frustrated and abandons the transaction.

[0013] Web sites are often implemented on a plurality of web servers which may or may not be running on separate hosting machines. Each web server handles a set of content and some load distribution software runs on top of the multiple web servers to direct requests to the web server that can handle the request. The multiple servers essentially act as peers with each server having a set of resources from which it serves requests. It is difficult to distribute load efficiently amongst the multiple servers, however. One or more servers may experience heavy traffic while other servers remain essentially idle, unable to assist the overworked servers. This results in poor performance.

[0014] Several load balancing solutions have been proposed for web sites including, for example, webQOS offered by Hewlett Packard Company. Another recent load balancing solution described in U.S. Pat. No. 5,894,554 uses a set of “page servers” to offload some page generation functionality from the originating web server. These page servers remain in the IP address domain of the originating web server and so are closely coupled to the web server itself. Such solutions implement a logical layer over all the available web servers that receive incoming requests and route the requests intelligently to available web servers at a rate that allows the web servers to operate more efficiently. However, these solutions are generally operative after a client request has been transported through the network and received at the web site and so do not affect load balancing through the network itself. Moreover, such solutions do not extend well to web servers that are geographically and/or toplogically distributed. A need exists for a load balancing mechanism that provides efficient load balancing throughout a communication network.

[0015] Also, current load balancing techniques operate on a request-by-request basis and so cannot readily direct requests in a manner that balances load at a more abstract level over whole systems and web sites. Although in many cases web sites are implemented in a stateless fashion so that requests can be handled on a request-by-request basis, some optimization can be obtained by stateful sessions. A need exists for load balancing systems, methods and software that work cooperatively with a web site to balance load on other than a request-by-request basis.

SUMMARY OF THE INVENTION

[0016] Briefly stated, the present invention involves a system for load balancing in a network environment including a plurality of network resources coupled to a network wherein at least some of the network resources provide redundant services. A client is coupled to the network and a gateway machine is coupled to the network in communication with the client. The gateway machine is configured to receive requests from the client, establish a communication channel through the network with a network resource specified by the client, and access the specified network resources to service the received client requests. The gateway machine includes or is coupled to mechanisms for selecting amongst providers of redundant services a particular provider for a received request so as to balance load amongst the plurality of resources providing redundant services.

BRIEF DESCRIPTION OF THE DRAWINGS

[0017]FIG. 1 illustrates a general distributed computing environment in which the present invention is implemented;

[0018]FIG. 2 shows in block-diagram form significant components of a system in accordance with the present invention;

[0019]FIG. 3 shows a domain name system used in an implementation of the present invention;

[0020]FIG. 4 shows front-end components of FIG. 2 in greater detail;

[0021]FIG. 5 shows back-end components of FIG. 2 in greater detail;

[0022]FIG. 6 shows a conceptual block diagram of the system of FIG. 2 in an alternative context; and

[0023]FIG. 7 illustrates an alternative embodiment of the present invention.

DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

[0024] The present invention is illustrated and described in terms of a distributed computing environment such as an enterprise computing system using public communication channels such as the Internet. However, an important feature of the present invention is that it is readily scaled upwardly and downwardly to meet the needs of a particular application. Accordingly, unless specified to the contrary, the present invention is applicable to significantly larger, more complex network environments, including wireless network environments, as well as small network environments such as conventional LAN systems.

[0025] In accordance with the present invention, load balancing is performed on the front-end, before a network connection. Prior systems load balance at the back-end after the connection, and cannot select a channel. One feature of the present invention is that it provides a way to improve use of multiple channels through the network. The load balancing decisions are made at an intermediary server that is located on the network topology at a location logically closer to the client or software application making the request than the web server itself. This enables the present invention to not only balance load amongst multiple web servers in a web site hosting center, but also to balance load across multiple available communication channels based on, for example, quality of service provided on the available channels.

[0026] One feature of the present invention is that the front-end servers are in separate IP address domains from the originating web server. A redirection mechanism is enabled to select from a pool of available front-end servers and direct client request packets from the originating web server to a selected front-end server. Preferably, the front-end server establishes and maintains an enhanced communication channel with the originating web server. By enhanced, it is meant that the channel offers improved quality of service, lower latency, prioritization services, higher security transport, or other features and services that improve upon the basic transport mechanisms (such as TCP) defined for Internet data transport.

[0027] In this manner, the load balancing functionality can be performed before the request is launched across the public network. A front-end that is logically close to the client process that is requesting service is selected from the pool of available front-end servers. The selected front-end is configured to provide the enhanced channel to the originating web site using, for example, a back-end server. An enhanced channel may already exist and such existence may be a criteria used to select a particular front-end server from the pool of front-end servers.

[0028] For purposes of this document, a web server is a computer or system of computers running server software coupled to the World Wide Web (i.e., “the web”) that delivers or serves web pages. The web server has a unique IP address and accepts connections in order to service requests by sending back responses. A web server differs from a proxy server or a gateway server in that a web server has resident a set of resources (i.e., software programs, data storage capacity, and/or hardware) that enable it to execute programs to provide an extensible range of functionality such as generating web pages, accessing remote network resources, analyzing contents of packets, reformatting request/response traffic and the like using the resident resources. In contrast, a proxy simply forwards request/response traffic on behalf of a client to resources that reside elsewhere, or obtains resources from a local cache if implemented. A web server in accordance with the present invention may reference external resources of the same or different type as the services requested by a user, and reformat and augment what is provided by the external resources in its response to the user. Commercially available web server software includes Microsoft Internet Information Server (IIS), Netscape Netsite, Apache, among others. Alternatively, a web site may be implemented with custom or semi-custom software that supports HTTP traffic.

[0029]FIG. 1 shows an exemplary computing environment 100 in which the present invention may be implemented. Environment 100 includes a plurality of local networks such as Ethernet network 102, FDDI network 103 and Token Ring network 104. Essentially, a number of computing devices and groups of devices are interconnected through a network 101. For example, local networks 102, 103 and 104 are each coupled to network 101 through routers 109. LANs 102, 103 and 104 may be implemented using any available topology and may implement one or more server technologies including, for example UNIX, Novell, or Windows NT networks, including both client-server and peer-to-peer type networks. Each network will include distributed storage implemented in each device and typically includes some mass storage device coupled to or managed by a server computer. Network 101 comprises, for example, a public network such as the Internet or another network mechanism such as a fiber channel fabric or conventional WAN technologies.

[0030] Local networks 102, 103 and 104 include one or more network appliances 107. One or more network appliances 107 may be configured as an application and/or file server. Each local network 102, 103 and 104 may include a number of shared devices (not shown) such as printers, file servers, mass storage and the like. Similarly, devices 111 may be shared through network 101 to provide application and file services, directory services, printing, storage, and the like. Routers 109 provide a physical connection between the various devices through network 101. Routers 109 may implement desired access and security protocols to manage access through network 101.

[0031] Network appliances 107 may also couple to network 101 through public switched telephone network 108 using copper or wireless connection technology. In a typical environment, an Internet service provider 106 supports a connection to network 101 as well as PSTN 108 connections to network appliances 107.

[0032] Network appliances 107 may be implemented as any kind of network appliance having sufficient computational function to execute software needed to establish and use a connection to network 101. Network appliances 107 may comprise workstation and personal computer hardware executing commercial operating systems such as Unix variants, Microsoft Windows, Macintosh OS, and the like. At the same time, some appliances 107 comprise portable or handheld devices using wireless connections through a wireless access provider such as personal digital assistants and cell phones executing operating system software such as PalmOS, WindowsCE, and the like. Moreover, the present invention is readily extended to network devices such as office equipment, vehicles, and personal communicators that make occasional connection through network 101.

[0033] Each of the devices shown in FIG. 1 may include memory, mass storage, and a degree of data processing capability sufficient to manage their connection to network 101. The computer program devices in accordance with the present invention are implemented in the memory of the various devices shown in FIG. 1 and enabled by the data processing capability of the devices shown in FIG. 1. In addition to local memory and storage associated with each device, it is often desirable to provide one or more locations of shared storage such as disk farm (not shown) that provides mass storage capacity beyond what an individual device can efficiently use and manage. Selected components of the present invention may be stored in or implemented in shared mass storage.

[0034] The present invention operates in a manner akin to a private network 200 implemented within the Internet infrastructure. Private network 200 expedites and prioritizes communications between a client 205 and a web site 210. In the specific examples herein, client 205 comprises a network-enabled graphical user interface such as a World Wide Web browser. However, the present invention is readily extended to client software other than conventional web browser software. Any client application that can access a standard or proprietary user level protocol for network access is a suitable equivalent. Examples include client applications for file transfer protocol (FTP) services, voice over Internet protocol (VOIP) services, network news protocol (NNTP) services, multi-purpose internet mail extensions (MIME) services, post office protocol (POP) services, simple mail transfer protocol (SMTP) services, as well as Telnet services. In addition to network protocols, the client application may access a network application such as a database management system (DBMS) in which case the client application generates query language (e.g., structured query language or “SQL”) messages. In wireless appliances, a client application may communicate via a wireless application protocol (WAP) or the like.

[0035] For convenience, the term “web site” is used interchangeably with “web server” in the description herein, although it should be understood that a web site comprises a collection of content, programs and processes implemented on one or more web servers. A web site is owned by the content provider such as an e-commerce vendor whereas a web server refers to set of programs running on one or more machines coupled to an Internet node. The web site 210 may be hosted on the site owner's own web server, or hosted on a web server owned by a third party. A web hosting center is an entity that implements one or more web sites on one or more web servers using shared hardware and software resources across the multiple web sites. In a typical web infrastructure, there are many web browsers, each of which has a TCP connection to the web server in which a particular web site is implemented. The present invention adds two components to the infrastructure: a front-end 201 and back-end 203. Front-end 201 and back-end 203 are coupled by a managed data communication link 202 that forms, in essence, a private network.

[0036] Front-end mechanism 201 serves as an access point for client-side communications. Front-end 201 implements a gateway that functions as a proxy for the web server(s) implementing web site 210 (i.e., from the perspective of client 205, front-end 201 appears to be the web site 210). Front-end 201 comprises, for example, a computer that sits “close” to clients 205. By “close”, it is meant that the average latency associated with a connection between a client 205 and a front-end 201 is less than the average latency associated with a connection between a client 205 and a web site 210. Desirably, front-end computers have as fast a connection as possible to the clients 205. For example, the fastest available connection may be implemented in point of presence (POP) of an Internet service provider (ISP) 106 used by a particular client 205. However, the placement of the front-ends 201 can limit the number of browsers that can use them. Because of this, in some applications it is more practical to place one front-end computer in such a way that several POPs can connect to it. Greater distance between front-end 201 and clients 205 may be desirable in some applications as this distance will allow for selection amongst a greater number front-ends 201 and thereby provide significantly different routes to a particular back-end 203. This may offer benefits when particular routes and/or front-ends become congested or otherwise unavailable.

[0037] Transport mechanism 202 is implemented by cooperative actions of the front-end 201 and back-end 203. Back-end 203 processes and directs data communication to and from web site 210. Transport mechanism 202 communicates data packets using a proprietary protocol over the public Internet infrastructure in the particular example. Hence, the present invention does not require heavy infrastructure investments and automatically benefits from improvements implemented in the general purpose network 101. Unlike the general purpose Internet, front-end 201 and back-end 203 are programmably assigned to serve accesses to a particular web site 210 at any given time.

[0038] It is contemplated that any number of front-end and back-end mechanisms may be implemented cooperatively to support the desired level of service required by the web site owner. The present invention implements a many-to-many mapping of front-ends to back-ends. Because the front-end to back-end mappings can by dynamically changed, a fixed hardware infrastructure can be logically reconfigured to map more or fewer front-ends to more or fewer back-ends and web sites or servers as needed.

[0039] Front-end 201 together with back-end 203 function to reduce traffic across the transport morphing protocol™ (TMP™) link 202 and to improve response time for selected browsers. Transport morphing protocol and TMP are trademarks or registered trademarks of Circadence Corporation in the United States and other countries. Traffic across the TMP link 202 is reduced by compressing data and serving browser requests from cache for fast retrieval. Also, the blending of request datagrams results in fewer request:acknowledge pairs across the TMP link 202 as compared to the number required to send the packets individually between front-end 201 and back-end 203. This action reduces the overhead associated with transporting a given amount of data, although conventional request:acknowledge traffic is still performed on the links coupling the front-end 201 to client 205 and back-end 203 to a web server. Moreover, resend traffic is significantly reduced further reducing the traffic. Response time is further improved for select privileged users and for specially marked resources by determining the priority for each HTTP transmission.

[0040] In one embodiment, front-end 201 and back-end 203 are closely coupled to the Internet backbone. This means they have high bandwidth connections, can expect fewer hops, and have more predictable packet transit time than could be expected from a general-purpose connection. Although it is preferable to have low latency connections between front-ends 201 and back-ends 203, a particular strength of the present invention is its ability to deal with latency by enabling efficient transport and traffic prioritization. Hence, in other embodiments front-end 201 and/or back-end 203 may be located farther from the Internet backbone and closer to clients 205 and/or web servers 210. Such an implementation reduces the number of hops required to reach a front-end 201 while increasing the number of hops within the TMP link 202 thereby yielding control over more of the transport path to the management mechanisms of the present invention.

[0041] Clients 205 no longer conduct all data transactions directly with the web server 210. Instead, clients 205 conduct some and preferably a majority of transactions with front-ends 201, which simulate the functions of web server 210. Client data is then sent, using TMP link 202, to the back-end 203 and then to the web server 210. Running multiple clients 205 over one large connection provides several advantages:

[0042] Since all client data is mixed, each client can be assigned a priority. Higher priority clients, or clients requesting higher priority data, can be given preferential access to network resources so they receive access to the channel sooner while ensuring low-priority clients receive sufficient service to meet their needs.

[0043] The large connection between a front-end 201 and back-end 203 can be permanently maintained, shortening the many TCP/IP connection sequences normally required for many clients connecting and disconnecting.

[0044] Using a proprietary protocol allows the use of more effective techniques to improve data throughput and makes better use of existing bandwidth during periods when the network is congested.

[0045] A particular advantage of the architecture shown in FIG. 2 is that it is readily scaled. Any number of client machines 205 may be supported. In a similar manner, a web site owner may choose to implement a site using multiple web servers 210 that are co-located or distributed throughout network 101. To avoid congestion, additional front-ends 201 may be implemented or assigned to particular web sites. Client traffic is dynamically directed to available front-ends 201 to provide load balancing. Hence, when quality of service drops because of a large number of client accesses, an additional front-end 201 can be assigned to the web site and subsequent client requests directed to the newly assigned front-end 201 to distribute traffic across a broader base.

[0046] In the particular examples, this is implemented by a front-end manager component 207 that communicates with multiple front-ends 201 to provide administrative and configuration information to front-ends 201. Each front-end 201 includes data structures for storing the configuration information, including information identifying the IP addresses of web servers 210 to which they are currently assigned. Other administrative and configuration information stored in front-end 201 may include information for prioritizing data from and to particular clients, quality of service information, and the like.

[0047] Similarly, additional back-ends 203 can be assigned to a web site to handle increased traffic. Back-end manager component 209 couples to one or more back-ends 203 to provide centralized administration and configuration service. Back-ends 203 include data structures to hold current configuration state, quality of service information and the like. In the particular examples front-end manager 207 and back-end manager 209 serve multiple web sites 210 and so are able to manipulate the number of front-ends and back-ends assigned to each web site 210 by updating this configuration information. When the congestion for the site subsides, the front-end 201 and back-end 203 can be reassigned to other, busier web sites. These and similar modifications are equivalent to the specific examples illustrated herein.

[0048] In the case of web-based environments, front-end 201 is implemented using custom or off-the-shelf web server software. Front-end 201 is readily extended to support other, non-web-based protocols, however, and may support multiple protocols for varieties of client traffic. Front-end 201 processes the data traffic it receives, regardless of the protocol of that traffic, to a form suitable for transport by TMP 202 to a back-end 203. Hence, most of the functionality implemented by front-end 201 is independent of the protocol or format of the data received from a client 205. Hence, although the discussion of the exemplary embodiments herein relates primarily to front-end 201 implemented as a web server, it should be noted that, unless specified to the contrary, web-based traffic management and protocols are merely examples and not a limitation of the present invention.

[0049] As shown in FIG. 2, in accordance with the present invention a web site is implemented using an originating web server 210 operating cooperatively with the web server of front-end 201. More generally, any network service (e.g., FTP, VoIP, NNTP, MIME, SMTP, Telnet, DBMS) can be implemented using a combination of an originating server working cooperatively with a front-end 201 configured to provide a suitable interface (e.g., FTP, VoIP, NNTP, MIME, SMTP, Telnet, DBMS, WAP) for the desired service. In contrast to a simple front-end cache or proxy software, implementing a server in front-end 201 enables portions of the web site (or other network service) to actually be implemented in and served from both locations. The actual web pages or service being delivered comprises a composite of the portions generated at each server. Significantly, however, the web server in front-end 201 is close to the browser in a client 205 whereas the originating web server is close to all resources available at the web hosting center at which web site 210 is implemented. In essence the web site 210 is implemented by a tiered set of web servers comprising a front-end server 201 standing in front of an originating web server.

[0050] This difference enables the web site or other network service to be implemented so as to take advantage of the unique topological position each entity has with respect to the client 205. By way of a particular example, assume an environment in which the front-end server 201 is located at the location of an ISP used by a particular set of clients 205. In such an environment, clients 205 can access the front-end server 205 without actually traversing the network 101.

[0051] In order for a client 205 to obtain service from a front-end 201, it must first be directed to a front-end 201 that can provide the desired service. Preferably, client 205 does not need to be aware of the location of front-end 201, and initiates all transactions as if it were contacting the originating server 210. FIG. 3 illustrates a domain name server (DNS) redirection mechanism that illustrates how a client 205 is connected to a front-end 201. The DNS systems is defined in a variety of Internet Engineering Task Force (IETF) documents such as RFC0883, RFC 1034 and RFC 1035 which are incorporated by reference herein. In a typical environment, a client 205 executes a browser 301, TCP/IP stack 303, and a resolver 305. For reasons of performance and packaging, browser 301, TCP/IP stack 303 and resolver 305 are often grouped together as routines within a single software product.

[0052] Browser 301 functions as a graphical user interface to implement user input/output (I/O) through monitor 311 and associated keyboard, mouse, or other user input device (not shown). Browser 301 is usually used as an interface for web-based applications, but may also be used as an interface for other applications such as email and network news, as well as special-purpose applications such as database access, telephony, and the like. Alternatively, a special-purpose user interface may be substituted for the more general-purpose browser 301 to handle a particular application.

[0053] TCP/IP stack 303 communicates with browser 301 to convert data between formats suitable for browser 301 and IP format suitable for Internet traffic. TCP/IP stack also implements a TCP protocol that manages transmission of packets between client 205 and an Internet service provider (ISP) or equivalent access point. IP protocol requires that each data packet include, among other things, an IP address identifying a destination node. In current implementations the IP address comprises a 32-bit value that identifies a particular Internet node. Non-IP networks have similar node addressing mechanisms. To provide a more user-friendly addressing system, the Internet implements a system of domain name servers that map alpha-numeric domain names to specific IP addresses. This system enables a name space that provides a more consistent reference between nodes on the Internet and avoids the need for users to know network identifiers, addresses, routes and similar information in order to make a connection.

[0054] The domain name service is implemented as a distributed database managed by domain name servers (DNSs) 307 such as DNS_A, DNS_B and DNS_C shown in FIG. 3. Each DNS relies on <domain name:IP> address mapping data stored in master files scattered through the hosts that use the domain system. These master files are updated by local system administrators. Master files typically comprise text files that are read by a local name server, and hence become available through the name servers 307 to users of the domain system.

[0055] The user programs (e.g., clients 205) access name servers through standard programs such as resolver 305. Resolver 305 includes an address of a DNS 307 that serves as a primary name server. When presented with a reference to a domain name (e.g., http://www.circadence.com), resolver 305 sends a request to the primary DNS (e.g., DNS_A in FIG. 3). The primary DNS 307 returns either the IP address mapped to that domain name, a reference to another DNS 307 which has the mapping information (e.g., DNS_B in FIG. 3), or a partial IP address together with a reference to another DNS that has more IP address information. Any number of DNS-to-DNS references may be required to completely determine the IP address mapping.

[0056] In this manner, the resolver 305 becomes aware of the IP address mapping which is supplied to TCP/IP component 303. Client 205 may cache the IP address mapping for future use. TCP/IP component 303 uses the mapping to supply the correct IP address in packets directed to a particular domain name so that reference to the DNS system need only occur once per connection to a web site.

[0057] In accordance with the present invention, at least one DNS server 307 is owned and controlled by system components of the present invention. When a user accesses a network resource (e.g., a web site), browser 301 contacts the public DNS system to resolve the requested domain name into its related IP address in a conventional manner. In a first embodiment, the public DNS performs a conventional DNS resolution directing the browser to an originating server 210 and server 210 performs a redirection of the browser to the system owned DNS server (i.e., DNC_C in FIG. 3). In a second embodiment, domain:address mappings within the DNS system are modified such that resolution of the of the originating server's domain automatically return the address of the system-owned DNS server (DNS_C). Once a browser is redirected to the system-owned DNS server, it begins a process of further redirecting the browser 301 to the best available front-end 201.

[0058] Unlike a conventional DNS server, however, the system-owned DNS_C in FIG. 3 receives domain:address mapping information from a redirector component 309. Redirector 309 is in communication with front-end manager 207 and back-end manager 209 to obtain information on current front-end and back-end assignments to a particular server 210. A conventional DNS is intended to be updated infrequently by reference to its associated master file. In contrast, the master file associated with DNS_C is dynamically updated by redirector 309 to reflect current assignment of front-end 201 and back-end 203. In operation, a reference to web server 210 (e.g., http://www.circadence.com) may result in an IP address returned from DNS_C that points to any selected front-end 201 that is currently assigned to web site 210. Likewise, web site 210 may identify a currently assigned back-end 203 by direct or indirect reference to DNS_C.

[0059] Front-end 201 typically receives information directly from front-end manager 207 about the address of currently assigned back-ends 203. Similarly, back-end 203 is aware of the address of a front-end 201 associated with each data packet. Hence, reference to the domain system is not required to map a front-end 201 to its appropriate back-end 203.

[0060]FIG. 4 illustrates principle functional components of an exemplary front-end 201 in greater detail. Primary functions of the front-end 201 include translating transmission control protocol (TCP) packets from client 205 into TMP packets used in the system in accordance with the present invention. It is contemplated that the various functions described in reference to the specific examples may be implemented using a variety of data structures and programs operating at any location in a distributed network. For example, a front-end 201 may be operated on a network appliance 107 or server within a particular network 102, 103, or 104 shown in FIG. 1. The present invention is readily adapted to any application where multiple clients are coupling to a centralized resource. Moreover, other transport protocols may be used, including proprietary transport protocols.

[0061] TCP component 401 includes devices for implementing physical connection layer and IP layer functionality. Current IP standards are described in IETF documents RFC0791, RFC0950, RFC0919, RFC0922, RFC792, RFC1112 that are incorporated by reference herein. For ease of description and understanding, these mechanisms are not described in great detail herein. Where protocols other than TCP/IP are used to couple to a client 205, TCP component 401 is replaced or augmented with an appropriate network protocol process.

[0062] TCP component 401 communicates TCP packets with one or more clients 205. Received packets are coupled to parser 402 where the IP layer information is extracted. TCP is described in IETF RFC0793, which is incorporated herein by reference. Each TCP packet includes header information that indicates addressing and control variables, and a payload portion that holds the user-level data being transported by the TCP packet. The user-level data in the payload portion typically comprises a user-level network protocol datagram.

[0063] Parser 402 analyzes the payload portion of the TCP packet. In the examples herein, HTTP is employed as the user-level protocol because of its widespread use and the advantage that currently available browser software is able to readily use the HTTP protocol. In this case, parser 402 comprises an HTTP parser. More generally, parser 402 can be implemented as any parser-type logic implemented in hardware or software for interpreting the contents of the payload portion. Parser 402 may implement file transfer protocol (FTP), mail protocols such as simple mail transport protocol (SMTP), structured query language (SQL) and the like. Any user-level protocol, including proprietary protocols, may be implemented within the present invention using appropriate modification of parser 402.

[0064] To improve performance, front-end 201 optionally includes a caching mechanism 403. Cache 403 may be implemented as a passive cache that stores frequently and/or recently accessed web pages or as an active cache that stores network resources that are anticipated to be accessed. In non-web applications, cache 403 may be used to store any form of data representing database contents, files, program code, and other information. Upon receipt of a TCP packet, HTTP parser 402 determines if the packet is making a request for data within cache 403. If the request can be satisfied from cache 403, the data is supplied directly without reference to web server 210 (i.e., a cache hit). Cache 403 implements any of a range of management functions for maintaining fresh content. For example, cache 403 may invalidate portions of the cached content after an expiration period specified with the cached data or by web sever 210. Also, cache 403 may proactively update the cache contents even before a request is received for particularly important or frequently used data from web server 210. Cache 403 evicts information using any desired algorithm such as least recently used, least frequently used, first in/first out, or random eviction. When the requested data is not within cache 403, a request is processed to web server 210, and the returned data may be stored in cache 403.

[0065] Several types of packets will cause parser 402 to forward a request towards web server 210. For example, a request for data that is not within cache 403 (or if optional cache 403 is not implemented) will require a reference to web server 210. Some packets will comprise data that must be supplied to web server 210 (e.g., customer credit information, form data and the like). In these instances, HTTP parser 402 couples to data blender 404.

[0066] Optionally, front-end 201 implements security processes, compression processes, encryption processes and the like to condition the received data for improved transport performance and/or provide additional functionality. These processes may be implemented within any of the functional components (e.g., data blender 404) or implemented as separate functional components within front-end 201. Also, parser 402 may implement a prioritization program to identify packets that should be given higher priority service. A prioritization program requires only that parser 402 include a data structure associating particular clients 205 or particular TCP packet types or contents with a prioritization value. Based on the prioritization value, parser 402 may selectively implement such features as caching, encryption, security, compression and the like to improve performance and/or functionality. The prioritization value is provided by the owners of web site 210, for example, and may be dynamically altered, statically set, or updated from time to time to meet the needs of a particular application.

[0067] Blender 404 slices and/or coalesces the data portions of the received packets into more desirable “TMP units” that are sized for transport through the TMP mechanism 202. The data portion of TCP packets may range in size depending on client 205 and any intervening links coupling client 205 to TCP component 401. Moreover, where compression is applied, the compressed data will vary in size depending on the compressibility of the data. Data blender 404 receives information from front-end manager 207 that enables selection of a preferable TMP packet size. Alternatively, a fixed TMP packet size can be set that yields desirable performance across TMP mechanism 202. Data blender 404 also marks the TMP units so that they can be re-assembled at the receiving end.

[0068] Data blender 404 also serves as a buffer for storing packets from all clients 205 that are associated with front-end 201. Blender 404 mixes data packets coming into front-end 201 into a cohesive stream of TMP packets sent to back-end 203 over TMP link 202. In creating a TMP packet, blender 404 is able to pick and choose amongst the available data packets so as to prioritize some data packets over others.

[0069] In an exemplary implementation, a “TMP connection” comprises a plurality of “TCP connection buffers”, logically arranged in multiple “rings”. Each TCP socket maintained between the front-end 201 and a client 205 corresponds to a TCP connection buffer. When a TCP connection buffer is created, it is assigned a priority. For purposes of the present invention, any algorithm or criteria may be used to assign a priority. Each priority ring is associated with a number of TCP connection buffers having similar priority. In a specific example, five priority levels are defined corresponding to five priority rings. Each priority ring is characterized by the number of connection buffers it holds (nSockets), the number of connection buffers it holds that have data waiting to be sent (nReady) and the total number of bytes of data in all the connection buffers that it holds (nBytes).

[0070] When composing TMP data packets, the blender goes into a loop comprising the steps:

[0071] 1) Determine the number of bytes available to be sent from each ring (nBytes), and the number of TCP connections that are ready to send (nReady)

[0072] 2) Determine how many bytes should be sent from each ring. This is based on a weight parameter for each priority. The weight can be thought of as the number of bytes that should be sent at each priority this time through the loop.

[0073] 3) The nSend value computed in the previous step reflects the weighted proportion that each ring will have in a blended TMP packet, but the values of nSend do not reflect how many bytes need to be selected to actually empty most or all of the data waiting to be sent a single round. To do this, the nSend value is normalized to the ring having the most data waiting (e.g., nBytes=nSendNorm). This involves a calculation of a factor: S=nBytes/(Weight*nReady) for the ring with the greatest nReady. Then, for each ring, calculate nReady*S*Weight to get the normalized value (nSendNorm) for each priority ring.

[0074] 4) Send sub-packets from the different rings. This is done by taking a sub-packet from the highest priority ring and adding it to a TMP packet, then adding a sub-packet from each of the top two queues, then the top three, and so on.

[0075] 5) Within each ring, sub-packets are added round robin. When a sub-packet is added from a TCP connection buffer the ring is rotated so the next sub-packet the ring adds will come from a different TCP connection buffer. Each sub-packet can be up to 512 bytes in a particular example. If the connection buffer has less than 512 bytes waiting, the data available is added to the TMP packet.

[0076] 6) When a full TMP packet (roughly 1.5 kB in a particular example)is built, it is sent. This can have three or more sub packets, depending on their size. The TMP packet will also be sent when there is no more data ready.

[0077] TMP mechanism 405 implements the TMP protocol in accordance with the present invention. TMP is a TCP-like protocol adapted to improve performance for multiple channels operating over a single connection. Front-end TMP mechanism 405 and a corresponding back-end TMP mechanism 505 shown in FIG. 5 are computer processes that implement the end points of TMP link 202. The TMP mechanism in accordance with the present invention creates and maintains a stable connection between two processes for high-speed, reliable, adaptable communication.

[0078] TMP is not merely a substitute for the standard TCP environment. TMP is designed to perform particularly well in heterogeneous network environments such as the Internet. TMP connections are made less often than TCP connections. Once a TMP connection is made, it remains up unless there is some kind of direct intervention by an administrator or there is some form of connection breaking network error. This reduces overhead associated with setting up, maintaining and tearing down connections normally associated with TCP.

[0079] Another feature of TMP is its ability to channel numerous TCP connections through a single TMP pipe 202. The environment in which TMP resides allows multiple TCP connections to occur at one end of the system. These TCP connections are then mapped into a single TMP connection. The TMP connection is then broken down at the other end of the TMP pipe 202 in order to traffic the TCP connections to their appropriate destinations. TMP includes mechanisms to ensure that each TMP connection gets enough of the available bandwidth to accommodate the multiple TCP connections that it is carrying.

[0080] Another advantage of TMP as compared to traditional protocols is the amount of information about the quality of the connection that a TMP connection conveys from one end to the other of a TMP pipe 202. As often happens in a network environment, each end has a great deal of information about the characteristics of the connection in one direction, but not the other. By knowing about the connection as a whole, TMP can better take advantage of the available bandwidth.

[0081] In contrast with conventional TCP mechanisms, the behavior implemented by TMP mechanism 405 is constantly changing. Because TMP obtains bandwidth to host a variable number of TCP connections and because TMP is responsive to information about the variable status of the network, the behavior of TMP is preferably continuously variable. One of the primary functions of TMP is being able to act as a conduit for multiple TCP connections. As such, a single TMP connection cannot behave in the same manner as a single TCP connection. For example, imagine that a TMP connection is carrying 100 TCP connections. At this time, it loses one packet (from any one of the TCP connections) and quickly cuts its window size in half (as specified for TCP). This is a performance reduction on 100 connections instead of just on the one that lost the packet.

[0082] Each TCP connection that is passed through the TMP connection must get a fair share of the bandwidth, and should not be easily squeezed out. To allow this to happen, every TMP becomes more aggressive in claiming bandwidth as it accelerates. Like TCP, the bandwidth available to a particular TMP connection is measured by its window size (i.e., the number of outstanding TCP packets that have not yet been acknowledged). Bandwidth is increased by increasing the window size, and relinquished by reducing the window size. Up to protocol specified limits, each time a packet is successfully delivered and acknowledged, the window size is increased until the window size reaches a protocol specified maximum. When a packet is dropped (e.g., no acknowledge received or a resend packet response is received), the bandwidth is decreased by backing off the window size. TMP also ensures that it becomes more and more resistant to backing off (as compared to TCP) with each new TCP connection that it hosts. A TMP should not go down to a window size of less than the number of TCP connections that it is hosting.

[0083] In a particular implementation, every time a TCP connection is added to (or removed from) what is being passed through the TMP connection, the TMP connection behavior is altered. It is this adaptation that ensures successful connections using TMP. Through the use of the adaptive algorithms discussed above, TMP is able to adapt the amount of bandwidth that it uses. When a new TCP connection is added to the TMP connection, the TMP connection becomes more aggressive. When a TCP connection is removed from the TMP connection, the TMP connection becomes less aggressive.

[0084] TMP pipe 202 provides improved performance in its environment as compared to conventional TCP channels, but it is recognized that TMP pipe 202 resides on the open, shared Internet in the preferred implementations. Hence, TMP must live together with many protocols and share the pipe efficiently in order to allow the other transport mechanisms fair access to the shared communication bandwidth. Since TMP takes only the amount of bandwidth that is appropriate for the number of TCP connections that it is hosting (and since it monitors the connection and controls the number of packets that it puts on the line), TMP will exist cooperatively with TCP traffic. Furthermore, since TMP does a better job at connection monitoring than TCP and TMP is better suited to throughput and bandwidth management than TCP.

[0085] Also shown in FIG. 4 are data filter component 406 and HTTP reassemble component 407 that process incoming (with respect to client 205) data. TMP mechanism 405 receives TMP packets from TMP pipe 202 and extracts the TMP data units. Using the appended sequencing information, the extracted data units are reassembled into HTTP data packet information by HTTP reassembler 407. Data filter component 406 may also implement data decompression where appropriate, decryption, and handle caching when the returning data is of a cacheable type.

[0086]FIG. 5 illustrates principle functional components of an exemplary back-end 203 in greater detail. Primary functions of the back-end 203 include translating transmission control protocol (TCP) packets from web server 210 into TMP packets as well as translating TMP packets received from a front-end 201 into the one or more corresponding TCP packets to be send to server 210.

[0087] TMP unit 505 receives TMP packets from TMP pipe 202 and passes them to HTTP reassemble unit 507 where they are reassembled into the corresponding TCP packets. Data filter 506 may implement other functionality such as decompression, decryption, and the like to meet the needs of a particular application. The reassembled data is forwarded to TCP component 501 for communication with web server 210.

[0088] TCP data generated by the web server process are transmitted to TCP component 501 and forwarded to HTTP parse mechanism 502. Parser 502 operates in a manner analogous to parser 402 shown in FIG. 4 to extract the data portion from the received TCP packets, perform optional compression, encryption and the like, and forward those packets to data blender 504. Data blender 504 operates in a manner akin to data blender 404 shown in FIG. 4 to buffer and prioritize packets in a manner that is efficient for TMP transfer. Priority information is received by, for example, back-end manager 209 based upon criteria established by the web site owner. TMP data is streamed into TMP unit 505 for communication on TMP pipe 202.

[0089] Returning again to FIG. 2, in a particular example, each front-end servers 201 will each maintain persistent connections to a number of back-ends, each of which is associated with a destination server. Front-ends 201 maintain a list of alternate connection addresses for back-ends 203 that support various destination sites. These alternate connections can be initiated when traffic on another route to the same destination site has reached capacity, when the connection route used by the current connection has deteriorated to the point that user performance is degraded beyond a specified acceptable limit, based on quality of service monitoring of the connection, or when the alternate route is expected to provide better Quality of Service (QOS). QOS monitoring comprises, for example, monitoring the percentage of lost packets and round trip time (e.g., time elapsed between sending a packet and receive an acknowledge packet). In a first case both routes will be used simultaneously, with the front-end 201 performing load balancing using performance data reported from the protocol QOS reporting functionality and considering current and expected loads on each route.

[0090]FIG. 6 illustrates an alternative embodiment in which the load balancing functionality is implemented in a single intermediary server 601 located at a front-end (e.g., on the client side of network 101). In an example operation shown in FIG. 6, front-end 601 may be supported in a network service provider's location 602. Front-end 601 supports each web server 610 that together implement a particular web site such as web site 210 shown in FIG. 2. Each front-end 601 may support multiple web sites as well, although multiple web site support is not shown for ease of description and understanding. Each web server 610 may have its own globally unique network address and so be directly accessible via network 101. Alternatively, one or more servers 610 may be coupled together by a WAN or LAN having private addresses so that connections must be funneled through a single access point.

[0091] QOS monitor 604 monitors the channel(s) between front-end 601 and each web server 610. Additionally, front-end 601 may have a preselected high water mark value indicating a maximum number of connections that can be in flight to a given web server 610. In yet another alternative, front-end 601 may conduct in-band or out-of-band communication with web servers 610 to determine their status with respect to current load. In any case, front-end 601 uses the QOS and server load information to select channels through network 101.

[0092] In a particular example, some or all of web servers 610 provide some set of redundant services such that a given request may be satisfied by more than one server 610. Front-end 601 selects which of the capable servers 610 to send a particular request based upon the server load availability and/or QOS data.

[0093] In another example, front-end 601 receives and buffers multiple requests for services directed to web servers 610. The buffered requests may come from a single client or from multiple browser clients 605. Front-end 601 selects buffered requests for transmission across network 101. In accordance with the present invention, the order in which buffered requests are selected is determined, at least in part, by the requirements of load balancing amongst the multiple available web servers 610. In other words, if a front-end 601 holds buffered requests for two web servers 610, preference will be given to launching requests to an non-busy server while a buffered request to a busy server may remain in the buffer for a longer time.

[0094]FIG. 7 illustrates an alternative implementation in which load balancing functionality is implemented in a single intermediary server 701 located at the back-end (e.g., the server side of network 101). Back-end server 701 receives requests from a variety of clients and/or front-end web servers 201. Some or all of web servers 610 provide a set of redundant services such that a given request may be satisfied by more than one server 610. back-end 701 selects which of the capable servers 610 to send a particular request based upon the server load and/or QOS data. Back-end 701 can directly monitor server load as it is aware of how many requests are pending at any particular serve 610. Moreover, back-end 701 can be programmed to be aware of the capabilities of each web server 610. Hence, when the number of outstanding requests to a particular server reaches a preselected high water mark, requests can be routed to other servers 610.

[0095] Preferably, back-end server 701 routes requests based not only on volume, but also based on type of request. Requests for database access, multimedia content delivery, dynamic web page generation, and static web page delivery vary significantly on the resource load of a server. Back-end server 701 monitors the type of request being made based upon, for example, header information in the received packet and/or information determined by parsing the request itself. Using this information, a server with no pending multimedia requests may be favored over an alternative server 610 with one or two pending multimedia requests.

[0096] As shown in FIG. 7, back-end 701 may include a queuing data structure 702. Queue 702 holds requests so that they can be applied to servers 610 in a manner that improves performance of server 610. The size, timing, and type of request may be used to determine when a request is released from queue 702 and applied to a server 610. Also, requests can be queued to maintain a substantially consistent number of pending requests to any given server 610 to improve performance of that server.

[0097] In any of the load balancing implementations disclosed herein, the load balancing decision (i.e., which server will receive a given request) can be made not only based upon relative load of the available servers, but also based upon other criteria. These other criteria might be, for example, which available server has the freshest content. Sometimes mirrored web sites, for example, have a disparity between mirror copies such that some are maintained more frequently than others. The load balancing web servers in accordance with the present invention, are capable of selecting amongst the mirrors by using knowledge of which mirror is designed to have the freshest contentment. Hence, the “original” web site may be given a disproportionate volume of traffic from a pure load balancing standpoint up until it is unable to efficiently process the volume of requests. After this point is reached, the mirror sites may be used.

[0098] Although the invention has been described and illustrated with a certain degree of particularity, it is understood that the present disclosure has been made only by way of example, and that numerous changes in the combination and arrangement of parts can be resorted to by those skilled in the art without departing from the spirit and scope of the invention, as hereinafter claimed. For example, while devices supporting HTTP data traffic are used in the examples, the HTTP devices may be replaced or augmented to support other public and proprietary protocols and languages including FTP, NNTP, SMTP, SQL and the like. In such implementations the front-end 201 and/or back-end 203 are modified to implement the desired protocol. Moreover, front-end 201 and back-end 203 may support different protocols and languages such that the front-end 201 supports, for example, HTTP traffic with a client and the back-end supports a DBMS protocol such as SQL. Such implementations not only provide the advantages of the present invention, but also enable a client to access a rich set of network resources with minimal client software.

Referenced by
Citing PatentFiling datePublication dateApplicantTitle
US7155475 *Feb 15, 2002Dec 26, 2006Sony CorporationSystem, method, and computer program product for media publishing request processing
US7185094 *Sep 26, 2001Feb 27, 2007Sandcherry, Inc.Media session framework using a control module to direct and manage application and service servers
US7346693 *Dec 8, 2003Mar 18, 2008Fujitsu LimitedLoad balancing server to determine optimum value of number of permanent connections with server
US7379966Aug 28, 2006May 27, 2008Sony CroporationSystem, method, and computer program product for media publishing request processing
US7451309Sep 17, 2004Nov 11, 2008At&T Intellectual Property L.P.Signature specification for encrypted packet streams
US7454500 *Sep 26, 2000Nov 18, 2008Foundry Networks, Inc.Global server load balancing
US7475106Oct 25, 2006Jan 6, 2009Sony CorporationSystem, method, and computer program product for media publishing request processing
US7478166Aug 28, 2006Jan 13, 2009Sony CorporationSystem, method, and computer program product for media publishing request processing
US7496651May 6, 2004Feb 24, 2009Foundry Networks, Inc.Configurable geographic prefixes for global server load balancing
US7574508Aug 7, 2002Aug 11, 2009Foundry Networks, Inc.Canonical name (CNAME) handling for global server load balancing
US7581009Apr 27, 2007Aug 25, 2009Foundry Networks, Inc.Global server load balancing
US7584301May 6, 2004Sep 1, 2009Foundry Networks, Inc.Host-level policies for global server load balancing
US7649876 *Aug 31, 2006Jan 19, 2010Berg Mitchell TMethod and system for communicating an information packet through multiple router devices
US7676576Feb 28, 2003Mar 9, 2010Foundry Networks, Inc.Method and system to clear counters used for statistical tracking for global server load balancing
US7730205Mar 24, 2004Jun 1, 2010Intel CorporationOS agnostic resource sharing across multiple computing platforms
US7730519Sep 17, 2004Jun 1, 2010At&T Intellectual Property I, L.P.Detection of encrypted packet streams using feedback probing
US7761705Sep 17, 2004Jul 20, 2010At&T Intellectual Property I, L.P.Detection of encrypted packet streams
US7782889 *Sep 30, 2004Aug 24, 2010Nokia Siemens Networks Gmbh & Co. KgMethod for regulating the transmission parameters of broadband transmission channels assembled to form a group
US7818422 *Dec 31, 2007Oct 19, 2010Microsoft CorporationReal-time monitoring of a routing server
US7899998 *Oct 19, 2007Mar 1, 2011Sprint Communications Company L.P.Conflict avoidance in data store replication
US7912893Oct 25, 2006Mar 22, 2011Sony CorporationSystem, method, and computer program product for media publishing request processing
US7979569Dec 1, 2006Jul 12, 2011Firestar Software, Inc.System and method for exchanging information among exchange applications
US8077609 *Dec 15, 2005Dec 13, 2011At&T Intellectual Property I, L.P.Method for providing quality-of-service based services in a packet network
US8239535 *Dec 20, 2005Aug 7, 2012Adobe Systems IncorporatedNetwork architecture with load balancing, fault tolerance and distributed querying
US8248928Nov 8, 2007Aug 21, 2012Foundry Networks, LlcMonitoring server load balancing
US8250313Jan 19, 2011Aug 21, 2012Sprint Communications Company L.P.Conflict avoidance in data store replication
US8316231Nov 10, 2008Nov 20, 2012At&T Intellectual Property I, L.P.Signature specification for encrypted packet streams
US8332938Sep 17, 2004Dec 11, 2012At&T Intellectual Property I, L.P.Detection of encrypted packet streams using a timer
US8379534May 13, 2010Feb 19, 2013At&T Intellectual Property I, L.P.Detection of encrypted packet streams using feedback probing
US8554790 *Dec 18, 2007Oct 8, 2013Red Hat, Inc.Content based load balancer
US8620989Dec 1, 2006Dec 31, 2013Firestar Software, Inc.System and method for exchanging information among exchange applications
US8645686Aug 8, 2012Feb 4, 2014At&T Intellectual Property I, L.P.Detection of encrypted packet streams using a timer
US20060274761 *Dec 20, 2005Dec 7, 2006Error Christopher RNetwork architecture with load balancing, fault tolerance and distributed querying
US20110158160 *Dec 31, 2009Jun 30, 2011Elster Electricity, LlcIp encapsulation and routing over wireless radio networks
US20140024383 *Jul 23, 2012Jan 23, 2014At&T Intellectual Property I, L.P.System and method for quality of service in a wireless network environment
EP1629357A2 *Jun 4, 2004Mar 1, 2006The Trustees of The University of PennsylvaniaNdma scalable archive hardware/software architecture for load balancing, independent processing, and querying of records
WO2005006186A2 *Jun 9, 2004Jan 20, 2005Intel CorpOs agnostic resource sharing across multiple computing platforms
Classifications
U.S. Classification1/1, 707/999.001
International ClassificationG06F17/30, G06F15/16, G06F9/50, H04L12/56, H04L29/12, H04L29/08, H04L29/06
Cooperative ClassificationY10S707/99944, Y10S707/99945, H04L47/193, G06F9/5055, H04L2029/06054, H04L63/0272, H04L63/1458, H04L47/10, G06F9/5027, H04L63/1416, H04L47/14, G06F9/5088, H04L47/24, H04L29/12066, H04L61/1511, H04L63/0428, H04L63/1408, G06F9/505, H04L67/2814, H04L67/1029, H04L67/303, H04L69/329, H04L67/142, H04L67/1008, H04L67/1021, H04L67/1002, H04L67/2804, H04L67/02, H04L67/1031, H04L67/2842, H04L67/2819, H04L67/2823, H04L67/2847, H04L67/1034, H04L67/2852, H04L67/101, H04L67/322, H04L29/06
European ClassificationH04L29/08N9A1B, H04L29/08N9A1C, H04L47/14, H04L29/08N9A7, H04L29/06, H04L29/08N27E, H04L61/15A1, G06F9/50A6, H04L47/19A, H04L47/10, H04L47/24, H04L63/14A1, H04L29/08N13B, H04L29/08N9A11, G06F9/50L2, H04L29/08N27S, H04L29/08N9A1H, G06F9/50A6S, H04L29/08N27S2, H04L29/08N27F, H04L29/08N27A, H04L63/14D2, H04L63/04B, H04L29/08N9A9, H04L63/14A, H04L29/08N27S4, H04L29/12A2A1, H04L29/08N29T, H04L29/08N9A, H04L29/08N1, G06F9/50A6L, H04L29/08N27D
Legal Events
DateCodeEventDescription
Aug 13, 2001ASAssignment
Owner name: CIRCADENCE CORPORATION, COLORADO
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VANGE, MARK;REEL/FRAME:012099/0340
Effective date: 20010627