Publication number | US20080267182 A1 |

Publication type | Application |

Application number | US 12/165,825 |

Publication date | Oct 30, 2008 |

Filing date | Jul 1, 2008 |

Priority date | Mar 11, 2004 |

Also published as | US7397796 |

Publication number | 12165825, 165825, US 2008/0267182 A1, US 2008/267182 A1, US 20080267182 A1, US 20080267182A1, US 2008267182 A1, US 2008267182A1, US-A1-20080267182, US-A1-2008267182, US2008/0267182A1, US2008/267182A1, US20080267182 A1, US20080267182A1, US2008267182 A1, US2008267182A1 |

Inventors | Aleksandra Smiljanic |

Original Assignee | Aleksandra Smiljanic |

Export Citation | BiBTeX, EndNote, RefMan |

Referenced by (7), Classifications (6) | |

External Links: USPTO, USPTO Assignment, Espacenet | |

US 20080267182 A1

Abstract

The present invention provides a method for balancing unicast or multicast flows in a multistage non-blocking fabric, wherein the fabric comprises at least one internal switching element (SE) stage, an input SE stage and an output SE stage. The method comprises: (a) receiving cells into the fabric wherein each cell is associated with an input subset and associated with an output subset according to the source and destination address of the cell, (b) assigning each cell to a flow, wherein cells sourced from the same input subset, and bound for the same output subset, or multiple output subsets, are assigned to the same flow, and (c) transmitting flows through the internal SE stage wherein cells of a particular flow are distributed among the internal switching elements, wherein the cells of each particular flow transmitted at each internal SE differs by at most h, wherein h is positive, whereby the flow in the fabric is balanced.

Claims(1)

(a) receiving cells into the fabric wherein each cell is associated with an input subset and associated with an output subset according to the source and destination address of the cell,

(b) assigning each cell to a flow, wherein cells sourced from the same input subset, and bound for the same output subset, or multiple output subsets, are assigned to the same flow, and

(c) transmitting flows through the internal SE stage wherein cells of a particular flow are distributed among the internal switching elements, wherein the cells of each particular flow transmitted at each internal SE differs by at most h, wherein h is positive,

whereby the flow in the fabric is balanced.

Description

This application claims benefit from U.S. provisional Application Ser. No. 60/496,978, filed on Aug. 21, 2003, which application is incorporated herein by reference in its entirety.

The invention relates generally to methods, and apparatuses, for balancing data flows through multistage networks.

Clos circuit switch has been proposed by Clos in 1953 at Bell Labs (C. Clos, “A study of non-blocking switching networks,” *Bell Systems Technology Journal *32:406-424 (1953)). *Switching and Traffic Theory for Integrated Broadband Networks*, Kluwer Academic Press 1990; F. K. Hwang, *The mathematical theory of nonblocking switching networks*, World Scientific, 1998). Here, all connections have the same bandwidths. It has been shown that a circuit can be established through the Clos switching fabric without rearranging existing circuits as long as the number of SEs in the second stage is at least twice the number of inputs of an SE in the first stage, i.e. l≧2 n. It has also been shown that a circuit can be established through the Clos switching fabric as long as the number of SEs in the second stage is no less than the number of inputs of an SE in the first stage, i.e. l≧n. In the latter case, the number of required SEs and their total capacity are smaller due to the fact that the existing circuits can be rearranged. While the complexity of the switching fabric hardware is reduced, the complexity of the algorithm for a circuit setup is increased. In both cases, non-blocking property of the Clos architecture has been proven assuming the specific algorithms for circuit setup (F. K. Hwang, World Scientific, 1998). Various implications of Clos findings have been examined in W. Kabacinski et al. “50th anniversary of Clos networks,” *IEEE Communication Magazine, *41(10): 26-64 (October 2003).

The Clos switching fabric can be used for increasing capacity of packet switches as well. The interconnection of SEs would be the same as in the circuit switch case. However, these SEs should be reconfigured in each cell time slot based on the outputs of outstanding cells. Here, packets are split into cells of a fixed duration which is typically 50 ns (64 bytes at 10 Gb/s). Algorithms for circuit setup in Clos circuit switches cannot be readily applied in Clos packet switches. First, all SEs should be synchronized on a cell-by-cell basis. Then, an implementation of the algorithm that rearranges connections on a cell-by-cell basis in SEs of a rearrangeable non-blocking Clos switch would be prohibitively complex (J. Hui, Kluwer Academic Press 1990). So, the Clos fabric with the larger hardware, l=2n, is needed for a non-blocking packet switch. A scheduling algorithm that would provide non-blocking in a Clos packet switch would require higher processing complexity than its counterpart designed for a cross-bar switch (A. Smiljanić, “Flexible bandwidth allocation in terabit packet switches,” *Proceedings of IEEE Conference on High Performance Switching and Routing*, June 2000, pp. 233-241; A. Smiljanić, “Flexible Bandwidth Allocation in High-Capacity Packet Switches,” *IEEE/ACM Transactions on Networking*, April 2002, pp. 287-293). Few heuristics have been proposed to configure SEs in Clos packet switches without assessment of their blocking nature (McDermott et al., “Large-scale IP router using a high-speed optical switch element,” *OSA Journal on Optical Networking, www.osa*-*jon.org*, July 2003, pp. 228-241; Oki et al., “Concurrent round-robin-based dispatching schemes for Clos-network switches,” *IEEE/ACM Transactions on Networking, *10(6):830-844 (December 2002)).

On the other side, it has been recognized that a Clos packet switch in which the traffic load is balanced across the SEs provides non-blocking, i.e. with sufficiently large buffers it passes all the traffic if the outputs are not over-loaded. Such an architecture has been described in Chaney et al., “Design of a gigabit ATM switch,” *Proceedings of INFOCOM *1997, 1:2-11 (1997) and J. S. Turner, “An optimal nonblocking multicast virtual circuit switch,” *Proceeding of INFOCOM *1994, 1:298-305 (1994). Turner showed that the architecture is non-blocking if the traffic of each multicast session is balanced over the SEs in a Benes packet switch. Here the multicast session carries the information between end users in the network.

However, the delay that packets experience through the Clos switch has not been assessed. Delay guarantees are important for various applications, for example, interactive voice and video, web browsing, streaming etc. In previous work, flows of data belonging to individual multicast sessions were balanced over switching elements (SEs) in the middle stage. The delay for such load balancing mechanism is too long. In order to guarantee acceptable delays for sensitive applications, the utilization of the mechanisms that balances loads of individual sessions decreases unacceptably with switch size (A. Smiljanic, “Performance load balancing algorithm in Clos packet switches,” Proceedings of IEEE Workshop on High Performance Switching and Routing, 2004; A. Smiljanic, “Load balancing algorithm in Clos packet switches,” Proceedings of IEEE International Conference on Communications, 2004). Accordingly, a challenge in the field is providing a minimum required delay guarantee without unacceptably decreasing fabric utilization.

The present invention pertains to load balancing algorithms for non-blocking multistage packet switches. These algorithms allow for maximization of fabric utilization while providing a guaranteed delay.

In one embodiment, the present invention provides a method for balancing unicast or multicast data flow in a multistage non-blocking fabric. The fabric comprises at least one internal switching element (SE) stage, wherein the stage has l internal switching elements, and wherein each internal switching element is associated with a unique numerical identifier.

In the method, the input ports of the fabric are grouped into input sets whereby each input set consists of input ports that transmit through the same input SE. The input sets are further divided into input subsets, designated by i. The output ports of the fabric are also grouped into output sets whereby each output set consists of output ports that receive cells through the same output SE. The output sets are further divided into output subsets, designated by j.

Data cells are received into the fabric. If a cell is a unicast cell, then the cell is associated with an input subset i and associated with an output subset j based on the input port and the output port of the cell. On the other hand, if a cell is a multicast cell, then the cell is associated with an input subset and associated with multiple output subsets based on the input port and the output ports of the cell. Each cell is then assigned a flow. If the cells are unicast cells, then the cells which are associated with the same input subset i and associated with the same output subset j are assigned to the same flow. On the other hand, if the cells are multicast cells, then the cells which are associated with the same input subset and associated with the output subsets of the same output sets are assigned to the same flow.

The flows are then transmitted through the internal SE stage wherein cells of a particular flow are distributed among the internal switching elements. The quantity of the cells of each particular flow transmitted at each internal SE differs by at most h, wherein h is positive, preferably equal to one.

In this method, the number of subsets of at least one input set or at least one output set is less than n, wherein n is the number of ports of that input SE or of that output SE. N is the total number of input ports and output ports. N_{f}, is the maximum number of flows whose cells pass any given link. The variables of n, N, N_{f}, h, i, j and l are natural numbers. One or more flows are received by the fabric simultaneously.

Preferably, the flows are distributed among the internal SE stage by using a counter. For example, a unique counter is associated with each flow, designated as c_{ij}. The counter for each flow is initialized with a number less than or equal to l. A cell from a particular flow is transmitted through the internal switching element associated with a numerical identifier which is equal to the numerical value of the counter. After the cell has been transmitted through that internal switching element, the numerical value of the counter is changed by decrementing or incrementing the counter modulus l. Thus, if another cell of the particular flow is received, then the cell will be transmitted through the internal switching element associated with the updated numerical value of the counter, i.e. through a different internal SE. Then, after transmission, the counter is again changed by decrementing or incrementing the counter modulus l. Such process continues until there are no longer any cells received for the particular flow. The process is performed for cells of each flow.

The counters can be varied in any way which would allow for a sufficient variation of the internal switching elements used to transmit cells of the same flow. Preferably, the counter is varied by the following formula: (c_{ij}+1) mod l, wherein l is the number of SEs in the internal SE stage.

In another embodiment, the present invention provides a flow control device which embodies the methods of the invention.

In a further embodiment, the present invention provides a multistage non-blocking fabric which embodies the methods of the invention.

For a better understanding of the present invention, reference is made to the following description, taken in conjunction with the accompanying drawings, and the scope of the invention set forth in the claims.

The present invention pertains to load balancing algorithms for balancing data flow in a multistage non-blocking fabric (e.g. packet switching networks). A non-blocking fabric is defined as a fabric in which all the traffic for a given output gets through to its destination as long as the output port is not overloaded. These algorithms allow for maximization of fabric utilization while providing for a guaranteed delay. In these algorithms, either inputs or input SEs may balance traffic, and flows to either output SE or outputs may be balanced separately. A fabric comprises packet switches. A packet switch is a system that is connected to multiple transmission links and does the central processing for the activity of a packet switching network where the network consists of switches, transmission links and terminals. The transmission links are connected to network equipment, such as multiplexers (MUX) and demultiplexers (DMUX). A terminal can be connected to the MUX/DMUX or it can be connected to the packet switch system. Generally, the packet switch consists of input and output transmission link controllers and the switching fabric. The input and output link controllers perform the protocol termination traffic management and system administration related to transmission jobs and packet transmission. These controllers also process the packets to help assist in the control of the internal switching of the switching fabric. The switching fabric of the packet switch performs space-division switching which switches each packet from its source link to its destination link.

A multistage fabric for the purposes of this specification comprises several switching element (SE) stages with a web of interconnections between adjacent stages. There is at least one internal switching element (SE) stage, wherein the stage has l internal switching elements, and wherein each internal switching element is associated with a unique numerical identifier. An internal SE stage is a stage that is between the input SE stage and the output SE stage.

Each SE stage consists of several basic switching elements where the switching elements perform the switching operation on individual cells. So, each cell is to be processed by the distributed switching elements without a central control scheme, and thus high throughput switching can be done.

The methods of the present invention can be applied to packets of variable length or packets of fixed length. If the packets received from the input links are of variable length, they are fragmented into fixed-size cells. Variable-length packets are preferably transmitted according to Ethernet protocol. If the packets arriving to the switch all have a fixed length, no fragmentation is required. Such packets are transmitted in accordance with asynchronous transfer mode (ATM) protocol. For the purposes of this invention, a packet of fixed length or a packet of variable length is referred to as a cell.

In the algorithms, the input ports of the fabric are grouped into input sets whereby each input set consists of input ports that transmit through the same input SE. The input sets are divided into input subsets. The output ports of the fabric are also grouped into output sets whereby each output set consists of output ports that receive cells through the same output SE. The output sets are divided into output subsets. Sets can be divided so that each input port and/or each output port belong to only one subset. Alternatively, sets can be divided so that each input port and/or each output port belong to more than one subset. The grouping into sets and division into subsets is made in any efficient manner as would be known by a skilled artisan.

For example, a fabric which comprises 100 input ports and 100 output ports can have the ports grouped into sets of five, i.e. input ports **1**-**5** belong to set one, and output ports **1**-**5** belong to set one; input ports **6**-**10** belong to set two, and output ports **6**-**10** belong to set two; etc. Then the input sets and output sets can be divided into subsets of, for example, even and odd numbered ports. So, in this example, input subsets would be (1,3,5), (2,4), (6,8,10), (7,9) etc.

In one preferred embodiment, each input port belongs to one subset. In another preferred embodiment, one or more of the input ports belong to at least two input subsets. Analogously, in one embodiment, each output port belongs to one subset. In another embodiment, one or more of the output ports belong to at least two input subsets.

Preferably, the number of subsets, and so the number of flows is as small as possible. For example if SEs are cross-bars, the input subsets can be equal to the input ports themselves; and output subsets can be equal to the output sets themselves. Or if SEs are shared buffers, input subsets can be equal to either input ports or input sets, while output subsets can be equal to the output sets.

In some algorithms, input subsets can be equal to either input ports or input sets, while output subsets can be equal to either output ports or output sets. In a first load balancing algorithm of the invention, cells from some input port bound for the particular output SE are spread equally among internal SEs. In a second case, cells from some input port bound for the particular output port are spread equally among internal SEs. Then, the load is balanced by input SEs, e.g., an arbiter associated with each input SE determines to which internal SE a cell will be transmitted. In a third algorithm, cells transmitted from an input SE to some output SE are spread equally across the internal SEs. In a fourth algorithm, cells transmitted from an input SE to some output port are spread equally across the internal SEs.

The methods of the invention are used for both unicast and multicast cells. Cells are received into the fabric. Characteristics of cells being transmitted according to the Internet Protocol (IP) are identified from the packet headers. The packet header contains the source IP address, and the destination IP address. From these addresses, the i, j designation of the cell is obtained, where i is the designation of input subset and j is the designation of the output subset. Based on the i, j designations, each cell is assigned a flow by the following algorithms of the invention. A flow can contain an indefinite number of cells.

If a cell is a unicast cell, then the cell is associated with an input subset and associated with an output subset j based on the input port and the output port of the cell. Then the cells which are associated with the same input subset and associated with the same output subset are assigned to the same flow.

Alternatively, if a cell is a multicast cell, then the cell is associated with an input subset i and associated with multiple output subsets {j} based on the input port and the multiple output ports of the cell, wherein {j} designates a set of output subsets. Then the cells which are associated with the same input subset and associated with the output subsets of the same output sets are assigned to the same flow.

As a way of illustration, using the example above, unicast cells that have the following input ports (x), and output port (y) are assigned to the same flow: (2, 1), (2, 3), (2, 5), (4, 1), (4, 3), (4, 5). As another example, cells that have the following i, j designations are assigned to the same flow: (2, 2), (2, 4), (4, 2), (4, 4).

The number of subsets of at least one input set or at least one output set is less than n, wherein n is the number of ports of that input SE or of that output SE. N is the total number of input ports and output ports. N_{f }is the maximum number of flows whose cells pass any given link. The variables of n, N, N_{f}, h, i, j and l are natural numbers. These variables are defined by the particular fabric with which the invention is used as would be known by a skilled artisan. One or more flows are received by the fabric simultaneously.

The flows are transmitted through the internal SE stage wherein cells of a particular flow are distributed among the internal switching elements. The quantity of the cells of each particular flow transmitted at each internal SE differs by at most h, wherein h is positive. Preferably, h is less than 50, less than 25, less than 20, less than 15, less than 10, or less than 5. Most preferably h is equal to one.

An alternate manner by which to generally define flow follows. Two cells of the same unicast flow must be sourced by the same input set or be bound to the same output sets. Two cells of the same multicast flow must be sourced by the same input sets or be bound to the same sets of output sets.

Preferably, the flows are distributed among the internal SE stage by using a counter. For example, a unique counter is associated with each flow, designated as c_{ij}, wherein i is the numerical identifier of an associated input subset and j is the numerical identifier of an associated output subset;

The counter for each flow is initialized with a number less than or equal to l. A cell from a particular flow is transmitted through the internal switching element associated with a numerical identifier which is equal to the numerical value of the counter. After the cell has been transmitted through that internal switching element, the numerical value of the counter is changed by decrementing or incrementing the counter modulus l. Thus, if another cell of the particular flow is received, then the cell will be transmitted through the internal switching element associated with the updated numerical value of the counter, i.e. through a different internal SE. Then, after transmission, the counter is again changed by decrementing or incrementing the counter modulus l. Such process continues until there are no longer any cells received for the particular flow. The process is performed for cells of each flow. The variable c_{ij }is a natural number.

A counter can be varied in any way which would allow for a sufficient distribution of cells of the same flow among the internal switching elements. The counter is varied by the following formula: (c_{ij}+p) mod l wherein gcd(p,l)=1, wherein gcd means greatest common divisor. Preferably, the counter is varied by the following formula: (c_{ij}+1) mod 1, wherein l is the number of SEs in the internal SE stage. Alternatively, the counters can be varied in a random fashion

In the first load balancing algorithm, input port i, 0≦i≦N, has m different counters associated with different output SEs, c_{ij}, 0≦j≦m. Here N=nm is the number of switch input and output ports. A cell arriving to input port i and bound for the jth output SE is marked to be transmitted through the c_{ij}th output of its SE, i.e. to be transmitted through the c_{ij}th center SE. Then, the counter in question is varied. For example, the counter is incremented modulo l, namely c_{ij}←(c_{ij}+1) mod l.

In the second load balancing algorithm, input i, 0≦i<N, stores N counters associated with different switch outputs, c_{ij}, 0≦j<N. A cell arriving to input port i and bound for the jth switch output port is marked to be transmitted through the c_{ij}th output of its SE, i.e. to be transmitted through the c_{ij}th center SE. Then, the counter in question is varied, e.g., incremented modulo l.

In the third load balancing algorithm, input SE i, 0≦i<m, stores m different counters associated with different output SEs, c_{ij}, 0≦j≦m. A cell arriving to input SE i and bound for the jth output SE is marked to be transmitted through the c_{ij}th output of its SE, i.e. to be transmitted through the c_{ij}th center SE. Then, the counter in question is varied, e.g., incremented modulo 1.

In the fourth load balancing algorithm, input SE i, 0≦i<m, stores N counters associated with different switch outputs, c_{ij}, 0≦j<N. A cell arriving to input SE i and bound for the jth switch output port is marked to be transmitted through the c_{ij}th output of its SE, i.e. to be transmitted through the c_{ij}th center SE. Then, the counter in question is incremented modulo l.

In certain preferred embodiments of the invention, the method further comprises grouping cell time slots into frames of length F. In some of such embodiments, the counter of each flow is set at the beginning of each frame. The counter is set to c_{ij}=(i+j) mod l, where i may be either an input or an input SE, and j may be either an output or an output SE.

In the embodiments wherein cell time slots are grouped into frames of length F, preferably, each frame input port (i) can transmit up to a_{ij }cells to output port (j). The following boundaries hold:

where S is the switching fabric speedup. Preferably, in this embodiment, the fabric speedup is defined as:

wherein:

In this case, the utilization of the fabric is maximized. In this embodiment, with fabric speedup defined in any manner, preferably, at each stage only cells that have arrived in the same frame are transmitted to the next stage, wherein F=D/3T_{c}, or F=D/4T_{c }if cells are reordered at the outputs, wherein D is the maximum tolerable delay and T_{c }is cell time slot duration. Namely, cells passing through different center SEs may lose correct ordering, i.e. a cell that is transmitted earlier through some center SE may arrive to the output later than a cell that is transmitted later through another center SE. For this reason, cell reordering may be required at the switch outputs. In certain preferred embodiments of the invention, the number of flows should fulfill inequality

*N* _{f}≦(*S−U*)·*D/T* _{c},

where S is switching fabric speedup, U is targeted utilization of the switching fabric, D is the maximum tolerable delay and T_{c }is cell time slot duration.

In a further embodiment wherein cell time slots are grouped into frames of length F, and wherein each frame can transmit a_{ij }cells from input port (i) to output port (j), preferably, the number of flows sourced by an input SE or bound for an output SE that are balanced starting from different internal SEs differ by at most one, wherein:

where S is the switching fabric speedup. In this embodiment, speedup is preferably defined as follows:

and wherein

whereby utilization of the fabric is maximized. Preferably, in this embodiment, wherein speedup is defined in any manner, at each stage only cells that have arrived in the same frame are transmitted to the next stage, wherein F=D/3T_{c}, or F=D/4T_{c }if cells are reordered at the outputs, wherein D is the maximum tolerable delay and T_{c }is cell time slot duration.

In one embodiment, in the methods of the present invention the number of flows sourced by an input SE or bound for an output SE that are balanced starting from different internal SEs differs by at most 1, wherein N_{f}, fulfills:

where S is the switching fabric speedup, U is targeted utilization of the switching fabric, D is the maximum tolerable delay and T_{c }is cell time slot duration. Preferably, flow synchronization is achieved by resetting counters each frame. In some proposed algorithms, counters are set in each frame to c_{ij}=(i+j) mod l, where i maybe either input or input SE, and j may be either output or output SE.

The methods of the present invention are analyzed in the present specification by means of theorems and proofs thereof, and by means of examples.

Theorem 1: Non-blocking is provided without link speedup if l≧n.

Proof: Let SE_{ij }denote the jth SE in stage i throughout this specification. In all algorithms, each input, or input SE, will transmit the traffic at equal rates through the connections from input (first stage) to center (second stage) SEs, and, consequently the rate transmitted through any of these connections is:

where s_{i′}, is the rate at which input i′ sends the traffic. If r_{i′k′} denotes the rate at which input i′ sends the traffic to output k′, then the rate transmitted through a connection from a center (second stage) SE to an output (third stage) SE, say SE_{3k}, is:

wherein the outputs are not overloaded. So, the maximum rate supported by a connection in the fabric should fulfill:

because equality may be reached in (1,2). So, non-blocking is provided without link speedup, i.e. with S=1, if l≧n.

Traffic of each individual flow is balanced independently across the SEs. If there are many flows that transmit cells across some SE at the same time, the cells will experience long delay. Many applications, e.g. voice and video, require rate and delay guarantees. The worst case utilizations for balancing algorithms that provide rate and delay guarantees has been assessed.

Time is divided into frames of F cells, and each input-output pair is guaranteed a specified number of time slots per frame, for example a_{ij }time slots are guaranteed to input-output pair (i, j), 0≦i, j<N. Each input, and each output can be assigned at most F_{u }time slots per frame, i.e.

F_{u }is evaluated in terms of F, N, N_{f }for various load balancing algorithms, under the assumption that that l=n. Here N_{f }is the maximum number of flows passing through some connection that are separately balanced.

It is assumed that there is a coarse synchronization in a switch, i.e. that at some point of time the input ports schedule cells belonging to the same frame. A possible implementation for such a coarse synchronization is described later. The coarse synchronization may introduce an additional delay smaller than the frame duration, but may also simplify the controller implementation. Otherwise, SEs should give priority to the earlier frames which complicates their schedulers; also cell resequencing becomes more complex because the maximum jitter is increased. The delay that a cell may experience through Clos switch is three times the frame duration D=3FTc, or D=4FT_{c }if cells are reordered at the outputs.

The number of cells per frame sent from a given input SE through a given center SE (F′_{c}≦F) in terms of F_{u}, and the maximal utilization of the connections from input ports to center SEs (F_{u}/F) is calculated. Because of the symmetry, utilization is the same for the connections from center to output SEs, as shown below. Note that all lemmas and theorems hold in large switches where n>10.

Lemma 1: Let F′_{c}, denote the maximum number of cells per frame sent from a given input SE through a given center SE. It holds that:

*F* _{c} *′≧F* _{u} *+N* _{f} *′−n,* (5)

where N_{f}′ denotes the number of flows sourced by SE_{1i }that pass through the links from this SE to center SEs.

Proof: Let f_{ig}′, 0≦g<N_{f}′, denote the number of time slots per frame that are guaranteed to the individual flows sourced by SE_{1i}. It follows:

where ┌x┐ is the smallest integer no less than x, i.e. ┌x┐<x+1. The maximum number of cells sourced by SE_{1i }that may happen to be transmitted through the given center SE, say SE_{2j}, has been found. It was assumed that out of N_{f}′ flows sourced by SE_{1i}, N_{f}′−n flows are assigned one time slot per frame, and the remaining n flows are assigned max(0, nF_{u}−(N_{f}′−n)) time slots per frame. If it happens that first cells in a frame of all flows are sent through SE_{2j}, the total number of cells per frame transmitted through SE_{2j }from SE_{1i }will be:

Note that in this case F_{c}′ almost reaches the upper bound in (6) for n>10, because n<<N≦F_{u}, and claim of the lemma follows.

Lemma 2: Maximum utilization of the links from input ports to center SEs is:

Proof. Since F_{c}′≦SF for any of the internal connections in the fabric, from Lemma 1 it follows that:

*F* _{u} *≦SF−N* _{f}′. (9)

If (9) holds, all cells pass from SE_{1i }to center SEs within designated frames. So, the maximum utilization of the links from input to center SEs is:

where the last approximation holds for large switches for which n>10.

Lemma 3: Let f_{c}″ denote the maximum number of cells per frame sent to a given output SE through a given center SE. It holds that:

*F* _{c} *″≧F* _{u} *+N* _{f}″. (10)

where N_{f}″ denotes the number of flows bound to SE_{3k }that pass through the links from center SEs to this output SE.

Proof. Let f_{kg}″, 0≦g<N_{f}″, denote the number of time slots per frame that are guaranteed to the individual flows bound for SE_{3k}. Similarly, as in the proof of Lemma 1, it holds that:

*F* _{c} *″<F* _{u} *+N* _{f}″. (11)

Similarly, as in the proof of Lemma 1, out of N_{f}″ flows bound for SE_{3k}, N_{f}″−n flows may transmit one cell per frame that pass through SE_{2j}, and *n *flows may transmit remaining max(0, nF_{u}−N_{f}′+n) cells. If it happens that first cells in a frame of all flows are sent through SE_{2j}, the upper bound in (11) is almost reached, and claim of the lemma follows.

Lemma 4: Maximum utilization of the links from center to output SEs is:

Proof. Maximum utilization of the links from center to output SEs can be derived from Lemma 3 as:

Theorem 2: Maximum utilization of any internal link in the fabric under which all cells pass it within designated frames is:

where N_{f }is the maximum number of flows sourced by any input SE or bound for any output SE, i.e. the maximum number of flows that are passing through some internal link of the fabric.

Proof Maximum utilization of any internal link in the fabric under which all cells pass it within designated frames can be derived from Lemmas 2 and 4:

where N_{f }is the maximum number of flows sourced by any input SE or bound to any output SE, i.e. the maximum number of flows that are passing through some internal link of the fabric.

Note that Theorem 2 holds for Benes network with an arbitrary number of stages as described in Chaney et al., *Proceedings of INFOCOM*1997 1:2-11 and J. S. Turner *Proceedings of INFOCOM*1994 1:298-305. In that case, the latter definition of N_{f }holds, i.e. N_{f }is the maximum number of flows that are passing through some internal link of the fabric.

The maximum utilization when different flows bound for the same SE are not properly synchronized was calculated, so they might send cells within a given frame starting from the same center SE. Alternatively, equal numbers of flows are balanced starting from different center SEs in each frame. For example, flow g of SE_{1i }resets its counter at the beginning of a frame to c_{ig}=(i+g) mod n. Or, flow g bound to SE_{3k }resets its counter at the beginning of a frame to c_{kg}=(k+g) mod n. It is assumed that N_{f}>10n in order to simplify the analysis of load balancing algorithms with synchronized counters.

Lemma 5: In load balancing algorithms with synchronized counters, if:

Proof. The maximum number of cells that are transmitted from SE_{1 }through SE_{2(n-1) }in the middle stage is calculated, and the same result holds for any other center SE. Let f′_{jg }denote the number of cells in flow g which is balanced starting from SE_{2j }at the beginning of each frame, where j=(i+g) mod n. Then, the number of cells in flow g transmitted from SE_{1i }through SE_{2(n-1) }is

where └x┘ is the smallest integer not greater than x i.e. └x┘≦x. So, the number of cells from SE_{1i }through SE_{2(n-1) }is:

for n>10 and N_{f}>10n. Note that inequality (18) holds for n>10 and N_{f }mod n=0 as well. Equality in (18) is reached if:

*f* _{ig} *′=n−*(*i+g*)*mod n+n·y* _{ig}′. (19)

where y_{fg}′≧0 are integers. Values f_{ig}′ that satisfy condition (19) exist if it holds that:

for n>10 and N_{f}>10n.

Note that inequality (20) holds for n>10 and N_{f }mod n=0 as well. When inequality (20) holds, equality in (18) may be reached, and:

If inequality (20) does not hold:

where 0≦z<n is an integer. For

F_{c}′ is maximal for:

If 10N_{f}′/(8N)≦F_{u}<N_{f}′/2 from (18, 23, 24):

Lemma 6: Maximum utilization of the links from input to center SEs, when the counters are synchronized is:

Proof: Since F_{c}′≦SF, from Lemma 5 it follows that for F_{u}≧N_{f}/2,

So, the maximum utilization when counters are reset each frame is:

From equations (27, 29), it follows that:

because N_{f}′≦F and N>>1, so range

is not of a practical interest and was omitted in the final formula.

Lemma 7. In load balancing algorithms with synchronized counters, if:

Proof. First the maximum number of cells that are transmitted to SE_{3k }through SE_{2(n-1) }in the middle stage is calculated, and the same result holds for any other center SE. Let f_{kg}″ denote the number of cells in flow g transmitted to SE_{3k }that are balanced starting from SE_{2j }at the beginning of each frame, where j=(k+g) mod n. Then, the number of cells in flow g transmitted to SE_{3k }through SE_{2(n-1) }is └(f_{kg}″+(k+g) mod n)/n┘. Similarly, as in the proof of Lemma 5, it holds that:

holds, equality in (33) may be reached, so:

Similarly, as in the proof of Lemma 5, if it holds that:

Lemma 8: Maximum utilization of the links from center to output SEs when the counters are reset each frame is:

Proof: Since F_{c}″≦SF, from Lemma 7 it follows that for F_{u}≧N_{f}″/2:

So, maximum utilization of the links from center to output SEs is:

From equations (39, 41), it follows that:

Theorem 3: In the algorithms where balancing of different flows is synchronized, maximum utilization of any internal link in the fabric under which all cells pass it within designated frames is:

Proof: Maximum utilization of any internal link in the fabric under which all cells pass it within designated frames is derived from Lemmas 6 and 8 to be:

Note that Theorem 3 provides the maximum utilization when both balancing of flows sourced by an input SE, and balancing of flows bound for an output SE are synchronized. This assumption holds in all the algorithms.

Often, signal transmission over the fibers connecting distant routers requires the most complex and costly hardware. Therefore, it is important to provide the highest utilization of the fiber transmission capacity. For this reason, switching fabrics with the speedup have been previously proposed. Namely, internal links of the fabric have higher capacity than the external links:

where R is a bit-rate at which data is transmitted through the fibers, and R_{c }is a bit-rate at which data is transmitted through the fabric connections.

Theorem 4: The speedup S required to pass all incoming packets with a tolerable delay when counters are not synchronized is:

and the speedup when counters are synchronized is:

Proof: It should hold that F_{u}=F while F_{c}≦SF, where F_{c }is the number of cells passing through some internal link per frame. When the counters are not synchronized from Lemmas 1 and 3 it follows that:

*S* _{a} *F*≧max(*F* _{c} *′,F* _{c}″)=*F+N* _{f }

and so:

When the counters are synchronized, from Lemmas 5 and 7 it follows that:

because F≧N_{f}>10 N_{f}/(8N), since N≧2. Note that the speedup smaller than 1 means that no speedup is really needed.

The performance of a load balancing algorithm depends on the number of flows that are separately balanced. Let N_{f }denote the maximum number of balanced flows passing through some internal link. As noted before, N_{f }is equal to the maximum number of flows sourced by some input SE or bound to some output SE. In the first algorithm N_{f}=N, because any input SE sources n^{2}=N flows, and each of N inputs balances one flow for any output SE. In the second algorithm, N_{f}=nN, because any input SE sources nN flows, and each of N inputs balances n flows bound for any output SE. In the third algorithm, N_{f}=n because any input SE sources n flows, and each of n input SEs balances one flow for any output SE. In the fourth algorithm, N_{f}=N because any input SE sources N flows, and each of n input SEs balances n flows for any output SE.

Under the assumption of no speedup, i.e. S=1, the maximum utilizations for described load balancing algorithms by substituting N_{f }in formula (14) are obtained:

Thus, the second load balancing algorithm is least efficient, while the third algorithm is most efficient.

In order to increase the efficiency of the load balancing algorithms, in one embodiment of the present invention, the frame length is increased. The cell delay is proportional to the frame length. So the maximum frame length is determined by the delay that could be tolerated by the applications, such as interactive voice and video. Assume that the maximum delay that can be tolerated by interactive applications is D, and the cell time slot duration is T_{c}, then

One way packet delay that can be tolerated by interactive applications is around 150 ms, but only 50-60 ms of this allowed delay can be budgeted for the queueing. The switch delay as low as 3 ms may be required for various reasons. For example, packets might pass multiple packet switches from their sources to the destinations, and packet delays through these switches would add. Also, in order to provide flexible multicasting, the ports should forward packets multiple times through the packet switch, and the packet delay is prolonged accordingly (Chaney et al., *Proceedings of INFOCOM *1997, 1:2-11 (1997); A. Smiljanić, “Scheduling of Multicast Traffic in High-Capacity Packet Switches,” *IEICE/IEEE Workshop on High*-*Performance Switching and Routing*, May 2002, pp. 29-33; A. Smiljanić, “Scheduling of Multicast Traffic in High-Capacity Packet Switches,” *IEEE Communication Magazine*, November 2002, pp. 72-77; and J. S. Turner, *Proceeding of INFOCOM *1994, 1:298-305 (1994)).

*a*) T_{c}=50 ns, while in *b*) T_{c}=100 ns. The solid curves represent the first and fourth algorithms (N_{f}=N), while the dashed curves correspond to the second algorithm (N_{f}=nN). The efficiency of the second balancing algorithm might decrease unacceptably as the switch size increases. For example, the utilization of a fabric with 1000 ports drops below 10% for a tolerable delay of 3 ms and T_{c}=50 ns. On the other side, for the same tolerable delay and cell duration, the utilization of a fabric with 4000 ports is 90% if the first or the fourth load balancing algorithm is applied. Note that utilizations are lower in *b*) when the cell duration is longer T_{c}=100 ns. Thus, the first and fourth load balancing algorithms (for which N_{f}=N) provide a superior performance. Flows balanced starting from different center SEs improve the efficiency of load balancing. Namely, at the beginning of each frame, counters are set to the appropriate values, e.g. c_{ij}=(i+j) mod n, where 0≦i<N, 0≦j<n for the first load balancing algorithm, 0≦i, j<N for the second algorithm, 0≦i<n, 0≦j<N for the fourth algorithm. (Efficiency of the third algorithm is already close to 100%.) Because in all these cases N_{f}≧N>10n and n>10, the guaranteed utilizations for the enhanced load balancing algorithms is derived by substituting N_{f }in formula (43) as follows:

where D is the maximum delay that can be tolerated, and again it is assumed that there is no speedup, i.e. that S=1.

*a*) T_{c}=50 ns, while in *b*) T_{c}=100 ns. The solid curves correspond to the first and fourth algorithms (N_{f}=N), while the dashed curves correspond to the second algorithm (N_{f}=nN). The efficiency of the second load balancing algorithm is improved, but, it is still low in large switches where cells bound for the particular output are spread equally across the center SEs. For example, the utilization of a fabric with 1000 ports drops below 30% for a tolerable delay of 3 ms and T_{c}=50 ns, and again drops below 10% in a switch with 4000 ports. The efficiency of the first and fourth load balancing algorithms is improved too, i.e. for the same tolerable delay and cell duration the utilization of a fabric with 4000 ports is 90%. Note that utilizations are lower in *b*) when the cell duration is longer, T_{c}=100 ns. Again, the first and fourth load balancing algorithms provide much better performance than the second load balancing algorithm.

In another embodiment of the present invention, the utilization of the transmission capacity is maximized to 100% by implementing the switching fabric with a speedup. The speedup required to provide non-blocking varies for different load balancing algorithms. In the simple case when different counters are not synchronized, required speedups can be obtained from formula (46) to be:

When the counters are synchronized, required speedups are decreased and are obtained from formula (47) as follows:

Speedups required to pass the packets with a tolerable delay of D can be calculated from formula (55):

When the counters are synchronized, required speedups are decreased and are obtained from formula (56) as follows:

*a*) T_{c}=50 ns, while in *b*) T_{c}=100 ns. The solid curves represent the first and fourth algorithms (N_{f}=N), while the dashed curves correspond to the second algorithm (N_{f}=nN). If the cell duration is 50 ns, the second load balancing algorithm requires the speedups larger than 2 and 10, in order to provide the delay less than 3 ms through a switch with 1000 and 4000 ports, respectively. If the cell duration is 100 ns, the second load balancing algorithm requires the speedups larger than 4 and 11, in order to provide the delay less than 3 ms through a switch with 1000 and 4000 ports, respectively. On the other side, the speedup required when the first and fourth load balancing algorithms are applied is close to 1 for all switch parameters.

*a*) T_{c}=50 ns, while in *b*) T_{c}=100 ns. The solid curves represent the first and fourth algorithms (N_{f}=N), while the dashed curves correspond to the second algorithm (N_{f}=nN). If the cell duration is 50 ns, the second load balancing algorithm requires the speedups larger than 2 and 7, in order to provide the delay less than 3 ms through a switch with 1000 and 4000 ports, respectively. If the cell duration is 100 ns, the second load balancing algorithm requires the speedups larger than 2 and 10, in order to provide the delay less than 3 ms through a switch with 1000 and 4000 ports, respectively. Thus, the required speedup is sometimes decreased when the counters are synchronized. No speedup is needed when the first and fourth load balancing algorithms are applied and the counters are synchronized.

Therefore, it is preferred that cells bound for the output SE are spread equally across center SEs, or that input SEs spread cells across center SEs (N_{f}<N). Since the performance improves as the number of balanced flows decreases, all algorithms for which N_{f}≦N perform well. However, the implementation of the algorithms where input SEs balance the traffic may be more complex, and, consequently, less scalable. First, inputs have to exchange the information with the SE arbiter. Secondly, counters of the arbiter should be updated n times per cell time slot, which may require advanced processing capability, and may limit the number of SE ports, i.e. the total switch capacity. Also, these algorithms assume the SEs with the shared buffers whose capacity was shown to be smaller than the capacity of crossbar SEs. Note that in the Turner article (J. S. Turner, *Proceeding of INFOCOM *1994, 1:298-305), it was proposed that the end-to-end sessions are separately balanced in a switch. In that case N_{f}≧nN; and consequently the performance is poorer than in the cases that were examined in this specification.

In some cases, there is a coarse synchronization in a switch during the flow of data, i.e. at some point of time the input ports schedule cells belonging to the same frame. In one embodiment of the present invention, if the frames at different ports are not synchronized, the correct switch operation can be accomplished in the following way. Frames are delineated by designated packets. One extra bit per packet, FB, is set at the port to denote its frame, and is toggled in each frame. In a given frame the switch arbiter will schedule only packets received before such frame with FB equal to the specified switch bit, SB. SB toggles in each frame as well. *a*) shows the switch frame boundaries, while the lower axes in *b*) and (*c*) show the port frame boundaries. At the beginning of each switch frame, SB toggles, and at the beginning of each port frame, FB toggles, as shown. Thus, only packets with FB=SB=0 that have arrived before the switch frame k+2 in *a*) will be scheduled in the switch frame k+2; and these are packets of the upper port frame m+1 in *b*). Similarly, packets of the port frame m+2 will be scheduled in the switch frame k+2 etc. In *b*), the port is synchronized properly, while in *c*), it is not. Namely, packets arriving at the end of the port frame m and packets arriving at the beginning of the port frame m+2 are eligible for scheduling in the switch frame k+3. So, the number of packets bound for some output that will be scheduled in frame k+3 might exceed negotiations, and would be blocked. Thus, SB and FBs have to be properly synchronized: an arbiter sets FB=1−SB if the switch frame boundary preceded the previous port frame boundary (delineation packet), or FB=SB otherwise, where FB is the frame bit of the first packet arriving as the synchronization process started. Although the coarse synchronization may introduce an additional delay smaller than the frame duration, the synchronization simplifies the controller implementation.

Multiple priorities can be served in the switch. In each SE, high priority cells are first served and their number is limited according to the various admission control conditions that were described above. On the other side, there are no limits for low priority cells which are served if they can get through after the high-priority cells are served. By limiting the number of high-priority cells with the above equation, they are served with the guaranteed delay. If there is any resource left, namely time slots in which some input and output are idle, and there are lower priority cells between them, the lower priority cells are served without any delay guarantees.

Multicasting

A significant amount of traffic on the Internet is multicast in nature; i.e. it carries the information from one source to multiple destinations. Scheduling of multicast packets in switches is a complicated task. If a multicast packet is scheduled to be simultaneously transmitted to all destination outputs, it may be unacceptably delayed. On the other side, if the multicast packet is scheduled to be separately transmitted to all destination outputs, its transmission may consume an unacceptably large portion of the input transmission capacity.

It has been proposed earlier that multicast packet should be forwarded through high-capacity switches (Chaney et al., *Proceedings of INFOCOM *1997, 1:2-11 (1997); A. Smiljanić, *IEICE/IEEE Workshop on High*-*Performance Switching and Routing*, May 2002, pp. 29-33; A. Smiljanic, IEEE Communication Magazine, November 2002, pp. 72-77; J. S. Turner, “An optimal nonblocking multicast virtual circuit switch,” *Proceeding of INFOCOM *1994, vol. 1, pp. 298-305). Namely, a multicast input sends multicast packets to a limited number of destinations, and each multicast destination output that received the packets will forward them to a limited number of destination outputs who did not received them yet, and such forwarding continues until all destination outputs received all the packets. By choosing appropriate forwarding fan-out P, i.e. the number of destination outputs to which a packet is forwarded from one port, the switch utilization and the guaranteed delay could be selected (A. Smiljanić, *IEICE/IEEE Workshop on High*-*Performance Switching and Routing*, May 2002, pp. 29-33; A. Smiljanić, *IEEE Communication Magazine*, November 2002, pp. 72-77).

Packets can be forwarded in two ways. In the first case, a port separately transmits a multicast packet to its destination ports. Then, the packet flow is determined solely based on its input and output ports as in the case of unicast packets. In the second case, a port transmits only one copy of a multicast packet to the Clos network. The multicast packet is transmitted through the network until the last SE from which it can reach some destination port where it is replicated and its copies are routed separately through the remainder of the network. So, the multicast flow is balanced in stages before the packet replication starts. In this case, the packet flow is determined by its input port and its multiple destinations of ports. Obviously, the number of flows is increased in this way, and the performance of load balancing is degraded. On the other side, the port transmission capacity required for forwarding is less. It was shown earlier that P=2 is the most practical choice; then, the port transmission capacity improvement is less than the utilization degradation due to imperfect load balancing, so the first multicasting scheme is recommended. In any case, the performance of the second multicasting scheme is improved when the number of flows is minimized.

Again, various load balancing algorithms can be performed depending on the definition of the flows that are separately balanced. Similarly, as for unicast transmission, four basic algorithms are provided.

In the first algorithm, all cells sourced by some input and bound to some set of P output SEs define one flow. So, for each multicast cell, its output SEs are determined, and the flow is determined by the found set of output SEs. There are N_{f}=nn(nn−1)/2≈nN/2 of such flows that are balanced through and link from input port to center SE. Remember that the corresponding utilization U_{a}=1−N_{f}/F=1−nN/(2F) has been shown to be unsatisfactory.

In the second algorithm, all cells sourced by some input and bound to some set of P outputs define one flow. There is an enormous number, N_{f}=nN(N−1)/2≈nN^{2}/2, of such flows that are balanced through and link from input port to center SE, and this algorithm should be avoided by all means.

In the third algorithm, all cells sourced by some input SE and bound to some set of P output SEs define one flow. There are N_{f}=n(n−1)/2≈N/2 of such flows that are balanced through and link from input to center SE. Thus, the performance of the third algorithm will be fine as shown before.

In the fourth algorithm, all cells sourced by some input SE and bound to some set of P outputs define one flow. There is again an enormous number, N_{f}=N(N−1)/2≈N^{2}/2, of such flows that are balanced through and link from input to center SE. The fourth algorithm should be by all means avoided. The only well performing algorithm is more complex for the implementation, and it assumes the SEs with shared buffers which have the smaller capacity than the cross-bar SEs.

Improvement in the performance of load balancing of unicast and multicast flows in a fabric can be accomplished by increasing the frame length, balancing flows among different internal SEs, implementing the fabric with a speedup, or combinations thereof.

Implementation

The methods of the present invention can be implemented by an article of manufacture which comprises a machine readable medium containing one or more programs which when executed implement the steps of the methods of the present invention.

For example, the methods of the present invention can be implemented using a conventional microprocessor programmed according to the teachings of the present specification, as will be apparent to those skilled in the computer art. Appropriate software coding can readily be prepared by skilled programmers based on the teachings of the present disclosure, as will be apparent to those skilled in the software art. The invention may also be implemented by the preparation of application specific units, such as integrated circuits (ASIC), configurable logic blocks, field programmable gate arrays, or by interconnecting an appropriate network of conventional circuit components, as will be readily apparent to those skilled in the art.

The article of manufacture can comprise a storage medium can include, but is not limited to Random-Access Memory (RAMs) for storing lookup tables. In one embodiment, the assignment of cells to a flow comprise inputting the i, j designation of a cell into a lookup table which table assigns to the cell an input and output set, an input and output subset, and the flow of the cell.

The methods of the present invention can be implemented by an apparatus which comprises: a flow control device configured to perform the steps of the invention. The apparatus can also comprise a counter module configured to assign counters to each flow pursuant to the methods of the invention.

The present invention also includes a multistage non-blocking fabric which comprises a network of switches that perform the method steps of the invention. The fabric comprises at least one internal switching element (SE) stage, wherein the stage has l internal switching elements, an input SE stage, an output SE stage, input ports which are divided into input sets wherein each input set consists of input ports that transmit through the same input SE, and wherein the input sets are further divided into input subsets, and output ports which are divided into output sets wherein each output set consists of output ports that receive cells through the same output SE, and wherein the output sets are further divided into output subsets, and a flow assignment module wherein the module assigns cells which are received into the fabric to a flow. The assignment module comprises a lookup table.

Thus, while there have been described what are presently believed to be the preferred embodiments of the invention, those skilled in the art will realize that changes and modifications may be made thereto without departing from the spirit of the invention, and it is intended to claim all such changes and modifications as fall within the true scope of the invention.

Referenced by

Citing Patent | Filing date | Publication date | Applicant | Title |
---|---|---|---|---|

US7639684 * | Dec 23, 2004 | Dec 29, 2009 | Infineon Technologies Ag | Modified ethernet switch |

US7873056 * | Nov 18, 2005 | Jan 18, 2011 | Nec Corporation | Switch device, switching method and switch control program |

US7907624 * | Jun 4, 2008 | Mar 15, 2011 | Oracle America, Inc. | Switch matrix |

US8897133 * | Jul 14, 2010 | Nov 25, 2014 | Huawei Technologies Co., Ltd. | Multi-stage switch fabric |

US20100050105 * | Apr 8, 2008 | Feb 25, 2010 | Eric Denis Dufosse | Centralized work flow monitoring |

US20110032934 * | Jul 14, 2010 | Feb 10, 2011 | Huawei Technologies Co., Ltd. | Multi-stage switch fabric |

US20120320917 * | Dec 20, 2012 | Electronics And Telecommunications Research Institute | Apparatus and method for forwarding scalable multicast packet for use in large-capacity switch |

Classifications

U.S. Classification | 370/390 |

International Classification | H04L12/28 |

Cooperative Classification | H04L49/254, H04L49/1515 |

European Classification | H04L49/00, H04L49/15C |

Rotate