WO2002098064A1 - Distributed control of data flow in a network switch - Google Patents

Distributed control of data flow in a network switch Download PDF

Info

Publication number
WO2002098064A1
WO2002098064A1 PCT/US2002/016288 US0216288W WO02098064A1 WO 2002098064 A1 WO2002098064 A1 WO 2002098064A1 US 0216288 W US0216288 W US 0216288W WO 02098064 A1 WO02098064 A1 WO 02098064A1
Authority
WO
WIPO (PCT)
Prior art keywords
ingress
egress
data
network switch
interfaces
Prior art date
Application number
PCT/US2002/016288
Other languages
French (fr)
Inventor
Steve West
Dirk Brandis
Russ Smith
Frank Marrone
Original Assignee
Turin Networks
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Turin Networks filed Critical Turin Networks
Publication of WO2002098064A1 publication Critical patent/WO2002098064A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/30Peripheral units, e.g. input or output ports
    • H04L49/3027Output queuing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/50Queue scheduling
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/50Queue scheduling
    • H04L47/52Queue scheduling by attributing bandwidth to queues
    • H04L47/521Static queue service slot or fixed bandwidth allocation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/50Queue scheduling
    • H04L47/56Queue scheduling implementing delay-aware scheduling
    • H04L47/564Attaching a deadline to packets, e.g. earliest due date first
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/50Queue scheduling
    • H04L47/62Queue scheduling characterised by scheduling criteria
    • H04L47/621Individual queue per connection or flow, e.g. per VC
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/10Packet switching elements characterised by the switching fabric construction
    • H04L49/112Switch control, e.g. arbitration
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/30Peripheral units, e.g. input or output ports
    • H04L49/3018Input queuing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/30Peripheral units, e.g. input or output ports
    • H04L49/3081ATM peripheral units, e.g. policing, insertion or extraction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/40Constructional details, e.g. power supply, mechanical construction or backplane
    • H04L49/405Physical details, e.g. power supply, mechanical construction or backplane of ATM switches
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/60Software-defined switches
    • H04L49/606Hybrid ATM switches, e.g. ATM&STM, ATM&Frame Relay or ATM&IP
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/90Buffering arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/90Buffering arrangements
    • H04L49/9063Intermediate storage in different physical parts of a node or terminal
    • H04L49/9068Intermediate storage in different physical parts of a node or terminal in the network interface card
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5638Services, e.g. multimedia, GOS, QOS
    • H04L2012/5646Cell characteristics, e.g. loss, delay, jitter, sequence integrity
    • H04L2012/5651Priority, marking, classes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5678Traffic aspects, e.g. arbitration, load balancing, smoothing, buffer management
    • H04L2012/5679Arbitration or scheduling
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/10Packet switching elements characterised by the switching fabric construction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/20Support for services

Definitions

  • the invention relates to network switches. More specifically, the invention relates to distributed control of data flow in network switches.
  • cross-connects are designed to provide the peak bandwidth to all data sources because specific data rates of specific data sources are not known when the cross-connect is designed. As a result, all data paths through the cross-connect provide the peak bandwidth, which may not be consumed by some or even most of the data sources.
  • a further disadvantage of synchronous switching architectures is that centralized switching control and interconnections grow exponentially as the input/output paths grow. Therefore, large switching architectures are complex and require complex control algorithms and techniques.
  • the network switch includes ingress cards to receive data from sources external to the switch and egress cards to transmit data to devices external to the switch.
  • the ingress cards have an ingress buffer to temporarily store data, an ingress scheduler coupled to the ingress buffer, and a set of ports coupled to the ingress scheduler.
  • the ingress scheduler reads data from the ingress buffer and selectively transfers the data to one of the set of ports.
  • the egress cards have a set of ports coupled to receive data from respective ingress card ports.
  • the egress cards also have an egress buffer coupled to the set of egress card ports.
  • the egress buffer selectively reads data from the ports and stores the data.
  • An egress scheduler is coupled to the egress buffer. The egress scheduler reads data from the egress buffer and transmits data to the external devices.
  • Figure 1 illustrates one embodiment of a network architecture having multiple network switches.
  • Figure 2 illustrates one embodiment of an interconnection of cards within a network switch.
  • FIG. 3 conceptually illustrates one embodiment of an ingress scheduler.
  • Figure 4 conceptually illustrates one embodiment of a egress cache scheduling of egress card ports.
  • the network switch described herein provides a cell/packet switching architecture that switches between line interface cards across a meshed backplane.
  • the switching can be accomplished at, or near, line speed in a protocol independent manner.
  • the protocol independent switching provides support for various applications including, but not limited to, Asynchronous Transfer Mode (ATM) switching, Internet Protocol (IP) switching, Multiprotocol Label Switching (MPLS) switching, Ethernet switching and frame relay switching.
  • ATM Asynchronous Transfer Mode
  • IP Internet Protocol
  • MPLS Multiprotocol Label Switching
  • Ethernet switching Ethernet switching and frame relay switching.
  • the architecture allows the network switch to provision service on a per port basis.
  • the network switch provides a non-blocking topology with both input and output queuing and per flow queuing at both ingress and egress.
  • Per flow flow-control can be provided between egress and ingress scheduling. Strict priority, round robin, weighted round robin and earliest deadline first scheduling can be provided.
  • cell/packet discard is provided only at the ingress side of the switch.
  • early packet discard (EPD), partial packet discard (PPD) and random early discard (RED) are provided.
  • Figure 1 illustrates one embodiment of a network architecture having multiple network switches. While the switches of Figure 1 are illustrated as coupled to only router/hosts and networks, any type of device that generates and/or receives data that can be carried by a wide area network can be used.
  • the router/hosts and networks are intended to illustrate data devices and statistical multiplexing devices.
  • Switch 110 is coupled to router/host 100, network 102, network 104 and router/host 104. Any number of devices can be coupled to switch 1 10 in any manner known in the art.
  • router/host 122, network 124, network 126 and router/host 128 are coupled to switch 120. Any number of device can be coupled to switch 120 in any manner known in the art.
  • Switch 110 and 120 are coupled to switch 130.
  • Switch 110 and 120 can also be coupled to other switch or other network devices (not shown in Figure 1).
  • Switch 130 is also coupled to network 140, which can include any type and any number of network elements including additional switches.
  • Switches 110, 120 and 130 receive data from multiple devices including router/hosts, local area networks and other switches.
  • the switches can aggregate multiple data sources into a single data stream.
  • Statistical aggregation allows multiple sources of packet/cell data to share a link or port. For example, 24 sources, each with sustained bandwidth of 64 kbps could share a DS1 (1.544 Mbps) link.
  • Statistical aggregation allows sources of data to burst to bandwidth higher than their sustained rate, based on availability of bandwidth capacity of the link.
  • STS-1 (51.840 Mbps) signals from three networks can be received and combined into an OC-3 (155.520 Mbps) signal.
  • the OC-3 signal can be transmitted to another switch for routing and/or further aggregation.
  • the switches of figure 1 include multiple cards that are interconnected by a switching fabric.
  • the cards have both an ingress data path and an egress data path.
  • the ingress data path is used to receive data from the network and transmit the data to an appropriate card within the switch.
  • the ingress data path schedules transmission of data across the switching fabric.
  • the egress data path is used to receive data from the switching fabric and transmit the data across the network.
  • the egress data path schedules transmission of data out of the switch across the network.
  • the ingress and egress data paths interact to prevent overflow of data within the network switch.
  • Figure 2 illustrates one embodiment of an interconnection of cards within a network switch.
  • the switch of Figure 2 can be, for example, any of switches 110, 120 or 130 of Figure 1.
  • the switch of Figure 2 is illustrated with four ingress cards and four egress cards for reasons of simplicity only.
  • a switch can have any number of ingress cards and any number of egress cards.
  • data flow can be bi-directional. That is, the cards can also provide both egress and ingress functionality.
  • Each ingress card includes an ingress buffer that receives data from an external source (not shown in Figure 2).
  • Data can be in any format, for example, IP packets or ATM cells.
  • the ingress buffers are coupled to ingress schedulers.
  • the ingress schedulers dispatch data to egress cards via a set of ingress ports.
  • each ingress card has a port for each egress card to which the ingress card is coupled. For example, ingress card 0 is coupled to egress card 0 through port 0, to egress card 1 through port 1 , to egress card 2 through port 2, and to egress card 3 through port 3.
  • each ingress card is coupled to each egress card, the interconnection between the ingress cards and the egress cards has n 2 connections where n is the number of ingress/egress cards.
  • the interconnection is referred to as an " n 2 mesh,” or an " n 2 switching fabric.”
  • the number of ingress cards is not equal to the number of egress cards, which is referred to as a " nxm mesh.” The mesh is described in greater detail in U.S. Patent application number 09/746,212, entitled “A FULL MESH INTERCONNECT BACKPLANE ARCHITECTURE,” filed December 22, 2000, which is assigned to the corporate assignee of the present application and incorporated by reference.
  • Data is transmitted by the ingress cards without reference to a bus or mesh clock or frame synchronization signal.
  • a protocol for use in communicating over the mesh is described in greater detail in U.S. Patent application number 09/745,982, entitled “A BACKPLANE PROTOCOL,” filed December 22, 2000, which is assigned to the corporate assignee of the present invention and incorporated by reference.
  • Each egress card includes a port for each ingress card to which the ingress card is coupled. For example, egress card 0 is coupled to ingress card 0 through port 0, to ingress card 1 through port 1 , to ingress card 2 through port 2, and to ingress card 3 through port 3.
  • the ports of the egress card are coupled to an egress buffer.
  • the egress buffer is coupled to an egress scheduler that outputs data to a device external to the egress card (not shown in Figure 2).
  • the architecture illustrated in Figure 2 allows scheduling duties to be distributed between ingress and egress cards. Because the scheduling duties are distributed, a centralized scheduler is not required and transmission of data between ingress cards and egress cards can be accomplished in an asynchronous manner. This allows simpler control of data switching and more efficient use of switching fabric bandwidth.
  • the ingress scheduler When data is received by an ingress card the data is temporarily stored in the ingress buffer on the card.
  • the ingress scheduler extracts data from the ingress buffer and sends the data to the appropriate port. For example data to be transmitted to egress card 2 are sent to port 2.
  • the ingress scheduler reads and sends data according to an earliest deadline first scheduling scheme. In alternate embodiments, strict priority scheduling, round robin scheduling, weighted round robin scheduling, or other scheduling techniques can be used.
  • Data that is transferred between ingress and egress cards can be variable in size.
  • the data can be transmitted as a group of fixed length cells or as one or more variable length packets.
  • the packets on the ingress side compete with each other on a packet basis. Each packet competes against the other packets to be selected by the ingress scheduler. In one embodiment, when one packet is selected, all of the entire packet is moved across the switch fabric. Once the ingress scheduler selects a packet of a given priority, the packet is transmitted before another packet is selected.
  • n ingress sources can potentially contend for a single egress destination.
  • the switch is required to transfer the data from ingress to egress such that the contracts of the individual data flows are honored.
  • the contracts specify bandwidth, latency, jitter, loss and burst tolerance. All contracts must be honored under all traffic contention conditions.
  • the ingress scheduler provides all cell/packet discard functionality.
  • Cell/packet discard can include, for example, early packet discard (EPD), partial packet discard (PPD), random early discard (RED), each of which is known in the art. Additional and/or different cell/packet discard procedures can also be used.
  • EPD early packet discard
  • PPD partial packet discard
  • RED random early discard
  • the ingress schedulers independently schedule packets and cells to the egress side.
  • n separate buffers are provided on each egress, one per ingress. These independent cache buffers have sufficient bandwidth to allow simultaneous egress-side arrivals from all ingress devices.
  • the egress side sends "backpressure" messages to the ingress control access to the n independent cache buffers. Data in the n buffers is transferred to a larger egress buffer, from where it is scheduled to the egress ports.
  • the egress buffer receives data from the ports of the egress card in a predetermined manner. For example, data can be extracted from the ports in a round robin fashion, or data can be extracted from the ports on a priority basis. Also, a combination of round robin and priority-based extraction can also be used.
  • the egress buffer includes a cache for each link (i.e., link between ingress card 2 and egress card 2, link between ingress card 3 and egress card 2).
  • Each cache includes a queue for each class of data. By including a queue for each class of data, the egress buffer can provide quality of service functionality.
  • the egress scheduler extracts data from the egress buffer and transmits the data according to the appropriate network protocol to an external device (not shown in Figure 2).
  • the egress scheduler extracts data from the egress buffer based on priority to provide quality of service functionality.
  • the egress scheduler can extract data from the egress buffer using earliest deadline first scheduling.
  • Figure 3 conceptually illustrates one embodiment of an ingress scheduler.
  • incoming data is categorized into one of three ingress traffic categories (ITCs). Having only three ITCs simplifies the architecture of the ingress scheduler.
  • ITCs ingress traffic categories
  • One embodiment of a mapping of ITCs to ATM and IP traffic is set forth in the following table. Other types of network traffic can be mapped to the three ITCs in a similar manner.
  • servicing of the three ITCs is accomplished according to the following priority: 1) RT, 2) MC, and 3) NRT, assuming no backpressure signals are active. If backpressure signals are active, the transmission of the corresponding category of data is stopped to avoid egress port buffer overflow. Lower priority data can be transmitted when the backpressure signal is active for higher priority data.
  • Ingress router 300 reads data from the ingress buffer (not shown in Figure 3) and sends the data to the appropriate queue based on the ITC mapping described above. Real time data is sent to one of the RT group queues (e.g., 305, 310), multicast data is sent to MC queue 320, and non-real time data is sent to one of the NRT queues (e.g., 330, 335).
  • the ingress scheduler includes 16 RT queues; however, any number of RT queues can be provided.
  • data is read out of the RT group queues in a round robin fashion.
  • the data is selected from the multiple RT group queues on a priority basis.
  • data is read out of the NRT class queues in a round robin fashion.
  • the data is selected from the multiple NRT class queues in a weighted round robin fashion.
  • Data from the three categories of data (RT, MC, NRT) is selected based on a priority basis to be sent to the appropriate ingress port 390.
  • FIG. 4 conceptually illustrates one embodiment of a egress cache scheduling of egress card ports.
  • each egress card port has three associated FIFO buffers for real time data, multicast data and non-real time data, respectively.
  • the data when data is received by egress port 400 the data is sent to one of three queues.
  • the queues correspond to the ITCs.
  • Real time data is stored in real time queue 410
  • multicast data is stored in multicast queue 420
  • non- real time data is stored in non-real time queue 430.
  • Data is read out of the queues in a round robin fashion.
  • the queue from which data is transmitted is selected on a priority basis.

Abstract

The network switch (110) described herein provides a cell/packet switching architecture (Fig. 1) that switches between line interface cards across a meshed backplane. In one embodiment, the switching can be accomplished at, or near, line speed in a protocol independent manner. The protocol independent switching provides support for various applications including Asynchronous Transfer Mode (ATM) switching Internet Protocol (IP) switching. Multiprotocol Label Switching (MPLS) switching, Ethernet switching and frame relay switching. The architecture allows the network switch (110) to provision service on a per port basis. In one embodiment, the network switch (110) provides a non-blocking topology with both input and output queuing and per flow queuing at both ingress and egress. Per flow flow-control can be provided between ingress and egress scheduling. Strict priority, round robin, weighted round robin and earliest deadline first scheduling can be provided.

Description

DISTRIBUTED CONTROL OF DATA FLOW IN A NETWORK SWITCH
FIELD OF THE INVENTION
The invention relates to network switches. More specifically, the invention relates to distributed control of data flow in network switches.
BACKGROUND OF THE INVENTION
In high bandwidth networks such as fiber optic networks, lower bandwidth services such as voice communications are aggregated and carried over a single fiber optic link. However, because the aggregated data can have different destinations some mechanism for switching the aggregated components is required. Switching can be performed at different levels of aggregation.
Current switching is accomplished in a synchronous manner. Signals are routed to a cross-connect or similar switching device that switch and route signals at some predetermined granularity level, for example, byte by byte. Synchronous switching in a cross-connect is a logically straight forward method for switching. However, because data flow between network nodes is not necessarily consistent, switching bandwidth may not be used optimally in a synchronous cross-connect. One source of data may use all available bandwidth while a second source of data may transmit data sporadically.
In order to support data sources that transmit at or near peak bandwidth, cross-connects are designed to provide the peak bandwidth to all data sources because specific data rates of specific data sources are not known when the cross-connect is designed. As a result, all data paths through the cross-connect provide the peak bandwidth, which may not be consumed by some or even most of the data sources.
A further disadvantage of synchronous switching architectures is that centralized switching control and interconnections grow exponentially as the input/output paths grow. Therefore, large switching architectures are complex and require complex control algorithms and techniques. SUMMARY OF THE INVENTION
A network switch is described. The network switch includes ingress cards to receive data from sources external to the switch and egress cards to transmit data to devices external to the switch. The ingress cards have an ingress buffer to temporarily store data, an ingress scheduler coupled to the ingress buffer, and a set of ports coupled to the ingress scheduler. The ingress scheduler reads data from the ingress buffer and selectively transfers the data to one of the set of ports. The egress cards have a set of ports coupled to receive data from respective ingress card ports. The egress cards also have an egress buffer coupled to the set of egress card ports. The egress buffer selectively reads data from the ports and stores the data. An egress scheduler is coupled to the egress buffer. The egress scheduler reads data from the egress buffer and transmits data to the external devices. BRIEF DESCRIPTION OF THE DRAWINGS
The invention is illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings in which like reference numerals refer to similar elements.
Figure 1 illustrates one embodiment of a network architecture having multiple network switches.
Figure 2 illustrates one embodiment of an interconnection of cards within a network switch.
Figure 3 conceptually illustrates one embodiment of an ingress scheduler.
Figure 4 conceptually illustrates one embodiment of a egress cache scheduling of egress card ports. DETAILED DESCRIPTION
Techniques for distributed control of data flow in a network switch are described. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the invention. It will be apparent, however, to one skilled in the art that the invention can be practiced without these specific details. In other instances, structures and devices are shown in block diagram form in order to avoid obscuring the invention.
Reference in the specification to "one embodiment" or "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the invention. The appearances of the phrase "in one embodiment" in various places in the specification are not necessarily all referring to the same embodiment.
The network switch described herein provides a cell/packet switching architecture that switches between line interface cards across a meshed backplane. In one embodiment, the switching can be accomplished at, or near, line speed in a protocol independent manner. The protocol independent switching provides support for various applications including, but not limited to, Asynchronous Transfer Mode (ATM) switching, Internet Protocol (IP) switching, Multiprotocol Label Switching (MPLS) switching, Ethernet switching and frame relay switching. The architecture allows the network switch to provision service on a per port basis.
In one embodiment, the network switch provides a non-blocking topology with both input and output queuing and per flow queuing at both ingress and egress. Per flow flow-control can be provided between egress and ingress scheduling. Strict priority, round robin, weighted round robin and earliest deadline first scheduling can be provided. In one embodiment, cell/packet discard is provided only at the ingress side of the switch. In one embodiment, early packet discard (EPD), partial packet discard (PPD) and random early discard (RED) are provided.
Figure 1 illustrates one embodiment of a network architecture having multiple network switches. While the switches of Figure 1 are illustrated as coupled to only router/hosts and networks, any type of device that generates and/or receives data that can be carried by a wide area network can be used. The router/hosts and networks are intended to illustrate data devices and statistical multiplexing devices. Switch 110 is coupled to router/host 100, network 102, network 104 and router/host 104. Any number of devices can be coupled to switch 1 10 in any manner known in the art. Similarly, router/host 122, network 124, network 126 and router/host 128 are coupled to switch 120. Any number of device can be coupled to switch 120 in any manner known in the art.
Switch 110 and 120 are coupled to switch 130. Switch 110 and 120 can also be coupled to other switch or other network devices (not shown in Figure 1). Switch 130 is also coupled to network 140, which can include any type and any number of network elements including additional switches.
Switches 110, 120 and 130 receive data from multiple devices including router/hosts, local area networks and other switches. The switches can aggregate multiple data sources into a single data stream. Statistical aggregation allows multiple sources of packet/cell data to share a link or port. For example, 24 sources, each with sustained bandwidth of 64 kbps could share a DS1 (1.544 Mbps) link. Statistical aggregation allows sources of data to burst to bandwidth higher than their sustained rate, based on availability of bandwidth capacity of the link. Similarly, STS-1 (51.840 Mbps) signals from three networks can be received and combined into an OC-3 (155.520 Mbps) signal. The OC-3 signal can be transmitted to another switch for routing and/or further aggregation.
In one embodiment, the switches of figure 1 include multiple cards that are interconnected by a switching fabric. In one embodiment, the cards have both an ingress data path and an egress data path.
The ingress data path is used to receive data from the network and transmit the data to an appropriate card within the switch. The ingress data path schedules transmission of data across the switching fabric.
The egress data path is used to receive data from the switching fabric and transmit the data across the network. The egress data path schedules transmission of data out of the switch across the network. The ingress and egress data paths interact to prevent overflow of data within the network switch. Figure 2 illustrates one embodiment of an interconnection of cards within a network switch. The switch of Figure 2 can be, for example, any of switches 110, 120 or 130 of Figure 1. The switch of Figure 2 is illustrated with four ingress cards and four egress cards for reasons of simplicity only. A switch can have any number of ingress cards and any number of egress cards. Also, data flow can be bi-directional. That is, the cards can also provide both egress and ingress functionality.
Each ingress card includes an ingress buffer that receives data from an external source (not shown in Figure 2). Data can be in any format, for example, IP packets or ATM cells. The ingress buffers are coupled to ingress schedulers. The ingress schedulers dispatch data to egress cards via a set of ingress ports. In one embodiment, each ingress card has a port for each egress card to which the ingress card is coupled. For example, ingress card 0 is coupled to egress card 0 through port 0, to egress card 1 through port 1 , to egress card 2 through port 2, and to egress card 3 through port 3.
In one embodiment, each ingress card is coupled to each egress card, the interconnection between the ingress cards and the egress cards has n2 connections where n is the number of ingress/egress cards. Thus, the interconnection is referred to as an " n2 mesh," or an " n2 switching fabric." In another embodiment, the number of ingress cards is not equal to the number of egress cards, which is referred to as a " nxm mesh." The mesh is described in greater detail in U.S. Patent application number 09/746,212, entitled "A FULL MESH INTERCONNECT BACKPLANE ARCHITECTURE," filed December 22, 2000, which is assigned to the corporate assignee of the present application and incorporated by reference.
In one embodiment, traffic crosses the mesh, or switching fabric, in an asynchronous manner in that no central clock signal drives data across the mesh. Data is transmitted by the ingress cards without reference to a bus or mesh clock or frame synchronization signal. A protocol for use in communicating over the mesh is described in greater detail in U.S. Patent application number 09/745,982, entitled "A BACKPLANE PROTOCOL," filed December 22, 2000, which is assigned to the corporate assignee of the present invention and incorporated by reference. Each egress card includes a port for each ingress card to which the ingress card is coupled. For example, egress card 0 is coupled to ingress card 0 through port 0, to ingress card 1 through port 1 , to ingress card 2 through port 2, and to ingress card 3 through port 3. The ports of the egress card are coupled to an egress buffer. The egress buffer is coupled to an egress scheduler that outputs data to a device external to the egress card (not shown in Figure 2).
The architecture illustrated in Figure 2 allows scheduling duties to be distributed between ingress and egress cards. Because the scheduling duties are distributed, a centralized scheduler is not required and transmission of data between ingress cards and egress cards can be accomplished in an asynchronous manner. This allows simpler control of data switching and more efficient use of switching fabric bandwidth.
When data is received by an ingress card the data is temporarily stored in the ingress buffer on the card. The ingress scheduler extracts data from the ingress buffer and sends the data to the appropriate port. For example data to be transmitted to egress card 2 are sent to port 2. In one embodiment, the ingress scheduler reads and sends data according to an earliest deadline first scheduling scheme. In alternate embodiments, strict priority scheduling, round robin scheduling, weighted round robin scheduling, or other scheduling techniques can be used.
Data that is transferred between ingress and egress cards can be variable in size. The data can be transmitted as a group of fixed length cells or as one or more variable length packets. In one embodiment, the packets on the ingress side compete with each other on a packet basis. Each packet competes against the other packets to be selected by the ingress scheduler. In one embodiment, when one packet is selected, all of the entire packet is moved across the switch fabric. Once the ingress scheduler selects a packet of a given priority, the packet is transmitted before another packet is selected.
Data must be transferred from the ingress side of the switch to the egress side of the switch through the switching fabric. In an n2 mesh, n ingress sources can potentially contend for a single egress destination. The switch is required to transfer the data from ingress to egress such that the contracts of the individual data flows are honored. The contracts specify bandwidth, latency, jitter, loss and burst tolerance. All contracts must be honored under all traffic contention conditions.
In one embodiment the ingress scheduler provides all cell/packet discard functionality. Cell/packet discard can include, for example, early packet discard (EPD), partial packet discard (PPD), random early discard (RED), each of which is known in the art. Additional and/or different cell/packet discard procedures can also be used.
In one embodiment, the ingress schedulers independently schedule packets and cells to the egress side. To allow this independent scheduling to function, n separate buffers are provided on each egress, one per ingress. These independent cache buffers have sufficient bandwidth to allow simultaneous egress-side arrivals from all ingress devices. In one embodiment, the egress side sends "backpressure" messages to the ingress control access to the n independent cache buffers. Data in the n buffers is transferred to a larger egress buffer, from where it is scheduled to the egress ports.
The egress buffer receives data from the ports of the egress card in a predetermined manner. For example, data can be extracted from the ports in a round robin fashion, or data can be extracted from the ports on a priority basis. Also, a combination of round robin and priority-based extraction can also be used.
Data received from the egress card ports in stored in the egress buffer. In one embodiment, the egress buffer includes a cache for each link (i.e., link between ingress card 2 and egress card 2, link between ingress card 3 and egress card 2). Each cache includes a queue for each class of data. By including a queue for each class of data, the egress buffer can provide quality of service functionality.
The egress scheduler extracts data from the egress buffer and transmits the data according to the appropriate network protocol to an external device (not shown in Figure 2). In one embodiment, the egress scheduler extracts data from the egress buffer based on priority to provide quality of service functionality. In alternate embodiments, the egress scheduler can extract data from the egress buffer using earliest deadline first scheduling. Figure 3 conceptually illustrates one embodiment of an ingress scheduler. In one embodiment, incoming data is categorized into one of three ingress traffic categories (ITCs). Having only three ITCs simplifies the architecture of the ingress scheduler. One embodiment of a mapping of ITCs to ATM and IP traffic is set forth in the following table. Other types of network traffic can be mapped to the three ITCs in a similar manner.
Figure imgf000010_0001
Ingress Traffic Category Mapping
In one embodiment, servicing of the three ITCs is accomplished according to the following priority: 1) RT, 2) MC, and 3) NRT, assuming no backpressure signals are active. If backpressure signals are active, the transmission of the corresponding category of data is stopped to avoid egress port buffer overflow. Lower priority data can be transmitted when the backpressure signal is active for higher priority data.
Ingress router 300 reads data from the ingress buffer (not shown in Figure 3) and sends the data to the appropriate queue based on the ITC mapping described above. Real time data is sent to one of the RT group queues (e.g., 305, 310), multicast data is sent to MC queue 320, and non-real time data is sent to one of the NRT queues (e.g., 330, 335). In one embodiment, the ingress scheduler includes 16 RT queues; however, any number of RT queues can be provided.
In one embodiment, data is read out of the RT group queues in a round robin fashion. In another embodiment, the data is selected from the multiple RT group queues on a priority basis. Similarly, data is read out of the NRT class queues in a round robin fashion. The data is selected from the multiple NRT class queues in a weighted round robin fashion. Data from the three categories of data (RT, MC, NRT) is selected based on a priority basis to be sent to the appropriate ingress port 390.
Data flow control is described in greater detail in U.S. Patent application number 09/812,985 (Atty. Docket No. P017) filed March 19, 2001, entitled "METHOD AND SYSTEM FOR SWITCH FABRIC FLOW CONTROL," which is assigned to the corporate assignee of the present U.S. Patent application and incorporated by reference herein.
Figure 4 conceptually illustrates one embodiment of a egress cache scheduling of egress card ports. In one embodiment, each egress card port has three associated FIFO buffers for real time data, multicast data and non-real time data, respectively.
In one embodiment, when data is received by egress port 400 the data is sent to one of three queues. The queues correspond to the ITCs. Real time data is stored in real time queue 410, multicast data is stored in multicast queue 420, and non- real time data is stored in non-real time queue 430. Data is read out of the queues in a round robin fashion. The queue from which data is transmitted is selected on a priority basis.
In the foregoing specification, the invention has been described with reference to specific embodiments thereof. It will, however, be evident that various modifications and changes can be made thereto without departing from the broader spirit and scope of the invention. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.

Claims

CLAIMSWhat is claimed is:
1. A network switch having an asynchronous mesh to transfer data from ingress interfaces to egress interfaces, the ingress interfaces to receive data from external sources and to selectively transmit the data across the asynchronous mesh to the egress interfaces, the egress interfaces to receive data from the asynchronous mesh and to transmit the data to external destinations.
2. The network switch of claim 1 wherein the ingress interfaces schedule respective data transmissions across the mesh and the egress interfaces schedule respective data transmissions to the external destinations.
3. The network switch of claim 2 comprising N ingress interfaces, each of the egress interfaces further comprising N independent cache buffers coupled to N respective ingress interfaces to receive data from the respective N ingress interfaces.
4. The network switch of claim 2 comprising N ingress interfaces, each of the N ingress interfaces having N independent cache buffers, each of the N independent cache buffers coupled to one of N respective egress interfaces.
5. The network switch of claim 2 wherein one or more of the N ingress interfaces segregates incoming data into queues based on one or more of: a flow identifier, a user identifier, a session identifier, a quality of service (QoS), a priority, a deadline, and a service class.
6. The network switch of claim 3 in which the egress interfaces generate a flow control signal to prevent access to one or more of the N buffers of the respective egress interfaces.
7. The network switch of claim 3 wherein the egress interfaces generate a flow control signal to prevent transmission to one or more of the N buffers of the respective egress interfaces.
8. The network switch of claim 3 wherein the N ingress interfaces transfer data to a shared egress buffer and further wherein the egress interfaces schedule and retrieve the data stored in the shared egress buffer prior to transmitting the data to the external destinations.
9. The network switch of claim 5 in which the egress interfaces generate a flow control signal to prevent access by one or more of the queues at the ingress interfaces to the egress buffer.
10. The network switch of claim 3 in which the N ingress interfaces concurrently transmit fixed-length cells and variable-length packets across the mesh to the egress interfaces.
1 1. A network switch comprising: a plurality of ingress cards, the plurality of ingress cards having an ingress buffer to temporarily store data, an ingress scheduler coupled to the ingress buffer, and a plurality of ports coupled to the ingress scheduler, the ingress scheduler to read data from the ingress buffer and to selectively transfer the data to one of the plurality of ports; and a plurality of egress cards, the plurality of egress cards having a plurality of ports coupled to receive data from respective ingress card ports, an egress buffer coupled to the plurality of ports, the egress buffer to selectively read data from the plurality of ports and to store the data, and an egress scheduler coupled to the egress buffer, the egress scheduler to read data from the egress buffer and to transmit data from the egress card.
12. The network switch of claim 11 wherein the ingress scheduler transfers data from the ingress buffer in a first in/first out (FIFO) manner.
13. The network switch of claim 1 1 wherein the ingress scheduler transfers data from the ingress buffer according to priorities associated with the data.
14. The network switch of claim 1 1 wherein the ingress buffer receives telecommunications data.
15. The network switch of claim 11 wherein the plurality of ports of the egress cards further comprise one or more buffers to temporarily store data received from the respective ingress cards, and further wherein if a buffer of a port of an egress card is full the buffer and the port refuse data transmitted from the associated ingress card.
16. The network switch of claim 15 wherein the associated ingress card from which data was refused retransmits the refused data until the associated egress port and buffer accept the previously refused data.
17. The network switch of claim 11 wherein the egress buffers include a data store for each of the plurality of ports of the egress card.
18. The network switch of claim 17 wherein the data store for each of the plurality of ports of the egress card stores the data according to an associated class.
19. A network switch comprising:
N ingress cards coupled to receive data from external sources, the N ingress cards having a plurality of ports to transmit data, wherein each of the N ingress cards comprises an ingress scheduler coupled to the ports of the ingress card, the ingress scheduler to cause data to be selectively and asynchronously transmitted via the ports of the ingress card; and
M egress cards having ports coupled to receive data from one or more of the plurality of ports of the N ingress cards, the egress cards coupled to transmit data to external destinations, wherein each of the M egress cards comprises an egress scheduler coupled to the ports of the egress card, the egress scheduler to cause data to be selectively transmitted to the external destinations.
20. The network switch of claim 19 wherein N and M are equal.
21. The network switch of claim 19 wherein one or more of the ingress interfaces segregates incoming data into queues based on one or more of: a flow identifier, a user identifier, a session identifier, a quality of service (QoS), a priority, a deadline, and a service class.
PCT/US2002/016288 2001-05-31 2002-05-22 Distributed control of data flow in a network switch WO2002098064A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US09/872,125 US7006438B2 (en) 2001-05-31 2001-05-31 Distributed control of data flow in a network switch
US09/872,125 2001-05-31

Publications (1)

Publication Number Publication Date
WO2002098064A1 true WO2002098064A1 (en) 2002-12-05

Family

ID=25358893

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2002/016288 WO2002098064A1 (en) 2001-05-31 2002-05-22 Distributed control of data flow in a network switch

Country Status (2)

Country Link
US (1) US7006438B2 (en)
WO (1) WO2002098064A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1592182A1 (en) * 2003-02-03 2005-11-02 Nippon Telegraph and Telephone Corporation Data transmission device and data transmission system
CN111351669A (en) * 2018-12-20 2020-06-30 比亚迪股份有限公司 Fault diagnosis method and device, storage medium and vehicle

Families Citing this family (52)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7310348B2 (en) * 2001-09-19 2007-12-18 Bay Microsystems, Inc. Network processor architecture
US7349403B2 (en) * 2001-09-19 2008-03-25 Bay Microsystems, Inc. Differentiated services for a network processor
US7039061B2 (en) * 2001-09-25 2006-05-02 Intel Corporation Methods and apparatus for retaining packet order in systems utilizing multiple transmit queues
US7248593B2 (en) * 2001-09-25 2007-07-24 Intel Corporation Method and apparatus for minimizing spinlocks and retaining packet order in systems utilizing multiple transmit queues
US20030079018A1 (en) * 2001-09-28 2003-04-24 Lolayekar Santosh C. Load balancing in a storage network
US20030084219A1 (en) * 2001-10-26 2003-05-01 Maxxan Systems, Inc. System, apparatus and method for address forwarding for a computer network
US7145914B2 (en) * 2001-12-31 2006-12-05 Maxxan Systems, Incorporated System and method for controlling data paths of a network processor subsystem
US7085846B2 (en) * 2001-12-31 2006-08-01 Maxxan Systems, Incorporated Buffer to buffer credit flow control for computer network
CA2369432A1 (en) * 2002-01-24 2003-07-24 Alcatel Canada Inc. System and method for reassembling packets in a network element
US7406038B1 (en) 2002-04-05 2008-07-29 Ciphermax, Incorporated System and method for expansion of computer network switching system without disruption thereof
US7295561B1 (en) 2002-04-05 2007-11-13 Ciphermax, Inc. Fibre channel implementation using network processors
US20030195956A1 (en) * 2002-04-15 2003-10-16 Maxxan Systems, Inc. System and method for allocating unique zone membership
US20030200330A1 (en) * 2002-04-22 2003-10-23 Maxxan Systems, Inc. System and method for load-sharing computer network switch
US20030202510A1 (en) * 2002-04-26 2003-10-30 Maxxan Systems, Inc. System and method for scalable switch fabric for computer network
US7408876B1 (en) * 2002-07-02 2008-08-05 Extreme Networks Method and apparatus for providing quality of service across a switched backplane between egress queue managers
US7286552B1 (en) 2002-07-26 2007-10-23 Extreme Networks Method and apparatus for providing quality of service across a switched backplane for multicast packets
US7599292B1 (en) 2002-08-05 2009-10-06 Extreme Networks Method and apparatus for providing quality of service across a switched backplane between egress and ingress queue managers
US20040030766A1 (en) * 2002-08-12 2004-02-12 Michael Witkowski Method and apparatus for switch fabric configuration
KR100499277B1 (en) * 2002-11-18 2005-07-01 한국전자통신연구원 Apparatus and method for controlling flows of cell data
US7830889B1 (en) * 2003-02-06 2010-11-09 Juniper Networks, Inc. Systems for scheduling the transmission of data in a network device
US8451718B2 (en) * 2003-06-27 2013-05-28 Broadcom Corporation Flow control and congestion control in a data distribution network
US7385984B2 (en) * 2004-03-30 2008-06-10 Extreme Networks, Inc. Packet processing system architecture and method
US7613209B1 (en) * 2004-03-30 2009-11-03 Extreme Networks, Inc. System and method for egress packet marking
US7292591B2 (en) 2004-03-30 2007-11-06 Extreme Networks, Inc. Packet processing system architecture and method
US8161270B1 (en) 2004-03-30 2012-04-17 Extreme Networks, Inc. Packet data modification processor
US7936770B1 (en) * 2005-03-08 2011-05-03 Enterasys Networks, Inc. Method and apparatus of virtual class of service and logical queue representation through network traffic distribution over multiple port interfaces
JP2007259073A (en) * 2006-03-23 2007-10-04 Nec Corp Scheduling apparatus in packet switch
US7844761B1 (en) * 2006-05-10 2010-11-30 Altera Corporation Flexible on-chip datapath interface for facilitating communication between first and second interfaces with different interface properties
US7948883B1 (en) * 2007-07-12 2011-05-24 Cisco Technology, Inc. Applying router quality of service on a cable modem interface on a per-service-flow basis
CN101355430B (en) * 2007-07-27 2012-02-29 华为技术有限公司 Exchange frame, cluster router
US7852486B2 (en) * 2008-02-07 2010-12-14 Board Of Regents, The University Of Texas System Wavelength and intensity monitoring of optical cavity
US7936669B2 (en) 2008-06-04 2011-05-03 Entropic Communications, Inc. Systems and methods for flow control and quality of service
US8837462B2 (en) * 2008-12-15 2014-09-16 Embraer S.A. Switch usage for routing ethernet-based aircraft data buses in avionics systems
US8811413B2 (en) * 2009-06-01 2014-08-19 Infinera Corporation Scalable multi-hop mesh packet switch fabric
US8605732B2 (en) 2011-02-15 2013-12-10 Extreme Networks, Inc. Method of providing virtual router functionality
US8707454B1 (en) 2012-07-16 2014-04-22 Wickr Inc. Multi party messaging
US9830089B1 (en) 2013-06-25 2017-11-28 Wickr Inc. Digital data sanitization
US10567349B2 (en) 2013-06-25 2020-02-18 Wickr Inc. Secure time-to-live
US9866591B1 (en) 2013-06-25 2018-01-09 Wickr Inc. Enterprise messaging platform
US10129260B1 (en) 2013-06-25 2018-11-13 Wickr Inc. Mutual privacy management
US20150058441A1 (en) * 2013-08-20 2015-02-26 Samsung Electronics Co., Ltd. Efficient content caching management method for wireless networks
US9419902B1 (en) 2013-12-30 2016-08-16 Google Inc. Method and system for network micro flow control
US9800509B2 (en) * 2014-02-20 2017-10-24 Uvic Industry Partnerships Inc. System and method for efficient transport of large data files
US9698976B1 (en) 2014-02-24 2017-07-04 Wickr Inc. Key management and dynamic perfect forward secrecy
US9600191B2 (en) * 2014-06-02 2017-03-21 Micron Technology, Inc. Systems and methods for reordering packet transmissions in a scalable memory system protocol
US9584530B1 (en) 2014-06-27 2017-02-28 Wickr Inc. In-band identity verification and man-in-the-middle defense
US9654288B1 (en) 2014-12-11 2017-05-16 Wickr Inc. Securing group communications
US9584493B1 (en) 2015-12-18 2017-02-28 Wickr Inc. Decentralized authoritative messaging
US10291607B1 (en) 2016-02-02 2019-05-14 Wickr Inc. Providing real-time events to applications
US9596079B1 (en) 2016-04-14 2017-03-14 Wickr Inc. Secure telecommunications
US9602477B1 (en) 2016-04-14 2017-03-21 Wickr Inc. Secure file transfer
GB2612619B (en) * 2021-11-05 2024-02-21 Graphcore Ltd Fair arbitration between multiple sources targeting a destination

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5889778A (en) * 1995-07-17 1999-03-30 Pmc-Sierra Ltd. ATM layer device
US6151300A (en) * 1996-05-10 2000-11-21 Fujitsu Network Communications, Inc. Method and apparatus for enabling flow control over multiple networks having disparate flow control capability
US6185209B1 (en) * 1997-07-11 2001-02-06 Telefonaktiebolaget Lm Ericsson VC merging for ATM switch
US6201809B1 (en) * 1998-05-28 2001-03-13 3Com Corporation Port redundancy and backpressure translation table apparatus

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5812527A (en) * 1996-04-01 1998-09-22 Motorola Inc. Simplified calculation of cell transmission rates in a cell based netwook
US6130889A (en) * 1996-10-02 2000-10-10 International Business Machines Corporation Determining and maintaining hop-count for switched networks
GB2322515A (en) * 1997-02-21 1998-08-26 Northern Telecom Ltd Adaptation layer switching
US6058116A (en) * 1998-04-15 2000-05-02 3Com Corporation Interconnected trunk cluster arrangement
WO2000003516A1 (en) * 1998-07-08 2000-01-20 Broadcom Corporation Network switching architecture with multiple table synchronization, and forwarding of both ip and ipx packets
ATE290743T1 (en) * 1998-07-31 2005-03-15 Cit Alcatel METHOD, SEQUENCER, INTELLIGENT BUFFER MEMORY, PROCESSOR AND TELECOMMUNICATIONS SYSTEM FOR DISTRIBUTING AVAILABLE BANDWIDTH
US6411617B1 (en) * 1998-12-10 2002-06-25 Nokia Telecommunications, Oy System and method for managing data traffic associated with various quality of service principles using a conventional network node switch
US6198725B1 (en) * 1999-03-16 2001-03-06 Fujitsu Network Communications, Inc. Network element delay allocation for efficient use of network resources
US6449255B1 (en) * 1999-04-26 2002-09-10 Cisco Technology, Inc. Method and apparatus for managing packets using a real-time feedback signal
US20020085567A1 (en) * 2000-12-28 2002-07-04 Maple Optical Systems, Inc. Metro switch and method for transporting data configured according to multiple different formats

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5889778A (en) * 1995-07-17 1999-03-30 Pmc-Sierra Ltd. ATM layer device
US6151300A (en) * 1996-05-10 2000-11-21 Fujitsu Network Communications, Inc. Method and apparatus for enabling flow control over multiple networks having disparate flow control capability
US6185209B1 (en) * 1997-07-11 2001-02-06 Telefonaktiebolaget Lm Ericsson VC merging for ATM switch
US6201809B1 (en) * 1998-05-28 2001-03-13 3Com Corporation Port redundancy and backpressure translation table apparatus

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1592182A1 (en) * 2003-02-03 2005-11-02 Nippon Telegraph and Telephone Corporation Data transmission device and data transmission system
EP1592182A4 (en) * 2003-02-03 2010-05-12 Nippon Telegraph & Telephone Data transmission device and data transmission system
US7797444B2 (en) 2003-02-03 2010-09-14 Nippon Telegraph And Telephone Corporation Data transfer apparatus and data transfer system
CN111351669A (en) * 2018-12-20 2020-06-30 比亚迪股份有限公司 Fault diagnosis method and device, storage medium and vehicle
CN111351669B (en) * 2018-12-20 2022-03-15 比亚迪股份有限公司 Fault diagnosis method and device, storage medium and vehicle

Also Published As

Publication number Publication date
US7006438B2 (en) 2006-02-28
US20020186703A1 (en) 2002-12-12

Similar Documents

Publication Publication Date Title
US7006438B2 (en) Distributed control of data flow in a network switch
Chiussi et al. Low-cost scalable switching solutions for broadband networking: the ATLANTA architecture and chipset
EP1093266B1 (en) Telecommunications switch and method
US6714517B1 (en) Method and apparatus for interconnection of packet switches with guaranteed bandwidth
JP3814393B2 (en) Cell scheduling method and apparatus
US6654374B1 (en) Method and apparatus to reduce Jitter in packet switched networks
CA2224753C (en) An atm switch queuing system
US6970424B2 (en) Method and apparatus to minimize congestion in a packet switched network
US5557607A (en) Methods and apparatus for enqueueing and dequeueing data cells in an ATM switch fabric architecture
US7242686B1 (en) System and method for communicating TDM traffic through a packet switch fabric
US7940786B2 (en) System and method for communicating data using a common switch fabric
US6151301A (en) ATM architecture and switching element
KR100328642B1 (en) Arrangement and method relating to packet flow control
US20110019544A1 (en) Systems for scheduling the transmission of data in a network device
JP2001217836A (en) Weighted round robin engine used in case of scheduing distribution of atm cell
JPH10285187A (en) Distributed buffering system for atm switch
EP0977404A2 (en) Scalable scheduled cell switch and method for switching
US6947380B1 (en) Guaranteed bandwidth mechanism for a terabit multiservice switch
AU736780B2 (en) Method for providing delays independent of switch size in a crossbar switch with speedup
US7450503B1 (en) System and method to multicast guaranteed and best-effort traffic in a communications network
US7068672B1 (en) Asynchronous receive and transmit packet crosspoint
Briem et al. Traffic management for an ATM switch with per-VC queuing: Concept and implementation
Smiljanic Rate and delay guarantees provided by clos packet switches with load balancing
Smiljanic Scheduling of multicast traffic in high-capacity packet switches
Chiussi et al. Providing QoS guarantees in packet switches

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ OM PH PL PT RO RU SD SE SG SI SK SL TJ TM TN TR TT TZ UA UG US UZ VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase

Ref country code: JP

WWW Wipo information: withdrawn in national office

Country of ref document: JP