WO2014200707A2 - Ingress switch multicast distribution in a fabric switch - Google Patents
Ingress switch multicast distribution in a fabric switch Download PDFInfo
- Publication number
- WO2014200707A2 WO2014200707A2 PCT/US2014/040060 US2014040060W WO2014200707A2 WO 2014200707 A2 WO2014200707 A2 WO 2014200707A2 US 2014040060 W US2014040060 W US 2014040060W WO 2014200707 A2 WO2014200707 A2 WO 2014200707A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- switch
- multicast
- identifier
- packet
- fabric
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/02—Details
- H04L12/16—Arrangements for providing special services to substations
- H04L12/18—Arrangements for providing special services to substations for broadcast or conference, e.g. multicast
- H04L12/185—Arrangements for providing special services to substations for broadcast or conference, e.g. multicast with management of multicast group membership
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/02—Details
- H04L12/16—Arrangements for providing special services to substations
- H04L12/18—Arrangements for providing special services to substations for broadcast or conference, e.g. multicast
- H04L12/1886—Arrangements for providing special services to substations for broadcast or conference, e.g. multicast with traffic restrictions for efficiency improvement, e.g. involving subnets or subdomains
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L45/00—Routing or path finding of packets in data switching networks
- H04L45/16—Multipoint routing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L49/00—Packet switching elements
- H04L49/20—Support for services
- H04L49/201—Multicast operation; Broadcast operation
Definitions
- the present disclosure relates to communication networks. More specifically, the present disclosure relates to multicast distribution in a fabric switch.
- a fabric switch is a collection of individual member switches. These member switches form a single, logical switch that can have an arbitrary number of ports and an arbitrary topology. As demands grow, customers can adopt a "pay as you grow" approach to scale up the capacity of the fabric switch.
- TRILL Transparent Interconnection of Lots of Links
- the switch includes an inter-switch multicast module and an edge multicast module.
- the inter-switch multicast module identifies for a first replication of a multicast packet an egress inter- switch port in a multicast tree rooted at the switch.
- the multicast tree is identified by an identifier of the switch.
- the edge multicast module identifies an egress edge port for a second replication of the multicast packet based on a multicast group identifier.
- the multicast group identifier is local within the switch.
- the inter-switch multicast module identifies the inter-switch port based on a bit value corresponding to the inter-switch port.
- the bit value is in an inter-switch bitmap associated with the multicast tree.
- the inter-switch bitmap is included in an entry in a multicast switch identifier table.
- the entry in the multicast switch identifier table corresponds to the identifier of the switch.
- the switch also includes a selection module which selects the multicast switch identifier table from a plurality of multicast switch identifier table instances based on a multicast group of the multicast packet.
- a respective multicast switch identifier table instance is associated with a respective multicast group.
- the edge multicast module identifies the edge port based on a bit value corresponding to the edge port.
- the bit value is in an edge bitmap associated with the multicast group identifier.
- the edge bitmap is included in an entry in a multicast group identifier table.
- the entry in the multicast switch identifier table corresponds to the multicast group identifier.
- the multicast group identifier is mapped to a virtual local area network (VLAN) identifier of the multicast packet in a mapping table.
- VLAN virtual local area network
- the switch also includes a fabric switch management module which maintains a membership in a fabric switch.
- the fabric switch accommodates a plurality of switches and operates as a single switch.
- the first replication of the multicast packet is encapsulated in a fabric encapsulation of the fabric switch.
- the inter-switch multicast module also identifies for a third replication of the multicast packet an egress inter-switch port in a second multicast tree rooted at a second switch. This second multicast tree is identified by an identifier of the second switch.
- the edge multicast module also determines whether the multicast group identifier is associated with the multicast packet based on a VLAN identifier of the multicast packet.
- FIG. 1A illustrates an exemplary fabric switch with ingress switch multicast tree support, in accordance with an embodiment of the present invention.
- FIG. IB illustrates exemplary ingress switch multicast trees in a fabric switch, in accordance with an embodiment of the present invention.
- FIG. 2 illustrates exemplary packet headers for multicast traffic distribution via an ingress switch multicast tree in a fabric switch, in accordance with an embodiment of the present invention.
- FIG. 3A illustrates an exemplary replication of multicast traffic received from an edge port via an ingress switch multicast tree in a fabric switch, in accordance with an embodiment of the present invention.
- FIG. 3B illustrates an exemplary replication of multicast traffic received from an inter-switch (IS) port via an ingress switch multicast tree in a fabric switch, in accordance with an embodiment of the present invention.
- IS inter-switch
- FIG. 3C illustrates an exemplary replication of multicast traffic via one of multiple ingress switch multicast tree instances in a fabric switch, in accordance with an embodiment of the present invention.
- FIG. 4A presents a flowchart illustrating the process of a switch in a fabric switch forwarding multicast traffic received from an edge port via an ingress switch multicast tree, in accordance with an embodiment of the present invention.
- FIG. 4B presents a flowchart illustrating the process of a switch in a fabric switch forwarding multicast traffic received from an IS port via an ingress switch multicast tree, in accordance with an embodiment of the present invention.
- FIG. 5 illustrates exemplary presence-based ingress switch multicast trees in a fabric switch, in accordance with an embodiment of the present invention.
- FIG. 6 illustrates an exemplary architecture of a switch with ingress switch multicast tree support, in accordance with an embodiment of the present invention.
- the problem of efficient multicast traffic distribution in a fabric switch is solved by facilitating a multicast distribution tree, which is referred to as an ingress switch multicast tree, at a respective member switch of the fabric switch.
- a multicast distribution tree which is referred to as an ingress switch multicast tree
- the switch Upon receiving a multicast packet, the switch forwards the packet via its own ingress switch multicast tree.
- a fabric switch has a finite number of multicast distribution trees.
- the member switches forward traffic belonging to all multicast groups using these trees to distribute traffic.
- the member switches can forward multicast traffic belonging to a large number of multicast groups via a single tree. This can congest the links in the multicast tree, leading to inefficient forwarding and greater delay. This problem can be further aggravated when this number is small.
- this finite number can typically be one (i.e., the fabric switch typically can have one multicast distribution tree). In that case, all member switches forward all multicast traffic via the same tree and cause the links of the tree to congest.
- a respective member switch computes its own ingress switch multicast tree and forwards multicast traffic via that tree.
- the multicast traffic load of the fabric switch becomes distributed among the ingress switch multicast trees of the corresponding member switches. Since different ingress switch multicast trees comprise different links of the fabric switch, the multicast traffic load becomes distributed across the links of the fabric switch instead of a few links of a finite number of trees.
- multicast packet replication is performed in two stages.
- the member switch replicates a multicast packet to its edge ports based on a multicast group identifier (MGID) representing the edge multicast replication of the switch.
- MGID multicast group identifier
- this MGID is local to the switch and operates as a local multicast replication identifier for the switch.
- this multicast group identifier is distinct from a multicast group address of a multicast packet, which is not local and specific to a multicast group.
- the switch replicates the packet to inter-switch (IS) ports for other member switches based on the egress switch identifier of the packet.
- IS inter-switch
- any number of switches coupled in an arbitrary topology may logically operate as a single switch.
- the fabric switch can be an Ethernet fabric switch or a virtual cluster switch (VCS), which can operate as a single Ethernet switch. Any member switch may join or leave the fabric switch in "plug-and-play" mode without any manual configuration.
- a respective switch in the fabric switch is a Transparent Interconnection of Lots of Links (TRILL) routing bridge (RBridge).
- TRILL Transparent Interconnection of Lots of Links
- a respective switch in the fabric switch is an Internet Protocol (IP) routing-capable switch (e.g., an IP router).
- IP Internet Protocol
- a fabric switch is not the same as conventional switch stacking.
- switch stacking multiple switches are interconnected at a common location (often within the same rack), based on a particular topology, and manually configured in a particular way. These stacked switches typically share a common address, e.g., an IP address, so they can be addressed as a single switch externally.
- switch stacking requires a significant amount of manual configuration of the ports and inter-switch links. The need for manual configuration prohibits switch stacking from being a viable option in building a large-scale switching system.
- the topology restriction imposed by switch stacking also limits the number of switches that can be stacked. This is because it is very difficult, if not impossible, to design a stack topology that allows the overall switch bandwidth to scale adequately with the number of switch units.
- a fabric switch can include an arbitrary number of switches with individual addresses, can be based on an arbitrary topology, and does not require extensive manual configuration.
- the switches can reside in the same location, or be distributed over different locations.
- the automatic and dynamic configurability of the fabric switch allows a network operator to build its switching system in a distributed and "pay-as-you-grow" fashion without sacrificing scalability.
- the fabric switch's ability to respond to changing network conditions makes it an ideal solution in a virtual computing environment, where network loads often change with time.
- the term "fabric switch” refers to a number of interconnected physical switches which form a single, scalable logical switch. These physical switches are referred to as member switches of the fabric switch. In a fabric switch, any number of switches can be connected in an arbitrary topology, and the entire group of switches functions together as one single, logical switch. This feature makes it possible to use many smaller, inexpensive switches to construct a large fabric switch, which can be viewed as a single logical switch externally.
- the present disclosure is presented using examples based on a fabric switch, embodiments of the present invention are not limited to a fabric switch. Embodiments of the present invention are relevant to any computing device that includes a plurality of devices operating as a single device.
- multicast is used in a generic sense, and can refer to any traffic forwarding toward a plurality of recipients. Any traffic forwarding that creates and forwards more than one copy of the same packet in a fabric switch can be a referred to as "multicast.”
- multicast traffic examples include, but are not limited to, broadcast, unknown unicast, and multicast traffic.
- end device can refer to any device external to a fabric switch.
- an end device examples include, but are not limited to, a host machine, a conventional layer-2 switch, a layer-3 router, or any other type of network device. Additionally, an end device can be coupled to other switches or hosts further away from a layer-2 or layer-3 network. An end device can also be an aggregation point for a number of network devices to enter the fabric switch.
- switch is used in a generic sense, and it can refer to any standalone or fabric switch operating in any network layer. "Switch” should not be interpreted as limiting embodiments of the present invention to layer-2 networks. Any device that can forward traffic to an external device or another switch can be referred to as a "switch.” Any physical or virtual device (e.g., a virtual machine/switch operating on a computing device) that can forward traffic to an end device can be referred to as a "switch.” Examples of a “switch” include, but are not limited to, a layer-2 switch, a layer-3 router, a TRILL RBridge, or a fabric switch comprising a plurality of similar or heterogeneous smaller physical and/or virtual switches.
- edge port refers to a port on a fabric switch which exchanges data frames with a network device outside of the fabric switch (i.e., an edge port is not used for exchanging data frames with another member switch of a fabric switch).
- inter-switch port refers to a port which sends/receives data frames among member switches of a fabric switch.
- interface and "port” are used interchangeably.
- switch identifier refers to a group of bits that can be used to identify a switch. Examples of a switch identifier include, but are not limited to, a media access control (MAC) address, an Internet Protocol (IP) address, and an RBridge identifier.
- MAC media access control
- IP Internet Protocol
- RBridge ID (RBridge identifier) to denote a 48-bit intermediate- system- to-intermediate-system (IS-IS) System ID assigned to an RBridge
- RBridge nickname to denote a 16-bit value that serves as an abbreviation for the "RBridge ID.”
- switch identifier is used as a generic term, is not limited to any bit format, and can refer to any format that can identify a switch.
- RBridge identifier is also used in a generic sense, is not limited to any bit format, and can refer to "RBridge ID,” “RBridge nickname,” or any other format that can identify an RBridge.
- Packet refers to a group of bits that can be transported together across a network. "Packet” should not be interpreted as limiting embodiments of the present invention to layer-3 networks. "Packet” can be replaced by other terminologies referring to a group of bits, such as “message,” “frame,” “cell,” or “datagram.” Network Architecture
- FIG. 1A illustrates an exemplary fabric switch with ingress switch multicast tree support, in accordance with an embodiment of the present invention.
- a fabric switch 100 includes member switches 101, 102, 103, 104, and 105. Switches 103 and 105 are coupled to end devices 112 and 114, respectively.
- fabric switch 100 is a TRILL network and a respective member switch of fabric switch 100, such as switch 105, is a TRILL RBridge.
- fabric switch 100 is an IP network and a respective member switch of fabric switch 100, such as switch 105, is an IP-capable switch, which calculates and maintains a local IP routing table (e.g., a routing information base or RIB), and is capable of forwarding packets based on its IP addresses.
- a local IP routing table e.g., a routing information base or RIB
- Switches in fabric switch 100 use edge ports to communicate with end devices (e.g., non-member switches) and inter-switch ports to communicate with other member switches.
- switch 105 is coupled to end device 114 via an edge port and to switches 101, 102, and 104 via inter-switch ports and one or more links.
- Data communication via an edge port can be based on Ethernet and via an inter-switch port can be based on IP and/or TRILL protocol.
- control message exchange via inter-switch ports can be based on a different protocol (e.g., Internet Protocol (IP) or Fibre Channel (FC) protocol).
- IP Internet Protocol
- FC Fibre Channel
- TRILL networks by inventors Shunjia Yu, Nagarajan Venkatesan, Anoop Ghanwani, Phanidhar Koganti, Mythilikanth Raman, Rajiv Krishnamurthy, and Dilip Chatwani, the disclosure of which is incorporated herein in its entirety.
- switch 103 receives a multicast packet from end device 112. Switch 103 is then the ingress switch of fabric switch 100 for that multicast packet.
- fabric switch 100 has a finite number of multicast distribution trees.
- switch 101 Upon receiving the multicast packet, switch 103 forwards the packet to switch 101, which in turn, forwards that packet to switches 102, 104, and 105 via the tree. Similarly, upon receiving a multicast packet from end device 114, switch 105 forwards the packet to switch 101, which in turn, forwards that packet to switches 102, 103, and 104 via the tree.
- switch 105 Using the same tree to forward multicast traffic from different ingress switches can congest the links in the multicast tree, leading to inefficient forwarding and greater delay.
- a respective member switch of fabric switch 100 computes its own ingress switch multicast tree and forwards multicast traffic via that tree. For example, upon receiving a multicast packet, switch 103 forwards the packet via its ingress switch multicast tree. Similarly, upon receiving a multicast packet, switch 105 forwards the packet via its ingress switch multicast tree. As a result, these multicast packets become distributed in fabric switch 100 among the ingress switch multicast trees rooted at switches 103 and 105. Since different ingress switch multicast trees comprise different links of fabric switch 100, the multicast traffic load becomes distributed across the links of fabric switch 100 instead of a few links of a finite number of trees.
- FIG. IB illustrates exemplary ingress switch multicast trees in a fabric switch, in accordance with an embodiment of the present invention.
- switches 101, 102, 103, 104, and 105 compute ingress switch multicast trees 131, 132, 133, 134, and 135, respectively.
- trees 131, 132, 133, 134, and 135 have switches 101, 102, 103, 104, and 105 as their root node (RN), respectively.
- RN root node
- switch 103 receives a multicast packet from end device 112 and forwards the packet via ingress switch multicast tree 133.
- Ingress switch multicast tree 133 includes links 121, 122, 123, and 124.
- Switch 103 replicates the multicast packet and forwards the replicated packets via links 121, 122, and 123.
- switch 104 Upon receiving the replicated packet, switch 104 further replicates the packet and forwards the replicated packet via link 124.
- switch 104 is coupled to an end device 116, which is a receiver of the multicast group of the packet. Switch 104 then replicates the packet and forwards the packet via the edge port coupling end device 116.
- switch 105 upon receiving a multicast packet from end device 114, switch 105 forwards the packet via ingress switch multicast tree 135.
- Ingress switch multicast tree 135 includes links 124, 125, 126, and 123.
- Switch 105 replicates the multicast packet and forwards the replicated packets via links 124, 125, and 126.
- switch 104 Upon receiving the replicated packet, switch 104 further replicates the packet and forwards the replicated packet via link 123. If end device 116 is a receiver of the multicast group of the packet, switch 104 replicates the packet and forwards the packet via the edge port coupling end device 116.
- FIG. 2 illustrates exemplary packet headers for multicast traffic distribution via an ingress switch multicast tree in a fabric switch, in accordance with an embodiment of the present invention.
- the ingress switch multicast tree is ingress switch multicast tree 133.
- switch 103 receives a multicast packet 202 from end device 112.
- This packet includes a layer-2 header 210 (e.g., an Ethernet header) comprising a source MAC addresses, a destination MAC addresses, and a payload.
- This payload can include an IP packet.
- packet 202 can be a broadcast or unknown unicast packet as well.
- Switch 103 encapsulates packet 202 in a fabric encapsulation 212 to generate fabric-encapsulated packet 204.
- fabric encapsulation 212 include, but are not limited to, TRILL encapsulation and IP encapsulation.
- Fabric encapsulation 212 includes the identifier of switch 103 as both the ingress and the egress identifier. Examples of a switch identifier include, but are not limited to, an RBridge identifier, an IP version 4 address, and an IP version 6 address.
- Examples of fabric-encapsulated packet 204 include, but are not limited to, a TRILL frame and an IP packet.
- Fabric encapsulation 212 can also include an outer layer-2 header comprising an all-MAC address 220, which indicates that this packet is destined to all recipients in fabric switch 100. The outer layer-2 header also includes the MAC address of switch 103 as the source MAC address.
- Switch 103 forwards fabric-encapsulated packet 204 via ingress switch multicast tree 133.
- Switches 101, 102, and 104 receive fabric-encapsulated packet 204, identify all-MAC address 220, and determine that this packet is a multicast packet.
- Switches 101, 102, and 104 also identify the identifier of switch 103 as egress switch identifier (which is also the ingress switch identifier), and recognize that the packet should be forwarded via ingress switch multicast tree 133 of switch 103.
- Switches 101 and 102 identify themselves as leaf nodes of ingress switch multicast tree 133.
- switch 101 is coupled to end device 222, which is a recipient of packet 202.
- Switch 101 then removes fabric encapsulation 212, replicates inner packet 202, and forwards packet 202 to end device 222 via the corresponding edge port.
- switch 104 detects that it is coupled to another downstream switch of ingress switch multicast tree 133. Switch 104 then replicates fabric-encapsulated packet 204 to generate fabric-encapsulated packet 206. However, because switch 104 is forwarding the packet, switch 104 changes the source MAC address of the outer layer-2 header to the MAC address of switch 104 to generate fabric encapsulation 214, and forwards fabric- encapsulated packet 206 to switch 105.
- end device 116 is a recipient of packet 202.
- Switch 104 then also removes fabric encapsulation 212, replicates inner packet 202, and forwards packet 202 to end device 116 via the corresponding edge port.
- Switch 105 receives fabric-encapsulated packet 206, identifies all-MAC address
- Switch 105 also identifies the identifier of switch 103 as the egress switch identifier (which is also the ingress switch identifier), and recognizes that the packet should be forwarded via ingress switch multicast tree 133 of switch 103. Switch 105 identifies itself as a leaf node of ingress switch multicast tree 133.
- Switch 105 then removes fabric encapsulation 214, replicates inner packet 202, and forwards packet 202 to end device 114 via the corresponding edge port.
- switch 104 replicates a multicast packet via an edge port
- multicast packet replication is performed in two stages. In the first stage, switch 104 replicates a multicast packet to its edge ports based on an MGID representing the edge multicast replication of switch 104. In some embodiments, this MGID is local to switch 104 (i.e., is not included in a packet and unique only within switch 104) and operates as a local multicast replication identifier for switch 104. In the second stage, switch 104 replicates the packet to IS ports for other member switches based on the egress switch identifier of the packet. In the example in FIG. 2, the IS port is for switch 105 and the egress switch identifier is the identifier of switch 103.
- FIG. 3A illustrates an exemplary replication of multicast traffic received from an edge port via an ingress switch multicast tree in a fabric switch, in accordance with an
- mapping table 302. maps an MGID with a virtual local area network (VLAN) identifier (VLAN ID).
- VLAN ID virtual local area network
- This VLAN identifier can be in the header of a packet (e.g., packet 202) and/or in the header of an
- mapping table 302 can also map one or more fields of a packet to an MGID, such as source and/or destination IP addresses, source and/or destination MAC addresses, source and/or destination ports, and a service and/or client VLANs.
- a respective entry in mapping table 302 can include the mapping or can be indexed based on VLAN identifiers.
- Mapping table 302 includes mapping for MGIDs 312-1, 312-2, ..., 312-w. The switch uses the VLAN identifier of the packet to obtain the corresponding MGID from mapping table 302.
- the switch uses this MGID to obtain an edge port bitmap from MGID table 304.
- the edge port bitmap represents the edge ports to which the packet should be replicated.
- An edge port is represented by a bit in the edge port bitmap, and a set (or unset) bit can indicate that the packet should be replicated and forwarded via the corresponding edge port. For example, a bitmap of "11000" can indicate that a packet should be replicated to the first two edge ports of the switch. It should be noted that the length of the edge port bitmap (i.e., the number of bits in the bitmap) can be equal to or greater than the number of edge ports of the switch.
- MGID table 304 includes edge port bitmaps 314-1, 314-2, ..., 314-m. It should be noted that m and n can be different. A respective entry in MGID table 304 can include a mapping between an edge port bitmap and an MGID, or can be indexed based on MGIDs. Upon obtaining an edge port bitmap, the switch replicates and forwards the packet via the edge ports indicated by the bitmap.
- the switch uses its local switch identifier (e.g., an RBridge identifier or an IP address) to obtain an IS port bitmap from a multicast switch identifier (MSID) table 306.
- the IS port bitmap represents the IS ports to which the packet should be replicated.
- An IS port is represented by a bit in the IS port bitmap, and a set (or unset) bit can indicate that the packet should be replicated and forwarded via the corresponding IS port. For example, a bitmap of "11000" can indicate that a packet should be replicated to the first two IS ports of the switch.
- the length of the IS port bitmap can be equal to or greater than the number of switches in a relevant network (e.g., in a fabric switch).
- MSID table 306 includes IS port bitmaps 316-1, 316-2, ..., 316-fc. It should be noted that each of m, n, and k can be different.
- a respective entry in MSID table 306 can include a mapping between an IS port bitmap and a switch identifier, or can be indexed based on the egress switch identifiers. Because the ingress and egress switch identifiers of a fabric encapsulation identify the root node of an ingress switch multicast tree, such indexing leads to the IS port bitmap corresponding to that ingress switch multicast tree.
- the switch Upon obtaining an IS port bitmap, the switch encapsulates the packet in fabric encapsulation, replicates the fabric-encapsulated packet, and forwards the fabric- encapsulated packets via the IS ports indicated by the bitmap. It should be noted that if a switch has no IS port to which the fabric-encapsulated packet should be replicated, a respective bit in the corresponding IS port bitmap can be unset (or set).
- switch 103 receives a multicast packet 202 via an edge port.
- Switch 103 obtain an MGID from its mapping table based on a VLAN identifier of packet 202, and obtains an edge port bitmap from its MGID table based on the MGID. Based on this edge port bitmap, switch 103 determines whether packet 202 should be replicated to any other local edge ports.
- Switch 103 also obtains an IS port bitmap from its MSID table based on the identifier of switch 103. Based on the IS port bitmap, switch 103 determines that fabric - encapsulated packet 204 should be replicated to the local IS ports which couple switches 101, 102, and 104.
- FIG. 3B illustrates an exemplary replication of multicast traffic received from an IS port via an ingress switch multicast tree in a fabric switch, in accordance with an embodiment of the present invention.
- a switch determines whether a valid VLAN identifier is included in the fabric-encapsulated packet.
- a valid VLAN identifier can be for a VLAN which is configured at the switch. If the switch identifies a valid VLAN identifier, the switch obtains an MGID from its mapping table based on the VLAN identifier and obtains an edge port bitmap from its MGID table 304 based on the MGID. Otherwise, the switch does not obtain an MGID from mapping table 302, and therefore does not obtain an edge port bitmap for the fabric-encapsulated packet.
- an edge port bitmap with all bits unset (or set) indicates that the packet should not be replicated to local edge ports.
- the switch removes the fabric encapsulation and forwards the inner packet via the edge ports indicated by the edge port bitmap.
- the switch also obtains an IS port bitmap from its MSID table 306 based on the egress switch identifier in the fabric encapsulation. Based on the IS port bitmap, the switch determines the IS ports to which the fabric-encapsulated packet should be replicated.
- switch 104 receives a fabric-encapsulated multicast packet via an IS port.
- Switch 104 determines whether a valid VLAN identifier is included in fabric-encapsulated packet 204.
- a valid VLAN identifier can be for a VLAN which is configured at switch 104. If switch 104 identifies a valid VLAN identifier, switch 104 obtains an MGID from its mapping table based on the VLAN identifier, and obtains an edge port bitmap from its MGID table based on the MGID. Based on this edge port bitmap, switch 104 determines that packet 202 should be replicated to the local edge port which couples end device 116.
- Switch 104 also obtains an IS port bitmap from its MSID table based on the identifier of switch 103 in fabric encapsulation 212 of packet 204. Based on the IS port bitmap, switch 104 determines that fabric- encapsulated packet 206 should be replicated to the local IS port which couples switch 105.
- FIG. 3C illustrates an exemplary replication of multicast traffic via one of multiple ingress switch multicast tree instances in a fabric switch, in accordance with an embodiment of the present invention.
- a respective member switch of the fabric switch can support multiple ingress switch multicast tree instances. If there are i such instances, these instances can be represented by a corresponding MSID tables 308-1, 308-2, ..., 308-i, respectively.
- an instance corresponds to a multicast group. If a multicast group has presence in a switch, the corresponding MSID table is included in that switch.
- a switch selects an MSID table from MSID tables 308-1,
- switch 103 can examine multicast group information in packet 202.
- the payload of packet 202 includes an IP packet comprising a multicast group address.
- Switch 103 can examine the address to select an MSID table.
- fabric encapsulation 212 can include multicast group information.
- Switch 104 can determine the multicast group from the header information of fabric-encapsulated packet 204.
- Switch 104 can also examine the outer MAC address (e.g., all-MAC address 220) of fabric-encapsulated packet 204 to select the corresponding MSID table.
- the outer MAC address can be a multicast MAC address mapped to a multicast group address. It should be noted that when switch 104 receives packet 204, upon selecting a MSID table from its MSID tables, switch 104 obtains an IS port bitmap from the MSID table based on the identifier of switch 103 in fabric encapsulation 212 of packet 204, as described in conjunction with FIG. 3B.
- switch 103 receives multicast packet 202 from an edge port, encapsulates packet 202 in fabric encapsulation 212, and forwards fabric-encapsulated packet 204 via its ingress switch multicast tree 133.
- switch 104 receives fabric-encapsulated packet 204 via an IS port, generates updated fabric encapsulation 214, and further forwards fabric-encapsulated packet 206 via ingress switch multicast tree 133 based on the egress switch identifier, which is the identifier of switch 103, of fabric-encapsulated packet 206.
- FIG. 4A presents a flowchart illustrating the process of a switch in a fabric switch forwarding multicast traffic received from an edge port via an ingress switch multicast tree, in accordance with an embodiment of the present invention.
- the switch receives a multicast packet via an edge port (operation 402).
- the edge port is an Ethernet port and the packet is an Ethernet frame.
- the switch obtains the VLAN identifier from the packet (operation 404) and obtains an MGID from a mapping table based on the VLAN identifier (operation 406).
- the switch can obtain the MGID from an entry in the mapping table comprising a mapping between the VLAN identifier and the MGID, or by using the VLAN identifier as an index of the mapping table.
- the switch obtains an edge port bitmap from an MGID table based on the obtained MGID (operation 408).
- the switch can obtain the edge port bitmap from an entry in the MGID table comprising a mapping between the MGID and the edge port bitmap, or by using the MGID as an index of the MGID table.
- the switch identifies the edge ports corresponding to the obtained edge port bitmap (operation 410), as described in conjunction with FIG. 3A, and replicates and forwards the packet via the identified edge ports (operation 412). If the switch has multiple MSID table instances, the switch selects an MSID table instance based on the layer-2 and/or layer-3 forwarding decision (operation 414), as described in conjunction with FIG. 3C.
- the switch obtains an IS port bitmap from an MSID table based on the local switch identifier (operation 416).
- the switch can obtain the IS port bitmap from an entry in the MSID table comprising a mapping between the switch identifier and the IS port bitmap, or by using the switch identifier as an index of the MSID table. Examples of the switch identifier include, but are not limited to, a TRILL RBridge identifier, a MAC address, and an IP address.
- the switch identifies the IS ports corresponding to the obtained IS port bitmap (operation 418), as described in conjunction with FIG. 3 A.
- FIG. 4B presents a flowchart illustrating the process of a switch in a fabric switch forwarding multicast traffic received from an IS port via an ingress switch multicast tree, in accordance with an embodiment of the present invention.
- the switch receives a fabric-encapsulated multicast packet via an IS port (operation 452).
- an IS port include, but are not limited to, a TRILL port (e.g., capable of receiving TRILL-encapsulated frames) and a layer-3 port (e.g., configured with an IP address).
- the switch obtains the VLAN identifier from the packet (operation 454) and checks whether the VLAN identifier is valid (operation 456), as described in conjunction with FIG. 3B.
- a valid VLAN identifier can be for a VLAN which is configured at the switch. If the switch identifies a valid VLAN identifier, the switch obtains an MGID from a mapping table based on the VLAN identifier (operation 458). The switch can obtain the MGID from an entry in the mapping table comprising a mapping between the VLAN identifier and the MGID, or by using the VLAN identifier as an index of the mapping table.
- the switch obtains an edge port bitmap from an MGID table based on the obtained MGID (operation 460).
- the switch can obtain the edge port bitmap from an entry in the MGID table comprising a mapping between the MGID and the edge port bitmap, or by using the MGID as an index of the MGID table.
- the switch identifies the edge ports corresponding to the obtained edge port bitmap (operation 462), as described in conjunction with FIG. 3A.
- the switch obtains the inner packet by removing the fabric encapsulation (operation 464), and replicates and forwards the inner packet via the identified edge ports (operation 466).
- the switch can select an MSID table instance based on the layer-2 and/or layer-3 forwarding decision (operation 468), as described in conjunction with FIG. 3C.
- the switch obtains an IS port bitmap from an MSID table based on the egress switch identifier in the fabric encapsulation (operation 470), as described in conjunction with FIG. 2.
- the switch can obtain the IS port bitmap from an entry in the MSID table comprising a mapping between the switch identifier and the IS port bitmap, or by using the switch identifier as an index of the MSID table.
- the switch identifier include, but are not limited to, a TRILL RBridge identifier, a MAC address, and an IP address.
- the switch identifies the IS ports corresponding to the obtained IS port bitmap (operation 472), as described in conjunction with FIG. 3A.
- the switch then updates the packet encapsulation in fabric encapsulation (e.g., changes the source MAC address of the outer Ethernet header) (operation 474), as described in conjunction with FIG. 2, and replicates and forwards the updated fabric-encapsulated packet via the identified IS ports (operation 476).
- FIG. 5 illustrates exemplary presence-based ingress switch multicast trees in a fabric switch, in accordance with an embodiment of the present invention.
- a presence-based ingress switch multicast tree spans only the member switches of the fabric switch with a corresponding multicast presence.
- ingress switch multicast trees 502 and 504 are rooted at switch 101.
- ingress switch multicast trees 502 and 504 represent corresponding MSID table instances, as described in conjunction with FIG. 3A.
- Ingress switch multicast trees 502 and 504 represent multicast groups 512 and 514, respectively.
- Switch 102 is coupled to an end device 520, which is a receiver of traffic of multicast group 512.
- multicast group 512 does not have presence in switch 103.
- switch 103 is not included in ingress switch multicast tree 502. Consequently, multicast group 512 does not need hardware resources on switch 103, which does not include the MSID table instance corresponding to ingress switch multicast tree 502.
- multicast group 514 does not have presence in switch 102.
- switch 102 is not included in ingress switch multicast tree 504. Consequently, multicast group 514 does not need hardware resources on switch 102, which does not include the MSID table instance corresponding to ingress switch multicast tree 504.
- a switch uses its hardware resources only for the multicast groups which are present in that switch. This allows efficient scaling of multicast groups in fabric switch 100.
- switch 102 becomes unavailable (e.g., due to a link or node failure, or reboot event).
- traffic of multicast group 514 does not have any impact on such unavailability.
- this unavailability of switch 102 hinders forwarding traffic of multicast group 512 to end device 520.
- switch 102 can start receiving traffic of multicast group 512 and start forwarding that traffic to end device 520.
- FIG. 6 illustrates an exemplary architecture of a switch with ingress switch multicast tree support, in accordance with an embodiment of the present invention.
- a switch 600 includes a number of communication ports 602, a packet processor 610, an inter-switch multicast module 630, an edge multicast module 640, and a storage device 650. Packet processor 610 extracts and processes header information from the received frames.
- inter-switch multicast module 630 identifies for a first replication of a multicast packet an egress inter-switch port of communication ports 602 in an ingress switch multicast tree of switch 600.
- Edge multicast module 640 identifies an egress edge port of communication ports 602 for a second replication of the multicast packet based on a local MGID of switch 600.
- switch 600 also includes a selection module 632, which selects an MSID table from a plurality of MSID table instances, as described in conjunction with FIG. 3C.
- switch 600 may maintain a membership in a fabric switch, as described in conjunction with FIG. 1A, wherein switch 600 also includes a fabric switch management module 660.
- Fabric switch management module 660 maintains a configuration database in storage device 650 that maintains the configuration state of every switch within the fabric switch.
- Fabric switch management module 660 maintains the state of the fabric switch, which is used to join other switches.
- switch 600 can be configured to operate in conjunction with a remote switch as an Ethernet switch. Under such a scenario, the first replication of the multicast packet can be encapsulated in a fabric encapsulation of the fabric switch.
- edge multicast module 640 also determines whether MGID is associated with the multicast packet based on a VLAN identifier of the multicast packet.
- Communication ports 602 can include inter-switch communication channels for communication within a fabric switch. This inter-switch communication channel can be implemented via a regular communication port and based on any open or proprietary format. Communication ports 602 can include one or more TRILL ports capable of receiving frames encapsulated in a TRILL header. Communication ports 602 can also include one or more IP ports capable of receiving IP packets. An IP port is capable of receiving an IP packet and can be configured with an IP address. Packet processor 610 can process TRILL-encapsulated frames and/or IP packets.
- modules can be implemented in hardware as well as in software.
- these modules can be embodied in computer-executable instructions stored in a memory, which is coupled to one or more processors in switch 600. When executed, these instructions cause the processor(s) to perform the aforementioned functions.
- the switch includes an inter-switch multicast module and an edge multicast module.
- the inter-switch multicast module identifies for a first replication of a multicast packet an egress inter-switch port in a multicast tree rooted at the switch.
- the multicast tree is identified by an identifier of the switch.
- the edge multicast module identifies an egress edge port for a second replication of the multicast packet based on a multicast group identifier.
- the multicast group identifier is local within the switch.
- the methods and processes described herein can be embodied as code and/or data, which can be stored in a computer-readable non-transitory storage medium.
- code and/or data can be stored in a computer-readable non-transitory storage medium.
- the computer system When a computer system reads and executes the code and/or data stored on the computer-readable non- transitory storage medium, the computer system performs the methods and processes embodied as data structures and code and stored within the medium.
- the methods and processes described herein can be executed by and/or included in hardware modules or apparatus.
- These modules or apparatus may include, but are not limited to, an application- specific integrated circuit (ASIC) chip, a field-programmable gate array (FPGA), a dedicated or shared processor that executes a particular software module or a piece of code at a particular time, and/or other programmable-logic devices now known or later developed.
- ASIC application-specific integrated circuit
- FPGA field-programmable gate array
- a dedicated or shared processor that executes a particular software module or a piece of code at a particular time
- other programmable-logic devices now known or later developed.
Abstract
One embodiment of the present invention provides a switch. The switch includes an inter-switch multicast module and an edge multicast module. The inter-switch multicast module identifies for a first replication of a multicast packet an egress inter-switch port in a multicast tree rooted at the switch. The multicast tree is identified by an identifier of the switch. The edge multicast module identifies an egress edge port for a second replication of the multicast packet based on a multicast group identifier. The multicast group identifier is local within the switch.
Description
INGRESS SWITCH MULTICAST DISTRIBUTION IN A
FABRIC SWITCH
BACKGROUND
Field
[0001] The present disclosure relates to communication networks. More specifically, the present disclosure relates to multicast distribution in a fabric switch.
Related Art
[0002] The exponential growth of the Internet has made it a popular delivery medium for a variety of applications running on physical and virtual devices. Such applications have brought with them an increasing demand for bandwidth. As a result, equipment vendors race to build larger and faster switches with versatile capabilities, such as distributed multicast traffic management, to move more traffic efficiently. However, the size of a switch cannot grow infinitely. It is limited by physical space, power consumption, and design complexity, to name a few factors. Furthermore, switches with higher capability are usually more complex and expensive. More importantly, because an overly large and complex system often does not provide economy of scale, simply increasing the size and capability of a switch may prove economically unviable due to the increased per-port cost.
[0003] A flexible way to improve the scalability of a switch system is to build a fabric switch. A fabric switch is a collection of individual member switches. These member switches form a single, logical switch that can have an arbitrary number of ports and an arbitrary topology. As demands grow, customers can adopt a "pay as you grow" approach to scale up the capacity of the fabric switch.
[0004] Meanwhile, layer-2 (e.g., Ethernet) switching technologies continue to evolve. More routing-like functionalities, which have traditionally been the characteristics of layer-3 (e.g., Internet Protocol or IP) networks, are migrating into layer-2. Notably, the recent development of the Transparent Interconnection of Lots of Links (TRILL) protocol allows Ethernet switches to function more like routing devices. TRILL overcomes the inherent
inefficiency of the conventional spanning tree protocol, which forces layer-2 switches to be coupled in a logical spanning-tree topology to avoid looping. TRILL allows routing bridges (RBridges) to be coupled in an arbitrary topology without the risk of looping by implementing routing functions in switches and including a hop count in the TRILL header.
[0005] While a fabric switch brings many desirable features to a network, some issues remain unsolved in facilitating efficient multicast traffic distribution for a large number of virtual servers.
SUMMARY
[0006] One embodiment of the present invention provides a switch. The switch includes an inter-switch multicast module and an edge multicast module. The inter-switch multicast module identifies for a first replication of a multicast packet an egress inter- switch port in a multicast tree rooted at the switch. The multicast tree is identified by an identifier of the switch. The edge multicast module identifies an egress edge port for a second replication of the multicast packet based on a multicast group identifier. The multicast group identifier is local within the switch.
[0007] In a variation on this embodiment, the inter-switch multicast module identifies the inter-switch port based on a bit value corresponding to the inter-switch port. The bit value is in an inter-switch bitmap associated with the multicast tree.
[0008] In a further variation, the inter-switch bitmap is included in an entry in a multicast switch identifier table. The entry in the multicast switch identifier table corresponds to the identifier of the switch.
[0009] In a further variation, the switch also includes a selection module which selects the multicast switch identifier table from a plurality of multicast switch identifier table instances based on a multicast group of the multicast packet. A respective multicast switch identifier table instance is associated with a respective multicast group.
[0010] In a variation on this embodiment, the edge multicast module identifies the edge port based on a bit value corresponding to the edge port. The bit value is in an edge bitmap associated with the multicast group identifier.
[0011] In a further variation, the edge bitmap is included in an entry in a multicast group identifier table. The entry in the multicast switch identifier table corresponds to the multicast group identifier.
[0012] In a variation on this embodiment, the multicast group identifier is mapped to a virtual local area network (VLAN) identifier of the multicast packet in a mapping table.
[0013] In a variation on this embodiment, the switch also includes a fabric switch management module which maintains a membership in a fabric switch. The fabric switch accommodates a plurality of switches and operates as a single switch.
[0014] In a further variation, the first replication of the multicast packet is encapsulated in a fabric encapsulation of the fabric switch. The inter-switch multicast module also identifies for a third replication of the multicast packet an egress inter-switch port in a second multicast tree rooted at a second switch. This second multicast tree is identified by an identifier of the second switch.
[0015] In a further variation, the edge multicast module also determines whether the multicast group identifier is associated with the multicast packet based on a VLAN identifier of the multicast packet. BRIEF DESCRIPTION OF THE FIGURES
[0016] FIG. 1A illustrates an exemplary fabric switch with ingress switch multicast tree support, in accordance with an embodiment of the present invention.
[0017] FIG. IB illustrates exemplary ingress switch multicast trees in a fabric switch, in accordance with an embodiment of the present invention.
[0018] FIG. 2 illustrates exemplary packet headers for multicast traffic distribution via an ingress switch multicast tree in a fabric switch, in accordance with an embodiment of the present invention.
[0019] FIG. 3A illustrates an exemplary replication of multicast traffic received from an edge port via an ingress switch multicast tree in a fabric switch, in accordance with an embodiment of the present invention.
[0020] FIG. 3B illustrates an exemplary replication of multicast traffic received from an inter-switch (IS) port via an ingress switch multicast tree in a fabric switch, in accordance with an embodiment of the present invention.
[0021] FIG. 3C illustrates an exemplary replication of multicast traffic via one of multiple ingress switch multicast tree instances in a fabric switch, in accordance with an embodiment of the present invention.
[0022] FIG. 4A presents a flowchart illustrating the process of a switch in a fabric switch forwarding multicast traffic received from an edge port via an ingress switch multicast tree, in accordance with an embodiment of the present invention.
[0023] FIG. 4B presents a flowchart illustrating the process of a switch in a fabric switch forwarding multicast traffic received from an IS port via an ingress switch multicast tree, in accordance with an embodiment of the present invention.
[0024] FIG. 5 illustrates exemplary presence-based ingress switch multicast trees in a fabric switch, in accordance with an embodiment of the present invention.
[0025] FIG. 6 illustrates an exemplary architecture of a switch with ingress switch multicast tree support, in accordance with an embodiment of the present invention.
[0026] In the figures, like reference numerals refer to the same figure elements.
DETAILED DESCRIPTION
[0027] The following description is presented to enable any person skilled in the art to make and use the invention, and is provided in the context of a particular application and its requirements. Various modifications to the disclosed embodiments will be readily apparent to those skilled in the art, and the general principles defined herein may be applied to other embodiments and applications without departing from the spirit and scope of the present invention. Thus, the present invention is not limited to the embodiments shown, but is to be accorded the widest scope consistent with the claims.
Overview
[0028] In embodiments of the present invention, the problem of efficient multicast traffic distribution in a fabric switch is solved by facilitating a multicast distribution tree, which is referred to as an ingress switch multicast tree, at a respective member switch of the fabric switch. Upon receiving a multicast packet, the switch forwards the packet via its own ingress switch multicast tree.
[0029] With existing technologies, a fabric switch has a finite number of multicast distribution trees. The member switches forward traffic belonging to all multicast groups using these trees to distribute traffic. As a result, the member switches can forward multicast traffic belonging to a large number of multicast groups via a single tree. This can congest the links in the multicast tree, leading to inefficient forwarding and greater delay. This problem can be
further aggravated when this number is small. For example, this finite number can typically be one (i.e., the fabric switch typically can have one multicast distribution tree). In that case, all member switches forward all multicast traffic via the same tree and cause the links of the tree to congest.
[0030] To solve this problem, a respective member switch computes its own ingress switch multicast tree and forwards multicast traffic via that tree. As a result, the multicast traffic load of the fabric switch becomes distributed among the ingress switch multicast trees of the corresponding member switches. Since different ingress switch multicast trees comprise different links of the fabric switch, the multicast traffic load becomes distributed across the links of the fabric switch instead of a few links of a finite number of trees.
[0031] In some embodiments, in a member switch, multicast packet replication is performed in two stages. In the first stage, the member switch replicates a multicast packet to its edge ports based on a multicast group identifier (MGID) representing the edge multicast replication of the switch. In some embodiments, this MGID is local to the switch and operates as a local multicast replication identifier for the switch. It should be noted that this multicast group identifier is distinct from a multicast group address of a multicast packet, which is not local and specific to a multicast group. In the second stage, the switch replicates the packet to inter-switch (IS) ports for other member switches based on the egress switch identifier of the packet.
[0032] In a fabric switch, any number of switches coupled in an arbitrary topology may logically operate as a single switch. The fabric switch can be an Ethernet fabric switch or a virtual cluster switch (VCS), which can operate as a single Ethernet switch. Any member switch may join or leave the fabric switch in "plug-and-play" mode without any manual configuration. In some embodiments, a respective switch in the fabric switch is a Transparent Interconnection of Lots of Links (TRILL) routing bridge (RBridge). In some further embodiments, a respective switch in the fabric switch is an Internet Protocol (IP) routing-capable switch (e.g., an IP router).
[0033] It should be noted that a fabric switch is not the same as conventional switch stacking. In switch stacking, multiple switches are interconnected at a common location (often within the same rack), based on a particular topology, and manually configured in a particular way. These stacked switches typically share a common address, e.g., an IP address, so they can be addressed as a single switch externally. Furthermore, switch stacking requires a significant amount of manual configuration of the ports and inter-switch links. The need for manual configuration prohibits switch stacking from being a viable option in building a large-scale
switching system. The topology restriction imposed by switch stacking also limits the number of switches that can be stacked. This is because it is very difficult, if not impossible, to design a stack topology that allows the overall switch bandwidth to scale adequately with the number of switch units.
[0034] In contrast, a fabric switch can include an arbitrary number of switches with individual addresses, can be based on an arbitrary topology, and does not require extensive manual configuration. The switches can reside in the same location, or be distributed over different locations. These features overcome the inherent limitations of switch stacking and make it possible to build a large "switch farm," which can be treated as a single, logical switch. Due to the automatic configuration capabilities of the fabric switch, an individual physical switch can dynamically join or leave the fabric switch without disrupting services to the rest of the network.
[0035] Furthermore, the automatic and dynamic configurability of the fabric switch allows a network operator to build its switching system in a distributed and "pay-as-you-grow" fashion without sacrificing scalability. The fabric switch's ability to respond to changing network conditions makes it an ideal solution in a virtual computing environment, where network loads often change with time.
[0036] In this disclosure, the term "fabric switch" refers to a number of interconnected physical switches which form a single, scalable logical switch. These physical switches are referred to as member switches of the fabric switch. In a fabric switch, any number of switches can be connected in an arbitrary topology, and the entire group of switches functions together as one single, logical switch. This feature makes it possible to use many smaller, inexpensive switches to construct a large fabric switch, which can be viewed as a single logical switch externally. Although the present disclosure is presented using examples based on a fabric switch, embodiments of the present invention are not limited to a fabric switch. Embodiments of the present invention are relevant to any computing device that includes a plurality of devices operating as a single device.
[0037] The term "multicast" is used in a generic sense, and can refer to any traffic forwarding toward a plurality of recipients. Any traffic forwarding that creates and forwards more than one copy of the same packet in a fabric switch can be a referred to as "multicast."
Examples of "multicast" traffic include, but are not limited to, broadcast, unknown unicast, and multicast traffic.
[0038] The term "end device" can refer to any device external to a fabric switch.
Examples of an end device include, but are not limited to, a host machine, a conventional layer-2 switch, a layer-3 router, or any other type of network device. Additionally, an end device can be coupled to other switches or hosts further away from a layer-2 or layer-3 network. An end device can also be an aggregation point for a number of network devices to enter the fabric switch.
[0039] The term "switch" is used in a generic sense, and it can refer to any standalone or fabric switch operating in any network layer. "Switch" should not be interpreted as limiting embodiments of the present invention to layer-2 networks. Any device that can forward traffic to an external device or another switch can be referred to as a "switch." Any physical or virtual device (e.g., a virtual machine/switch operating on a computing device) that can forward traffic to an end device can be referred to as a "switch." Examples of a "switch" include, but are not limited to, a layer-2 switch, a layer-3 router, a TRILL RBridge, or a fabric switch comprising a plurality of similar or heterogeneous smaller physical and/or virtual switches.
[0040] The term "edge port" refers to a port on a fabric switch which exchanges data frames with a network device outside of the fabric switch (i.e., an edge port is not used for exchanging data frames with another member switch of a fabric switch). The term "inter-switch port" refers to a port which sends/receives data frames among member switches of a fabric switch. The terms "interface" and "port" are used interchangeably.
[0041] The term "switch identifier" refers to a group of bits that can be used to identify a switch. Examples of a switch identifier include, but are not limited to, a media access control (MAC) address, an Internet Protocol (IP) address, and an RBridge identifier. Note that the TRILL standard uses "RBridge ID" (RBridge identifier) to denote a 48-bit intermediate- system- to-intermediate-system (IS-IS) System ID assigned to an RBridge, and "RBridge nickname" to denote a 16-bit value that serves as an abbreviation for the "RBridge ID." In this disclosure, "switch identifier" is used as a generic term, is not limited to any bit format, and can refer to any format that can identify a switch. The term "RBridge identifier" is also used in a generic sense, is not limited to any bit format, and can refer to "RBridge ID," "RBridge nickname," or any other format that can identify an RBridge.
[0042] The term "packet" refers to a group of bits that can be transported together across a network. "Packet" should not be interpreted as limiting embodiments of the present invention to layer-3 networks. "Packet" can be replaced by other terminologies referring to a group of bits, such as "message," "frame," "cell," or "datagram."
Network Architecture
[0043] FIG. 1A illustrates an exemplary fabric switch with ingress switch multicast tree support, in accordance with an embodiment of the present invention. As illustrated in FIG. 1A, a fabric switch 100 includes member switches 101, 102, 103, 104, and 105. Switches 103 and 105 are coupled to end devices 112 and 114, respectively. In some embodiments, fabric switch 100 is a TRILL network and a respective member switch of fabric switch 100, such as switch 105, is a TRILL RBridge. In some further embodiments, fabric switch 100 is an IP network and a respective member switch of fabric switch 100, such as switch 105, is an IP-capable switch, which calculates and maintains a local IP routing table (e.g., a routing information base or RIB), and is capable of forwarding packets based on its IP addresses.
[0044] Switches in fabric switch 100 use edge ports to communicate with end devices (e.g., non-member switches) and inter-switch ports to communicate with other member switches. For example, switch 105 is coupled to end device 114 via an edge port and to switches 101, 102, and 104 via inter-switch ports and one or more links. Data communication via an edge port can be based on Ethernet and via an inter-switch port can be based on IP and/or TRILL protocol. It should be noted that control message exchange via inter-switch ports can be based on a different protocol (e.g., Internet Protocol (IP) or Fibre Channel (FC) protocol). Supporting multiple multicast trees in a TRILL network is specified in U.S. Patent Application No. 13/030,688 (Attorney Docket No. BRCD-3041.1.US.NP), titled "Supporting multiple multicast trees in
TRILL networks," by inventors Shunjia Yu, Nagarajan Venkatesan, Anoop Ghanwani, Phanidhar Koganti, Mythilikanth Raman, Rajiv Krishnamurthy, and Dilip Chatwani, the disclosure of which is incorporated herein in its entirety.
[0045] During operation, switch 103 receives a multicast packet from end device 112. Switch 103 is then the ingress switch of fabric switch 100 for that multicast packet. With existing technologies, fabric switch 100 has a finite number of multicast distribution trees.
Suppose that one of these trees is rooted at switch 101. Upon receiving the multicast packet, switch 103 forwards the packet to switch 101, which in turn, forwards that packet to switches 102, 104, and 105 via the tree. Similarly, upon receiving a multicast packet from end device 114, switch 105 forwards the packet to switch 101, which in turn, forwards that packet to switches 102, 103, and 104 via the tree. Using the same tree to forward multicast traffic from different
ingress switches can congest the links in the multicast tree, leading to inefficient forwarding and greater delay.
[0046] To solve this problem, a respective member switch of fabric switch 100 computes its own ingress switch multicast tree and forwards multicast traffic via that tree. For example, upon receiving a multicast packet, switch 103 forwards the packet via its ingress switch multicast tree. Similarly, upon receiving a multicast packet, switch 105 forwards the packet via its ingress switch multicast tree. As a result, these multicast packets become distributed in fabric switch 100 among the ingress switch multicast trees rooted at switches 103 and 105. Since different ingress switch multicast trees comprise different links of fabric switch 100, the multicast traffic load becomes distributed across the links of fabric switch 100 instead of a few links of a finite number of trees.
[0047] FIG. IB illustrates exemplary ingress switch multicast trees in a fabric switch, in accordance with an embodiment of the present invention. During operation, switches 101, 102, 103, 104, and 105 compute ingress switch multicast trees 131, 132, 133, 134, and 135, respectively. In other words, trees 131, 132, 133, 134, and 135 have switches 101, 102, 103, 104, and 105 as their root node (RN), respectively. During operation, switch 103 receives a multicast packet from end device 112 and forwards the packet via ingress switch multicast tree 133.
Ingress switch multicast tree 133 includes links 121, 122, 123, and 124. Switch 103 replicates the multicast packet and forwards the replicated packets via links 121, 122, and 123. Upon receiving the replicated packet, switch 104 further replicates the packet and forwards the replicated packet via link 124. Suppose that switch 104 is coupled to an end device 116, which is a receiver of the multicast group of the packet. Switch 104 then replicates the packet and forwards the packet via the edge port coupling end device 116.
[0048] Similarly, upon receiving a multicast packet from end device 114, switch 105 forwards the packet via ingress switch multicast tree 135. Ingress switch multicast tree 135 includes links 124, 125, 126, and 123. Switch 105 replicates the multicast packet and forwards the replicated packets via links 124, 125, and 126. Upon receiving the replicated packet, switch 104 further replicates the packet and forwards the replicated packet via link 123. If end device 116 is a receiver of the multicast group of the packet, switch 104 replicates the packet and forwards the packet via the edge port coupling end device 116.
Packet Headers
[0049] In some embodiments, in the example in FIG. IB, switch 103 can encapsulate a received multicast packet forwarding to other member switches. FIG. 2 illustrates exemplary packet headers for multicast traffic distribution via an ingress switch multicast tree in a fabric switch, in accordance with an embodiment of the present invention. In this example, the ingress switch multicast tree is ingress switch multicast tree 133. During operation, switch 103 receives a multicast packet 202 from end device 112. This packet includes a layer-2 header 210 (e.g., an Ethernet header) comprising a source MAC addresses, a destination MAC addresses, and a payload. This payload can include an IP packet. It should be noted that packet 202 can be a broadcast or unknown unicast packet as well.
[0050] Switch 103 encapsulates packet 202 in a fabric encapsulation 212 to generate fabric-encapsulated packet 204. Examples of fabric encapsulation 212 include, but are not limited to, TRILL encapsulation and IP encapsulation. Fabric encapsulation 212 includes the identifier of switch 103 as both the ingress and the egress identifier. Examples of a switch identifier include, but are not limited to, an RBridge identifier, an IP version 4 address, and an IP version 6 address. Examples of fabric-encapsulated packet 204 include, but are not limited to, a TRILL frame and an IP packet. Fabric encapsulation 212 can also include an outer layer-2 header comprising an all-MAC address 220, which indicates that this packet is destined to all recipients in fabric switch 100. The outer layer-2 header also includes the MAC address of switch 103 as the source MAC address.
[0051] Switch 103 forwards fabric-encapsulated packet 204 via ingress switch multicast tree 133. Switches 101, 102, and 104 receive fabric-encapsulated packet 204, identify all-MAC address 220, and determine that this packet is a multicast packet. Switches 101, 102, and 104 also identify the identifier of switch 103 as egress switch identifier (which is also the ingress switch identifier), and recognize that the packet should be forwarded via ingress switch multicast tree 133 of switch 103. Switches 101 and 102 identify themselves as leaf nodes of ingress switch multicast tree 133. Suppose that switch 101 is coupled to end device 222, which is a recipient of packet 202. Switch 101 then removes fabric encapsulation 212, replicates inner packet 202, and forwards packet 202 to end device 222 via the corresponding edge port.
[0052] On the other hand, switch 104 detects that it is coupled to another downstream switch of ingress switch multicast tree 133. Switch 104 then replicates fabric-encapsulated packet 204 to generate fabric-encapsulated packet 206. However, because switch 104 is forwarding the packet, switch 104 changes the source MAC address of the outer layer-2 header to
the MAC address of switch 104 to generate fabric encapsulation 214, and forwards fabric- encapsulated packet 206 to switch 105. Suppose that end device 116 is a recipient of packet 202. Switch 104 then also removes fabric encapsulation 212, replicates inner packet 202, and forwards packet 202 to end device 116 via the corresponding edge port.
[0053] Switch 105 receives fabric-encapsulated packet 206, identifies all-MAC address
220, and determines that this packet is a multicast packet. Switch 105 also identifies the identifier of switch 103 as the egress switch identifier (which is also the ingress switch identifier), and recognizes that the packet should be forwarded via ingress switch multicast tree 133 of switch 103. Switch 105 identifies itself as a leaf node of ingress switch multicast tree 133.
Suppose that end device 114 is a recipient of packet 202. Switch 105 then removes fabric encapsulation 214, replicates inner packet 202, and forwards packet 202 to end device 114 via the corresponding edge port.
Multicast Replication
[0054] In the example in FIG. 2, switch 104 replicates a multicast packet via an edge port
(e.g., for end device 116) and an IS port (e.g., for switch 105). In some embodiments, in switch 104, multicast packet replication is performed in two stages. In the first stage, switch 104 replicates a multicast packet to its edge ports based on an MGID representing the edge multicast replication of switch 104. In some embodiments, this MGID is local to switch 104 (i.e., is not included in a packet and unique only within switch 104) and operates as a local multicast replication identifier for switch 104. In the second stage, switch 104 replicates the packet to IS ports for other member switches based on the egress switch identifier of the packet. In the example in FIG. 2, the IS port is for switch 105 and the egress switch identifier is the identifier of switch 103.
[0055] FIG. 3A illustrates an exemplary replication of multicast traffic received from an edge port via an ingress switch multicast tree in a fabric switch, in accordance with an
embodiment of the present invention. When a multicast packet is received via the edge port of a switch, the switch obtains an MGID for the packet from a mapping table 302. Mapping table 302 maps an MGID with a virtual local area network (VLAN) identifier (VLAN ID). This VLAN identifier can be in the header of a packet (e.g., packet 202) and/or in the header of an
encapsulated packet (e.g., fabric-encapsulated packet 204). For example this VLAN identifier can be a service VLAN (S-VLAN) and/or a customer VLAN (C-VLAN) identifier.
[0056] Mapping table 302 can also map one or more fields of a packet to an MGID, such as source and/or destination IP addresses, source and/or destination MAC addresses, source and/or destination ports, and a service and/or client VLANs. A respective entry in mapping table 302 can include the mapping or can be indexed based on VLAN identifiers. Mapping table 302 includes mapping for MGIDs 312-1, 312-2, ..., 312-w. The switch uses the VLAN identifier of the packet to obtain the corresponding MGID from mapping table 302.
[0057] The switch uses this MGID to obtain an edge port bitmap from MGID table 304. The edge port bitmap represents the edge ports to which the packet should be replicated. An edge port is represented by a bit in the edge port bitmap, and a set (or unset) bit can indicate that the packet should be replicated and forwarded via the corresponding edge port. For example, a bitmap of "11000" can indicate that a packet should be replicated to the first two edge ports of the switch. It should be noted that the length of the edge port bitmap (i.e., the number of bits in the bitmap) can be equal to or greater than the number of edge ports of the switch. MGID table 304 includes edge port bitmaps 314-1, 314-2, ..., 314-m. It should be noted that m and n can be different. A respective entry in MGID table 304 can include a mapping between an edge port bitmap and an MGID, or can be indexed based on MGIDs. Upon obtaining an edge port bitmap, the switch replicates and forwards the packet via the edge ports indicated by the bitmap.
[0058] The switch uses its local switch identifier (e.g., an RBridge identifier or an IP address) to obtain an IS port bitmap from a multicast switch identifier (MSID) table 306. The IS port bitmap represents the IS ports to which the packet should be replicated. An IS port is represented by a bit in the IS port bitmap, and a set (or unset) bit can indicate that the packet should be replicated and forwarded via the corresponding IS port. For example, a bitmap of "11000" can indicate that a packet should be replicated to the first two IS ports of the switch. It should be noted that the length of the IS port bitmap can be equal to or greater than the number of switches in a relevant network (e.g., in a fabric switch). MSID table 306 includes IS port bitmaps 316-1, 316-2, ..., 316-fc. It should be noted that each of m, n, and k can be different.
[0059] A respective entry in MSID table 306 can include a mapping between an IS port bitmap and a switch identifier, or can be indexed based on the egress switch identifiers. Because the ingress and egress switch identifiers of a fabric encapsulation identify the root node of an ingress switch multicast tree, such indexing leads to the IS port bitmap corresponding to that ingress switch multicast tree. Upon obtaining an IS port bitmap, the switch encapsulates the packet in fabric encapsulation, replicates the fabric-encapsulated packet, and forwards the fabric-
encapsulated packets via the IS ports indicated by the bitmap. It should be noted that if a switch has no IS port to which the fabric-encapsulated packet should be replicated, a respective bit in the corresponding IS port bitmap can be unset (or set).
[0060] In the example in FIG. 2, switch 103 receives a multicast packet 202 via an edge port. Switch 103 obtain an MGID from its mapping table based on a VLAN identifier of packet 202, and obtains an edge port bitmap from its MGID table based on the MGID. Based on this edge port bitmap, switch 103 determines whether packet 202 should be replicated to any other local edge ports. Switch 103 also obtains an IS port bitmap from its MSID table based on the identifier of switch 103. Based on the IS port bitmap, switch 103 determines that fabric - encapsulated packet 204 should be replicated to the local IS ports which couple switches 101, 102, and 104.
[0061] FIG. 3B illustrates an exemplary replication of multicast traffic received from an IS port via an ingress switch multicast tree in a fabric switch, in accordance with an embodiment of the present invention. Upon receiving a fabric-encapsulated multicast packet, a switch determines whether a valid VLAN identifier is included in the fabric-encapsulated packet. A valid VLAN identifier can be for a VLAN which is configured at the switch. If the switch identifies a valid VLAN identifier, the switch obtains an MGID from its mapping table based on the VLAN identifier and obtains an edge port bitmap from its MGID table 304 based on the MGID. Otherwise, the switch does not obtain an MGID from mapping table 302, and therefore does not obtain an edge port bitmap for the fabric-encapsulated packet.
[0062] In some embodiments, an edge port bitmap with all bits unset (or set) indicates that the packet should not be replicated to local edge ports. The switch removes the fabric encapsulation and forwards the inner packet via the edge ports indicated by the edge port bitmap. The switch also obtains an IS port bitmap from its MSID table 306 based on the egress switch identifier in the fabric encapsulation. Based on the IS port bitmap, the switch determines the IS ports to which the fabric-encapsulated packet should be replicated.
[0063] In the example in FIG. 2, switch 104 receives a fabric-encapsulated multicast packet via an IS port. Switch 104 determines whether a valid VLAN identifier is included in fabric-encapsulated packet 204. A valid VLAN identifier can be for a VLAN which is configured at switch 104. If switch 104 identifies a valid VLAN identifier, switch 104 obtains an MGID from its mapping table based on the VLAN identifier, and obtains an edge port bitmap from its MGID table based on the MGID. Based on this edge port bitmap, switch 104 determines that
packet 202 should be replicated to the local edge port which couples end device 116. Switch 104 also obtains an IS port bitmap from its MSID table based on the identifier of switch 103 in fabric encapsulation 212 of packet 204. Based on the IS port bitmap, switch 104 determines that fabric- encapsulated packet 206 should be replicated to the local IS port which couples switch 105.
[0064] FIG. 3C illustrates an exemplary replication of multicast traffic via one of multiple ingress switch multicast tree instances in a fabric switch, in accordance with an embodiment of the present invention. A respective member switch of the fabric switch can support multiple ingress switch multicast tree instances. If there are i such instances, these instances can be represented by a corresponding MSID tables 308-1, 308-2, ..., 308-i, respectively. In some embodiments, an instance corresponds to a multicast group. If a multicast group has presence in a switch, the corresponding MSID table is included in that switch.
Otherwise, that instance is not included in the switch, thereby saving hardware resources. This also allows the MSID tables to scale well in the fabric switch for a large number of multicast groups.
[0065] In some embodiments, a switch selects an MSID table from MSID tables 308-1,
308-2, ..., 308-i based on a layer-2 or layer-3 forwarding decision. In the example in FIG. 2, switch 103 can examine multicast group information in packet 202. In some embodiments, the payload of packet 202 includes an IP packet comprising a multicast group address. Switch 103 can examine the address to select an MSID table. In some embodiments, fabric encapsulation 212 can include multicast group information. Switch 104 can determine the multicast group from the header information of fabric-encapsulated packet 204. Switch 104 can also examine the outer MAC address (e.g., all-MAC address 220) of fabric-encapsulated packet 204 to select the corresponding MSID table. The outer MAC address can be a multicast MAC address mapped to a multicast group address. It should be noted that when switch 104 receives packet 204, upon selecting a MSID table from its MSID tables, switch 104 obtains an IS port bitmap from the MSID table based on the identifier of switch 103 in fabric encapsulation 212 of packet 204, as described in conjunction with FIG. 3B.
Multicast Forwarding
[0066] In the example in FIG. 2, switch 103 receives multicast packet 202 from an edge port, encapsulates packet 202 in fabric encapsulation 212, and forwards fabric-encapsulated packet 204 via its ingress switch multicast tree 133. On the other hand, switch 104 receives
fabric-encapsulated packet 204 via an IS port, generates updated fabric encapsulation 214, and further forwards fabric-encapsulated packet 206 via ingress switch multicast tree 133 based on the egress switch identifier, which is the identifier of switch 103, of fabric-encapsulated packet 206.
[0067] FIG. 4A presents a flowchart illustrating the process of a switch in a fabric switch forwarding multicast traffic received from an edge port via an ingress switch multicast tree, in accordance with an embodiment of the present invention. During operation, the switch receives a multicast packet via an edge port (operation 402). In some embodiments, the edge port is an Ethernet port and the packet is an Ethernet frame. The switch obtains the VLAN identifier from the packet (operation 404) and obtains an MGID from a mapping table based on the VLAN identifier (operation 406). The switch can obtain the MGID from an entry in the mapping table comprising a mapping between the VLAN identifier and the MGID, or by using the VLAN identifier as an index of the mapping table.
[0068] The switch obtains an edge port bitmap from an MGID table based on the obtained MGID (operation 408). The switch can obtain the edge port bitmap from an entry in the MGID table comprising a mapping between the MGID and the edge port bitmap, or by using the MGID as an index of the MGID table. The switch identifies the edge ports corresponding to the obtained edge port bitmap (operation 410), as described in conjunction with FIG. 3A, and replicates and forwards the packet via the identified edge ports (operation 412). If the switch has multiple MSID table instances, the switch selects an MSID table instance based on the layer-2 and/or layer-3 forwarding decision (operation 414), as described in conjunction with FIG. 3C.
[0069] If the switch has selected an MSID table instance (operation 414) and/or has replicated the packet via the edge ports (operation 412), the switch obtains an IS port bitmap from an MSID table based on the local switch identifier (operation 416). The switch can obtain the IS port bitmap from an entry in the MSID table comprising a mapping between the switch identifier and the IS port bitmap, or by using the switch identifier as an index of the MSID table. Examples of the switch identifier include, but are not limited to, a TRILL RBridge identifier, a MAC address, and an IP address. The switch identifies the IS ports corresponding to the obtained IS port bitmap (operation 418), as described in conjunction with FIG. 3 A. The switch encapsulates the packet in fabric encapsulation (operation 420), as described in conjunction with FIG. 2, and replicates and forwards the fabric-encapsulated packet via the identified IS ports (operation 422).
[0070] FIG. 4B presents a flowchart illustrating the process of a switch in a fabric switch forwarding multicast traffic received from an IS port via an ingress switch multicast tree, in accordance with an embodiment of the present invention. During operation, the switch receives a fabric-encapsulated multicast packet via an IS port (operation 452). Examples of an IS port include, but are not limited to, a TRILL port (e.g., capable of receiving TRILL-encapsulated frames) and a layer-3 port (e.g., configured with an IP address). The switch obtains the VLAN identifier from the packet (operation 454) and checks whether the VLAN identifier is valid (operation 456), as described in conjunction with FIG. 3B. A valid VLAN identifier can be for a VLAN which is configured at the switch. If the switch identifies a valid VLAN identifier, the switch obtains an MGID from a mapping table based on the VLAN identifier (operation 458). The switch can obtain the MGID from an entry in the mapping table comprising a mapping between the VLAN identifier and the MGID, or by using the VLAN identifier as an index of the mapping table.
[0071] The switch obtains an edge port bitmap from an MGID table based on the obtained MGID (operation 460). The switch can obtain the edge port bitmap from an entry in the MGID table comprising a mapping between the MGID and the edge port bitmap, or by using the MGID as an index of the MGID table. The switch identifies the edge ports corresponding to the obtained edge port bitmap (operation 462), as described in conjunction with FIG. 3A. The switch obtains the inner packet by removing the fabric encapsulation (operation 464), and replicates and forwards the inner packet via the identified edge ports (operation 466). If the packet does not include a valid VLAN identifier (operation 456) or has replicated the packet via the edge ports (operation 466), and if the switch has multiple MSID table instances, the switch can select an MSID table instance based on the layer-2 and/or layer-3 forwarding decision (operation 468), as described in conjunction with FIG. 3C.
[0072] If the switch has selected an MSID table instance (operation 468) and/or has replicated the packet via the edge ports (operation 466), the switch obtains an IS port bitmap from an MSID table based on the egress switch identifier in the fabric encapsulation (operation 470), as described in conjunction with FIG. 2. The switch can obtain the IS port bitmap from an entry in the MSID table comprising a mapping between the switch identifier and the IS port bitmap, or by using the switch identifier as an index of the MSID table. Examples of the switch identifier include, but are not limited to, a TRILL RBridge identifier, a MAC address, and an IP address. The switch identifies the IS ports corresponding to the obtained IS port bitmap (operation 472),
as described in conjunction with FIG. 3A. The switch then updates the packet encapsulation in fabric encapsulation (e.g., changes the source MAC address of the outer Ethernet header) (operation 474), as described in conjunction with FIG. 2, and replicates and forwards the updated fabric-encapsulated packet via the identified IS ports (operation 476).
Presence-Based Multicast Trees
[0073] FIG. 5 illustrates exemplary presence-based ingress switch multicast trees in a fabric switch, in accordance with an embodiment of the present invention. A presence-based ingress switch multicast tree spans only the member switches of the fabric switch with a corresponding multicast presence. In the example in FIG. 5, ingress switch multicast trees 502 and 504 are rooted at switch 101. In some embodiments, ingress switch multicast trees 502 and 504 represent corresponding MSID table instances, as described in conjunction with FIG. 3A. Ingress switch multicast trees 502 and 504 represent multicast groups 512 and 514, respectively. Switch 102 is coupled to an end device 520, which is a receiver of traffic of multicast group 512.
[0074] Suppose that multicast group 512 does not have presence in switch 103. As a result, switch 103 is not included in ingress switch multicast tree 502. Consequently, multicast group 512 does not need hardware resources on switch 103, which does not include the MSID table instance corresponding to ingress switch multicast tree 502. Similarly, suppose that multicast group 514 does not have presence in switch 102. As a result, switch 102 is not included in ingress switch multicast tree 504. Consequently, multicast group 514 does not need hardware resources on switch 102, which does not include the MSID table instance corresponding to ingress switch multicast tree 504. In this way, a switch uses its hardware resources only for the multicast groups which are present in that switch. This allows efficient scaling of multicast groups in fabric switch 100.
[0075] Suppose that switch 102 becomes unavailable (e.g., due to a link or node failure, or reboot event). Under such a scenario, traffic of multicast group 514 does not have any impact on such unavailability. However, this unavailability of switch 102 hinders forwarding traffic of multicast group 512 to end device 520. When switch 102 becomes available again, switch 102 can start receiving traffic of multicast group 512 and start forwarding that traffic to end device 520.
Exemplary Switch
[0076] FIG. 6 illustrates an exemplary architecture of a switch with ingress switch multicast tree support, in accordance with an embodiment of the present invention. In this example, a switch 600 includes a number of communication ports 602, a packet processor 610, an inter-switch multicast module 630, an edge multicast module 640, and a storage device 650. Packet processor 610 extracts and processes header information from the received frames.
[0077] As described in conjunction with FIG. 2, inter-switch multicast module 630 identifies for a first replication of a multicast packet an egress inter-switch port of communication ports 602 in an ingress switch multicast tree of switch 600. Edge multicast module 640 identifies an egress edge port of communication ports 602 for a second replication of the multicast packet based on a local MGID of switch 600. In some embodiments, switch 600 also includes a selection module 632, which selects an MSID table from a plurality of MSID table instances, as described in conjunction with FIG. 3C.
[0078] In some embodiments, switch 600 may maintain a membership in a fabric switch, as described in conjunction with FIG. 1A, wherein switch 600 also includes a fabric switch management module 660. Fabric switch management module 660 maintains a configuration database in storage device 650 that maintains the configuration state of every switch within the fabric switch. Fabric switch management module 660 maintains the state of the fabric switch, which is used to join other switches. In some embodiments, switch 600 can be configured to operate in conjunction with a remote switch as an Ethernet switch. Under such a scenario, the first replication of the multicast packet can be encapsulated in a fabric encapsulation of the fabric switch. In some embodiments, edge multicast module 640 also determines whether MGID is associated with the multicast packet based on a VLAN identifier of the multicast packet.
[0079] Communication ports 602 can include inter-switch communication channels for communication within a fabric switch. This inter-switch communication channel can be implemented via a regular communication port and based on any open or proprietary format. Communication ports 602 can include one or more TRILL ports capable of receiving frames encapsulated in a TRILL header. Communication ports 602 can also include one or more IP ports capable of receiving IP packets. An IP port is capable of receiving an IP packet and can be configured with an IP address. Packet processor 610 can process TRILL-encapsulated frames and/or IP packets.
[0080] Note that the above-mentioned modules can be implemented in hardware as well as in software. In one embodiment, these modules can be embodied in computer-executable
instructions stored in a memory, which is coupled to one or more processors in switch 600. When executed, these instructions cause the processor(s) to perform the aforementioned functions.
[0081] In summary, embodiments of the present invention provide a switch and a method for facilitating ingress switch multicast trees in a fabric switch. In one embodiment, the switch includes an inter-switch multicast module and an edge multicast module. The inter-switch multicast module identifies for a first replication of a multicast packet an egress inter-switch port in a multicast tree rooted at the switch. The multicast tree is identified by an identifier of the switch. The edge multicast module identifies an egress edge port for a second replication of the multicast packet based on a multicast group identifier. The multicast group identifier is local within the switch.
[0082] The methods and processes described herein can be embodied as code and/or data, which can be stored in a computer-readable non-transitory storage medium. When a computer system reads and executes the code and/or data stored on the computer-readable non- transitory storage medium, the computer system performs the methods and processes embodied as data structures and code and stored within the medium.
[0083] The methods and processes described herein can be executed by and/or included in hardware modules or apparatus. These modules or apparatus may include, but are not limited to, an application- specific integrated circuit (ASIC) chip, a field-programmable gate array (FPGA), a dedicated or shared processor that executes a particular software module or a piece of code at a particular time, and/or other programmable-logic devices now known or later developed. When the hardware modules or apparatus are activated, they perform the methods and processes included within them.
[0084] The foregoing descriptions of embodiments of the present invention have been presented only for purposes of illustration and description. They are not intended to be exhaustive or to limit this disclosure. Accordingly, many modifications and variations will be apparent to practitioners skilled in the art. The scope of the present invention is defined by the appended claims.
Claims
What Is Claimed Is: 1. A switch, comprising:
an inter-switch multicast module adapted to identify for a first replication of a multicast packet an egress inter-switch port in a multicast tree rooted at the switch, wherein the multicast tree is identified by an identifier of the switch; and
an edge multicast module adapted to identify an egress edge port for a second replication of the multicast packet based on a multicast group identifier, wherein the multicast group identifier is local within the switch.
2. The switch of claim 1, wherein the inter-switch multicast module identifies the inter-switch port based on a bit value corresponding to the inter-switch port, wherein the bit value is in an inter-switch bitmap associated with the multicast tree.
3. The switch of claim 2, wherein the inter-switch bitmap is included in an entry in a multicast switch identifier table, wherein the entry in the multicast switch identifier table corresponds to the identifier of the switch.
4. The switch of claim 3, further comprising a selection module adapted to select the multicast switch identifier table from a plurality of multicast switch identifier table instances based on a multicast group of the multicast packet, wherein a respective multicast switch identifier table instance is associated with a respective multicast group.
5. The switch of claim 1, wherein the edge multicast module identifies the edge port based on a bit value corresponding to the edge port, wherein the bit value is in an edge bitmap associated with the multicast group identifier.
6. The switch of claim 5, wherein the edge bitmap is included in an entry in a multicast group identifier table, wherein the entry in the multicast switch identifier table corresponds to the multicast group identifier.
7. The switch of claim 1, wherein the multicast group identifier is mapped to a virtual local area network (VLAN) identifier of the multicast packet in a mapping table.
8. The switch of claim 1, further comprising a fabric switch management module adapted to maintain a membership in a fabric switch, wherein the fabric switch is adapted to accommodate a plurality of switches and operates as a single switch.
9. The switch of claim 8, wherein the first replication of the multicast packet is encapsulated in a fabric encapsulation of the fabric switch; and
wherein the inter-switch multicast module is further adapted to identify for a third replication of the multicast packet an egress inter- switch port in a second multicast tree rooted at a second switch in the fabric switch, wherein the second multicast tree is identified by an identifier of the second switch.
10. The switch of claim 9, wherein the edge multicast module is further adapted to determine whether the multicast group identifier is associated with the multicast packet based on a VLAN identifier of the multicast packet.
11. A computer-executable method, comprising:
identifying for a first replication of a multicast packet an egress inter- switch port of a switch in a multicast tree rooted at the switch, wherein the multicast tree is identified by an identifier of the switch; and
identifying an egress edge port for a second replication of the multicast packet based on a multicast group identifier, wherein the multicast group identifier is local within the switch.
12. The method of claim 11, wherein the inter-switch port is identified based on a bit value corresponding to the inter- switch port, wherein the bit value is in an inter- switch bitmap associated with the multicast tree.
13. The method of claim 12, wherein the inter-switch bitmap is included in an entry in a multicast switch identifier table, wherein the entry in the multicast switch identifier table corresponds to the identifier of the switch.
14. The method of claim 13, further comprising selecting the multicast switch identifier table from a plurality of multicast switch identifier table instances based on a multicast group of the multicast packet, wherein a respective multicast switch identifier table instance is
associated with a respective multicast group.
15. The method of claim 11, wherein the edge port is identified based on a bit value corresponding to the edge port, wherein the bit value is in an edge bitmap associated with the multicast group identifier.
16. The method of claim 15, wherein the edge bitmap is included in an entry in a multicast group identifier table, wherein the entry in the multicast switch identifier table corresponds to the multicast group identifier.
17. The method of claim 11, wherein the multicast group identifier is mapped to a virtual local area network (VLAN) identifier of the multicast packet in a mapping table.
18. The method of claim 11, further comprising maintaining a membership in a fabric switch, wherein the fabric switch is adapted to accommodate a plurality of switches and operates as a single switch.
19. The method of claim 18, wherein the first replication of the multicast packet is encapsulated in a fabric encapsulation of the fabric switch; and
wherein the method further comprises identifying for a third replication of the multicast packet an egress inter-switch port in a second multicast tree rooted at a second switch in the fabric switch, wherein the second multicast tree is identified by an identifier of the second switch.
20. The method of claim 19, further comprising determining whether the multicast group identifier is associated with the multicast packet based on a VLAN identifier of the multicast packet.
21. A computing system, comprising:
a processor; and
a computer-readable storage medium storing instructions which when executed by the processor cause the processor to perform a method, the method comprising:
identifying for a first replication of a multicast packet an egress inter- switch port of a switch in a multicast tree rooted at the switch, wherein the multicast tree is identified by an identifier of the switch; and
identifying an egress edge port for a second replication of the multicast packet based on a multicast group identifier, wherein the multicast group identifier is local within the switch.
22. The computing system of claim 21, wherein the inter-switch port is identified based on a bit value corresponding to the inter-switch port, wherein the bit value is in an inter- switch bitmap associated with the multicast tree, and wherein the inter-switch bitmap is included in an entry in a multicast switch identifier table.
23. The computing system of claim 22, wherein the method further comprises selecting the multicast switch identifier table from a plurality of multicast switch identifier table instances based on a multicast group of the multicast packet, wherein a respective multicast switch identifier table instance is associated with a respective multicast group.
24. The computing system of claim 21, wherein the edge port is identified based on a bit value corresponding to the edge port, wherein the bit value is in an edge bitmap associated with the multicast group identifier, and wherein the edge bitmap is included in an entry in a multicast group identifier table.
25. The computing system of claim 21, wherein the first replication of the multicast packet is encapsulated in a fabric encapsulation of a fabric switch; and
wherein the method further comprises identifying for a third replication of the multicast packet an egress inter-switch port in a second multicast tree rooted at a second switch in the fabric switch, wherein the second multicast tree is identified by an identifier of the second switch.
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361833385P | 2013-06-10 | 2013-06-10 | |
US61/833,385 | 2013-06-10 | ||
US14/284,212 US9565028B2 (en) | 2013-06-10 | 2014-05-21 | Ingress switch multicast distribution in a fabric switch |
US14/284,212 | 2014-05-21 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2014200707A2 true WO2014200707A2 (en) | 2014-12-18 |
WO2014200707A3 WO2014200707A3 (en) | 2015-02-05 |
Family
ID=52005429
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2014/040060 WO2014200707A2 (en) | 2013-06-10 | 2014-05-29 | Ingress switch multicast distribution in a fabric switch |
Country Status (2)
Country | Link |
---|---|
US (1) | US9565028B2 (en) |
WO (1) | WO2014200707A2 (en) |
Families Citing this family (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8867552B2 (en) | 2010-05-03 | 2014-10-21 | Brocade Communications Systems, Inc. | Virtual cluster switching |
US9769016B2 (en) | 2010-06-07 | 2017-09-19 | Brocade Communications Systems, Inc. | Advanced link tracking for virtual cluster switching |
US9270486B2 (en) | 2010-06-07 | 2016-02-23 | Brocade Communications Systems, Inc. | Name services for virtual cluster switching |
US9807031B2 (en) | 2010-07-16 | 2017-10-31 | Brocade Communications Systems, Inc. | System and method for network configuration |
US9450870B2 (en) | 2011-11-10 | 2016-09-20 | Brocade Communications Systems, Inc. | System and method for flow management in software-defined networks |
US9374301B2 (en) | 2012-05-18 | 2016-06-21 | Brocade Communications Systems, Inc. | Network feedback in software-defined networks |
US9401872B2 (en) | 2012-11-16 | 2016-07-26 | Brocade Communications Systems, Inc. | Virtual link aggregations across multiple fabric switches |
US9565099B2 (en) | 2013-03-01 | 2017-02-07 | Brocade Communications Systems, Inc. | Spanning tree in fabric switches |
US9306804B2 (en) | 2013-04-16 | 2016-04-05 | Arista Networks, Inc. | Method and system for multichassis link aggregation in-service software update |
US9912612B2 (en) | 2013-10-28 | 2018-03-06 | Brocade Communications Systems LLC | Extended ethernet fabric switches |
US9548873B2 (en) | 2014-02-10 | 2017-01-17 | Brocade Communications Systems, Inc. | Virtual extensible LAN tunnel keepalives |
JP6349788B2 (en) * | 2014-03-05 | 2018-07-04 | 富士通株式会社 | Switch device, network system, and control method of switch device |
US10581758B2 (en) | 2014-03-19 | 2020-03-03 | Avago Technologies International Sales Pte. Limited | Distributed hot standby links for vLAG |
US10476698B2 (en) | 2014-03-20 | 2019-11-12 | Avago Technologies International Sales Pte. Limited | Redundent virtual link aggregation group |
US9281954B2 (en) * | 2014-04-29 | 2016-03-08 | Arista Networks, Inc. | Method and system for protocol independent multicasting in multichassis link aggregation domains |
US9722915B2 (en) | 2014-04-29 | 2017-08-01 | Dell Products L.P. | System and method for multicast routing using peer groups |
US10063473B2 (en) | 2014-04-30 | 2018-08-28 | Brocade Communications Systems LLC | Method and system for facilitating switch virtualization in a network of interconnected switches |
US9800471B2 (en) | 2014-05-13 | 2017-10-24 | Brocade Communications Systems, Inc. | Network extension groups of global VLANs in a fabric switch |
US10616108B2 (en) | 2014-07-29 | 2020-04-07 | Avago Technologies International Sales Pte. Limited | Scalable MAC address virtualization |
US9807007B2 (en) | 2014-08-11 | 2017-10-31 | Brocade Communications Systems, Inc. | Progressive MAC address learning |
US9942097B2 (en) | 2015-01-05 | 2018-04-10 | Brocade Communications Systems LLC | Power management in a network of interconnected switches |
US10003552B2 (en) | 2015-01-05 | 2018-06-19 | Brocade Communications Systems, Llc. | Distributed bidirectional forwarding detection protocol (D-BFD) for cluster of interconnected switches |
US10038592B2 (en) | 2015-03-17 | 2018-07-31 | Brocade Communications Systems LLC | Identifier assignment to a new switch in a switch group |
US10579406B2 (en) | 2015-04-08 | 2020-03-03 | Avago Technologies International Sales Pte. Limited | Dynamic orchestration of overlay tunnels |
US10439929B2 (en) | 2015-07-31 | 2019-10-08 | Avago Technologies International Sales Pte. Limited | Graceful recovery of a multicast-enabled switch |
US10171303B2 (en) | 2015-09-16 | 2019-01-01 | Avago Technologies International Sales Pte. Limited | IP-based interconnection of switches with a logical chassis |
US9912614B2 (en) | 2015-12-07 | 2018-03-06 | Brocade Communications Systems LLC | Interconnection of switches based on hierarchical overlay tunneling |
US20170310582A1 (en) * | 2016-04-21 | 2017-10-26 | Brocade Communications Systems, Inc. | Dynamic multi-destination traffic management in a distributed tunnel endpoint |
US10237090B2 (en) | 2016-10-28 | 2019-03-19 | Avago Technologies International Sales Pte. Limited | Rule-based network identifier mapping |
EP3622777B1 (en) | 2017-05-12 | 2021-07-07 | Telefonaktiebolaget LM Ericsson (Publ) | Local identifier locator network protocol (ilnp) breakout |
WO2020096594A1 (en) | 2018-11-07 | 2020-05-14 | Telefonaktiebolaget Lm Ericsson (Publ) | Local identifier locator network protocol (ilnp) breakout |
US10965589B2 (en) * | 2019-02-28 | 2021-03-30 | Cisco Technology, Inc. | Fast receive re-convergence of multi-pod multi-destination traffic in response to local disruptions |
CN113132257B (en) * | 2021-04-29 | 2022-04-26 | 杭州迪普信息技术有限公司 | Message processing method and device |
US11777838B2 (en) * | 2021-07-30 | 2023-10-03 | Avago Technologies International Sales Pte. Limited | Systems and methods for reducing bias in multicast replication sequence |
Family Cites Families (391)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US829529A (en) | 1906-01-31 | 1906-08-28 | John N Keathley | Cotton-chopper. |
US5309433A (en) | 1992-06-18 | 1994-05-03 | International Business Machines Corp. | Methods and apparatus for routing packets in packet transmission networks |
US5390173A (en) | 1992-10-22 | 1995-02-14 | Digital Equipment Corporation | Packet format in hub for packet data communications system |
US5802278A (en) | 1995-05-10 | 1998-09-01 | 3Com Corporation | Bridge/router architecture for high performance scalable networking |
US5684800A (en) | 1995-11-15 | 1997-11-04 | Cabletron Systems, Inc. | Method for establishing restricted broadcast groups in a switched network |
US5983278A (en) | 1996-04-19 | 1999-11-09 | Lucent Technologies Inc. | Low-loss, fair bandwidth allocation flow control in a packet switch |
US6085238A (en) | 1996-04-23 | 2000-07-04 | Matsushita Electric Works, Ltd. | Virtual LAN system |
US5878232A (en) | 1996-12-27 | 1999-03-02 | Compaq Computer Corporation | Dynamic reconfiguration of network device's virtual LANs using the root identifiers and root ports determined by a spanning tree procedure |
US20010005527A1 (en) | 1997-03-31 | 2001-06-28 | Kathleen Michelle Vaeth | Thin film fabrication |
US6331983B1 (en) * | 1997-05-06 | 2001-12-18 | Enterasys Networks, Inc. | Multicast switching |
US6041042A (en) | 1997-05-27 | 2000-03-21 | Cabletron Systems, Inc. | Remote port mirroring system and method thereof |
US5959968A (en) | 1997-07-30 | 1999-09-28 | Cisco Systems, Inc. | Port aggregation protocol |
US6185214B1 (en) | 1997-09-11 | 2001-02-06 | 3Com Corporation | Use of code vectors for frame forwarding in a bridge/router |
US7430164B2 (en) | 1998-05-04 | 2008-09-30 | Hewlett-Packard Development Company, L.P. | Path recovery on failure in load balancing switch protocols |
US5973278A (en) | 1998-05-07 | 1999-10-26 | Eaton Corporation | Snap acting charge/discharge and open/closed indicators displaying states of electrical switching apparatus |
IL125272A0 (en) | 1998-07-08 | 1999-03-12 | Galileo Technology Ltd | Vlan protocol |
US7065050B1 (en) | 1998-07-08 | 2006-06-20 | Broadcom Corporation | Apparatus and method for controlling data flow in a network switch |
WO2000003256A1 (en) | 1998-07-08 | 2000-01-20 | Broadcom Corporation | Network switch utilizing packet based per head-of-line blocking prevention |
DE69934644T2 (en) | 1998-10-05 | 2007-11-15 | Alcatel Lucent | Network switching equipment with distributed based on use diversion databases |
US6185241B1 (en) | 1998-10-29 | 2001-02-06 | Xerox Corporation | Metal spatial filter to enhance model reflectivity in a vertical cavity surface emitting laser |
US6438106B1 (en) | 1998-12-22 | 2002-08-20 | Nortel Networks Limited | Inter-class schedulers utilizing statistical priority guaranteed queuing and generic cell-rate algorithm priority guaranteed queuing |
US6771610B1 (en) | 1999-01-19 | 2004-08-03 | 3Com Corporation | Spanning tree with protocol for bypassing port state transition timers |
US6542266B1 (en) | 1999-06-24 | 2003-04-01 | Qwest Communications International Inc. | System and method for providing broadband data service |
JP4148605B2 (en) | 1999-08-06 | 2008-09-10 | 富士通株式会社 | Network system and server |
US6498781B1 (en) | 1999-08-13 | 2002-12-24 | International Business Machines Corporation | Self-tuning link aggregation system |
EP2267009B1 (en) | 1999-08-25 | 2014-10-29 | Allergan, Inc. | Activatable recombinant neurotoxins |
US7061877B1 (en) | 1999-09-10 | 2006-06-13 | Georgia Tech Reseach Corporation | System and method for providing high speed wireless media access |
JP2003525000A (en) | 2000-02-22 | 2003-08-19 | トップ レイヤー ネットワークス,インク. | Data flow mirror processing system and method in network switch |
JP2001313670A (en) | 2000-04-28 | 2001-11-09 | Oki Electric Ind Co Ltd | Method for managing network |
US20020019904A1 (en) | 2000-05-11 | 2002-02-14 | Katz Abraham Yehuda | Three-dimensional switch providing packet routing between multiple multimedia buses |
JP4168574B2 (en) | 2000-06-02 | 2008-10-22 | 株式会社日立製作所 | Packet transfer apparatus, packet transfer control method, and packet transfer apparatus setting method |
AU7170301A (en) | 2000-06-29 | 2002-01-14 | Cachestream Corp | Virtual multicasting |
US7924837B1 (en) | 2000-07-31 | 2011-04-12 | Avaya Communication Israel Ltd. | IP multicast in VLAN environment |
US6633761B1 (en) | 2000-08-11 | 2003-10-14 | Reefedge, Inc. | Enabling seamless user mobility in a short-range wireless networking environment |
US8619793B2 (en) | 2000-08-21 | 2013-12-31 | Rockstar Consortium Us Lp | Dynamic assignment of traffic classes to a priority queue in a packet forwarding device |
US7373425B2 (en) | 2000-08-22 | 2008-05-13 | Conexant Systems, Inc. | High-speed MAC address search engine |
CA2355473A1 (en) | 2000-09-29 | 2002-03-29 | Linghsiao Wang | Buffer management for support of quality-of-service guarantees and data flow control in data switching |
US6937576B1 (en) | 2000-10-17 | 2005-08-30 | Cisco Technology, Inc. | Multiple instance spanning tree protocol |
JP2002135410A (en) | 2000-10-26 | 2002-05-10 | Kddi Research & Development Laboratories Inc | Access network system |
US6957269B2 (en) | 2001-01-03 | 2005-10-18 | Advanced Micro Devices, Inc. | Method and apparatus for performing priority-based flow control |
US6912592B2 (en) | 2001-01-05 | 2005-06-28 | Extreme Networks, Inc. | Method and system of aggregate multiple VLANs in a metropolitan area network |
EP1358559A4 (en) | 2001-01-31 | 2009-04-29 | Lancope Inc | Network port profiling |
US7016352B1 (en) | 2001-03-23 | 2006-03-21 | Advanced Micro Devices, Inc. | Address modification within a switching device in a packet-switched network |
US7450595B1 (en) | 2001-05-01 | 2008-11-11 | At&T Corp. | Method and system for managing multiple networks over a set of ports |
US7102996B1 (en) | 2001-05-24 | 2006-09-05 | F5 Networks, Inc. | Method and system for scaling network traffic managers |
US20070116422A1 (en) | 2001-06-06 | 2007-05-24 | Reynolds Thomas A | Photoresponsive polyimide based fiber |
US6956824B2 (en) | 2001-06-14 | 2005-10-18 | Tropic Networks Inc. | Extension of link aggregation protocols over the network |
US20040001433A1 (en) | 2001-07-18 | 2004-01-01 | Gram Charles Andrew | Interactive control of network devices |
US7382787B1 (en) | 2001-07-30 | 2008-06-03 | Cisco Technology, Inc. | Packet routing and switching device |
US20040165596A1 (en) | 2001-08-01 | 2004-08-26 | Garcia Javier Romero | Apparatus and method for flow scheduling based on priorities in a mobile network |
JP2003069573A (en) | 2001-08-23 | 2003-03-07 | Allied Tereshisu Kk | System and method for managing network equipment using information recording medium |
US7173934B2 (en) | 2001-09-10 | 2007-02-06 | Nortel Networks Limited | System, device, and method for improving communication network reliability using trunk splitting |
US20030084219A1 (en) | 2001-10-26 | 2003-05-01 | Maxxan Systems, Inc. | System, apparatus and method for address forwarding for a computer network |
US20070094465A1 (en) | 2001-12-26 | 2007-04-26 | Cisco Technology, Inc., A Corporation Of California | Mirroring mechanisms for storage area networks and network based virtualization |
US20030123393A1 (en) | 2002-01-03 | 2003-07-03 | Feuerstraeter Mark T. | Method and apparatus for priority based flow control in an ethernet architecture |
US7327748B2 (en) | 2002-01-28 | 2008-02-05 | Alcatel Lucent | Enterprise switching device and method |
US7464177B2 (en) | 2002-02-20 | 2008-12-09 | Mitsubishi Denki Kabushiki Kaisha | Mobile network that routes a packet without transferring the packet to a home agent server |
US7688960B1 (en) | 2002-02-26 | 2010-03-30 | Sprint Communications Company L.P. | Method and system for separating business and device logic in a computing network system |
US7092943B2 (en) | 2002-03-01 | 2006-08-15 | Enterasys Networks, Inc. | Location based data |
US20030174706A1 (en) | 2002-03-15 | 2003-09-18 | Broadcom Corporation | Fastpath implementation for transparent local area network (LAN) services over multiprotocol label switching (MPLS) |
US7315545B1 (en) | 2002-03-29 | 2008-01-01 | Nortel Networks Limited | Method and apparatus to support differential internet data packet treatment in a base station controller |
TW550902B (en) | 2002-04-03 | 2003-09-01 | Accton Technology Corp | Method of setting network configuration and device and system thereof |
US7209435B1 (en) | 2002-04-16 | 2007-04-24 | Foundry Networks, Inc. | System and method for providing network route redundancy across Layer 2 devices |
US20030208616A1 (en) | 2002-05-01 | 2003-11-06 | Blade Software, Inc. | System and method for testing computer network access and traffic control systems |
US20090279558A1 (en) | 2002-05-06 | 2009-11-12 | Ian Edward Davis | Network routing apparatus for enhanced efficiency and monitoring capability |
US7206288B2 (en) | 2002-06-12 | 2007-04-17 | Cisco Technology, Inc. | Methods and apparatus for characterizing a route in fibre channel fabric |
US20040003094A1 (en) | 2002-06-27 | 2004-01-01 | Michael See | Method and apparatus for mirroring traffic over a network |
FI113127B (en) | 2002-06-28 | 2004-02-27 | Ssh Comm Security Corp | Broadcast packet handling method for gateway computer, involves encapsulating packet into form acceptable for transmission over Internet protocol security protected connection and transmitting packet to logical network segment |
US7330897B2 (en) | 2002-07-09 | 2008-02-12 | International Business Machines Corporation | Methods and apparatus for storage area network component registration |
US7453888B2 (en) | 2002-08-27 | 2008-11-18 | Alcatel Lucent | Stackable virtual local area network provisioning in bridged networks |
US7316031B2 (en) | 2002-09-06 | 2008-01-01 | Capital One Financial Corporation | System and method for remotely monitoring wireless networks |
US20060062187A1 (en) | 2002-10-04 | 2006-03-23 | Johan Rune | Isolation of hosts connected to an access network |
US7647427B1 (en) | 2002-10-18 | 2010-01-12 | Foundry Networks, Inc. | Redundancy support for network address translation (NAT) |
US7292581B2 (en) | 2002-10-24 | 2007-11-06 | Cisco Technology, Inc. | Large-scale layer 2 metropolitan area network |
WO2004043046A1 (en) | 2002-11-08 | 2004-05-21 | Koninklijke Philips Electronics N.V. | Method and apparatus allowing remote access in data networks |
US7424014B2 (en) | 2002-11-12 | 2008-09-09 | Cisco Technology, Inc. | System and method for local packet transport services within distributed routers |
US7397794B1 (en) | 2002-11-21 | 2008-07-08 | Juniper Networks, Inc. | Systems and methods for implementing virtual switch planes in a physical switch fabric |
KR100480366B1 (en) | 2002-12-24 | 2005-03-31 | 한국전자통신연구원 | A system for VLAN configuration of E-PON and method thereof, its program stored recording medium |
US7417950B2 (en) | 2003-02-03 | 2008-08-26 | Ciena Corporation | Method and apparatus for performing data flow ingress/egress admission control in a provider network |
JP4256693B2 (en) | 2003-02-18 | 2009-04-22 | 株式会社日立製作所 | Computer system, I / O device, and virtual sharing method of I / O device |
US20040165595A1 (en) | 2003-02-25 | 2004-08-26 | At&T Corp. | Discovery and integrity testing method in an ethernet domain |
US7411973B2 (en) | 2003-03-11 | 2008-08-12 | Broadcom Corporation | System and method for interfacing with a management system |
US7486674B2 (en) | 2003-04-28 | 2009-02-03 | Alcatel-Lucent Usa Inc. | Data mirroring in a service |
US7370346B2 (en) | 2003-04-29 | 2008-05-06 | Hewlett-Packard Development Company, L.P. | Method and apparatus for access security services |
US7561590B1 (en) | 2003-05-05 | 2009-07-14 | Marvell International Ltd. | Network switch having virtual input queues for flow control |
US7516487B1 (en) | 2003-05-21 | 2009-04-07 | Foundry Networks, Inc. | System and method for source IP anti-spoofing security |
WO2004112327A1 (en) | 2003-06-11 | 2004-12-23 | Nec Corporation | Router and network connecting method |
US7480258B1 (en) | 2003-07-03 | 2009-01-20 | Cisco Technology, Inc. | Cross stack rapid transition protocol |
US7463579B2 (en) | 2003-07-11 | 2008-12-09 | Nortel Networks Limited | Routed split multilink trunking |
JP4123088B2 (en) | 2003-08-06 | 2008-07-23 | 株式会社日立製作所 | Storage network management apparatus and method |
US7380025B1 (en) | 2003-10-07 | 2008-05-27 | Cisco Technology, Inc. | Method and apparatus providing role-based configuration of a port of a network element |
US20050105538A1 (en) | 2003-10-14 | 2005-05-19 | Ananda Perera | Switching system with distributed switching fabric |
US8050180B2 (en) | 2003-10-31 | 2011-11-01 | Brocade Communications Systems, Inc. | Network path tracing method |
US8179808B2 (en) | 2003-10-31 | 2012-05-15 | Brocade Communication Systems, Inc. | Network path tracing method |
WO2005050950A1 (en) | 2003-11-13 | 2005-06-02 | Cryptek, Inc. | System and method for traversing access control metadata across multiple network domains |
US7558273B1 (en) | 2003-12-23 | 2009-07-07 | Extreme Networks, Inc. | Methods and systems for associating and translating virtual local area network (VLAN) tags |
US7333508B2 (en) | 2004-01-20 | 2008-02-19 | Nortel Networks Limited | Method and system for Ethernet and frame relay network interworking |
US8804728B2 (en) | 2004-01-20 | 2014-08-12 | Rockstar Consortium Us Lp | Ethernet differentiated services conditioning |
US7701948B2 (en) | 2004-01-20 | 2010-04-20 | Nortel Networks Limited | Metro ethernet service enhancements |
US7310664B1 (en) | 2004-02-06 | 2007-12-18 | Extreme Networks | Unified, configurable, adaptive, network architecture |
US7843906B1 (en) | 2004-02-13 | 2010-11-30 | Habanero Holdings, Inc. | Storage gateway initiator for fabric-backplane enterprise servers |
US7860097B1 (en) | 2004-02-13 | 2010-12-28 | Habanero Holdings, Inc. | Fabric-backplane enterprise servers with VNICs and VLANs |
US7843907B1 (en) | 2004-02-13 | 2010-11-30 | Habanero Holdings, Inc. | Storage gateway target for fabric-backplane enterprise servers |
WO2005083982A1 (en) | 2004-02-23 | 2005-09-09 | Sinett Corporation | Unified architecture for wired and wireless networks |
US7477894B1 (en) | 2004-02-23 | 2009-01-13 | Foundry Networks, Inc. | Methods and apparatus for handling wireless roaming among and across wireless area networks |
US7690040B2 (en) | 2004-03-10 | 2010-03-30 | Enterasys Networks, Inc. | Method for network traffic mirroring with data privacy |
US20050220096A1 (en) | 2004-04-06 | 2005-10-06 | Robert Friskney | Traffic engineering in frame-based carrier networks |
US7792920B2 (en) | 2004-04-30 | 2010-09-07 | Vulcan Inc. | Network-accessible control of one or more media devices |
JP4373271B2 (en) | 2004-05-14 | 2009-11-25 | 富士通株式会社 | Method and program for grasping network configuration of virtual LAN in node network |
GB2414623B (en) | 2004-05-27 | 2006-05-17 | 3Com Corp | Distributed bridging with synchronization of forwarding databases |
JP4397292B2 (en) | 2004-07-09 | 2010-01-13 | 富士通株式会社 | Control packet loop prevention method and bridge device using the same |
FR2873524B1 (en) | 2004-07-22 | 2006-10-27 | Alcatel Sa | LOCAL NETWORK WITH VIRTUAL GROUP (S) OF HEART EQUIPMENT WHICH IS CLEAR AT THE LEVEL TWO SWITCHING |
US7466712B2 (en) | 2004-07-30 | 2008-12-16 | Brocade Communications Systems, Inc. | System and method for providing proxy and translation domains in a fibre channel router |
US8819213B2 (en) | 2004-08-20 | 2014-08-26 | Extreme Networks, Inc. | System, method and apparatus for traffic mirror setup, service and security in communication networks |
US7463597B1 (en) | 2004-08-27 | 2008-12-09 | Juniper Networks, Inc. | Spanning tree protocol synchronization within virtual private networks |
US7558219B1 (en) | 2004-08-30 | 2009-07-07 | Juniper Networks, Inc. | Multicast trees for virtual private local area network (LAN) service multicast |
US8116307B1 (en) | 2004-09-23 | 2012-02-14 | Juniper Networks, Inc. | Packet structure for mirrored traffic flow |
US7764768B2 (en) | 2004-10-06 | 2010-07-27 | Alcatel-Lucent Usa Inc. | Providing CALEA/legal intercept information to law enforcement agencies for internet protocol multimedia subsystems (IMS) |
US7508757B2 (en) | 2004-10-15 | 2009-03-24 | Alcatel Lucent | Network with MAC table overflow protection |
US7801125B2 (en) | 2004-10-22 | 2010-09-21 | Cisco Technology, Inc. | Forwarding table reduction and multipath network forwarding |
US8238347B2 (en) | 2004-10-22 | 2012-08-07 | Cisco Technology, Inc. | Fibre channel over ethernet |
US7136550B2 (en) | 2004-10-28 | 2006-11-14 | Corning Incorporated | Single-fiber launch/receive system for biosensing applications |
GB2419701A (en) | 2004-10-29 | 2006-05-03 | Hewlett Packard Development Co | Virtual overlay infrastructure with dynamic control of mapping |
US8700799B2 (en) | 2004-11-12 | 2014-04-15 | Brocade Communications Systems, Inc. | Methods, devices and systems with improved zone merge operation by operating on a switch basis |
EP1657853A1 (en) | 2004-11-12 | 2006-05-17 | STMicroelectronics (Research & Development) Limited | Roaming network stations using a MAC address identifier to select a new access point |
US8005084B2 (en) | 2004-11-30 | 2011-08-23 | Broadcom Corporation | Mirroring in a network device |
US7808992B2 (en) | 2004-12-30 | 2010-10-05 | Cisco Technology, Inc. | Platform independent implementation of private VLANS |
US20070036178A1 (en) | 2005-02-02 | 2007-02-15 | Susan Hares | Layer 2 virtual switching environment |
US20060184937A1 (en) | 2005-02-11 | 2006-08-17 | Timothy Abels | System and method for centralized software management in virtual machines |
US7586895B2 (en) | 2005-04-01 | 2009-09-08 | Cisco Technology, Inc. | Performing extended lookups on MAC-based tables including level 3 multicast group destination addresses |
US7616578B2 (en) | 2005-04-11 | 2009-11-10 | Cisco Technology, Inc. | Forwarding traffic flow information using an intelligent line card |
US7673068B2 (en) | 2005-04-18 | 2010-03-02 | Alcatel Lucent | Method and system for implementing a high availability VLAN |
GB2425681A (en) | 2005-04-27 | 2006-11-01 | 3Com Corporaton | Access control by Dynamic Host Configuration Protocol snooping |
US7835370B2 (en) | 2005-04-28 | 2010-11-16 | Cisco Technology, Inc. | System and method for DSL subscriber identification over ethernet network |
US8751649B2 (en) | 2005-06-07 | 2014-06-10 | Extreme Networks | Port management system |
US20060285499A1 (en) | 2005-06-17 | 2006-12-21 | Broadcom Corporation | Loop detection for a network device |
US7571447B2 (en) | 2005-06-20 | 2009-08-04 | International Business Machines Corporation | Loose coupling of web services |
KR100653634B1 (en) | 2005-06-23 | 2006-12-06 | 조창환 | System and method for controlling a traffic of a net-work |
GB0516158D0 (en) | 2005-08-05 | 2005-09-14 | Univ Montfort | An apparatus and method for `non-contact' electrical impedance imaging |
US7937756B2 (en) | 2005-08-19 | 2011-05-03 | Cpacket Networks, Inc. | Apparatus and method for facilitating network security |
US20070053294A1 (en) | 2005-09-02 | 2007-03-08 | Michael Ho | Network load balancing apparatus, systems, and methods |
US7821949B2 (en) | 2005-09-12 | 2010-10-26 | Nortel Networks Limited | Forwarding plane data communications channel for ethernet transport networks |
US9143841B2 (en) | 2005-09-29 | 2015-09-22 | Brocade Communications Systems, Inc. | Federated management of intelligent service modules |
DE102005048585A1 (en) | 2005-10-06 | 2007-04-12 | Robert Bosch Gmbh | Subscriber and communication controller of a communication system and method for implementing a gateway functionality in a subscriber of a communication system |
CN100442772C (en) | 2005-10-19 | 2008-12-10 | 华为技术有限公司 | Bridge-connection transmitting method |
US9497600B2 (en) | 2005-10-28 | 2016-11-15 | Hewlett Packard Enterprise Development Lp | Service chaining |
US7697528B2 (en) | 2005-11-01 | 2010-04-13 | Nortel Networks Limited | Multilink trunking for encapsulated traffic |
WO2007068266A1 (en) | 2005-12-12 | 2007-06-21 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and devices for specifying the quality of service in a transmission of data packets |
US7716240B2 (en) | 2005-12-29 | 2010-05-11 | Nextlabs, Inc. | Techniques and system to deploy policies intelligently |
US20070177597A1 (en) | 2006-02-02 | 2007-08-02 | Yu Ju | Ethernet connection-based forwarding process |
US7835378B2 (en) | 2006-02-02 | 2010-11-16 | Cisco Technology, Inc. | Root node redundancy for multipoint-to-multipoint transport trees |
US7639605B2 (en) | 2006-02-08 | 2009-12-29 | Cisco Technology, Inc. | System and method for detecting and recovering from virtual switch link failures |
US8189575B2 (en) | 2006-03-13 | 2012-05-29 | Rockstar Bidco, L.P. | Modular scalable switch architecture |
US7948977B2 (en) | 2006-05-05 | 2011-05-24 | Broadcom Corporation | Packet routing with payload analysis, encapsulation and service module vectoring |
US8160080B1 (en) | 2006-05-08 | 2012-04-17 | Marvell Israel (M.I.S.L.) Ltd. | Implementation of reliable synchronization of distributed databases |
JP2007318553A (en) | 2006-05-26 | 2007-12-06 | Fujitsu Ltd | Network managing method |
US8018938B1 (en) | 2006-06-02 | 2011-09-13 | World Wide Packets, Inc. | Translating between a switching format and a transport format |
JP4834493B2 (en) | 2006-08-25 | 2011-12-14 | アラクサラネットワークス株式会社 | Network relay device and method for controlling network relay device |
CN100583825C (en) | 2006-08-30 | 2010-01-20 | 华为技术有限公司 | Method of generating symmetrical tree in the shortest path bridge |
US8169912B2 (en) | 2006-08-31 | 2012-05-01 | Futurewei Technologies, Inc. | System for dynamic bandwidth adjustment and trading among peers |
US8396945B2 (en) | 2006-09-11 | 2013-03-12 | Alcatel Lucent | Network management system with adaptive sampled proactive diagnostic capabilities |
US20080080517A1 (en) | 2006-09-28 | 2008-04-03 | At & T Corp. | System and method for forwarding traffic data in an MPLS VPN |
US8208463B2 (en) | 2006-10-24 | 2012-06-26 | Cisco Technology, Inc. | Subnet scoped multicast / broadcast packet distribution mechanism over a routed network |
US7697556B2 (en) | 2006-10-26 | 2010-04-13 | Telefonaktiebolaget L M Ericsson (Publ) | MAC (media access control) tunneling and control and method |
US7720889B1 (en) | 2006-10-31 | 2010-05-18 | Netapp, Inc. | System and method for nearly in-band search indexing |
WO2008056838A1 (en) | 2006-11-08 | 2008-05-15 | Chang Hwan Cho | System and method for controlling network traffic |
US20080112400A1 (en) | 2006-11-15 | 2008-05-15 | Futurewei Technologies, Inc. | System for Providing Both Traditional and Traffic Engineering Enabled Services |
US7599901B2 (en) | 2006-12-06 | 2009-10-06 | Microsoft Corporation | Processing data-centric business models |
US20080181243A1 (en) | 2006-12-15 | 2008-07-31 | Brocade Communications Systems, Inc. | Ethernet forwarding in high performance fabrics |
US20080159277A1 (en) | 2006-12-15 | 2008-07-03 | Brocade Communications Systems, Inc. | Ethernet over fibre channel |
US8973098B2 (en) | 2007-01-11 | 2015-03-03 | International Business Machines Corporation | System and method for virtualized resource configuration |
US7706255B1 (en) | 2007-01-29 | 2010-04-27 | Solace Systems, Inc. | Communications apparatus with redundant switching or backpressure mechanism |
US20080181196A1 (en) | 2007-01-31 | 2008-07-31 | Alcatel Lucent | Link aggregation across multiple chassis |
US20100046471A1 (en) | 2007-02-06 | 2010-02-25 | Mitsubishi Electric Corporation | Communication system, communication apparatus, wireless base station, and wireless terminal station |
JP4259581B2 (en) | 2007-02-07 | 2009-04-30 | 日立電線株式会社 | Switching hub and LAN system |
IL189514A (en) | 2007-02-14 | 2011-12-29 | Marvell Israel Misl Ltd | Logical bridging system and method |
US9661112B2 (en) | 2007-02-22 | 2017-05-23 | International Business Machines Corporation | System and methods for providing server virtualization assistance |
US8140696B2 (en) | 2007-03-12 | 2012-03-20 | International Business Machines Corporation | Layering serial attached small computer system interface (SAS) over ethernet |
US8077721B2 (en) | 2007-03-15 | 2011-12-13 | Cisco Technology, Inc. | Methods and apparatus providing two stage tunneling |
US7916741B2 (en) | 2007-04-02 | 2011-03-29 | William Marsh Rice University | System and method for preventing count-to-infinity problems in ethernet networks |
US8078704B2 (en) | 2007-04-12 | 2011-12-13 | Hewlett-Packard Development Company, L.P. | Provisioning of a service environment using web services |
US8301686B1 (en) | 2007-04-16 | 2012-10-30 | Citrix Systems, Inc. | Systems and methods for decentralized computing |
US7873038B2 (en) | 2007-04-30 | 2011-01-18 | Hewlett-Packard Development Company, L.P. | Packet processing |
US7724674B2 (en) | 2007-05-16 | 2010-05-25 | Simula Innovations As | Deadlock free network routing |
JP4862743B2 (en) | 2007-05-17 | 2012-01-25 | 日本電気株式会社 | Node, communication method and node program |
US20080298248A1 (en) | 2007-05-28 | 2008-12-04 | Guenter Roeck | Method and Apparatus For Computer Network Bandwidth Control and Congestion Management |
US7945941B2 (en) | 2007-06-01 | 2011-05-17 | Cisco Technology, Inc. | Flexible access control policy enforcement |
US8054833B2 (en) | 2007-06-05 | 2011-11-08 | Hewlett-Packard Development Company, L.P. | Packet mirroring |
US20080310342A1 (en) | 2007-06-12 | 2008-12-18 | Cisco Technology, Inc. | Addressing Messages in a Two-Tier Network |
US7898959B1 (en) | 2007-06-28 | 2011-03-01 | Marvell Israel (Misl) Ltd. | Method for weighted load-balancing among network interfaces |
US8615008B2 (en) | 2007-07-11 | 2013-12-24 | Foundry Networks Llc | Duplicating network traffic through transparent VLAN flooding |
GB0713785D0 (en) | 2007-07-16 | 2007-08-22 | Cellfire Security Technologies | Voice over IP system |
US7836332B2 (en) | 2007-07-18 | 2010-11-16 | Hitachi, Ltd. | Method and apparatus for managing virtual ports on storage systems |
US20090044270A1 (en) | 2007-08-07 | 2009-02-12 | Asaf Shelly | Network element and an infrastructure for a network risk management system |
US7864712B2 (en) | 2007-07-20 | 2011-01-04 | Cisco Technology, Inc. | Preventing loops in networks operating different protocols to provide loop-free topology |
US8166205B2 (en) | 2007-07-31 | 2012-04-24 | Cisco Technology, Inc. | Overlay transport virtualization |
US7729296B1 (en) | 2007-09-07 | 2010-06-01 | Force 10 Networks, Inc. | Distributed BPDU processing for spanning tree protocols |
US20090080345A1 (en) | 2007-09-21 | 2009-03-26 | Ericsson, Inc. | Efficient multipoint distribution tree construction for shortest path bridging |
US8798056B2 (en) | 2007-09-24 | 2014-08-05 | Intel Corporation | Method and system for virtual port communications |
CA2926677C (en) | 2007-09-26 | 2020-07-14 | Nicira, Inc. | Network operating system for managing and securing networks |
US20090079560A1 (en) | 2007-09-26 | 2009-03-26 | General Electric Company | Remotely monitoring railroad equipment using network protocols |
US7751329B2 (en) * | 2007-10-03 | 2010-07-06 | Avaya Inc. | Providing an abstraction layer in a cluster switch that includes plural switches |
JP5030063B2 (en) | 2007-10-05 | 2012-09-19 | 本田技研工業株式会社 | Navigation device and navigation system |
US7975033B2 (en) | 2007-10-23 | 2011-07-05 | Virtudatacenter Holdings, L.L.C. | System and method for initializing and maintaining a series of virtual local area networks contained in a clustered computer system |
US7916647B2 (en) | 2007-11-07 | 2011-03-29 | Brocade Communications Systems, Inc. | Automatic adjustment of logical channels in a fibre channel network |
WO2009064407A1 (en) | 2007-11-16 | 2009-05-22 | Ericsson Ab | Method and system for telecommunications including self-organizing scalable ethernet using is-is hierarchy |
US8117495B2 (en) | 2007-11-26 | 2012-02-14 | Stratus Technologies Bermuda Ltd | Systems and methods of high availability cluster environment failover protection |
US20100272107A1 (en) | 2007-11-26 | 2010-10-28 | Oktavian Papp | Technique for address resolution in a data transmission network |
US8194674B1 (en) | 2007-12-20 | 2012-06-05 | Quest Software, Inc. | System and method for aggregating communications and for translating between overlapping internal network addresses and unique external network addresses |
US7796593B1 (en) | 2007-12-21 | 2010-09-14 | Juniper Networks, Inc. | Router using internal flood groups for flooding VPLS traffic |
US7860093B2 (en) | 2007-12-24 | 2010-12-28 | Cisco Technology, Inc. | Fast multicast convergence at secondary designated router or designated forwarder |
US8018841B2 (en) | 2007-12-31 | 2011-09-13 | Ciena Corporation | Interworking an ethernet ring network and an ethernet network with traffic engineered trunks |
JP2009187368A (en) | 2008-02-07 | 2009-08-20 | Hitachi Ltd | Method for controlling sharing of usb port |
US20090222879A1 (en) | 2008-03-03 | 2009-09-03 | Microsoft Corporation | Super policy in information protection systems |
US20090245137A1 (en) | 2008-03-03 | 2009-10-01 | Green Hills Software, Inc. | Highly available virtual stacking architecture |
US20110044352A1 (en) | 2008-03-04 | 2011-02-24 | France Telecom | Technique for determining a point-to-multipoint tree linking a root node to a plurality of leaf nodes |
US8230069B2 (en) | 2008-03-04 | 2012-07-24 | International Business Machines Corporation | Server and storage-aware method for selecting virtual machine migration targets |
US7801137B2 (en) | 2008-03-11 | 2010-09-21 | Cisco Technology, Inc. | Receiver-based construction of point-to-multipoint trees using path computation elements in a computer network |
US7792148B2 (en) | 2008-03-31 | 2010-09-07 | International Business Machines Corporation | Virtual fibre channel over Ethernet switch |
PT2615085E (en) | 2008-03-31 | 2015-10-09 | Vertex Pharma | Pyridyl derivatives as cftr modulators |
US8743740B2 (en) | 2008-04-08 | 2014-06-03 | At&T Intellectual Property I, L.P. | Methods and apparatus to implement a partial mesh virtual private local area network service |
US7911982B1 (en) | 2008-05-01 | 2011-03-22 | Juniper Networks, Inc. | Configuring networks including spanning trees |
US8625615B2 (en) | 2008-05-16 | 2014-01-07 | Nec Corporation | PCI express switch, PCI express system, and network control method |
US8195774B2 (en) | 2008-05-23 | 2012-06-05 | Vmware, Inc. | Distributed virtual switch for virtualized computer systems |
US8160063B2 (en) | 2008-06-09 | 2012-04-17 | Microsoft Corporation | Data center interconnect and traffic engineering |
US8565248B2 (en) | 2008-06-26 | 2013-10-22 | Cisco Technology, Inc. | Pure control-plane approach for on-path connection admission control operations in multiprotocol label switching virtual private networks |
US7873711B2 (en) | 2008-06-27 | 2011-01-18 | International Business Machines Corporation | Method, system and program product for managing assignment of MAC addresses in a virtual machine environment |
US7941539B2 (en) | 2008-06-30 | 2011-05-10 | Oracle America, Inc. | Method and system for creating a virtual router in a blade chassis to maintain connectivity |
KR101508794B1 (en) | 2008-07-09 | 2015-04-06 | 삼성전자주식회사 | Method for selectively securing records in a ndef message |
US8102791B2 (en) | 2008-07-25 | 2012-01-24 | Newport Media, Inc. | Interleaver address generation in turbo decoders for mobile multimedia multicast system communication systems |
US8102781B2 (en) | 2008-07-31 | 2012-01-24 | Cisco Technology, Inc. | Dynamic distribution of virtual machines in a communication network |
US9426095B2 (en) | 2008-08-28 | 2016-08-23 | International Business Machines Corporation | Apparatus and method of switching packets between virtual ports |
US8259569B2 (en) | 2008-09-09 | 2012-09-04 | Cisco Technology, Inc. | Differentiated services for unicast and multicast frames in layer 2 topologies |
US8134922B2 (en) | 2008-09-12 | 2012-03-13 | Cisco Technology, Inc. | Reducing flooding in a bridged network |
US8392606B2 (en) | 2008-09-23 | 2013-03-05 | Synapse Wireless, Inc. | Wireless networks and methods using multiple valid network identifiers |
US7944812B2 (en) | 2008-10-20 | 2011-05-17 | International Business Machines Corporation | Redundant intermediary switch solution for detecting and managing fibre channel over ethernet FCoE switch failures |
US8571052B2 (en) | 2008-10-24 | 2013-10-29 | International Business Machines Corporation | Determining the configuration of an ethernet fabric |
US9100269B2 (en) | 2008-10-28 | 2015-08-04 | Rpx Clearinghouse Llc | Provisioned provider link state bridging (PLSB) with routed back-up |
US7962647B2 (en) | 2008-11-24 | 2011-06-14 | Vmware, Inc. | Application delivery control module for virtual network switch |
US8392496B2 (en) | 2008-12-19 | 2013-03-05 | Watchguard Technologies, Inc. | Cluster architecture for network security processing |
US7929554B2 (en) | 2008-12-23 | 2011-04-19 | Cisco Technology, Inc. | Optimized forwarding for provider backbone bridges with both I and B components (IB-PBB) |
US8509248B2 (en) | 2008-12-29 | 2013-08-13 | Juniper Networks, Inc. | Routing frames in a computer network using bridge identifiers |
US8331362B2 (en) | 2008-12-30 | 2012-12-11 | Juniper Networks, Inc. | Methods and apparatus for distributed dynamic network provisioning |
US8054832B1 (en) | 2008-12-30 | 2011-11-08 | Juniper Networks, Inc. | Methods and apparatus for routing between virtual resources based on a routing location policy |
US8255496B2 (en) | 2008-12-30 | 2012-08-28 | Juniper Networks, Inc. | Method and apparatus for determining a network topology during network provisioning |
US7820853B2 (en) | 2008-12-31 | 2010-10-26 | Celanese International Corporation | Integrated process for the production of vinyl acetate from acetic acid via ethyl acetate |
US8336079B2 (en) | 2008-12-31 | 2012-12-18 | Hytrust, Inc. | Intelligent security control system for virtualized ecosystems |
JP5168166B2 (en) | 2009-01-21 | 2013-03-21 | 富士通株式会社 | Communication apparatus and communication control method |
US9043621B2 (en) | 2009-01-21 | 2015-05-26 | Hitachi, Ltd. | Power-saving network management server, network system, and method of determining supply of power |
US8098572B2 (en) | 2009-02-03 | 2012-01-17 | Google Inc. | Interface monitoring for link aggregation |
KR20110126670A (en) | 2009-02-13 | 2011-11-23 | 에이디씨 텔레커뮤니케이션스 인코포레이티드 | Inter-networking devices for use with physical layer information |
US8213336B2 (en) | 2009-02-23 | 2012-07-03 | Cisco Technology, Inc. | Distributed data center access switch |
US8274980B2 (en) | 2009-02-26 | 2012-09-25 | International Business Machines Corporation | Ethernet link aggregation |
US7787480B1 (en) | 2009-03-04 | 2010-08-31 | Juniper Networks, Inc. | Routing frames in a trill network using service VLAN identifiers |
US8238340B2 (en) | 2009-03-06 | 2012-08-07 | Futurewei Technologies, Inc. | Transport multiplexer—mechanisms to force ethernet traffic from one domain to be switched in a different (external) domain |
US8155150B1 (en) | 2009-03-11 | 2012-04-10 | Juniper Networks, Inc. | Cooperative MAC learning/aging in highly distributed forwarding system |
US7912091B1 (en) | 2009-03-18 | 2011-03-22 | Extreme Networks, Inc. | Traffic forwarding in a traffic-engineered link aggregation group |
US8665886B2 (en) | 2009-03-26 | 2014-03-04 | Brocade Communications Systems, Inc. | Redundant host connection in a routed network |
US8918631B1 (en) | 2009-03-31 | 2014-12-23 | Juniper Networks, Inc. | Methods and apparatus for dynamic automated configuration within a control plane of a switch fabric |
CA3002975C (en) | 2009-04-01 | 2020-07-14 | Nicira, Inc. | Method and apparatus for implementing and managing virtual switches |
US8213313B1 (en) | 2009-04-15 | 2012-07-03 | Tellabs Operations, Inc. | Methods and apparatus for shared layer 3 application card in multi-service router |
US8000336B2 (en) | 2009-04-21 | 2011-08-16 | Voltaire Ltd. | Spanning tree root selection in a hierarchical network |
US8116213B2 (en) | 2009-04-24 | 2012-02-14 | Verizon Patent And Licensing Inc. | Tracing routes and protocols |
US8027354B1 (en) | 2009-04-29 | 2011-09-27 | Cisco Technology, Inc. | Network consolidation for virtualized servers |
US8874709B2 (en) | 2009-05-01 | 2014-10-28 | Futurewei Technologies, Inc. | Automatic subnet creation in networks that support dynamic ethernet-local area network services for use by operation, administration, and maintenance |
US8429647B2 (en) | 2009-05-06 | 2013-04-23 | Vmware, Inc. | Virtual machine migration across network by publishing routes to the associated virtual networks via virtual router after the start of migration of the virtual machine |
US20100287262A1 (en) | 2009-05-08 | 2010-11-11 | Uri Elzur | Method and system for guaranteed end-to-end data flows in a local networking domain |
US9282057B2 (en) | 2009-05-11 | 2016-03-08 | Brocade Communication Systems, Inc. | Flexible stacking port |
US8351431B2 (en) * | 2009-05-13 | 2013-01-08 | Avaya Inc. | Method and apparatus for providing fast reroute of a multicast packet within a network element to an available port associated with a multi-link trunk |
US8472443B2 (en) | 2009-05-15 | 2013-06-25 | Cisco Technology | Port grouping for association with virtual interfaces |
US8165122B2 (en) | 2009-05-26 | 2012-04-24 | Alcatel Lucent | System and method for converting unicast client requests into multicast client requests |
US8170038B2 (en) | 2009-05-27 | 2012-05-01 | International Business Machines Corporation | Two-layer switch apparatus to avoid first layer inter-switch link data traffic in steering packets through bump-in-the-wire service applications |
US8174984B2 (en) | 2009-05-29 | 2012-05-08 | Oracle America, Inc. | Managing traffic on virtualized lanes between a network switch and a virtual machine |
US7944860B2 (en) | 2009-06-04 | 2011-05-17 | Cisco Technology, Inc. | Preventing loss of network traffic due to inconsistent configurations within the network |
US8199753B2 (en) | 2009-06-05 | 2012-06-12 | Juniper Networks, Inc. | Forwarding frames in a computer network using shortest path bridging |
US8102760B2 (en) | 2009-06-30 | 2012-01-24 | Alcatel Lucent | Method for reconvergence after failure in a dual-homing network environment |
US8351352B1 (en) | 2009-07-15 | 2013-01-08 | Eastlake Iii Donald E | Methods and apparatus for RBridge hop-by-hop compression and frame aggregation |
US8204061B1 (en) | 2009-07-23 | 2012-06-19 | Cisco Technology, Inc. | Virtual port channel switches with distributed control planes |
US8125928B2 (en) | 2009-07-24 | 2012-02-28 | Juniper Networks, Inc. | Routing frames in a shortest path computer network for a multi-homed legacy bridge node |
US8341725B2 (en) | 2009-07-30 | 2012-12-25 | Calix, Inc. | Secure DHCP processing for layer two access networks |
US8503329B2 (en) | 2009-08-05 | 2013-08-06 | Cisco Technology, Inc. | Signaling of attachment circuit status and automatic discovery of inter-chassis communication peers |
US8504690B2 (en) | 2009-08-07 | 2013-08-06 | Broadcom Corporation | Method and system for managing network power policy and configuration of data center bridging |
US8175107B1 (en) | 2009-08-18 | 2012-05-08 | Hewlett-Packard Development Company, L.P. | Network routing based on MAC address subnetting |
IL200504A0 (en) | 2009-08-20 | 2011-08-01 | Eci Telecom Ltd | Technique for dual homing interconnection between communication networks |
US8369332B2 (en) | 2009-08-21 | 2013-02-05 | Alcatel Lucent | Server-side load balancing using parent-child link aggregation groups |
US8706905B1 (en) | 2009-08-24 | 2014-04-22 | Qlogic, Corporation | Method and system for routing information in a network |
US8339994B2 (en) | 2009-08-27 | 2012-12-25 | Brocade Communications Systems, Inc. | Defining an optimal topology for a group of logical switches |
US8369347B2 (en) | 2009-09-14 | 2013-02-05 | Futurewei Technologies, Inc. | Fiber channel over Ethernet and fiber channel switching based on Ethernet switch fabrics |
US8599850B2 (en) | 2009-09-21 | 2013-12-03 | Brocade Communications Systems, Inc. | Provisioning single or multistage networks using ethernet service instances (ESIs) |
US8914598B2 (en) | 2009-09-24 | 2014-12-16 | Vmware, Inc. | Distributed storage resource scheduler and load balancer |
US8599864B2 (en) | 2009-10-08 | 2013-12-03 | Brocade Communications Systems, Inc. | Transit switches in a network of logical switches |
US20110085560A1 (en) | 2009-10-12 | 2011-04-14 | Dell Products L.P. | System and Method for Implementing a Virtual Switch |
US8693485B2 (en) | 2009-10-14 | 2014-04-08 | Dell Products, Lp | Virtualization aware network switch |
CN102474449B (en) | 2009-11-02 | 2016-05-04 | 马维尔国际贸易有限公司 | Switch based on virtual interface and method |
US8917625B2 (en) | 2009-11-10 | 2014-12-23 | Broadcom Corporation | Mapping quality of service (QOS) from a wireless network to a wired network |
US20110134802A1 (en) | 2009-12-09 | 2011-06-09 | Cisco Technology, Inc. | Determining A Routing Tree For Networks With Different Routing Protocols |
US8270420B2 (en) | 2009-12-15 | 2012-09-18 | Hewlett-Packard Development Company, L.P. | iSCSI to FCoE gateway |
WO2011074516A1 (en) | 2009-12-15 | 2011-06-23 | 日本電気株式会社 | Network system, method for controlling same, and controller |
US8705513B2 (en) | 2009-12-15 | 2014-04-22 | At&T Intellectual Property I, L.P. | Methods and apparatus to communicatively couple virtual private networks to virtual machines within distributive computing networks |
US8295291B1 (en) | 2009-12-21 | 2012-10-23 | Juniper Networks, Inc. | Computation of next hops within layer two networks |
US8161156B2 (en) | 2009-12-30 | 2012-04-17 | Verizon Patent And Licensing, Inc. | Feature delivery packets for peer-to-peer based feature network |
WO2011081020A1 (en) | 2010-01-04 | 2011-07-07 | 日本電気株式会社 | Network system, controller, network control method |
US8446817B2 (en) | 2010-01-19 | 2013-05-21 | Cisco Technology, Inc. | Distributed virtual fibre channel over Ethernet forwarder |
JP5493926B2 (en) | 2010-02-01 | 2014-05-14 | 日本電気株式会社 | Interface control method, interface control method, and interface control program |
US8619595B2 (en) | 2010-02-05 | 2013-12-31 | Cisco Technology, Inc. | Fault isolation in trill networks |
CN102158386B (en) | 2010-02-11 | 2015-06-03 | 威睿公司 | Distributed load balance for system management program |
US8996720B2 (en) | 2010-03-16 | 2015-03-31 | Brocade Communications Systems, Inc. | Method and apparatus for mirroring frames to a remote diagnostic system |
US8873401B2 (en) | 2010-03-16 | 2014-10-28 | Futurewei Technologies, Inc. | Service prioritization in link state controlled layer two networks |
US8369335B2 (en) | 2010-03-24 | 2013-02-05 | Brocade Communications Systems, Inc. | Method and system for extending routing domain to non-routing end stations |
JP5190084B2 (en) | 2010-03-30 | 2013-04-24 | 株式会社日立製作所 | Virtual machine migration method and system |
US8249069B2 (en) | 2010-03-30 | 2012-08-21 | Cisco Technology, Inc. | Forwarding multi-destination packets in a network with virtual port channels |
US8599854B2 (en) | 2010-04-16 | 2013-12-03 | Cisco Technology, Inc. | Method of identifying destination in a virtual environment |
US8611352B2 (en) | 2010-04-20 | 2013-12-17 | Marvell World Trade Ltd. | System and method for adapting a packet processing pipeline |
US8345692B2 (en) | 2010-04-27 | 2013-01-01 | Cisco Technology, Inc. | Virtual switching overlay for cloud computing |
US9461840B2 (en) | 2010-06-02 | 2016-10-04 | Brocade Communications Systems, Inc. | Port profile management for virtual cluster switching |
US9270486B2 (en) | 2010-06-07 | 2016-02-23 | Brocade Communications Systems, Inc. | Name services for virtual cluster switching |
US9231890B2 (en) | 2010-06-08 | 2016-01-05 | Brocade Communications Systems, Inc. | Traffic management for virtual cluster switching |
US8867552B2 (en) | 2010-05-03 | 2014-10-21 | Brocade Communications Systems, Inc. | Virtual cluster switching |
US8989186B2 (en) | 2010-06-08 | 2015-03-24 | Brocade Communication Systems, Inc. | Virtual port grouping for virtual cluster switching |
US8625616B2 (en) | 2010-05-11 | 2014-01-07 | Brocade Communications Systems, Inc. | Converged network extension |
US8520595B2 (en) | 2010-05-04 | 2013-08-27 | Cisco Technology, Inc. | Routing to the access layer to support mobility of internet protocol devices |
US8335236B2 (en) | 2010-05-06 | 2012-12-18 | Cisco Technology, Inc. | FCoE isolated port channels and FCoE session resynchronization in vPC/MCEC environments using DCBXP |
US8503307B2 (en) | 2010-05-10 | 2013-08-06 | Hewlett-Packard Development Company, L.P. | Distributing decision making in a centralized flow routing system |
US8724456B1 (en) | 2010-05-19 | 2014-05-13 | Juniper Networks, Inc. | Network path selection for multi-homed edges to ensure end-to-end resiliency |
US9491085B2 (en) | 2010-05-24 | 2016-11-08 | At&T Intellectual Property I, L.P. | Methods and apparatus to route control packets based on address partitioning |
US8667171B2 (en) | 2010-05-28 | 2014-03-04 | Microsoft Corporation | Virtual data center allocation with bandwidth guarantees |
CN102577331B (en) | 2010-05-28 | 2015-08-05 | 华为技术有限公司 | Virtual 2nd layer and make its extendible mechanism |
US9608833B2 (en) | 2010-06-08 | 2017-03-28 | Brocade Communications Systems, Inc. | Supporting multiple multicast trees in trill networks |
US20110299533A1 (en) | 2010-06-08 | 2011-12-08 | Brocade Communications Systems, Inc. | Internal virtual network identifier and internal policy identifier |
US10033650B2 (en) | 2010-06-08 | 2018-07-24 | Brocade Communication Systems Llc | Preserving quality of service across trill networks |
US9246703B2 (en) | 2010-06-08 | 2016-01-26 | Brocade Communications Systems, Inc. | Remote port mirroring |
US9806906B2 (en) | 2010-06-08 | 2017-10-31 | Brocade Communications Systems, Inc. | Flooding packets on a per-virtual-network basis |
US9628293B2 (en) | 2010-06-08 | 2017-04-18 | Brocade Communications Systems, Inc. | Network layer multicasting in trill networks |
US8446914B2 (en) | 2010-06-08 | 2013-05-21 | Brocade Communications Systems, Inc. | Method and system for link aggregation across multiple switches |
US8897134B2 (en) | 2010-06-25 | 2014-11-25 | Telefonaktiebolaget L M Ericsson (Publ) | Notifying a controller of a change to a packet forwarding configuration of a network element over a communication channel |
JP5830093B2 (en) | 2010-06-29 | 2015-12-09 | ホアウェイ・テクノロジーズ・カンパニー・リミテッド | Asymmetric network address encapsulation |
US8588081B2 (en) | 2010-07-14 | 2013-11-19 | Cisco Technology, Inc. | Monitoring a flow set to detect faults |
US8873551B2 (en) | 2010-07-30 | 2014-10-28 | Cisco Technology, Inc. | Multi-destination forwarding in network clouds which include emulated switches |
US8472447B2 (en) | 2010-08-04 | 2013-06-25 | Alcatel Lucent | IP multicast snooping and routing with multi-chassis link aggregation |
US8767735B2 (en) | 2010-08-04 | 2014-07-01 | Alcatel Lucent | System and method for multi-chassis link aggregation |
US9049098B2 (en) | 2010-08-05 | 2015-06-02 | Cisco Technology, Inc. | Discovery of services provided by application nodes in a network |
CN102404181B (en) | 2010-09-08 | 2014-10-08 | 华为技术有限公司 | Address corresponding relationship sending method of layer 2 protocol utilizing link state routing |
US8953621B2 (en) | 2010-09-10 | 2015-02-10 | Futurewei Technologies, Inc. | Specifying priority on a virtual station interface discovery and configuration protocol response |
US8665267B2 (en) | 2010-09-24 | 2014-03-04 | Adobe Systems Incorporated | System and method for generating 3D surface patches from unconstrained 3D curves |
US8705502B2 (en) | 2010-10-20 | 2014-04-22 | Cisco Technology, Inc. | Using encapsulation to enable 802.1 bridging across 802.11 links |
US20120099602A1 (en) | 2010-10-25 | 2012-04-26 | Brocade Communications Systems, Inc. | End-to-end virtualization |
US8634297B2 (en) | 2010-11-01 | 2014-01-21 | Cisco Technology, Inc. | Probing specific customer flow in layer-2 multipath networks |
US8583978B2 (en) | 2010-11-09 | 2013-11-12 | Cisco Technology, Inc. | Multicast message retransmission |
US8756602B2 (en) | 2010-11-14 | 2014-06-17 | Brocade Communications Systems, Inc. | Virtual machine and application migration over local and wide area networks without timeout |
US8762668B2 (en) | 2010-11-18 | 2014-06-24 | Hitachi, Ltd. | Multipath switching over multiple storage systems |
US8660005B2 (en) | 2010-11-30 | 2014-02-25 | Marvell Israel (M.I.S.L) Ltd. | Load balancing hash computation for network switches |
US8806031B1 (en) | 2010-12-15 | 2014-08-12 | Juniper Networks, Inc. | Systems and methods for automatically detecting network elements |
US8521884B2 (en) | 2010-12-15 | 2013-08-27 | Industrial Technology Research Institute | Network system and method of address resolution |
US20120163164A1 (en) | 2010-12-27 | 2012-06-28 | Brocade Communications Systems, Inc. | Method and system for remote load balancing in high-availability networks |
US8559335B2 (en) | 2011-01-07 | 2013-10-15 | Jeda Networks, Inc. | Methods for creating virtual links between fibre channel over ethernet nodes for converged network adapters |
US8776207B2 (en) | 2011-02-16 | 2014-07-08 | Fortinet, Inc. | Load balancing in a network with session information |
US8755383B2 (en) | 2011-03-21 | 2014-06-17 | Avaya, Inc. | Usage of masked ethernet addresses between transparent interconnect of lots of links (TRILL) routing bridges |
US8761005B2 (en) | 2011-04-26 | 2014-06-24 | Dell Products L.P. | Multi-chassis link aggregation on network devices |
US9054999B2 (en) | 2012-05-09 | 2015-06-09 | International Business Machines Corporation | Static TRILL routing |
US20120287785A1 (en) * | 2011-05-14 | 2012-11-15 | International Business Machines Corporation | Data traffic handling in a distributed fabric protocol (dfp) switching network architecture |
US8605626B2 (en) | 2011-05-18 | 2013-12-10 | Cisco Technology, Inc. | Method and apparatus for preserving extensions in multi-vendor trill networks |
US20120294192A1 (en) | 2011-05-19 | 2012-11-22 | Hitachi, Ltd. | Method and apparatus of connectivity discovery between network switch and server based on vlan identifiers |
WO2012166139A1 (en) | 2011-06-02 | 2012-12-06 | Hewlett-Packard Development Company, L.P. | Network virtualization |
EP2723020B1 (en) | 2011-06-17 | 2019-11-06 | Huawei Technologies Co., Ltd. | Method and ethernet switching device for detecting loop position in ethernet |
US9497073B2 (en) | 2011-06-17 | 2016-11-15 | International Business Machines Corporation | Distributed link aggregation group (LAG) for a layer 2 fabric |
US9736065B2 (en) | 2011-06-24 | 2017-08-15 | Cisco Technology, Inc. | Level of hierarchy in MST for traffic localization and load balancing |
US9288288B2 (en) | 2011-06-27 | 2016-03-15 | Marvell Israel (M.I.S.L) Ltd. | FCoE over trill |
US20130003738A1 (en) | 2011-06-29 | 2013-01-03 | Brocade Communications Systems, Inc. | Trill based router redundancy |
US8559302B2 (en) | 2011-06-29 | 2013-10-15 | Fujitsu Limited | Systems and methods for distributed service protection across plug-in units |
US8619635B2 (en) | 2011-06-29 | 2013-12-31 | Telefonaktiebolaget L M Ericsson (Publ) | E-tree using two pseudowires between edge routers with enhanced forwarding methods and systems |
US8467375B2 (en) | 2011-07-07 | 2013-06-18 | Ciena Corporation | Hybrid packet-optical private network systems and methods |
US8705551B2 (en) | 2011-07-27 | 2014-04-22 | Fujitsu Limited | Method and system for management of flood traffic over multiple 0:N link aggregation groups |
US20130034015A1 (en) | 2011-08-05 | 2013-02-07 | International Business Machines Corporation | Automated network configuration in a dynamic virtual environment |
US8966499B2 (en) | 2011-09-09 | 2015-02-24 | Microsoft Technology Licensing, Llc | Virtual switch extensibility |
US9185056B2 (en) | 2011-09-20 | 2015-11-10 | Big Switch Networks, Inc. | System and methods for controlling network traffic through virtual switches |
US8885643B2 (en) | 2011-11-04 | 2014-11-11 | Futurewei Technologies, Inc. | Method for multicast flow routing selection |
US9450870B2 (en) | 2011-11-10 | 2016-09-20 | Brocade Communications Systems, Inc. | System and method for flow management in software-defined networks |
CN103139037B (en) | 2011-11-30 | 2016-05-18 | 国际商业机器公司 | For realizing the method and apparatus of VLAN flexibly |
US8942139B2 (en) | 2011-12-07 | 2015-01-27 | International Business Machines Corporation | Support for converged traffic over ethernet link aggregation (LAG) |
US8995272B2 (en) | 2012-01-26 | 2015-03-31 | Brocade Communication Systems, Inc. | Link aggregation in software-defined networks |
KR101718824B1 (en) | 2012-02-22 | 2017-03-22 | 노키아 솔루션스 앤드 네트웍스 오와이 | Controlling access |
US9154416B2 (en) | 2012-03-22 | 2015-10-06 | Brocade Communications Systems, Inc. | Overlay tunnel in a fabric switch |
CN104221332B (en) | 2012-03-28 | 2017-12-19 | 富士通株式会社 | LAN multiplexer |
CN102594711B (en) | 2012-03-28 | 2014-11-26 | 杭州华三通信技术有限公司 | Message forwarding method and edge device therefor |
US9184995B2 (en) | 2012-04-11 | 2015-11-10 | Gigamon Inc. | Traffic visibility in an open networking environment |
US9081603B2 (en) | 2012-07-09 | 2015-07-14 | Cisco Technology, Inc. | Packet forwarding optimization with virtual machine mobility by comparing device identifiers to determine VM movement |
US9143439B2 (en) | 2012-07-23 | 2015-09-22 | Cisco Technology, Inc. | System and method for cluster link aggregation control in a network environment |
CN102801599B (en) | 2012-07-26 | 2015-09-30 | 华为技术有限公司 | A kind of communication means and system |
US8855117B2 (en) | 2012-08-08 | 2014-10-07 | Cisco Technology, Inc. | Scalable media access control protocol synchronization techniques for fabric extender based emulated switch deployments |
US8937865B1 (en) | 2012-08-21 | 2015-01-20 | Juniper Networks, Inc. | Scheduling traffic over aggregated bundles of links |
US9602430B2 (en) | 2012-08-21 | 2017-03-21 | Brocade Communications Systems, Inc. | Global VLANs for fabric switches |
-
2014
- 2014-05-21 US US14/284,212 patent/US9565028B2/en active Active
- 2014-05-29 WO PCT/US2014/040060 patent/WO2014200707A2/en active Application Filing
Also Published As
Publication number | Publication date |
---|---|
WO2014200707A3 (en) | 2015-02-05 |
US9565028B2 (en) | 2017-02-07 |
US20140362854A1 (en) | 2014-12-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9565028B2 (en) | Ingress switch multicast distribution in a fabric switch | |
US9660939B2 (en) | Protection switching over a virtual link aggregation | |
US10063473B2 (en) | Method and system for facilitating switch virtualization in a network of interconnected switches | |
US9871676B2 (en) | Scalable gateways for a fabric switch | |
US9887916B2 (en) | Overlay tunnel in a fabric switch | |
US9270572B2 (en) | Layer-3 support in TRILL networks | |
US9742693B2 (en) | Dynamic service insertion in a fabric switch | |
US9407533B2 (en) | Multicast in a trill network | |
US8885641B2 (en) | Efficient trill forwarding | |
US10075394B2 (en) | Virtual link aggregations across multiple fabric switches | |
US9806949B2 (en) | Transparent interconnection of Ethernet fabric switches | |
US10171303B2 (en) | IP-based interconnection of switches with a logical chassis | |
US20170310582A1 (en) | Dynamic multi-destination traffic management in a distributed tunnel endpoint | |
US9699117B2 (en) | Integrated fibre channel support in an ethernet fabric switch | |
US20150023359A1 (en) | Edge extension of an ethernet fabric switch |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14733875 Country of ref document: EP Kind code of ref document: A2 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 14733875 Country of ref document: EP Kind code of ref document: A2 |