CSC/ECE 506 Spring 2010/ch 12 PP: Difference between revisions

From Expertiza_Wiki
Jump to navigation Jump to search
No edit summary
Line 36: Line 36:




 
''Figure 1: Classification of Interconnection Networks''
===Shared-Medium Networks===
===Shared-Medium Networks===


Line 46: Line 46:
[[Image:Ring.gif]]
[[Image:Ring.gif]]


''Figure 2: Token Ring''


'''Implementation examples:'''  
'''Implementation examples:'''  
Line 59: Line 61:
[[Image:Token_bus.gif]]
[[Image:Token_bus.gif]]


''Figure3: Token Bus''


Nodes are physically connected as a bus, but logically form a ring with tokens passed around to determine the turns for sending. A token is passed around the network nodes and only the node possessing the token may transmit. If a node doesn't have anything to send, the token is passed on to the next node on the virtual ring.
Nodes are physically connected as a bus, but logically form a ring with tokens passed around to determine the turns for sending. A token is passed around the network nodes and only the node possessing the token may transmit. If a node doesn't have anything to send, the token is passed on to the next node on the virtual ring.
Line 71: Line 74:


[[Image:Backplane.jpg]]
[[Image:Backplane.jpg]]
''Figure3: Backplane Bus''


Simplest interconnection structure for bus-based parallel computers. Commonly used to connect processors and memory modules to form UMA architecture. Many techniques are used for information transfers like synchronous, asynchronous, multiplexed and non-multiplexed.  
Simplest interconnection structure for bus-based parallel computers. Commonly used to connect processors and memory modules to form UMA architecture. Many techniques are used for information transfers like synchronous, asynchronous, multiplexed and non-multiplexed.  
Line 90: Line 95:
----
----


Bus based systems are not '''scalable''' since bus becomes a bottleneck when more processors are added. '''Direct network or point-to-point network''' scales well with large number of processors. A direct network consists of a set of nodes, each directly connected to a subset of other nodes. Figure__ shows a generic node.  
Bus based systems are not '''scalable''' since bus becomes a bottleneck when more processors are added. '''Direct network or point-to-point network''' scales well with large number of processors. A direct network consists of a set of nodes, each directly connected to a subset of other nodes. ''Figure 4'' shows a generic node.  


[[Image:Node.jpg]]
[[Image:Node.jpg]]


''Figure4: A generic node''


Each '''Router''' is connected to its neighbor’s routers through a '''unidirectional''' or '''bidirectional''' channel and handles the message communication among nodes. '''Internal channels''' connect local memory/processor to the router. '''External channels''' are used for communication between the routers. A direct network is formed by connecting input channels of one node to output channels of other nodes. Every input channel is paired with a corresponding output channel and there are many ways to interconnect these nodes such that every node in the network is able to reach every other node.
Each '''Router''' is connected to its neighbor’s routers through a '''unidirectional''' or '''bidirectional''' channel and handles the message communication among nodes. '''Internal channels''' connect local memory/processor to the router. '''External channels''' are used for communication between the routers. A direct network is formed by connecting input channels of one node to output channels of other nodes. Every input channel is paired with a corresponding output channel and there are many ways to interconnect these nodes such that every node in the network is able to reach every other node.
Line 107: Line 114:


[[Image:2d3d.jpg]]
[[Image:2d3d.jpg]]
''Figure5: 2D & 3D Mesh topologies''


'''Implementation Examples:'''
'''Implementation Examples:'''
Line 122: Line 132:


[[Image:Torus_123D.jpg]]
[[Image:Torus_123D.jpg]]
''Figure6: 1D, 2D & 3D Torus topologies''




Line 136: Line 148:


[[Image:hyper.jpg]]
[[Image:hyper.jpg]]
''Figure7: Hypercube''


'''Implementation Examples:'''
'''Implementation Examples:'''
Line 151: Line 165:


[[Image:tree1.jpg]]
[[Image:tree1.jpg]]
''Figure8: Tree''


Disadvantage of this structure is that network is entirely dependent on the trunk which is the main backbone of the network. If that has to fail then the entire network would fail. Also, it gets difficult to configure after a certain point.
Disadvantage of this structure is that network is entirely dependent on the trunk which is the main backbone of the network. If that has to fail then the entire network would fail. Also, it gets difficult to configure after a certain point.
Line 158: Line 175:
[[Image:Fat_tree.jpg]]
[[Image:Fat_tree.jpg]]


''Figure9: Fat Tree''


=====Cube-Connected Cycles=====
=====Cube-Connected Cycles=====
Line 165: Line 183:
[[Image:Cubeconnectedcycles.jpg]]
[[Image:Cubeconnectedcycles.jpg]]


''Figure10: Cube-Connected Cycles''


=====de Bruijn and Star Graph Networks=====
=====de Bruijn and Star Graph Networks=====
Line 170: Line 189:


[[Image:De_star.jpg]]
[[Image:De_star.jpg]]
''Figure11: de Bruijn and Star Graph Networks''


De Bruijn and Star Graphs are two well known topologies proposed to minimize the network diameter for a given number of nodes and node degree.
De Bruijn and Star Graphs are two well known topologies proposed to minimize the network diameter for a given number of nodes and node degree.
Line 189: Line 211:
[[Image:Cross_bar.jpg]]
[[Image:Cross_bar.jpg]]


Each switch point for a crossbar with distributed control can have '''four states''' as shown in Figure __.  (a) shows that input from the row containing the switch has been granted access to the output while input from the upper row requesting the same output has been blocked. (b) shows that both inputs have been granted access since they requested different outputs. (c) shows that the upper input has been grated access and the same row input is blocked. (d) is required if multicasting (one-to-many communication) is supported.
''Figure12: Crossbar Networks''
 
Each switch point for a crossbar with distributed control can have '''four states''' as shown in ''Figure 13''.  (a) shows that input from the row containing the switch has been granted access to the output while input from the upper row requesting the same output has been blocked. (b) shows that both inputs have been granted access since they requested different outputs. (c) shows that the upper input has been grated access and the same row input is blocked. (d) is required if multicasting (one-to-many communication) is supported.


[[Image:Switch_states.jpg]]
[[Image:Switch_states.jpg]]
''Figure13: Four Switch States''


=====Multistage Interconnection Network=====
=====Multistage Interconnection Network=====
Line 202: Line 228:
[[Image:MIN.jpg]]
[[Image:MIN.jpg]]


''Figure14: A generic MIN''


MINs can be divided into 3 categories depending on the availability of paths:
MINs can be divided into 3 categories depending on the availability of paths:
Line 217: Line 244:
[[Image:Min1.jpg]]
[[Image:Min1.jpg]]


Two 16 x 16 unidirectional multistage interconnection networks: (a) '''Baseline Network''' (b) '''Butterfly Network'''
Figure15: Two 16 x 16 unidirectional multistage interconnection networks: (a) '''Baseline Network''' (b) '''Butterfly Network'''


[[Image:Min2.jpg]]
[[Image:Min2.jpg]]
Two 16 x 16 unidirectional multistage interconnection networks: (a) '''Cube Network''' (b) '''Omega Network'''
Figure16: Two 16 x 16 unidirectional multistage interconnection networks: (a) '''Cube Network''' (b) '''Omega Network'''


Following figures show an 8-node '''bidirectional butterfly''' MIN and alternate paths in the same.
Following figures show an 8-node '''bidirectional butterfly''' MIN and alternate paths in the same.
Line 226: Line 253:
[[Image:Bi_butterfly.jpg]]
[[Image:Bi_butterfly.jpg]]


''Figure17: Bidirectional Butterfly Multistage Interconnection Network ''


[[Image:Bi_butterfly2.jpg]]
[[Image:Bi_butterfly2.jpg]]


''Figure18: An Alternate path in a Bidirectional Butterfly Multistage Interconnection Network ''


'''Implementation examples of Indirect Networks:'''
'''Implementation examples of Indirect Networks:'''
Line 257: Line 286:
[[Image:Hybrid1.jpg]]
[[Image:Hybrid1.jpg]]


''Figure19: Multiple Backplane Buses ''


====Hierarchical Networks====
====Hierarchical Networks====
Line 263: Line 293:


[[Image:Hybrid2.jpg]]
[[Image:Hybrid2.jpg]]
''Figure20: Hierarchical Networks ''


====Cluster-Based Networks====
====Cluster-Based Networks====
Line 269: Line 301:


[[Image:Hybrid3.jpg]]
[[Image:Hybrid3.jpg]]
''Figure21: Cluster-Based Networks ''


====Other Hypergraph Topologies====
====Other Hypergraph Topologies====


Figure __ shows a distributed crossbar switch hypermesh. Each bus is driven by a single node. Bandwidth scales with the number of nodes.  
''Figure 22'' shows a distributed crossbar switch hypermesh. Each bus is driven by a single node. Bandwidth scales with the number of nodes.  


[[Image:Hybrid4.jpg]]
[[Image:Hybrid4.jpg]]
''Figure22: Distributed Crossbar Switch Hypermesh Networks ''


==References==
==References==

Revision as of 16:14, 24 April 2010

Interconnection Networks

Choosing the “Best” Network

While there is no best network that would work well for all applications, it greatly depends on the given application and the parallel system at hand. The following are some of the factors affect the choice of interconnection network.


1. Performance Requirements: Minimize message latency, avoid network from saturating (unable to deliver messages injected by nodes) and increase throughput of the network.

2. Scalability: Adding more processors should increase I/O bandwidth, Memory Bandwidth and Network Bandwidth should increase proportionally.

3. Incremental expandability: Should provide incremental expandability, allowing addition of a small number of nodes while minimizing resource wastage. For example, a network designed for number of processors to be a power of 4 makes it difficult to expand.

4. Partitionability: May be required for security reasons. If network can be partitioned into smaller systems, traffic produced by one user will not affect the others.

5. Simplicity: Simple Design lead to higher clock frequencies and better performance.

6. Distance Span: Maximum distance between nodes should be small.

7. Physical Constraints: Complexity of the connection is limited by the maximum wire density possible and by the maximum pin count. Factors like packaging, wiring, operating temperature should be taken into account since they pose many limitations on designs.

8. Reliability and Reparability: Should allow easy upgrades and repairs. Should minimize faults or detect them and correct them.

9. Expected Workloads: If kind of application is known in advance, network can be optimized for it. If not, network should be robust, design paramenters should be selected to perform well over a wide range of traffic conditions.

10. Cost Constraints: The “best” network might be too expensive. Alternative design considerations are important to meet cost constraints.


Classification of Interconnection networks

Figure 1 classifies Interconnection Networks based on network topologies. For each class, figure shows hierarchy of subclasses with some real implementations in parenthesis. In Shared-Medium Networks, transmission medium is shared by all devices. Direct Networks have point to point links directly connecting each communicating device to a subset of other devices in the network. Indirect networks connect devices using one or more switches. Hybrid networks are a combination of the above approaches.


Figure 1: Classification of Interconnection Networks

Shared-Medium Networks



Token Ring


Figure 2: Token Ring

Implementation examples:

  • IBM token ring: Supports bandwidth of 4 or 16 Mbps based on coaxial tube
  • FDDI: Bandwidth of 100Mbps using fiber optics


Token Bus

Figure3: Token Bus

Nodes are physically connected as a bus, but logically form a ring with tokens passed around to determine the turns for sending. A token is passed around the network nodes and only the node possessing the token may transmit. If a node doesn't have anything to send, the token is passed on to the next node on the virtual ring.

Implementation examples:

  • ARCNET: Attached Resource Computer NETwork


Backplane Bus

Figure3: Backplane Bus

Simplest interconnection structure for bus-based parallel computers. Commonly used to connect processors and memory modules to form UMA architecture. Many techniques are used for information transfers like synchronous, asynchronous, multiplexed and non-multiplexed.

Implementation examples:

  • Gigaplane used in Sun Ultra Enterprise X000 Server (ca. 1996): 2.68 Gbyte/s peak, 256 bits data, 42 bits address, split transaction protocol, 83.8 MHz clock
  • DEC AlphaServer8X00, i.e. 8200 & 8400 (ca. 1995): 2.1 Gbyte/s, 256 bits data, 40 bits address, split transaction protocol, 100 MHz clock (1 foot length)
  • SGI PowerPath -2 (ca. 1993): 1.2 Gbyte/s, 256 bits data, 40 bits address, 6 bits control, split transaction protocol, 47.5 MHz clock
  • HP9000 Multiprocessor Memory Bus (ca. 1993): 1Gbyte/s, 128 buts data, 64 bit address, 13 inches, pipelined bus, 60 MHz clock


Direct Networks


Bus based systems are not scalable since bus becomes a bottleneck when more processors are added. Direct network or point-to-point network scales well with large number of processors. A direct network consists of a set of nodes, each directly connected to a subset of other nodes. Figure 4 shows a generic node.


Figure4: A generic node

Each Router is connected to its neighbor’s routers through a unidirectional or bidirectional channel and handles the message communication among nodes. Internal channels connect local memory/processor to the router. External channels are used for communication between the routers. A direct network is formed by connecting input channels of one node to output channels of other nodes. Every input channel is paired with a corresponding output channel and there are many ways to interconnect these nodes such that every node in the network is able to reach every other node.

Network topology is Orthogonal if and only if nodes can be arranged in a orthogonal n-dimensional space, and every link can be arranged in such a way that it produces a displacement in a single dimension.

Strictly Orthogonal Topologies

In a strictly orthogonal topology, every node has at least one link crossing each dimension.

Mesh


Figure5: 2D & 3D Mesh topologies

Implementation Examples:

  • Intel Paragon: 16 rows and 16 columns of nodes, with a 2D mesh connecting them, 130 MB/sec for 1 MB messages, with a latency of around 100 usecs.
  • MIT Reliable Router: 2-D Mesh, 23-bit links (16- bit data), 200 MHz, 400 M bytes/s per link per direction, bidirectional signaling, reliable transmission.
  • MIT M-Machine: 3-D Mesh, 800 M bytes/s for each network channel.


Torus

Figure6: 1D, 2D & 3D Torus topologies


Implementation Examples:

  • Chaos Router: 2-D torus topology, bidirectional 8-bit links, 180 MHz, 260 Mbytes/s in each direction.
  • Cray T3E: Bidirectional 3-D torus, 14-bit data in each direction, 375 MHz link speed, 600 Mbytes/s in each direction.
  • Cray T3D: Bidirectional 3-D torus, up to 1024 nodes (8 x 16 x 8), 24-bit links (16-bit data, 8-bit control), 150 MHz, 300 Mbytes/s per link.


Hypercube

Figure7: Hypercube

Implementation Examples:

  • Intel iPSC-2 Hypercube: Binary hypercube topology, bit-serial channels at 2.8 Mbytes/s.


Other Topologies

In a Weakly Orthogonal topology, some nodes may not have any link in some dimensions.

Tree

Tree has a root node connected to a certain number of descendant nodes. A node with no descendants is a leaf node. If the distance from every node to the root is same, tree is balanced.


Figure8: Tree

Disadvantage of this structure is that network is entirely dependent on the trunk which is the main backbone of the network. If that has to fail then the entire network would fail. Also, it gets difficult to configure after a certain point.

Fat Tree, unlike the normal tree, which has "skinny" links all over, the links become "fatter" as one moves up the tree towards the root. By judiciously choosing the fatness of links, the network can be tailored to efficiently use any bandwidth.

Figure9: Fat Tree

Cube-Connected Cycles

Cube-Connected cycles can be considered as an n-dimensional hypercube of virtual nodes, where each node is a ring with n nodes, for a total of n2^n nodes.

Figure10: Cube-Connected Cycles

de Bruijn and Star Graph Networks


Figure11: de Bruijn and Star Graph Networks

De Bruijn and Star Graphs are two well known topologies proposed to minimize the network diameter for a given number of nodes and node degree.


Indirect Networks


Indirect networks do not provide a direct connection among nodes; instead communication has to be carried out through switches. Each node has a network adapter that connects to a network switch. Regular topologies have a regular connection pattern between switches whereas irregular topologies do not follow ant predefined pattern.

Regular Topologies

Crossbar Network

Crossbar networks allow many processors to communicate simultaneously without contention. It can be defined as a switching network with N inputs and M outputs, which allows up to min{N,M} one-to-one interconnections without contention.

Figure12: Crossbar Networks

Each switch point for a crossbar with distributed control can have four states as shown in Figure 13. (a) shows that input from the row containing the switch has been granted access to the output while input from the upper row requesting the same output has been blocked. (b) shows that both inputs have been granted access since they requested different outputs. (c) shows that the upper input has been grated access and the same row input is blocked. (d) is required if multicasting (one-to-many communication) is supported.

Figure13: Four Switch States

Multistage Interconnection Network

Multistage interconnection networks (MINs) connect devices through a number of switch stages, where each switch is a crossbar network. They number of stages and connection patterns determine the routing capacity.

Figure below shows a generic MIN with N inputs, M outputs and g stages.


Figure14: A generic MIN

MINs can be divided into 3 categories depending on the availability of paths:

  • Blocking: A connection between a free input/output is not always possible because of conlicts with the existing connections. But, multiple paths are provided to reduce conflicts and increase fault tolerance.
  • Nonblocking: Any input port can be connected to any free output port without affecting the existing connections.
  • Rearrangeable: Any input port can be connected to any free output port. However, existing connections mat require rearrangement of paths.

Depending on the kind of channels and switches, MINs can be divided into two classes: Unidirectional MINs and Bidirectional MINs.

Figures below shows 4 unidirectional MINs. They differ in connection patterns.

Figure15: Two 16 x 16 unidirectional multistage interconnection networks: (a) Baseline Network (b) Butterfly Network

Figure16: Two 16 x 16 unidirectional multistage interconnection networks: (a) Cube Network (b) Omega Network

Following figures show an 8-node bidirectional butterfly MIN and alternate paths in the same.

Figure17: Bidirectional Butterfly Multistage Interconnection Network

Figure18: An Alternate path in a Bidirectional Butterfly Multistage Interconnection Network

Implementation examples of Indirect Networks:

  • Myricom Myrinet: Supports regular and irregular topologies, 8 x 8 crossbar switch, 9-bit channels, full-duplex, 160 Mbytes/s per link.
  • Thinking Machines CM-5: Fat tree topology, 4-bit bidirectional channels at 40 MHz, aggregate bandwidth in each direction of 20 Mbytes/s.
  • Inmos C104: Supports regular and irregular topologies, 32 x 32 crossbar switch, serial links, 100 Mbits/s per link.
  • IBM SP2: Crossbar switches supporting Omega network topologies with bidirectional, 16-bit channels at 150 MHz, 300 Mbytes/s in each direction.
  • SGI SPIDER: Router with 20-bit bidirectional channels operating on both edges, 200 MHz clock aggregate row data rate of 1Gbyte/s. Offers support for configurations as nonblocking multistage network topologies as well as irregular topologies.
  • Tandem ServerNet: Irregular topologies, 8-bit bidirectional channels at 50 MHz, 50 Mbytes/s link


Hybrid Networks


Hybrid networks combine mechanisms from shared-medium and direct or indirect networks. They increase bandwidth with respect to shared-medium networks and reduce distance between nodes with respect to direct networks.

Multiple Backplane Buses

By having multiple buses as shown in Figure __, network bandwidth can be increased as compared to shared-medium networks. Shared-Medium networks pose a bottleneck since they can support only a small number of devices, have limited distance and are not scalable.

Figure19: Multiple Backplane Buses

Hierarchical Networks

Different buses are interconnected by routers and bridges to transfer information. Routers/bridges filter the network traffic by checking the destination address of the message. This approach is used in bridged LAN.

Figure20: Hierarchical Networks

Cluster-Based Networks

Cluster based network combine advantages of buses and point-to-point links by using buses at the lower level in the hierarchy to form clusters and a direct network topology connecting the clusters at the higher level.

Figure21: Cluster-Based Networks

Other Hypergraph Topologies

Figure 22 shows a distributed crossbar switch hypermesh. Each bus is driven by a single node. Bandwidth scales with the number of nodes.

Figure22: Distributed Crossbar Switch Hypermesh Networks

References

  • Interconnection Networks: An Engineering Approach by Duato, Jośe. Yalamanchili, Sudhakar. Ni, Lionel M.