Two FC-100 FC-AL hubs, with three installed GBICs in each hub

Connectivity

Cluster interconnect: The cluster connectivity uses Ethernet patch cables (no Ethernet switch required), with redundant qfe 100BASE-T ports (qfe0 and qfe4) on two separate SBus controllers to avoid a controller single point of failure.

Public networks: The cluster nodes main network connection is implemented using the on-board hme0 (100BASE-T) primary port with hme1 (100BASE-T) as the failover port. The main network connection has a failover interface on separate controllers. Two hme ports, hme2 and hme3, are available to implement additional production networks.

Storage access: Access to disk arrays is achieved through proper configuration of the FC-AL from each node.

Administration networks: Sun StorEdge™ Component Manager 2.1, included with your Cluster Platform 4500/3, is used to administer the disk arrays.

Network Administration

The administration network provides access to the Ethernet hub.

Miscellaneous Hardware

The following miscellaneous hardware is used:

Two one-meter RS-232C (DB-25/RJ-45) serial cables (Part No. 2151A) for the cluster nodes console (Part No. 530-2151)

One one-meter RS-232C (RJ-45/RJ-45) serial cable for the management server (Part No. 530-9524) to connect the terminal concentrator to the management server

Power cables provided with the power sequencers in the expansion cabinet

Terminal concentrator mounting bracket

Ethernet hub mounting brackets

By default, the I/O board slots in the servers are labeled 1, 3, 5, and 7 (FIGURE 2-2).

Chapter 2 The Cluster Platform 4500/3 System 9

Page 21
Image 21
Sun Microsystems 4500/3 manual Network Administration