HP Integrity NonStop NS1000 Planning Guide Abstract This guide describes the HP Integrity NonStop™ NS1000 system and provides examples of system configurations to assist in planning for installation of a new system. Product Version N.A. Supported Release Version Updates (RVUs) This publication supports H06.05.02 and all subsequent H-series RVUs until otherwise indicated by its replacement publication.
Document History Part Number Product Version Published 542527-001 N.A. March 2006 542527-004 N.A.
HP Integrity NonStop NS1000 Planning Guide Glossary Index What’s New in This Manual vii Manual Information vii New and Changed Information Figures vii About This Guide ix Who Should Use This Guide ix What’s in This Guide ix Where to Get More Information x Notation Conventions x 1. System Overview System Description 1-1 Hardware Enclosures and Configurations 1-3 Preparing for Other Than Integrity NonStop NS1000 Server Hardware 2.
3. System Installation Specifications Contents 3.
4. Integrity NonStop NS1000 System Description (continued) Contents 4. Integrity NonStop NS1000 System Description (continued) Default Startup Characteristics 4-9 Verify the Processor Type 4-9 Changing the CONFTEXT File 4-9 System Installation Document Packet 4-11 Tech Memo for the Factory-Installed Hardware Configuration Configuration Forms for the ServerNet Adapter 4-12 4-12 5.
. System Configuration Guidelines (continued) Contents 6.
B. Control, Configuration, and Maintenance Tools (continued) Contents B. Control, Configuration, and Maintenance Tools (continued) OSM B-13 System-Down OSM Low-Level Link AC Power Monitoring B-14 B-14 C. Guide to Integrity NonStop NS-Series Server Manuals Support and Service Library C-1 Safety and Compliance Glossary Index Figures Figure 1-1.
Contents HP Integrity NonStop NS1000 Planning Guide —542527-004 vi
What’s New in This Manual Manual Information HP Integrity NonStop NS1000 Planning Guide Abstract This guide describes the HP Integrity NonStop™ NS1000 system and provides examples of system configurations to assist in planning for installation of a new system. Product Version N.A. Supported Release Version Updates (RVUs) This publication supports H06.05.02 and all subsequent H-series RVUs until otherwise indicated by its replacement publication.
What’s New in This Manual New and Changed Information HP Integrity NonStop NS1000 Planning Guide —542527-004 viii
About This Guide Who Should Use This Guide This guide is written for those responsible for planning the installation, configuration, and maintenance of the server and the software environment at a particular site. Appropriate personnel must have completed HP training courses on system support for Integrity NonStop NS1000 servers. Note. Integrity NonStop NS1000 refers to hardware systems. H-series refers to release version updates (RVUs).
Where to Get More Information About This Guide Where to Get More Information For information about Integrity NonStop NS-series hardware, software, and operations, refer to Appendix C, Guide to Integrity NonStop NS-Series Server Manuals. Notation Conventions Hypertext Links Blue underline is used to indicate a hypertext link within text. By clicking a passage of text with a blue underline, you are taken to the location described.
1 System Overview This section provides an overview of the Integrity NonStop NS1000 system and covers these topics: Topic Page System Description 1-1 Hardware Enclosures and Configurations 1-3 Preparing for Other Than Integrity NonStop NS1000 Server Hardware 1-5 System Description The Integrity NonStop NS1000 system combines up to eight HP Integrity rx2620 servers with the NonStop operating system to create the NonStop value architecture (NSVA).
System Description System Overview Figure 1-1 shows a configuration example of a 2-processor Integrity NonStop NS1000 server. Figure 1-1.
Hardware Enclosures and Configurations System Overview Hardware Enclosures and Configurations Enclosures that house specific hardware components in an Integrity NonStop NS1000 system include: • • • • Blade element (HP Integrity rx2620 server) I/O adapter module (IOAM) Fibre Channel disk module (FCDM) Maintenance switch (Ethernet) A large number of enclosure combinations are possible within the modular cabinet of an Integrity NonStop NS1000 server.
Hardware Enclosures and Configurations System Overview This illustration shows the rear view of an example modular cabinet with eight blade elements and hardware for a complete system: Power Distribution Units (PDUs) IOAM Enclosure 42 42 41 41 40 40 39 39 38 38 37 37 36 36 35 35 34 34 33 33 32 32 31 31 30 30 29 29 28 28 27 27 26 26 25 25 24 24 23 23 22 Processor 7 Processor 6 Processor 5 Processor 4 Processor 3 Processor 2 Processor 1 Processor 0 System Console
System Overview Preparing for Other Than Integrity NonStop NS1000 Server Hardware Preparing for Other Than Integrity NonStop NS1000 Server Hardware This guide provides the specifications only for the Integrity NonStop NS1000 server modular cabinet and enclosures identified earlier in this section. For site preparation specifications for other HP hardware that will be installed at the site with the Integrity NonStop NS-series servers, consult with your HP account team.
System Overview Preparing for Other Than Integrity NonStop NS1000 Server Hardware HP Integrity NonStop NS1000 Planning Guide —542527-004 1 -6
2 Installation Facility Guidelines This section provides guidelines for preparing the installation site for Integrity NonStop NS1000 systems: Topic Page Modular Cabinet Power and I/O Cable Entry 2-1 Emergency Power-Off Switches 2-1 Electrical Power and Grounding Quality 2-2 Cooling and Humidity Control 2-3 Weight 2-3 Flooring 2-4 Dust and Pollution Control 2-4 Zinc Particulates 2-4 Space for Receiving and Unpacking the System 2-5 Operational Space 2-5 Modular Cabinet Power and I/O Ca
Installation Facility Guidelines Electrical Power and Grounding Quality Electrical Power and Grounding Quality Proper design and installation of a power distribution system for an Integrity NonStop NS1000 server requires specialized skills, knowledge, and understanding of appropriate electrical codes and the limitations of the power systems for computer and data processing equipment.
Installation Facility Guidelines Power Cables for European Union Power Cables for European Union If your site is in one of the European Union (EU) countries or in an area that complies with EU regulations, you can order modular cabinets with PDU power cables specifically manufactured and color coded for use in the EU. These are often called harmonized cables and comply with color coding and materials standards different from those used in North America and Japan.
Installation Facility Guidelines Flooring Flooring Integrity NonStop NS1000 servers can be installed either on the site’s floor with the cables entering from above the equipment or on raised flooring with power and I/O cables entering from underneath. Because cooling airflow through each enclosure in the modular cabinets is front-to-back, raised flooring is not required for system cooling.
Installation Facility Guidelines Space for Receiving and Unpacking the System Space for Receiving and Unpacking the System Identify areas that are large enough to receive and to unpack the system from its shipping cartons and pallets. Be sure to allow adequate space to remove the system equipment from the shipping pallets using supplied ramps. Also be sure adequate personnel are present to remove each cabinet from its shipping pallet and to safely move it to the installation site. WARNING.
Installation Facility Guidelines HP Integrity NonStop NS1000 Planning Guide —542527-004 2 -6 Operational Space
3 System Installation Specifications This section provides these specifications necessary for system installation site planning: Topic Page Processor Type and Memory Size 3-1 Modular Cabinet AC Input Power 3-2 Dimensions and Weights 3-8 Environmental Specifications 3-8 Calculating Specifications for Enclosure Combinations 3-9 Note. All specifications provided in this section assume that each enclosure in the modular cabinet is fully populated.
System Installation Specifications Modular Cabinet AC Input Power Modular Cabinet AC Input Power Topic Page North America and Japan: 208 V AC PDU Power Cords 3-2 North America and Japan: 200 to 240 V AC PDU Power Cords 3-2 International: 380 to 415 V AC PDU Power Cords 3-2 International: 200 to 240 V AC PDU Power Cords 3-3 Grounding 3-3 Branch Circuits and Circuit Breakers 3-3 Enclosure AC Input 3-4 Enclosure Power Loads 3-4 An AC power cable is factory-connected to each of two PDUs insi
System Installation Specifications International: 200 to 240 V AC PDU Power Cords International: 200 to 240 V AC PDU Power Cords For information about power cords for a 200 to 240 V AC input, single phase, 32A RMS modular cabinet, refer to International 200 to 240 V AC input, Single phase, 32A RMS Modular Cabinet on page 5-4. Grounding A safety ground/protective earth conductor is required for each AC service entrance to the NonStop server equipment.
Enclosure AC Input System Installation Specifications Enclosure AC Input Enclosures (such as blade element, IOAM, and so forth) require: Specification Value Nominal input voltage 200/208/220/230/240 V ac RMS Voltage range* 180-264 V ac Nominal line frequency 50 or 60 Hz Frequency ranges 47-53 Hz or 57-63 Hz Number of phases 1 * Voltage range for the maintenance switch is 200-240 V ac. Each PDU is factory wired to distribute the phases to its receptacles.
Dimensions and Weights System Installation Specifications AC Power Lines per Enclosure1 Apparent Power (volt-amps measured on single AC line with one line powered) Apparent Power (volt-amps measured on single AC line with both lines powered)5 Peak Inrush Current (amps) ( page 2 of 2) Fibre Channel disk module3 2 220 140 14 Maintenance switch (Ethernet)4 1 50 - Enclosure Type 280 4 1 Half of the plugs for an enclosure must be connected to the right-side PDU and the other half connected to t
Plan View of the Modular Cabinet System Installation Specifications Plan View of the Modular Cabinet 40 in. (102 cm) 75.5 in. (192 cm) 44 in. (112 cm) 23.5 in. (59.7 cm) VST102.vsd Service Clearances for the Modular Cabinet Aisles: 6 feet (182.9 centimeters) Front: 3 feet (91.4 centimeters) Rear: 3 feet (91.
Modular Cabinet Physical Specifications System Installation Specifications Modular Cabinet Physical Specifications Item Height Width Depth Weight in. cm in. cm in. cm Modular cabinet 78.5 199.4 23.5 59.7 44.0 111.8 Rack 78.5 199.4 23.5 59.7 40.0 101.9 Front door 78.5 199.4 23.5 59.7 3.0 7.6 Left-rear door 78.5 199.4 11.0 27.9 1.0 2.5 Right-rear door 78.5 199.4 12.0 30.5 1.0 2.5 Shipping (palletized) 83.5 212.0 39.0 99.0 48.0 121.
Environmental Specifications System Installation Specifications Environmental Specifications Topic Page Heat Dissipation Specifications and Worksheet 3-8 Operating Temperature, Humidity, and Altitude 3-8 Nonoperating Temperature, Humidity, and Altitude 3-9 Cooling Airflow Direction 3-9 Typical Acoustic Noise Emissions 3-9 Tested Electrostatic Immunity 3-9 Heat Dissipation Specifications and Worksheet Unit Heat (Btu/hour with single AC line powered) Unit Heat (Btu/hour with both AC lines pow
Nonoperating Temperature, Humidity, and Altitude System Installation Specifications Specification Operating Range1 Recommended Range1 Maximum Rate of Change per Hour (page 2 of 2) Humidity 15% to 80%, noncondensing 40% to 50%, noncondensing 6%, noncondensing Altitude2 0 to 10,000 feet (0 to 3,048 meters) - - 1. Operating and recommended ranges refer to the ambient air temperature and humidity measured 19.7 in. (50 cm) from the front of the air intake cooling vents. 2.
Calculating Specifications for Enclosure Combinations System Installation Specifications AC current calculations assume that one PDU delivers all power. In normal operation, the power is split equally between the two PDUs in the cabinet. However, calculate the power load to assume delivery from only one PDU to allow the system to continue to operate if one of the two AC power sources or PDUs fails.
Calculating Specifications for Enclosure Combinations System Installation Specifications This table shows the completed weight, power, and thermal calculations for the 4-processor sample configuration cabinet: Weight Component Volt-amps per AC feed Heat (BTU) AC line(s) powered AC line(s) powered Quantity Height (U) (lbs) (kg) Single Blade element (rx2620) with 4 GB memory 4 8 224 100 880 IOAM enclosure 1 11 200 90.7 Fibre Channel disk module 2 6 156 Console 1 2 Maint.
System Installation Specifications Calculating Specifications for Enclosure Combinations HP Integrity NonStop NS1000 Planning Guide —542527-004 3- 12
4 Integrity NonStop NS1000 System Description This section describes the Integrity NonStop NS1000 system and covers these topics: Topic Page NonStop Value Architecture 4-1 Blade Complex 4-2 Blade Element 4-3 Processor Element 4-4 ServerNet Fabric I/O 4-4 System Architecture 4-8 Modular Hardware 4-9 System Models 4-9 Default Startup Characteristics 4-9 System Installation Document Packet 4-11 For information about installing the Integrity NonStop NS1000 server hardware, refer to the No
Blade Complex Integrity NonStop NS1000 System Description Blade Complex The basic building block of the NSVA is the blade complex, which consists of rx2620 servers that are also referred to as blade elements. Each blade element houses one microprocessor called a processor element (PE). All input to and output from each blade element occurs through a ServerNet PCI adapter card located in each blade element. The ServerNet PCI adapter card interfaces with the ServerNet fabrics.
Integrity NonStop NS1000 System Description Blade Element In the term ServerNet fabric, the word fabric is significant because it contrasts with the concept of a bus. A bus provides a single, fixed communications path between start and end points. A fabric is a complex web of links between electronic routers that provide a large number of possible paths from one point to another. Two ServerNet communications fabrics, the X and Y, provide redundant, fault-tolerant communications pathways.
Processor Element Integrity NonStop NS1000 System Description the Integrity NonStop NS1000 server hardware, refer to the NonStop NS1000 Hardware Installation Manual. Processor Element The processor element (PE) in a blade element in an Integrity NonStop NS1000 system includes: • • A standard Intel Itanium microprocessor running at 1.
ServerNet Fabric I/O Integrity NonStop NS1000 System Description In the Integrity NonStop1000 server, fiber-optic links connect ServerNet PCI adapter cards in the blade elements to 4PSEs located in the IOAM enclosure. The 4PSEs provide the physical ServerNet connectivity to the blade elements by externalizing the ServerNet links from the IOAM enclosure midplane. Up to four 4PSEs can be installed in an IOAM enclosure, in slots 1 and 2 of modules 2 and 3.
ServerNet Fabric I/O Integrity NonStop NS1000 System Description IOAM Enclosure ServerNet Pathways This drawing shows the ServerNet communication pathways in the IOAM enclosure. Optic lines connect the ServerNet PCI adapter cards in the blade elements with the 4PSEs in each IOAM module. The 4PSEs in IOAM module 2 communicate with the X ServerNet switch board. The 4PSEs in IOAM module 3 communicate with the Y ServerNet switch board.
ServerNet Fabric I/O Integrity NonStop NS1000 System Description Example System ServerNet Pathways This drawing shows the redundant routing and connection of the ServerNet X and Y fabric within a simple example system. This example system includes: Four processors (0 through 3) contained in blade elements 1 through 4. One IOAM enclosure, group 100, connected to the ServerNet PCI adapter card in a blade element.
System Architecture Integrity NonStop NS1000 System Description If a cable, connection, router, or other failure occurs, only the system resources that are downstream of the failure on the same fabric are affected. Because of the redundant ServerNet architecture, communication takes the alternate path on the other fabric to the peer resources.
Modular Hardware Integrity NonStop NS1000 System Description Modular Hardware Hardware for Integrity NonStop NS1000 systems is implemented in modules, or enclosures that are installed in modular cabinets. For descriptions of the components and cabinets, see Modular Hardware Components on page 5-1. All Integrity NonStop NS1000 server components are field-replaceable units (FRUs) that can only be serviced by service providers trained by HP.
Changing the CONFTEXT File Integrity NonStop NS1000 System Description Each Integrity NonStop NS1000 system ships with these default startup characteristics: • $SYSTEM disks residing in a Fibre Channel disk module connected to IOAM enclosure group 100 with the disks in these locations: IOAM • • FCSA Fibre Channel Disk Module Path Group Module Slot SAC Shelf Bay Primary 100 2 3 1 1 1 Backup 100 3 3 1 1 1 Mirror 100 3 3 2 1 1 Backup 100 2 3 2 1 1 Configured system lo
System Installation Document Packet Integrity NonStop NS1000 System Description This illustration shows the system load paths: Fibre Channel Disk Module 2 1P 2 B 1 Fibre Channel Disk Module 2 2 MB 1 1M X Fabric Y Fabric ServerNet Switch Board ServerNet Switch Board 1 1 2 2 IOAM Enclosure Group 100 FCSA residing in slot 3 of module 2 X Y X FCSA residing in slot 3 of module 3 Y Processor 0 Processor 1 (Blade element 1) (Blade element 2) VST818.
Integrity NonStop NS1000 System Description Tech Memo for the Factory-Installed Hardware Configuration Tech Memo for the Factory-Installed Hardware Configuration Each new Integrity NonStop NS1000 system includes a document that describes: • • • • The cabinet included with the system Each hardware enclosure installed in the cabinet Cabinet U location of the bottom edge of each enclosure Each ServerNet cable with: ° ° ° Source and destination enclosure, component, and connector Cable part number Source
5 Modular System Hardware This section describes the hardware used in the Integrity NonStop NS1000 system: Topic Page Modular Hardware Components 5-1 Component Location and Identification 5-11 Modular Hardware Components These hardware components can be part of an Integrity NonStop NS1000 system: Topic Page Modular Cabinets 5-3 Power Distribution Unit (PDU) 5-5 Blade Element 5-5 I/O Adapter Module (IOAM) Enclosure and I/O Adapters 5-8 Fibre Channel Disk Module 5-9 Tape Drive and Interfac
Modular Hardware Components Modular System Hardware This example shows a modular cabinet with eight blade elements and hardware for a complete system (rear view): Power Distribution Units (PDUs) IOAM Enclosure 42 42 41 41 40 40 39 39 38 38 37 37 36 36 35 35 34 34 33 33 32 32 31 31 30 30 29 29 28 28 27 27 26 26 25 25 24 24 23 23 22 Processor 7 Processor 6 Processor 5 Processor 4 Processor 3 Processor 2 Processor 1 Processor 0 System Console FCDMs ServerNet Swi
Modular System Hardware Modular Cabinets Modular Cabinets Note. For current cabinet installation and configuration information, contact your HP product representative . The modular cabinet is a 19-inch, 42 U high, rack for mounting modular components. The modular cabinet comes equipped with front and rear doors. To mount a Power Distribution Unit (PDU) without occupying any U-space in the cabinet, the cabinet includes an extension, which makes it deeper than some industry-standard racks.
Modular System Hardware • • • Modular Cabinets ° Zero-U" rack design PDU input characteristics ° 200 to 240 V ac, single phase, 40A RMS, 3-wire ° 50/60Hz ° Non-NEMA Locking CS8265C, 50A input plug ° 6.
Modular System Hardware Power Distribution Unit (PDU) ° ° • 3 circuit-breaker-protected 20A load segments 36 AC receptacles per PDU (12 per segment) - IEC 320 C13 10A receptacle type ° 3 AC receptacles per PDU (1 per segment) - IEC 320 C19 12A receptacle type Includes front and rear doors Power Distribution Unit (PDU) Note. For current information about PDUs and AC power distribution, contact your HP product representative.
Blade Element Modular System Hardware This simplified example shows connections from two blade elements to the 4PSEs in an IOAM enclosure: 4 4 3 3 3 2 1 2 1 2 1 3 2 1 2 3 4 5 1 Slots G4SA 4 FCSA 4PSE 4 FCSA 4PSE G4SA 4PSE FCSA Module 3 Y-Fabric 4PSE FCSA Module 2 X-Fabric IOAM Enclosure 1 2 3 4 5 Processor 1 (Blade element 2) Slot 2 Port X Port Y Processor 0 (Blade element 1) Slot 2 Port X Port Y X Fabric Y Fabric VST015.
Blade Element Modular System Hardware • Front-to-rear cooling ServerNet PCI Adapter Card Each HP Integrity rx2620 server contains one ServerNet PCI adapter card (installed in PCI slot 2) to provide ServerNet connectivity. This illustration shows the rear of an HP Integrity rx2620 server equipped a ServerNet PCI adapter card: ServerNet Fabric Connections PCI Adapter Card VST011.vsd Front Panel Buttons and Indicator LEDs This illustration shows the front panel of the HP Integrity rx2620 server: VST070.
I/O Adapter Module (IOAM) Enclosure and I/O Adapters Modular System Hardware This table describes front panel button and indicator LEDs for the preceding illustration: Front Panel Illustration Reference Number 1 2 3 4 LED Indicator State Meaning e-buzzer icon On Indicates the position of the internal speaker Locator LED and button On Diagnostic LEDs On LAN LED On Off Identifies the rack position of the system Off Identifies system errors Off Indicates the system is communicating over the L
Modular System Hardware • • • • • • Fibre Channel Disk Module The maintenance entity on the ServerNet switch board supports direct connections to blade elements and ServerNet I/O adapters. Up to four 4PSEs provide ServerNet connectivity to the blade elements. Only 4PSEs are supported for installation in slots 2.1, 2.2, 3.1, and 3.2 of the IOAM enclosure for an Integrity NonStop NS1000 server. Up to six ServerNet I/O adapters are supported, and a minimum of two FCSAs and two G4SAs must be installed.
System Console Modular System Hardware Details about the use or implementation of the maintenance switch that are specific to an Integrity NonStop NS1000 server are presented here.
Component Location and Identification Modular System Hardware This illustration shows an example of connections between two IOAM enclosures and an ESS via the separate Fibre Channel switch: Fibre Channel Switch FCSA IOAM Enclosure IOAM Enclosure ESS FCSA Fibre Channel Switch VST 068.vsd For fault tolerance, the primary and backup paths to an ESS logical device (LDEV) must go through different Fibre Channel switches.
Terminology Modular System Hardware Terminology Terms specific to an Integrity NonStop NS1000 system that are used in locating and describing components are: Term Definition blade complex In an Integrity NonStop NS1000 system, OSM uses this term to hierarchically differentiate between each blade element. See Blade Complex on page 4-2. blade element An HP Integrity rx2620 server that contains the PE, power supplies, fan assemblies, and firmware.
Modular System Hardware Blade Element Group-Module-Slot Numbering Example: module 1 = any blade element ° 100-107 relates to processors 0-7 Example: module 102 = processor 2 ° In the OSM Low-Level Link, 100 through 107 relates to processors 0 through 7 Example: module 103 = processor 3 • Slot: ° In OSM Service Connection displays: ° 2 relates to the location of the ServerNet PCI adapter card in the PCI slots. Only slot 2 is used.
Blade Element Group-Module-Slot Numbering Modular System Hardware OSM Service Connection GMS Numbering: Module Processor ID Group* Blade Element Processor Slot 0 400 1 100 1 401 1 101 2 402 1 102 2, PCI adapter 3 403 1 103 4 404 1 104 5 405 1 105 6 406 1 106 7 407 1 107 Port (Slot 2 only) X Y 3-4, Power supplies 17-19, Fans OSM Low-Level GMS Numbering: Processor ID Group* Module Slot 0 400 100 1 1 400 101 1 2 400 102 1 3 400 103 1 4 400 104 1
Blade Element Group-Module-Slot Numbering Modular System Hardware This illustration shows the physical GMS numbering for a blade element: X Port Y Port Blade group (400 - 407) Blade element (Module 1) Processor (100-107) PCI Slot 2 Slots 17-19 Fan Assemblies Under Top Panel Blade element Rear Blade element Front Slots 3-4 Power Supplies Behind Front Bezel VST008.
Fibre Channel Disk Module (FCDM) Group-ModuleSlot Numbering Modular System Hardware This illustration shows the slot locations for the IOAM enclosure: LCD Maintenance Connector Display (100 BaseT RJ-45) ServerNet Switch Board (Module 2, Slot 14) Fans (Mod 2 Slot 17) 4PSEs Slots 1 and 2 Slot 5 Slot 4 Slot 3 Slot 5 Slot 1 Slot 4 Slot 3 Fans (Mod 2 Slot 16) Slot 1 Slot 2 Fans (Mod 3 Slot 16) ServerNet Switch Board (Module 3, Slot 14) IOAM (Module 3) IOAM (Module 2) Slot 2 Fans (Mod 3 Slot 17)
Fibre Channel Disk Module (FCDM) Group-ModuleSlot Numbering Modular System Hardware The form of GMS numbering for a disk in a Fibre Channel disk module is: 100.231.103 GMS number for disk drive FCDM bay housing disk FCDM shelf number ServerNet addressable controller (SAC) number in FCSA IOAM slot housing FCSA IOAM module housing FCSA IOAM group number housing FCSA VST505.
Modular System Hardware Fibre Channel Disk Module (FCDM) Group-ModuleSlot Numbering HP Integrity NonStop NS1000 Planning Guide —542527-004 5- 18
6 System Configuration Guidelines This section provides configuration guidelines for an Integrity NonStop NS1000 system: Topic Page Enclosure Locations in Cabinets 6-3 Internal ServerNet Interconnect Cabling 6-3 IOAM Enclosure and Disk Storage Considerations 6-9 Fibre Channel Devices 6-10 G4SAs to Networks 6-20 Default Naming Conventions 6-22 PDU Strapping Configurations 6-23 Integrity NonStop NS1000 systems use a flexible modular architecture.
System Configuration Guidelines This example shows one possible system configuration using eight blade elements: Power Distribution Units (PDUs) IOAM Enclosure 42 42 41 41 40 40 39 39 38 38 37 37 36 36 35 35 34 34 33 33 32 32 31 31 30 30 29 29 28 28 27 27 26 26 25 25 24 24 23 23 22 Processor 7 Processor 6 Processor 5 Processor 4 Processor 3 Processor 2 Processor 1 Processor 0 System Console FCDMs ServerNet Switch Boards 4-Port ServerNet Extenders ServerNet Ada
Enclosure Locations in Cabinets System Configuration Guidelines Enclosure Locations in Cabinets This table provides details about the location of Integrity NonStop NS1000 server enclosures and components within a cabinet. The enclosure location refers to the U location on the rack where the lower edge of the enclosure resides, such as the bottom of an HP Integrity rx2620 server at 28U.
Cable Labeling System Configuration Guidelines Topic Page Blade Elements to IOAM Enclosure 6-6 FCSA to Fibre Channel Disk Modules 6-9 FCSA to Tape Devices 6-9 For general information about internal ServerNet interconnect cabling, refer to the NonStop NS-Series Planning Guide. Details about internal ServerNet interconnect cabling specific to an Integrity NonStop NS1000 server are presented here.
Internal Interconnect Cables System Configuration Guidelines nn.nn Identifies the slot location and port connection of the component. Near Refers to the information for this end of this cable. Far Refers to the information for the other end of this cable. If you replace a cable or either install or move an enclosure, be sure to update information on the labels at both ends of the cables.
Dedicated Service LAN Cables System Configuration Guidelines This illustration shows an SC fiber cable connector pair: VST599.vsd Dedicated Service LAN Cables The Integrity NonStop NS1000 system uses Category 5, unshielded twisted pair Ethernet cables for the internal dedicated service LAN and for connections between the G4SA and the application LAN equipment. Cable Length Restrictions For a general description about cable length restrictions, refer to the NonStop NSSeries Planning Guide.
Blade Elements to IOAM Enclosure System Configuration Guidelines Blade Element to IOAM Enclosure and Processor IDs Each blade element contains one processor element. The maintenance entity (ME) firmware running in the IOAM enclosure assigns a number to each processor element based on its connection from the blade element to the 4PSEs in slots 2.1, 2.2, 3.1, and 3.2.
Blade Elements to IOAM Enclosure System Configuration Guidelines This cabling diagram illustrates the default configuration and connections for an 8-processor NonStop NS1000 Hardware Installation Manualprocessor Integrity NonStop NS1000 system. Four 4PSEs are required: two for the X-fabric and two for the Y-fabric The diagram is not for use in installing or cabling the system. For instructions on connecting the cables, see the NonStop NS-Series Hardware Installation Manual.
System Configuration Guidelines FCSA to Fibre Channel Disk Modules FCSA to Fibre Channel Disk Modules Fibre Channel disk modules (FCDMs) can be connected directly to FCSAs in an IOAM enclosure (see Blade Elements to IOAM Enclosure on page 6-6), with these exceptions: • • Only configurations with one IOAM enclosure are supported. A maximum of 16 FCDMs can be connected in the Integrity NonStop NS1000 system because only one IOAM enclosure containing a maximum of four FCSAs is supported.
Fibre Channel Devices System Configuration Guidelines Fibre Channel Devices This subsection describes Fibre Channel devices and covers these topics: Topic Page Factory-Default Locations for Disk Volumes 6-12 Configurations for Fibre Channel Devices 6-12 Configuration Restrictions for Fibre Channel Devices 6-12 Recommendations for Fibre Channel Device Configuration 6-13 Example Configurations of the IOAM Enclosure and Fibre Channel Disk Module 6-14 The only Fibre Channel device used internally
Fibre Channel Devices System Configuration Guidelines This illustration shows the locations of the hardware in the Fibre Channel disk module as well as the Fibre Channel port connectors at the back of the enclosure: FC-AL Port B2 FC-AL Port A2 Fibre Channel Disk Module (rear) EMU Port A1 Port B1 Fibre Channel Disk Module (front) Disk Drive Bays 1-14 VSD.503.vst Fibre Channel disk modules connect to Fibre Channel ServerNet adapters (FCSAs) via Fiber Channel arbitrated loop (FC-AL) cables.
Factory-Default Locations for Disk Volumes System Configuration Guidelines Factory-Default Locations for Disk Volumes This illustration shows where the factory-default locations for the primary and mirror system disk volumes reside in separate Fibre Channel disk modules: $SYSTEM (bay 1) $AUDIT (bay 3) Fibre Channel Disk Module (front) $DSMSCM (bay 2) $OSS (bay 4) VSD.082.vst FCSA location and cable connections vary depending on the various controller and Fibre Channel disk module combinations.
System Configuration Guidelines Recommendations for Fibre Channel Device Configuration Recommendations for Fibre Channel Device Configuration These recommendations apply to FCSA and Fibre Channel disk module configurations: • • • • • • • • • • • • • • Primary Fibre Channel disk module connects to the FCSA F-SAC 1. Mirror Fibre Channel disk module connects to the FCSA F-SAC 2. FC-AL port A1 is the incoming port from an FCSA or from another Fibre Channel disk module.
System Configuration Guidelines • • • • • Example Configurations of the IOAM Enclosure and Fibre Channel Disk Module Daisy-chain configurations follow the same configuration restrictions and rules that apply to configurations that are not daisy-chained. (See Daisy-Chain Configurations on page 6-17.) Fibre Channel disk modules containing mirrored volumes must be installed in separate daisy chains.
Example Configurations of the IOAM Enclosure and Fibre Channel Disk Module System Configuration Guidelines Two FCSAs, Two FCDMs, One IOAM Enclosure This illustration shows example cable connections between two FCSAs and the primary and mirror Fibre Channel disk modules: Mirror FCDM Primary FCDM Fibre Channel Cables FCSA Fibre Channel Cables FCSA VST088.
Example Configurations of the IOAM Enclosure and Fibre Channel Disk Module System Configuration Guidelines Four FCSAs, Four FCDMs, One IOAM Enclosure This illustration shows example cable connections between the four FCSAs and the two sets of primary and mirror Fibre Channel disk modules: Mirror FCDM 2 Primary FCDM 2 Mirror FCDM1 Primary FCDM1 FCSAs FCSAs VST089.
Example Configurations of the IOAM Enclosure and Fibre Channel Disk Module System Configuration Guidelines Daisy-Chain Configurations When planning for possible use of daisy-chained disks, consider: Daisy-Chained Disks Recommended for ... Daisy-Chained Disks Not Recommended for ... Requirements for Daisy-Chain1 Cost-sensitive storage and applications using low-bandwidth disk I/O. Many volumes in a large Fiber Channel loop.
Example Configurations of the IOAM Enclosure and Fibre Channel Disk Module System Configuration Guidelines This illustration shows an example of cable connections between the two FCSAs and four Fibre Channel disk modules in a single daisy-chain configuration. A single IOAM enclosure, two FCSAs, and four Fibre Channel disk modules with ID expander does not provide fault-tolerant mirrored disk storage.
Example Configurations of the IOAM Enclosure and Fibre Channel Disk Module System Configuration Guidelines Four FCSAs, Three FCDMs, One IOAM Enclosure This illustration shows example cable connections between the four FCSAs and three Fibre Channel disk modules, with the primary and mirror drives split within each Fibre Channel disk module: Primary 3 Mirror 1 FCDM 3 Primary 2 Mirror 3 FCDM 2 Primary 1 Mirror 2 FCDM 1 IOAM Enclosure FCSAs FCSAs VST085.
G4SAs to Networks System Configuration Guidelines This illustration shows the factory-default locations for the configurations of four FCSAs and three Fibre Channel disk modules where the primary system file disk volumes are in Fibre Channel disk module 1: $SYSTEM (bay 1) $AUDIT (bay 3) Fibre Channel Disk Module (front) $DSMSCM (bay 2) $OSS (bay 4) VSD.082.
G4SAs to Networks System Configuration Guidelines This illustration shows the G4SA: G4SA G4SA LC Connectors (Fiber) RJ-45 Connector (10/100/1000 Mbps) RJ-45 Connector (10/100 Mbps) VST002.
Default Naming Conventions System Configuration Guidelines Default Naming Conventions The Integrity NonStop NS1000 system implements default naming conventions in the same manner as other Integrity NonStop NS-series systems. With a few exceptions, default naming conventions are not necessary for the modular resources that make up an Integrity NonStop NS1000 system.
System Configuration Guidelines PDU Strapping Configurations No TFTP or WANBOOT process is configured for new NonStop systems. Note. Naming conventions or configurations for the dedicated service LAN TCP/IP are the same as the TCP/IP conventions used with G-series RVUs. The names are $ZTCP0 and $ZTCP1. OSM Service Connection provides the location of the resource by adding an identifying suffix to the names of all the system resources.
System Configuration Guidelines PDU Strapping Configurations HP Integrity NonStop NS1000 Planning Guide —542527-004 6- 24
7 Example Configurations This section shows example hardware component configurations within a modular cabinet for an Integrity NonStop NS1000 server. A number of other configurations are also possible because of the flexibility inherent to the NonStop value architecture and ServerNet. Note. Hardware configuration drawings in this appendix represent the physical arrangement of the modular enclosures but do not provide information about PDUs.
Typical Configurations Example Configurations Typical Configurations This subsection shows the U locations of modular hardware configurations that can be installed in the specified system.
2-processor System Example Configurations 2-processor System This 2-processor configuration has a maximum of two blade elements (HP Integrity rx2620 servers) with one IOAM enclosure and two Fibre Channel disk modules: 42 41 Configurable Space 40 39 38 37 Fibre Channel Disk Module 36 35 34 Fibre Channel Disk Module 33 32 31 30 29 28 27 IOAM Enclosure 26 25 24 23 22 21 Console 20 19 18 17 16 15 14 Configurable Space 13 12 11 10 09 08 07 06 05 Blade element Blade element 04 03 Configurable Spa
4-processor System Example Configurations 4-processor System This 4-processor configuration has a maximum of four blade elements (HP Integrity rx2620 servers) with one IOAM enclosure and two Fibre Channel disk modules: 42 41 Configurable Space 40 39 38 37 Fibre Channel Disk Module 36 35 34 Fibre Channel Disk Module 33 32 31 30 29 28 27 IOAM Enclosure 26 25 24 23 22 21 Console 20 19 18 17 16 Configurable Space 15 14 13 12 11 Blade element 10 09 08 07 06 05 Blade element Blade element Blade e
6-processor System Example Configurations 6-processor System This 6-processor configuration has a maximum of six blade elements (HP Integrity rx2620 servers) with one IOAM enclosure and four Fibre Channel disk modules: 42 41 40 Fibre Channel Disk Module 39 38 37 Fibre Channel Disk Module 36 35 34 Fibre Channel Disk Module 33 32 31 30 29 28 27 IOAM Enclosure 26 25 24 23 22 21 Console 20 19 18 Configurable Space 17 16 15 Blade element 14 13 12 11 Blade element Blade element 10 09 08 07 Blad
8-processor System Example Configurations 8-processor System This 8-processor configuration has a maximum of eight blade elements (HP Integrity rx2620 servers) with one IOAM enclosure, and four Fibre Channel disk modules: 42 41 40 Fibre Channel Disk Module 39 38 37 Fibre Channel Disk Module 36 35 34 Fibre Channel Disk Module 33 32 31 30 29 28 27 IOAM Enclosure 26 25 24 23 22 21 20 19 Console Blade element 18 17 16 15 Blade element Blade element 14 13 12 11 Blade element Blade element 10 09 0
Example U Locations for Modular Enclosures Example Configurations Example U Locations for Modular Enclosures This illustration lists the relative U location of each modular enclosure in an example 2-processor Integrity NonStop NS1000 system: IOAM Enclosure (U23) 42 42 41 41 40 40 39 39 38 38 37 37 36 36 35 35 34 34 33 33 32 32 31 31 30 30 29 29 28 28 27 27 26 26 25 25 24 24 23 23 22 21 Power Distribution Units (PDUs) Processor 1 Processor 0 Primary FCDM (U37)
Example of 2-processor System Cabling Example Configurations Example of 2-processor System Cabling This illustration shows an example of a 2-processor system. This conceptual representation shows the simplified X and Y ServerNet cabling between the blade elements and the IOAM enclosure and also the simplified cabling between the FCSAs and the Fibre Channel disk modules.
A Cables Internal Cables Available internal cables and their lengths are: Cable Type Connectors Length (meters) Length (feet) Product ID MMF LC-LC 2 7 M8900-02 5 16 M8900-05 15 49 M8900-15 40 131 M8900-40 80 262 M8900-80 100 328 M8900100 1251 4101 M8900125 2001 6561 M8900200 2501 8201 M8900250 10 33 M8910-10 20 66 M8910-20 50 164 M8910-50 100 328 TBD 1251 4101 M8910125 3 10 M8920-3 5 16 M8920-5 10 33 M8920-10 30 98 M8920-30 50 164 M8920-50 0.
Cable Length Restrictions Cables Cable Length Restrictions For a general description of cable length restrictions, refer to the NonStop NS-Series Planning Guide. Details about cable length restrictions that are specific to an Integrity NonStop NS1000 server are presented here.
B Control, Configuration, and Maintenance Tools This section introduces the control, configuration, and maintenance tools used in Integrity NonStop NS-series systems: Topic Page Support and Service Library B-1 System Console B-1 Maintenance Architecture B-6 Dedicated Service LAN B-6 OSM B-13 System-Down OSM Low-Level Link B-14 AC Power Monitoring B-14 Support and Service Library See Support and Service Library on page C-1.
System Console Configurations Control, Configuration, and Maintenance Tools Some system console hardware, including the PC system unit, monitor, and keyboard, can be mounted in the cabinet. Other PCs are installed outside the cabinet and require separate provisions or furniture to hold the PC hardware. System consoles communicate with Integrity NonStop NS-series servers over a dedicated service local area network (LAN) or a secure operations LAN.
System Console Configurations Control, Configuration, and Maintenance Tools One System Console Managing One System (Setup Configuration) DHCP DNS server (optional) Remote Service Provider Secure Operations LAN Modem Primary System Console Optional Connection to a Secure Operations LAN (One or Two Connections) 4PSE 4PSE FCSA FCSA G4SA 10/100 ENET Port, ServerNet Switch Boards 4PSE 4PSE FCSA FCSA G4SA Maintenance Switch 1 IOAM Enclosure VST073.
Control, Configuration, and Maintenance Tools System Console Configurations add a backup system console and additional system consoles. Note. Because the system console and maintenance switch are single points of failure that could disrupt access to OSM, this configuration is not recommended for operations that require high availability or fault tolerance. When you use this configuration, you do not need to change the preconfigured IP addresses.
System Console Configurations Control, Configuration, and Maintenance Tools DHCP DNS server (optional) Remote Service Provider Remote Service Provider Secure Operations LAN Modem Primary System Console Backup System Console Modem Optional Connection to a Secure Operations LAN (One or Two Connections) Maintenance Switch 2 4PSE 4PSE FCSA FCSA G4SA 10/100 ENET Port, ServerNet Switch Boards G4SA (X Fabric) Group 2, Slot 5, Port A 4PSE 4PSE FCSA FCSA G4SA Maintenance Switch 1 G4SA (Y Fabric) Group
Maintenance Architecture Control, Configuration, and Maintenance Tools Maintenance Architecture This simplified illustration shows the two elements of the maintenance architecture plus the OSM maintenance console applications: To Remote Support Center Maintenance Switch Maintenance LAN OSM Console ServerNet Fabric IOAM ME IOAM G4SA FCSA FCDM FC-AL Blade element PE EMU I/O and Fabric Functional Element Processor Functional Element VST517.
Control, Configuration, and Maintenance Tools Dedicated Service LAN This subsection includes: Topic Page IP Addresses B-9 Ethernet Cables B-11 SWAN Concentrator Restriction B-11 System-Up Dedicated Service LAN B-11 Dedicated Service LAN Links With One IOAM Enclosure B-12 Initial Configuration for a Dedicated Service LAN B-12 Operating Configurations for Dedicated Service LANs B-13 An Integrity NonStop NS1000 system requires a dedicated LAN for system maintenance through OSM.
Dedicated Service LAN Control, Configuration, and Maintenance Tools • Connect one G4SA to each maintenance switch. Caution. To avoid possible conflicts on the LAN: • • • If the configuration includes two maintenance switches, install and configure one switch completely, including assigning its IP address, before you install the other. Only one connection between the maintenance switches is permitted. More than one connection overloads network traffic, rendering the dedicated service LAN unusable.
IP Addresses Control, Configuration, and Maintenance Tools IP Addresses Integrity NonStop NS1000 servers require Internet protocol (IP) addresses for these components that are connected to the dedicated service LAN: • • • • ServerNet switch boards in the IOAM enclosure Maintenance switches System consoles G4SAs These components have default IP addresses that are preconfigured at the factory.
Control, Configuration, and Maintenance Tools • IP Addresses Dynamically Assigned IP Addresses Note. Be aware of possible conflicts with existing operations LANs. This guide cannot predict all possible configurations of existing LANs. If the site where you install the maintenance switch has network administrators and specialists, consult them for information about their current Dynamic Host Configuration Protocol (DHCP) configuration.
Ethernet Cables Control, Configuration, and Maintenance Tools IOAM enclosures Dynamic System consoles Dynamic *If both dynamic and static IP addresses are used in the LAN configuration, HP recommends that the maintenance switches be configured with static IP addresses. **You do not need to reconfigure the IP addresses for the TCP/IP processes. Ethernet Cables Ethernet connections for a dedicated service LAN require Category 5 unshielded twisted pair (UTP) cables.
Dedicated Service LAN Links With One IOAM Enclosure Control, Configuration, and Maintenance Tools dedicated service LAN, use the slower 10/100 Mbps PIF A rather than one of the highspeed 1000 Mbps Ethernet ports of PIF C or D.
Control, Configuration, and Maintenance Tools Operating Configurations for Dedicated Service LANs HP recommends that you change these preconfigured IP addresses to addresses appropriate for your LAN environment.
System-Down OSM Low-Level Link Control, Configuration, and Maintenance Tools System-Down OSM Low-Level Link In the Integrity NonStop NS1000 system, the maintenance entity (ME) in the IOAM enclosure provides dedicated service LAN services via the OSM Low-Level Link for both OS coldload, system management, and hardware configuration when hardware is powered up but the OS is not running.
Control, Configuration, and Maintenance Tools AC Power Monitoring This information relates to handling power failures: • • • For ride-through time, see the SCF Reference Manual for the Kernel Subsystem. For TACL SETTIME command, see the TACL Reference Manual. For setting system time programmatically, see the Guardian Procedure Calls Reference Manual.
Control, Configuration, and Maintenance Tools HP Integrity NonStop NS1000 Planning Guide —542527-004 B -16 AC Power Monitoring
C Guide to Integrity NonStop NS-Series Server Manuals These manuals support the Integrity NonStop NS-series systems. Category Purpose Title Reference Provide information about the manuals, the RVUs, and hardware that support NonStop NS-series servers NonStop Systems Introduction for H-Series RVUs Describe how to prepare for changes to software or hardware configurations Managing Software Changes Change planning and control H06.
Guide to Integrity NonStop NS-Series Server Manuals Support and Service Library Authorized service providers can also order the NTL Support and Service Library CD: • • HP employees: Subscribe at World on a Workbench (WOW). Subscribers automatically receive CD updates. Access the WOW order form at http://hps.knowledgemanagement.hp.com/wow/order.asp. HP Authorized Service Providers and Channel Partners: Send an inquiry to pubs.comments@hp.com.
Safety and Compliance This section contains three types of required safety and compliance statements: • • • Regulatory compliance Waste Electrical and Electronic Equipment (WEEE) Safety Regulatory Compliance Statements The following regulatory compliance statements apply to the products documented by this manual. FCC Compliance This equipment has been tested and found to comply with the limits for a Class A digital device, pursuant to part 15 of the FCC Rules.
Regulatory Compliance Statements Safety and Compliance Korea MIC Compliance Taiwan (BSMI) Compliance Japan (VCCI) Compliance This is a Class A product based on the standard or the Voluntary Control Council for Interference by Information Technology Equipment (VCCI). If this equipment is used in a domestic environment, radio disturbance may occur, in which case the user may be required to take corrective actions.
Regulatory Compliance Statements Safety and Compliance European Union Notice Products with the CE Marking comply with both the EMC Directive (89/336/EEC) and the Low Voltage Directive (73/23/EEC) issued by the Commission of the European Community.
SAFETY CAUTION Safety and Compliance SAFETY CAUTION The following icon or caution statements may be placed on equipment to indicate the presence of potentially hazardous conditions: DUAL POWER CORDS CAUTION: “THIS UNIT HAS MORE THAN ONE POWER SUPPLY CORD. DISCONNECT ALL POWER SUPPLY CORDS TO COMPLETELY REMOVE POWER FROM THIS UNIT." "ATTENTION: CET APPAREIL COMPORTE PLUS D'UN CORDON D'ALIMENTATION. DÉBRANCHER TOUS LES CORDONS D'ALIMENTATION AFIN DE COUPER COMPLÈTEMENT L'ALIMENTATION DE CET ÉQUIPEMENT".
Waste Electrical and Electronic Equipment (WEEE) Safety and Compliance HIGH LEAKAGE CURRENT To reduce the risk of electric shock due to high leakage currents, a reliable grounded (earthed) connection should be checked before servicing the power distribution unit (PDU).
Safety and Compliance Important Safety Information HP Integrity NonStop NS1000 Planning Guide —542527-004 Statements -6
Glossary For a glossary of Integrity NonStop NS-series terms, see the NonStop System Glossary in the NonStop Technical Library (NTL).
Glossary HP Integrity NonStop NS1000 Planning Guide —542527-004 Glossary- 2
Index Numbers C 4-port ServerNet extender card connections 5-6 description 1-1 installation slots 5-9 cabinet dimensions 3-6 cable labeling 6-4 cable length restrictions 6-6, A-2 cable specifications 6-5 cables, power 2-3 cabling blade element to IOAM enclosure 6-6 FCSA to FCDM 6-14 processors 0 to 7 6-8 calculation, heat 2-3, 3-8 calculation, weight 2-3, 3-7 clearances, service 3-6 configuration considerations Fibre Channel devices 6-12 IOAM enclosures 6-9 maximum number of enclosures 7-1 configuration
E Index documentation NonStop NS-series servers C-1 packet 4-11 ServerNet adapter configuration 4-12 dust and microscopic particles 2-4 dynamic IP addresses B-10 E electrical disturbances 2-2 electrical power loading 3-4 emergency power off (EPO) switches 2-1 enclosure combinations 1-3 dimensions 3-7 height in U 3-6 maximum number 7-1 power loading 3-4 types 1-3 weight 3-7 enclosure height in U 3-6 enclosure location 6-3 Enterprise Storage System (ESS) 5-10 European Union, harmonized cables 2-3 example h
I Index HP Integrity rx2620 server and blade element 4-3 described 5-6 LEDs 5-7 I indicator LEDs for blade element front panel 5-8 inrush current 2-2 Integrity NonStop NS1000 server characteristics 1-1 Integrity NonStop NS1000 system architecture 4-8 description 4-1 overview 1-1 internal cable product IDs A-1 internal interconnect cabling 6-5 internal ServerNet cabling 7-7 IOAM configuration considerations 6-9 ServerNet pathways 4-6 IOAM enclosure 5-8 IP addresses dynamic B-10 static B-10 L labeling, op
R Index power-fail monitoring B-14 processor interconnect cabling 6-7 processor element 4-2, 4-4 processor ID 5-5 Processor types 3-1 R startup characteristics, default 4-10 static IP addresses B-10 SWAN concentrator restriction B-11 system configurations, examples 3-9 system console configurations B-2 description B-1 overview 5-10 system disk location 4-10 rack offset 5-12 raised flooring 2-4 receiving and unpacking space 2-5 recovery, failure 4-4 related manuals for NonStop NS-series servers C-1 rest