HP StorageWorks X9300 Network Storage Gateway Administrator Guide Abstract This guide describes tasks related to cluster configuration and monitoring, system upgrade and recovery, hardware component replacement, and troubleshooting. It does not document X9000 file system features or standard Linux administrative tools and commands. For information about configuring and using X9000 Software file system features, see the HP StorageWorks X9000 File Serving Software File System User Guide.
Legal and notice information © Copyright 2010 Hewlett-Packard Development Company, L.P. Confidential computer software. Valid license from HP required for possession, use or copying. Consistent with FAR 12.211 and 12.212, Commercial Computer Software, Computer Software Documentation, and Technical Data for Commercial Items are licensed to the U.S. Government under vendor's standard commercial license. The information contained herein is subject to change without notice.
Contents 1 Product description ............................................................................. 9 HP X9300 Network Storage Gateway features ............................................................................... 9 System components ..................................................................................................................... 9 HP X9000 Software features ......................................................................................................
Setting up HBA monitoring .................................................................................................. Discovering HBAs ........................................................................................................ Identifying standby-paired HBA ports ............................................................................. Turning HBA monitoring on or off .................................................................................. Deleting standby port pairings ...
Viewing operating statistics for file serving nodes .......................................................................... 56 9 Maintaining the system ..................................................................... 57 Shutting down the system ........................................................................................................... Shutting down the X9000 Software ...................................................................................... Powering off the hardware ....
Installing firmware upgrades ...................................................................................................... 93 14 Troubleshooting ............................................................................. 95 Managing support tickets ........................................................................................................... Creating, viewing, and deleting support tickets ....................................................................... Support ticket states .
10 GbE (AW540A) ......................................................................................................... 10 GbE/IB (AW540B) ..................................................................................................... IB (AW541A) ........................................................................................................................ Base rack (AW546A) .............................................................................................................
Slovenian notice ............................................................................................................... 154 Spanish notice ................................................................................................................. 154 Swedish notice ................................................................................................................. 155 Glossary .......................................................................................... 157 Index ..
1 Product description The HP X9300 Network Storage Gateway is a flexible, scale-out solution that brings feature-rich gateway file services to HP MSA, EVA, P4000, or 3rd-party arrays or SANs.
• Front-end connectivity (chosen by the customer): • 10-GbE network • 1-GbE network • Infiniband network • Pre-installed software: • Red Hat Linux operating system • HP StorageWorks X9000 File Serving Software • Integrated Lights-Out 2 (iLO 2) remote management software For component and cabling diagrams, see Appendix A.
• Storage connection. For servers with HBA-protected Fibre Channel access, failure of the HBA triggers failover of the node to a designated standby server.
Product description
2 Getting started IMPORTANT: Do not modify any parameters of the operating system or kernel, or update any part of the X9300 Network Storage Gateway unless instructed to do so by HP; otherwise, the X9300 Network Storage Gateway could fail to operate properly. Setting up the X9300 Network Storage Gateway Installation steps The HP StorageWorks X9300 Network Storage Gateway is deployed at your site according to the terms in your Statement of Work. See your Statement of Work for details.
Management interfaces Cluster operations are managed through the X9000 Software management console, which provides both a GUI and a CLI. Most operations can be performed from either the GUI or the CLI.
The GUI dashboard opens in the same browser window. You can open multiple GUI windows as necessary. See the online help for information about all GUI displays and operations. The GUI dashboard enables you to monitor the entire cluster. There are three parts to the dashboard: System Status, Cluster Overview, and the Navigator. System Status The System Status section lists the number of cluster events that have occurred in the last 24 hours.
Alerts. Disruptive events that can result in loss of access to file system data. Examples are a segment that is unavailable or a server that cannot be accessed. Warnings. Potentially disruptive conditions where file system access is not lost, but if the situation is not addressed, it can escalate to an alert condition. Examples are a very high server CPU utilization level or a quota limit close to the maximum. Information. Normal events that change the cluster.
you select an object, a details page shows a summary for that object. The lower Navigator allows you to view details for the selected object, or to initiate a task. In the following example, we selected Cluster Configuration in the Navigator, and the Summary shows configuration information. In the lower Navigator, we selected NDMP Backup > Active Sessions to see details about the sessions. NOTE: When you perform an operation on the GUI, a spinning finger is displayed until the operation is complete.
Adding user accounts for GUI access X9000 Software supports administrative and user roles. When users log in under the administrative role, they can configure the cluster and initiate operations such as remote replication or snapshots. When users log in under the user role, they can view the cluster configuration and status, but cannot make configuration changes or initiate operations. The default administrative user name is ibrix. The default regular username is ibrixuser.
Using the Windows X9000 client GUI The Windows X9000 client GUI is the client interface to the management console. To open the GUI, double-click the desktop icon or select the IBRIX Client program from the Start menu on the client. The client program contains tabs organized by function. NOTE: The Windows X9000 client application can be started only by users with Administrative privileges. • Status.
Getting started
3 Configuring virtual interfaces for client access X9000 Software uses a cluster network interface to carry management console traffic and traffic between file serving nodes. This network is configured as bond0 when the cluster is installed. For clusters with an agile management console configuration, a virtual interface is also created for the cluster network interface to provide failover support for the console.
3. Assign an IP address to the bond1:1 VIFs on each node: # # # # ibrix_nic ibrix_nic ibrix_nic ibrix_nic –c –c –c –c –n –n –n –n bond1:1 bond1:1 bond1:1 bond1:1 –h –h –h –h node1 node2 node3 node4 –I –I –I –I 16.123.200.201 16.123.200.202 16.123.200.203 16.123.200.204 –M –M –M –M 255.255.255.0 255.255.255.0 255.255.255.0 255.255.255.0 Configuring standby backup nodes Assign standby backup nodes for the bond1:1 interface. The backup nodes should be configured in pairs.
clients could connect to bond1 on either host, as these clients do not support or require NIC failover. (The following sample output shows only the relevant fields.
Configuring virtual interfaces for client access
4 Configuring failover This chapter describes how to configure failover for agile management consoles, file serving nodes, network interfaces, and HBAs. Agile management consoles The management console maintains the cluster configuration and provides graphical and command-line user interfaces for managing and monitoring the cluster. Typically, one active management console and one passive management console are installed when the cluster is installed.
• Management console GUI. You will need to reconnect to the management console VIF after the failover. Failing over the management console manually To fail over the active management console manually, place the console into maintenance mode. Enter the following command on the node hosting the console: ibrix_fm -m maintenance The command takes effect immediately.
When automated failover is enabled, the management console listens for heartbeat messages that the file serving nodes broadcast at one-minute intervals. The management console automatically initiates failover when it fails to receive five consecutive heartbeats or, if HBA monitoring is enabled, when a heartbeat message indicates that a monitored HBA or pair of HBAs has failed.
• • • • The management console must have access to both the primary server and its standby. The same file system must be mounted on both the primary server and its standby. A server identified as a standby must be able to see all segments that might fail over to it. In a SAN environment, a primary server and its standby must use the same storage infrastructure to access a segment’s physical volumes (for example, a multiported RAID array).
APC power source. To identify an APC power source, use the following command: /bin/ibrix_powersrc -a -t {apc|apc_msp} -h POWERSRCNAME -n NUMSLOTS -I IPADDR For example, to identify an eight-port APC power source named ps1 at IP address 192.168.3.150: /bin/ibrix_powersrc -a -t apc -h ps1 -n 8 -I 192.168.3.150 For APC power sources, you must also associate file serving nodes to power source slots.
/bin/ibrix_powersrc -d -h POWERSRCLIST Turning automated failover on and off Automated failover is turned off by default. When automated failover is turned on, the management console starts monitoring heartbeat messages from file serving nodes. You can turn automated failover on and off for all file serving nodes or for selected nodes.
After failing back the node, determine whether the failback completed fully. If the failback is not complete, contact HP Support for assistance. NOTE: A failback might not succeed if the time period between the failover and the failback is too short, and the primary server has not fully recovered. HP recommends ensuring that both servers are up and running and then waiting 60 seconds before starting the failback. Use the ibrix_server -l command to verify that the primary server is up and running.
Identifying standbys To protect a network interface, you must identify a standby for it on each file serving node that connects to the interface. The following restrictions apply when identifying a standby network interface: • The standby network interface must be unconfigured and connected to the same switch (network) as the primary interface. • The file serving node that supports the standby network interface must have access to the file system that the clients on that interface will mount.
/bin/ibrix_nic -m -h MONHOST -D DESTHOST/IFNAME Deleting standbys To delete a standby for a network interface, use the following command: /bin/ibrix_nic -b -U HOSTNAME1/IFNAME1 For example, to delete the standby that was assigned to interface eth2 on file serving node s1.hp.com: /bin/ibrix_nic -b -U s1.hp.com/eth2 Setting up HBA monitoring You can configure High Availability to initiate automated failover upon detection of a failed HBA.
Identifying standby-paired HBA ports Identifying standby-paired HBA ports to the configuration database allows the management console to apply the following logic when they fail: • If one port in a pair fails, do nothing. Traffic will automatically switch to the surviving port, as configured by the vendor or the software. • If both ports in a pair fail, fail over the server’s segments to the standby server.
/bin/ibrix_hba -l [-h HOSTLIST] The following table describes the fields in the output. Field Description Host Server on which the HBA is installed. Node WWN This HBA’s WWNN. Port WWN This HBA’s WWPN. Port State Operational state of the port. Backup Port WWN WWPN of the standby port for this port (standby-paired HBAs only). Monitoring Whether HBA monitoring is enabled for this port.
/bin/ibrix_haconfig -l -h xs01.hp.com,xs02.hp.com Host HA Configuration Power Sources Backup Servers Auto Failover Nics Monitored Standby Nics HBAs Monitored xs01.hp.com FAILED PASSED PASSED PASSED FAILED PASSED FAILED xs02.hp.
5 Configuring cluster event notification Setting up email notification of cluster events You can set up event notifications by event type or for one or more specific events. To set up automatic email notification of cluster events, associate the events with email recipients and then configure email settings to initiate the notification process.
Turning email notifications on or off After configuration is complete, use the -m on option to turn on email notifications. To turn off email notifications, use the -m off option.
NOTE: Users of software versions earlier than 4.3 should be aware that the single ibrix_snmp command has been replaced by two commands, ibrix_snmpagent and ibrix_snmptrap. If you have scripts that include ibrix_snmp, be sure to edit them to include the correct commands. Whereas SNMPV2 security was enforced by use of community password strings, V3 introduces the USM and VACM. Discussion of these models is beyond the scope of this document. Refer to RFCs 3414 and 3415 at http://www.ietf.
The update command for SNMPv1 and v2 uses optional community names. By convention, the default READCOMMUNITY name used for read-only access and assigned to the agent is public. No default WRITECOMMUNITY name is set for read-write access (although the name private is often used). The following command updates a v2 agent with the write community name private, the agent’s system name, and that system’s physical location: ibrix_snmpagent -u –v 2 -w private -n agenthost.domain.
/bin/ibrix_event -c -y SNMP [-e ALERT|INFO|EVENTLIST] -m TRAPSINK For example, to associate all Alert events and two Info events with a trapsink at IP address 192.168.2.32, enter: /bin/ibrix_event -c -y SNMP -e ALERT,server.registered, filesystem.created -m 192.168.2.
ibrix_snmpgroup -c -g GROUPNAME [-s {noAuthNoPriv|authNoPriv|authPriv}] [-r READVIEW] [-w WRITEVIEW] [-x CONTEXT_NAME] [-m {exact|prefix}] For example, to create the group group2 to require authorization, no encryption, and read access to the hp view, enter: ibrix_snmpgroup -c -g group2 -s authNoPriv -r hp The format to create a user and add that user to a group follows: ibrix_snmpuser -c -n USERNAME -g GROUPNAME [-j {MD5|SHA}] [-k AUTHORIZATION_PASSWORD] [-y {DES|AES}] [-z PRIVACY_PASSWORD] Authenticati
6 Configuring system backups Backing up the management console configuration The management console configuration is automatically backed up whenever the cluster configuration changes. The backup takes place on the node hosting the active management console (or on the Management Server, if a dedicated management console is configured). The backup file is stored at /tmp/fmbackup.zip on the machine where it was created.
• Three-way NDMP operations between two X9000 systems Each file serving node functions as an NDMP Server and runs the NDMP Server daemon (ndmpd) process. When you start a backup or restore operation on the DMA, you can specify the node and tape device to be used for the operation.
NDMP process management Normally all NDMP actions are controlled from the DMA. However, if the DMA cannot resolve a problem or you suspect that the DMA may have incorrect information about the NDMP environment, take the following actions from the X9000 Software management console GUI or CLI: • Cancel one or more NDMP sessions on a file serving node. Canceling a session kills all spawned sessions processes and frees their resources if necessary. • Reset the NDMP server on one or more file serving nodes.
Viewing or rescanning tape and media changer devices To view the tape and media changer devices currently configured for backups, select Cluster Configuration from the Navigator, and then select NDMP Backup > Tape Devices. If you add a tape or media changer device to the SAN, click Rescan Device to update the list. If you remove a device and want to delete it from the list, you will need to reboot all of the servers to which the device is attached.
7 Creating hostgroups for X9000 clients A hostgroup is a named set of X9000 clients. Hostgroups provide a convenient way to centrally manage clients using the management console. You can put different sets of clients into hostgroups and then perform the following operations on all members of the group: • • • • • Create and delete mountpoints Mount file systems Prefer a network interface Tune host parameters Set allocation policies Hostgroups are optional.
hostgroups. To do this, mount ifs1 on the clients hostgroup, ifs2 on hostgroup A, ifs3 on hostgroup C, and ifs4 on hostgroup D, in any order. Then, set Tuning 1 on the clients hostgroup and Tuning 2 on hostgroup B. The end result is that all clients in hostgroup B will mount ifs1 and implement Tuning 2. The clients in hostgroup A will mount ifs2 and implement Tuning 1. The clients in hostgroups C and D respectively, will mount ifs3 and ifs4 and implement Tuning 1.
IP address that corresponds to a client network. Adding a domain rule to a hostgroup restricts its members to X9000 clients that are on the specified subnet. You can add a domain rule at any time. To add a domain rule to a hostgroup, use the ibrix_hostgroup command as follows: /bin/ibrix_hostgroup -a -g GROUPNAME -D DOMAIN For example, to add the domain rule 192.168 to the finance group: /bin/ibrix_hostgroup -a -g finance -D 192.
Creating hostgroups for X9000 clients
8 Monitoring cluster operations Monitoring the status of file serving nodes The dashboard on the management console GUI displays information about the operational status of file serving nodes, including CPU, I/O, and network performance information. To view status from the CLI, use the ibrix_server -l command.
• Alerts. A disruptive evens that can result in loss of access to file system data. For example, a segment is unavailable or a server is unreachable. • Warnings. A potentially disruptive condition where file system access is not lost, but if the situation is not addressed, it can escalate to an alert condition. Some examples are reaching a very high server CPU utilization or nearing a quota limit. • Information.
• Pings remote file serving nodes that share a network with the test hosts. Remote servers that are pingable might not be connected to a test host because of a Linux or X9000 Software issue. Remote servers that are not pingable might be down or have a network problem. • If test hosts are assigned to be network interface monitors, pings their monitored interfaces to assess the health of the connection. (For information on network interface monitoring, see “Using network interface monitoring” on page 31.
i080 PASSED lab13-116 PASSED Server Client Up Up Mon Apr 09 16:45:03 EDT 2007 Mon Apr 09 16:07:22 EDT 2007 Viewing a detailed health report To view a detailed health report, use the ibrix_health -i command: /bin/ibrix_health -i -h HOSTLIST [-f] [-s] [-v] The -f option displays results only for hosts that failed the check. The -s option includes information about the file system and its segments. The -v option includes details about checks that received a Passed or Warning result.
Check Results ============= Check : lab15-62 can ping remote segment server hosts ===================================================== Check Description Result Result Information ------------------------------- ------ -----------------Remote server lab15-61 pingable PASSED Check : Physical volumes are readable ===================================== Check Description Information ------------------------------------------------------------------------Physical volume 0ownQk-vYCm-RziC-OwRU-qStr-C6d5-ESrMIf read
Viewing operating statistics for file serving nodes Periodically, the file serving nodes report the following statistics to the management console: • Summary. General operational statistics including CPU usage, disk throughput, network throughput, and operational state. For information about the operational states, see Monitoring the status of file serving nodes, page 51. • IO. Aggregate statistics about reads and writes. • Network. Aggregate statistics about network inputs and outputs. • Memory.
9 Maintaining the system Shutting down the system To shut down the system completely, first shut down the X9000 Software, and then power off the system hardware. Shutting down the X9000 Software Use the following procedure to shut down the X9000 Software. Unless noted otherwise, run the commands from the dedicated Management Console or from the node hosting the active agile management console. 1. Disable HA for all file serving nodes: ibrix_server -m -U 2.
Powering off the hardware Power off the file serving nodes in any order. The step completely shuts down the cluster. Starting the system To start the system, first power on the file serving nodes, and then start the X900 Software. Starting the X9000 Software To start the X9000 Software, complete the following steps: 1. 2. 3. 4. Power on the dedicated Management Console or the node hosting the active agile management console.
Starting and stopping processes You can start, stop, and restart processes and can display status for the processes that perform internal X9000 Software functions. The following commands also control the operation of PostgreSQL on the machine. The PostgreSQL service is available at /usr/local/ibrix/init/. To start and stop processes and view process status on the management console, use the following command: /etc/init.
• To reset host parameters to their default values on nodes or hostgroups: /bin/ibrix_host_tune -U {-h HOSTLIST|-g GROUPLIST} [-n OPTIONS] To reset all options on all file serving nodes, hostgroups, and X9000 clients, omit the -h HOSTLIST and -n OPTIONS options. To reset host tunings on all clients, include the -g clients option. The values that are restored depend on the values specified for the -h HOSTLIST command: • File serving nodes.
Migrating specific segments Use the following command to migrate ownership of the segments in LVLIST on file system FSNAME to a new host and update the source host: /bin/ibrix_fs -m -f FSNAME -s LVLIST -h HOSTNAME [-M] [-F] [-N] To force the migration, include -M. To skip the source host update during the migration, include -F. To skip host health checks, include -N. The following command migrates ownership of ilv2 and ilv3 in file system ifs1 to s1.hp.
Segments from the lower Navigator. If segments with adequate space are not available, add segments to the file system. In this example, the data from segment 1 will be evacuated to segments 2 and 3. 3. If quotas are enabled on the file system, disable them: ibrix_fs -q -D -f FSNAME 4. Evacuate the segment. Select the file system on the management console GUI and then select Tasks > Rebalancer from the lower Navigator.
Maintaining networks Cluster and user network interfaces X9000 Software supports the following logical network interfaces: • Cluster network interface. This network interface carries management console traffic, traffic between file serving nodes, and traffic between file serving nodes and clients. A cluster can have only one cluster interface. For backup purposes, each file serving node and management console can have two cluster NICs. • User network interface.
Identifying a user network interface for a file serving node To identify a user network interface for specific file serving nodes, use the ibrix_nic command. The interface name (IFNAME) can include only alphanumeric characters and underscores, such as eth1. /bin/ibrix_nic -a -n IFNAME -h HOSTLIST If you are identifying a VIF, add the VIF suffix (:nnnn) to the physical interface name.
Preferring network interfaces After creating a user network interface for file serving nodes or X9000 clients, you will need to prefer the interface for those nodes and clients. (It is not necessary to prefer a network interface for NFS or CIFS clients, because they can select the correct user network interface at mount time.) When you prefer a user network interface for traffic from a source host to a destination host, traffic in the reverse direction remains defaulted to the cluster interface.
Making network changes This section describes how to change IP addresses, change the cluster interface, manage routing table entries, and delete a network interface. Changing the IP address for a Linux X9000 client After changing the IP address for a Linux X9000 client, you must update the X9000 Software configuration with the new information to ensure that the management console can communicate with the client. Use the following procedure: 1. Unmount the file system from the client. 2.
To specify a new cluster interface for a cluster with a dedicated management console, use the following command: /bin/ibrix_nic -t -n IFNAME -h HOSTNAME To specify a new virtual cluster interface for a cluster with an agile management console configuration, use the following command: /bin/ibrix_fm -c –d -n -v cluster [–I ] Managing routing table entries X9000 Software supports one route for ea
Field Description BACKUP HOST File serving node for the standby network interface. BACKUP-IF Standby network interface. HOST File serving node. An asterisk (*) denotes the management console. IFNAME Network interface on this file serving node. IP_ADDRESS IP address of this NIC. LINKMON Whether monitoring is on for this NIC. MAC_ADDR MAC address of this NIC. ROUTE IP address in routing table used by this NIC. STATE Network interface state. TYPE Network type (cluster or user).
10 Migrating to an agile managment console configuration The agile management console configuration provides one active management console and one passive management console installed on different nodes in the cluster. The migration procedure configures the current Management Server machine as a host for an agile management console and installs another instance of the agile management console on a file serving node.
1. The agile management console uses a virtual interface (VIF) IP address to enable failover and prevent any interruptions to file serving nodes and X9000 clients. The existing cluster NIC IP address becomes the permanent VIF IP address. Identify an unused IP address to use as the Cluster NIC IP address for the currently running management console. Then reconfigure the cluster and user networks: • Edit the /etc/sysconfig/network-scripts/ifcfg-bond0 file.
For example: ./ibrixinit -ts -C eth4 -i 172.16.3.100 -F 6. Register the Management Server hostname and IP address with the now “active” agile management console: ibrix_fm -R -I -a For example: ibrix_fm –R x109s1 –I 172.16.3.100 –a 172.16.3.1 7.
[root@x109s3 ibrix]# ibrix_fm -i FusionServer: x109s3 (passive, quorum is running) ============================= Command succeeded Removing the dedicated Management Server This procedure removes the dedicated Management Server machine from the cluster and installs another instance of the agile management console on a second file serving node. Complete the following steps: 1. On the Management Server machine, place the agile management console into maintenance mode: ibrix_fm -m maintenance 2.
x109s3 172.16.3.3 Command succeeded! 5. Remove the Management Server machine from the cluster database: ibrix_server -d -h HOSTNAME 6. To provide high availability for the management console, install a passive agile management console on another file serving node. In the command, the -F option forces the overwrite of the new_lvm2_uuid file that was installed with the X9000 Software.
Migrating to an agile managment console configuration
11 Upgrading the X9000 Software This chapter describes how to upgrade to the latest X9000 File Serving Software release. The management console and all file serving nodes must be upgraded to the new release at the same time. X9000 Clients are supported for one version beyond their release. For example, an X9000 5.3.2 client can run with a 5.4 X9000 server, but not with a 5.5 X9000 server. IMPORTANT: Do not start new remote replication jobs while a cluster upgrade is in progress.
2. On the current active management console, move the /ibrix directory used in the previous release installation to ibrix.old. For example, if you expanded the tarball in /root during the previous X9000 installation on this node, the installer is in /root/ibrix. 3. On the current active management console, expand the distribution tarball or mount the distribution DVD in a directory of your choice. Expanding the tarball creates a subdirectory named ibrix that contains the installer program.
file system access to continue. This procedure cannot be used for major upgrades, but is appropriate for minor and maintenance upgrades. • Offline upgrades. This procedure requires that you first unmount file systems and stop services. (Each file serving node may need to be rebooted if NFS or CIFS causes the unmount operation to fail.) You can then perform the upgrade. Clients will experience a short interruption to file system access while each file serving node is upgraded.
Upgrading file serving nodes After the management console has been upgraded, complete the following steps on each file serving node: 1. From the management console, manually fail over the file serving node: /bin/ibrix_server -f -p -h HOSTNAME The node reboots automatically. 2. Move the /ibrix directory used in the previous release installation to ibrix.old.
3. Upgrade X9000 Clients: • For Linux clients, see Upgrading Linux X9000 clients, page 88. • For Windows clients, see Upgrading Windows X9000 clients, page 89. 4. Verify that all version indicators match for file serving nodes and X9000 Clients. Run the following command from the management console: /bin/ibrix_version –l If there is a version mismatch, run the /ibrix/ibrixupgrade -f script again on the affected node, and then recheck the versions.
2. Move the /ibrix directory used in the previous release installation to ibrix.old. For example, if you expanded the tarball in /root during the previous X9000 installation on this node, the installer is in /root/ibrix. 3. Expand the distribution tarball or mount the distribution DVD in a directory of your choice. Expanding the tarball creates a subdirectory named ibrix that contains the installer program. For example, if you expand the tarball in /root, the installer is in /root/ibrix.
Completing the upgrade 1. Remount all file systems: /bin/ibrix_mount -f -m 2. From the management console, turn automated failover back on: /bin/ibrix_server -m 3. Confirm that automated failover is enabled: /bin/ibrix_server -l In the output, HA displays on. 4. From the management console, perform a manual backup of the upgraded configuration: /bin/ibrix_fm -B 5.
Upgrading the file serving nodes hosting the management console Complete the following steps: 1. On the node hosting the active management console, force a backup of the management console configuration: /bin/ibrix_fm -B The output is stored at /usr/local/ibrix/tmp/fmbackup.zip. Be sure to save this file in a location outside of the cluster. 2. On the active management console node, disable automated failover on all file serving nodes: /bin/ibrix_server -m -U 3.
11. On the node with the active agile management console, expand the distribution tarball or mount the distribution DVD in a directory of your choice. Expanding the tarball creates a subdirectory named ibrix that contains the installer program. For example, if you expand the tarball in /root, the installer is in /root/ibrix. nl 12. Change to the installer directory if necessary and run the upgrade: .
Also run the following command, which should report that the console is passive: /bin/ibrix_fm -i 22. Check /usr/local/ibrix/log/fusionserver.log for errors. 23. If the upgrade was successful, fail back the node. Run the following command on the node with the active agile management console: /bin/ibrix_server -f -U -h HOSTNAME 24.
ipfs1 102592 0 (unused) If either grep command returns empty, contact HP Support. 7. From the management console, verify that the new version of X9000 Software FS/IAS has been installed on the file serving node: /bin/ibrix_version -l –S 8. If the upgrade was successful, failback the file serving node: /bin/ibrix_server -f -U -h HOSTNAME 9. Repeat steps 1 through 8 for each remaining file serving node in the cluster.
NOTE: To determine which node is hosting the active management console, run the following command: /bin/ibrix_fm -i Preparing for the upgrade 1. On the active management console node, disable automated failover on all file serving nodes: /bin/ibrix_server -m -U 2. Verify that automated failover is off. In the output, the HA column should display off. /bin/ibrix_server -l 3.
/etc/init.d/ibrix_fusionmanager status The status command confirms whether the correct services are running. Output will be similar to the following: Fusion Manager Daemon (pid 18748) running... 7. Check /usr/local/ibrix/log/fusionserver.log for errors. 8. Upgrade the remaining management console node. Move the ibrix directory used in the previous release to ibrix.old. Then expand the distribution tarball or mount the distribution DVD in a directory of your choice.
6. From the active management console node, verify that the new version of X9000 Software FS/IAS is installed on the file serving nodes: /bin/ibrix_version -l –S Completing the upgrade 1. Remount the X9000 Software file systems: /bin/ibrix_mount -f -m 2. From the node hosting the active management console, turn automated failover back on: /bin/ibrix_server -m 3.
The IAD service should be running, as shown in the sample output above. If it is not, contact HP Support. Upgrading Windows X9000 clients Complete the following steps on each client: 1. 2. 3. 4. 5. Remove the old Windows X9000 client software using the Add or Remove Programs utility in the Control Panel. Copy the Windows X9000 client MSI file for the upgrade to the machine. Launch the Windows Installer and follow the instructions to complete the upgrade.
To do this, launch the X9000 Software Client User Interface on the Client. Go to the Registration Tab, enter the Management Console name, select Recover Registration, and then click Register. If you are prompted to overwrite the existing registration, select yes to complete the operation.
12 Licensing This chapter describes how to view your current license terms and how to obtain and install new X9000 Software product license keys. NOTE: For MSA2000 G2 licensing (for example, snapshots), see the MSA2000 G2 documentation. Viewing license terms The X9000 Software license file is stored in the installation directory on the management console. To view the license from the management console GUI, select Cluster Configuration in the Navigator and then select License.
3. Launch the AutoPass GUI: /usr/local/ibrix/bin/fusion-license-manager 4. In the AutoPass GUI, go to Tools, select Configure Proxy, and configure your proxy settings. 5. Click Retrieve/Install License > Key and then retrieve and install your license key. If the management console does not have an Internet connection, retrieve the license from a machine that does have a connection, deliver the file with the license to the management console machine, and then use the AutoPass GUI to import the license.
13 Upgrading firmware Upgradable firmware The HP X9300 system includes several components with upgradable firmware. The following table lists these components and specifies whether they can be upgraded online and in a nondisruptive manner.
Upgrading firmware
14 Troubleshooting Managing support tickets A support ticket includes system and X9000 software information useful for analyzing performance issues and node terminations. A support ticket is created automatically if a file serving node terminates unexpectedly. You can also create a ticket manually if your cluster experiences issues that need to be investigated by HP Support.
To view a support ticket on the GUI, select Support Tickets from the Navigator. On the CLI, use the following command to view all support tickets: /bin/ibrix_supportticket -l To view details for a specific support ticket, use the following command: /bin/ibrix_supportticket -v -n When you no longer need a support ticket, you can delete it. From the GUI, select Support Ticket from the Navigator.
NOTE: During the X9000 Software installation, the names of crash dumps in the /var/crash directory change to include _PROCESSED. For example, 2010-03-08-10:09 changes to 2010-03-08-10:09_PROCESSED. NOTE: Be sure to monitor the /var/crash directory and remove any unneeded processed crash dumps. Configuring shared ssh keys To configure one-way shared ssh keys on the cluster, complete the following steps: 1. On the management console, run the following commands as root: # mkdir -p $HOME/.
Troubleshooting specific issues Software services Cannot start services on the management console, a file serving node, or a Linux X9000 client SELinux might be enabled. To determine the current state of SELinux, use the getenforce command. If it returns enforcing, disable SELinux using either of these commands: setenforce Permissive setenforce 0 To permanently disable SELinux, edit its configuration file (/etc/selinux/config) and set SELINUX=parameter to either permissive or disabled.
operations. If, however, the network connection between a client and the management console is not active, the client cannot receive the updated map, resulting in client I/O errors. To fix the problem, restore the network connection between the clients and the management console. Windows X9000 clients Logged in but getting a “Permission Denied” message The X9000 client cannot access the Active Directory server because the domain name was not specified.
To run a health check on a file serving node, use the following command: /bin/ibrix_health -i -h HOSTLIST If the last line of the output reports Passed, the file system information on the file serving node and management console is consistent. To repair file serving node information, use the following command: /bin/ibrix_dbck -o -f FSNAME [-h HOSTLIST] To repair information on all file serving nodes, omit the -h HOSTLIST argument.
15 Replacing components Customer replaceable components WARNING! Before performing any of the procedures in this chapter, read the important warnings, precautions, and safety information in Appendix C and Appendix D. IMPORTANT: To avoid unintended consequences, HP recommends that you perform the procedures in this chapter during scheduled maintenance times. CAUTION: Be sure the replacement is available before removing a component or a blanking panel.
Parts-only warranty service Your HP Limited Warranty could include a parts-only warranty service. Under the terms of parts-only warranty service, HP provides replacement parts, free of charge. For parts-only warranty service, CSR part replacement is mandatory. If you request HP to replace these parts for you, you are charged for the travel and labor costs of this service.
4. Power on the server. a. Check the MAC address of the new NIC. Use the ifdown eth(n) command, and then use the ip addr command. b. Change the HWADDR= line in the relevant ifcfg-eth(n) file to be the MAC address from the previous step. c. Follow this procedure for every NIC that was affected by the physical board swapout (for example, if this was a dual NIC, then two files need to be updated). d. Restart the network subsystem with service network restart. e.
5. For a file serving node, fail back the server using the GUI or CLI: • On the GUI, select Servers from the Navigator pane, and then select the appropriate server from the Servers pane. Next, select the server name in the left pane, and click Failback. • On the CLI, execute the following command: ibrix_server -f –U –h Replacing a Fibre Channel HBA To replace a Fibre Channel HBA on a file serving node: 1.
8. Enable standby pairing and monitoring on the new HBA if desired, using the GUI or CLI: • On the GUI, select Servers from the Navigator pane, and then select the appropriate server from the Servers pane. Select HBAs in the left pane, select the appropriate Node WWN, and click Modify. On the Modify HBA Properties window, select Enable monitoring and Enable backup pair, and select the WWPN for the backup HBA.
Replacing components
16 Recovering a file serving node Use the following procedure to recover a failed file serving node. You will need to create a QuickRestore DVD, as described later, and then install it on the affected node. This step installs the operating system and X9000 Software on the node and launches a configuration wizard.
6. When the following screen appears, enter qr to recover the file serving node. The server reboots automatically after the software is installed. Remove the DVD from the DVD-ROM drive. 7. When your cluster was configured initially, the installer may have created a template for configuring file serving nodes. To use this template to configure the file serving node undergoing recovery, go to “Configuring a file serving node using the original template” on page 108.
3. The Configuration Wizard attempts to discover management consoles on the network and then displays the results. Select the appropriate management console for this cluster. NOTE: If the list does not include the appropriate management console, or you want to customize the cluster configuration for the file serving node, select Cancel. Go to Configuring a file serving node manually, page 112 for information about completing the configuration. 4.
5. The Verify Configuration window shows the configuration received from the management console. Select Accept to apply the configuration to the server and register the server with the management console. NOTE: If you select Reject, the wizard will exit and the shell prompt will be displayed. You can restart the Wizard by entering the command /usr/local/ibrix/autocfg/bin/ menu_ss_wizard or logging in to the server again. 6.
IMPORTANT: Configure a passive agile management console only if the following conditions are met: • The agile management console is enabled and an active agile management console is configured. • The passive agile management console is not configured on any other nodes in the cluster. If you have configured a user network, enter a VIF IP address and netmask.
Completing the restore on a file serving node Complete the following steps: 1. Ensure that you have root access to the node. The restore process sets the root password to hpinvent, the factory default. 2. The QuickRestore DVD enables the iptables firewall. Either make the firewall configuration match that of your other server blades to allow traffic on appropriate ports, or disable the service entirely by running the chkconfig iptables off and service iptables stop commands.
3. The Configuration Wizard attempts to discover management consoles on the network and then displays the results. Select Cancel to configure the node manually. 4. The file serving node Configuration Menu appears.
5. The Cluster Configuration Menu lists the configuration parameters that you will need to set. Use the Up and Down arrow keys to select an item in the list. When you have made your select, press Tab to move to the buttons at the bottom of the dialog box, and press Space to go to the next dialog box. 6. Select Management Console from the menu, and enter the IP address of the management console. This is typically the address of the management console on the cluster network.
7. Select Hostname from the menu, and enter the hostname of this server. 8. Select Time Zone from the menu, and then use Up or Down to select your time zone.
9. Select Default Gateway from the menu, and enter the IP Address of the host that will be used as the default gateway. 10. Select DNS Settings from the menu, and enter the IP addresses for the primary and secondary DNS servers that will be used to resolve domain names. Also enter the DNS domain name.
11. Select NTP Servers from the menu, and enter the IP addresses or hostnames for the primary and secondary NTP servers. 12. Select Networks from the menu. Select to create a bond for the cluster network.
You are creating a bonded interface for the cluster network; select Ok on the Select Interface Type dialog box. Enter a name for the interface (bond0 for the cluster interface) and specify the appropriate options and slave devices.
13. When the Configure Network dialog box reappears, select bond0.
14. To complete the bond0 configuration, enter a space to select the Cluster Network role. Then enter the IP address and netmask information that the network will use. Repeat this procedure to create a bonded user network (typically bond1) and any custom networks as required.
15. When you have completed your entries on the file serving node Configuration Menu, select Continue. 16. Verify your entries on the confirmation screen, and select Commit to apply the values to the file serving node and register it with the management console. 17. If the hostname specified for the node already exists in the cluster (the name was used by the node you are replacing), the Replace Existing Server window asks whether you want to replace the existing server with the node you are configuring.
If you configured a passive management console, enter the following command to verify the status of the console: ibrix_fm -i Next, go to Completing the restore on a file serving node, page 112.
17 Support and other resources Contacting HP For worldwide technical support information, see the HP support website: http://www.hp.
HP websites For additional information, see the following HP websites: • • • • • • • http://www.hp.com/go/X9000 http://www.hp.com http://www.hp.com/go/storage http://www.hp.com/service_locator http://www.hp.com/support/manuals http://www.hp.com/support/downloads http://www.hp.com/storage/whitepapers Rack stability Rack stability protects personnel and equipment. WARNING! To reduce the risk of personal injury or damage to equipment: • Extend leveling jacks to the floor.
A Component and cabling diagrams Component diagrams Front view of file serving node Item Description 1 Quick-release levers (2) 2 HP Systems Insight Manager display 3 Hard drive bays 4 SATA optical drive bay 5 Video connector 6 USB connectors (2) Rear view of file serving node X9300 Network Storage Gateway Administrator Guide 125
Item Description 1 PCI slot 5 2 PCI slot 6 3 PCI slot 4 4 PCI slot 2 5 PCI slot 3 6 PCI slot 1 7 Power supply 2 (PS2) 8 Power supply 1 (PS1) 9 USB connectors (2) 10 Video connector 11 NIC 1 connector 12 NIC 2 connector 13 Mouse connector 14 Keyboard connector 15 Serial connector 16 iLO 2 connector 17 NIC 3 connector 18 NIC 4 connector 126 Component and cabling diagrams
Server PCIe card PCI slot HP SC08Ge 3Gb SAS Host Bus Adapter 1 NC364T Quad 1Gb NIC 2 empty 3 empty 4 empty 5 empty 6 HP SC08Ge 3Gb SAS Host Bus Adapter 1 empty 2 empty 3 NC522SFP dual 10Gb NIC 4 empty 5 empty 6 HP SC08Ge 3Gb SAS Host Bus Adapter 1 NC364T Quad 1Gb NIC 2 empty 3 HP SC08Ge 3Gb SAS Host Bus Adapter 4 empty 5 empty 6 HP SC08Ge 3Gb SAS Host Bus Adapter 1 HP SC08Ge 3Gb SAS Host Bus Adapter 2 empty 3 NC522SFP dual 10Gb NIC 4 empty 5 empty 6 SATA 1G
Cabling diagrams Cluster network cabling diagram NOTE: The MSA controllers are optional in the X9300 models.
B Spare parts list This appendix lists spare parts (both customer replaceable and non customer replaceable) for the X9300 Network Storage Gateway components. Spare parts are categorized as follows: • Mandatory. Parts for which customer self repair is mandatory. If you request HP to replace these parts, you will be charged for the travel and labor costs of this service. • Optional. Parts for which customer self repair is optional. These parts are also designed for customer self-repair.
Description Spare part number Customer self repair SPS-CABLE, SAS BACKPLANE 496070-001 Mandatory SPS-CABLE, SATA DVD PWR 496071-001 Mandatory SPS-BD, SID 496073-001 Mandatory SPS-CAGE, HD, SFF 496074-001 Mandatory SPS-CAGE, DVD OPT DRIVE 496076-001 Mandatory SPS-BD, PCIX 496077-001 Optional SPS-BD, PCIE 496078-001 Optional SPS-BEZEL 496080-001 Mandatory SPS-BACKPLANE,SAS 507690-001 Optional SPS-PROC,NEHALEM EP 2.
Description Spare part number Customer self repair SPS-CAGE, PS BACKPLANE 496063-001 Optional SPS-HEATSINK, 80W 496064-001 Optional SPS-FAN 496066-001 Mandatory SPS-CAGE, FAN 496067-001 Mandatory SPS-BD,SYSTEM I/O, W/SUBPAN 496069-001 Optional SPS-CABLE, SAS BACKPLANE 496070-001 Mandatory SPS-CABLE, SATA DVD PWR 496071-001 Mandatory SPS-BD, SID 496073-001 Mandatory SPS-CAGE, HD, SFF 496074-001 Mandatory SPS-CAGE, DVD OPT DRIVE 496076-001 Mandatory SPS-BD, PCIX 496077-001 O
10 GbE spare parts 10 GbE (AW540A) Description Spare part number Customer self repair SPS-HOOD 496056-001 Mandatory SPS-CARD, RISER 496057-001 Optional SPS-KIT, MISC HARDWARE 496058-001 Mandatory SPS-CAGE, PCI 496060-001 Optional SPS-BAFFLE 496061-001 Mandatory SPS-BACKPLANE, PS 496062-001 Optional SPS-CAGE, PS BACKPLANE 496063-001 Optional SPS-HEATSINK, 80W 496064-001 Optional SPS-FAN 496066-001 Mandatory SPS-CAGE, FAN 496067-001 Mandatory SPS-BD,SYSTEM I/O, W/SUBPAN 49606
Description Spare part number Customer self repair SPS-TRAY, DVD 532390-001 Mandatory SPS-POWER SUPPLY, 750W 511778-001 Optional SPS-BD,NC522SFP+ 10 GIGABIT,SERVER ADPTR 468349-001 Optional SPS-HARDWARE MTG KIT 574765-001 Mandatory SPS-CORD,AC PWR IEC/IEC 6 FT 142258-001 Mandatory Description Spare part number Customer self repair SPS-HOOD 496056-001 Mandatory SPS-CARD, RISER 496057-001 Optional SPS-KIT, MISC HARDWARE 496058-001 Mandatory SPS-CAGE, PCI 496060-001 Optional SP
Description Spare part number Customer self repair SPS-PROC,NEHALEM EP 2.26 GHZ, 8M, 80W 490073-001 Optional SPS-DIMM,4GB PC310600R,256MX4,ROHS 501534-001 Mandatory SPS-DRV,HD,146GB,10K 2.
Description Spare part number Customer self repair SPS-BEZEL 496080-001 Mandatory SPS-BACKPLANE,SAS 507690-001 Optional SPS-PROC,NEHALEM EP 2.26 GHZ, 8M, 80W 490073-001 Optional SPS-DIMM,8GB PC310600R,512MX4,ROHS 501536-001 Optional SPS-DRV,HD,146GB,15K 2.
Description Spare part number Customer self repair SPS-SWITCH,SVR CNSL,KVM,0X2X16,USB 410529-001 Mandatory SPS-RACK,BUS BAR & WIRE TRAY 457015-001 Optional SPS-STICK,4X FIXED,C-13,OFFSET,WW 483915-001 Optional SPS-ASSY,RETENTION,PWR CD, LT&RT BRKT 490992-001 Mandatory SPS-SWITCH,KVM,SVR CNSL,0X2X16 517691-001 Optional CABLE, CONSOLE D-SUB9 - RJ45 L250 5188-6699 Mandatory PWR-CORD OPT-918 3-COND 2.0-M-LG ROHS 8120-4753 Not customer replaceable PWR-CORD OPT-927 3-COND 2.
C Warnings and precautions Electrostatic discharge information See Electrostatic discharge. Grounding methods There are several methods for grounding. Use one or more of the following methods when handling or installing electrostatic sensitive parts: • Use a wrist strap connected by a ground cord to a grounded workstation or computer chassis. Wrist straps are flexible straps with a minimum of 1 megohm ±10 percent resistance in the ground cords.
WARNING! Any RJ-45 receptacle marked with these symbols indicates a network interface connection. To reduce the risk of electrical shock, fire, or damage to the equipment, do not plug telephone or telecommunications connectors into this receptacle. WARNING! Any surface or area of the equipment marked with these symbols indicates the presence of a hot surface or hot component. Contact with this surface could result in injury.
WARNING! To reduce the risk of personal injury or damage to the equipment: • Observe local occupational safety requirements and guidelines for heavy equipment handling. • Obtain adequate assistance to lift and stabilize the product during installation or removal. • Extend the leveling jacks to the floor. • Rest the full weight of the rack on the leveling jacks. • Attach stabilizing feet to the rack if it is a single-rack installation. • Ensure the racks are coupled in multiple-rack installations.
Device warnings and precautions WARNING! To reduce the risk of electric shock or damage to the equipment: • Allow the product to cool before removing covers and touching internal components. • Do not disable the power cord grounding plug. The grounding plug is an important safety feature. • Plug the power cord into a grounded (earthed) electrical outlet that is easily accessible at all times. • Disconnect power from the device by unplugging the power cord from either the electrical outlet or the device.
CAUTION: To properly ventilate the system, you must provide at least 7.6 centimeters (3.0 inches) of clearance at the front and back of the device. CAUTION: Schedule physical configuration changes during periods of low or no activity. If the system is performing rebuilds, RAID migrations, array expansions LUN expansions, or experiencing heavy I/O, avoid physical configuration changes such as adding or replacing hard drives or hot-plugging a controller or any other component.
Warnings and precautions
D Regulatory compliance and safety Regulatory compliance identification numbers For the purpose of regulatory compliance certifications and identification, this product has been assigned a unique regulatory model number. The regulatory model number can be found on the product nameplate label, along with all required approval markings and information. When requesting compliance information for this product, always refer to this regulatory model number.
• Reorient or relocate the receiving antenna. • Increase the separation between the equipment and receiver. • Connect the equipment into an outlet on a circuit different from that to which the receiver is connected. • Consult the dealer or an experienced radio or television technician for help. Declaration of conformity for products marked with the FCC logo, United States only This device complies with Part 15 of the FCC Rules.
WARNING! Use of controls or adjustments, or performance of procedures other than those specified herein, or in the laser product's installation guide, could result in hazardous radiation exposure. To reduce the risk of exposure to hazardous radiation: • Do not try to open the module enclosure. There are no user-serviceable components inside. • Do not operate controls, make adjustments, or perform procedures to the laser device, other than those specified herein.
BSMI notice Japanese notice Korean notice (A&B) Class A equipment Class B equipment 146 Regulatory compliance and safety
Safety Battery Replacement notice WARNING! The computer contains an internal lithium manganese dioxide, a vanadium pentoxide, or an alkaline battery pack. A risk of fire and burns exists if the battery pack is not properly handled. To reduce the risk of personal injury: • Do not attempt to recharge the battery. • Do not expose the battery to temperatures higher than 60˚C (140˚F). • Do not disassemble, crush, puncture, short external contacts, or dispose of in fire or water.
Japanese Power Cord notice Electrostatic discharge To prevent damage to the system, be aware of the precautions you need to follow when setting up the system or handling parts. A discharge of static electricity from a finger or other conductor could damage system boards or other static-sensitive devices. This type of damage could reduce the life expectancy of the device.
Waste Electrical and Electronic Equipment directive Czechoslovakian notice Danish notice Dutch notice X9300 Network Storage Gateway Administrator Guide 149
English notice Estonian notice Finnish notice 150 Regulatory compliance and safety
French notice German notice Greek notice X9300 Network Storage Gateway Administrator Guide 151
Hungarian notice Italian notice Latvian notice 152 Regulatory compliance and safety
Lithuanian notice Polish notice Portuguese notice X9300 Network Storage Gateway Administrator Guide 153
Slovakian notice Slovenian notice Spanish notice 154 Regulatory compliance and safety
Swedish notice X9300 Network Storage Gateway Administrator Guide 155
Regulatory compliance and safety
Glossary ACE Access control entry. ACL Access control list. ADS Active Directory Service. ALB Advanced load balancing. BMC Baseboard Management Configuration. CIFS Common Internet File System. The protocol used in Windows environments for shared folders. CLI Command-line interface. An interface comprised of various commands which are used to control operating system responses. CSR Customer self repair. DAS Direct attach storage.
MTU Maximum Transmission Unit. NAS Network attached storage. NFS Network file system. The protocol used in most UNIX environments to share folders or mounts. NIC Network interface card. A device that handles communication between a device and other devices on a network. NTP Network Time Protocol. A protocol that enables the storage system’s time and date to be obtained from a network-attached server, keeping multiple hosts and storage devices synchronized. OA HP Onboard Administrator.
WWNN World wide node name. A globally unique 64-bit identifier assigned to each Fibre Channel node process. WWPN World wide port name. A unique 64-bit address used in a FC storage network to identify each device in a FC network.
Glossary
Index A agile management console, 25 AutoPass, 91 B backups file systems, 43 management console configuration, 43 NDMP applications, 43 C Class A equipment, 143 Class B equipment, 143 cluster events, monitor, 51 health checks, 52 license key, 91 license, view, 91 log files, 55 operating statistics, 56 version numbers, view, 97 cluster interface change IP address, 66 change network, 66 defined, 63 components returning, 101 contacting HP, 123 customer self repair, 124 D declaration of conformity, 144 docum
G I grounding methods, 137 international notices and statements, 145 IP address change for cluster interface, 66 change for X9000 client, 66 H hazardous conditions symbols on equipment, 137 HBAs delete HBAs, 34 delete standby port pairings, 34 discover, 33 identify standby-paired ports, 34 list information, 34 monitor for high availability, 33 monitoring, turn on or off, 34 health check reports, 53 help obtaining, 123 High Availability agile management console, 25 automated failover, turn on or off, 30
user network interface add, 63 configuration rules, 66 defined, 63 identify for X9000 clients, 64 modify, 64 prefer, 65 unprefer, 65 P passwords, change GUI password, 19 Q QuickRestoreDVD, 107 R W rack stability warning, 124 regulatory compliance, 143 related documentation, 123 returning components, 101 routing table entries add, 67 delete, 67 warning rack stability, 124 warnings loading rack, 138 Waste Electrical and Electronic Equipment directive, 149 websites customer self repair, 124 HP, 124 Windo