Dell™ PowerVault™ NAS Fibre Channel Cluster Systems Installation and Troubleshooting Guide Introduction Preparing Your Systems for Clustering Cabling Your Cluster Hardware Installing Your Cluster in a Direct-Attached Environment Installing Your Cluster in a SAN Environment Maintaining Your Cluster Using MSCS Troubleshooting Cluster Data Sheets Abbreviations and Acronyms NOTE: A NOTE indicates important information that helps you make better use of your computer.
Back to Contents Page Introduction Dell™ PowerVault™ NAS Fibre Channel Cluster Systems Installation and Troubleshooting Guide Intended Audience Configuring Active and Passive Cluster Nodes Obtaining Technical Assistance Failover and Failback Support Overview Failover Policies NAS Fibre Channel Cluster Solution Minimum System Requirements Optional Cluster Configurations Other Documents You May Need This guide provides information for installing a Dell™ PowerVault™ NAS Fibre Cha
Obtaining Technical Assistance Dell Enterprise Training and Certification is available; see www.dell.com/training for more information. This service may not be offered in all locations. Overview This section provides an overview of clustering and the major cluster components used in the Windows Server 2003, Enterprise Edition operating system. NOTE: In this guide, Microsoft Cluster Service for Windows Storage Server 2003, Enterprise Edition is also referred to as MSCS.
Failover and Failback If one of the cluster nodes should fail for any reason, MSCS moves (or fails over) the virtual server to another cluster node. After the cluster node is repaired and brought online, MSCS may move (or fail back) the virtual server to the original cluster node, if required. This failover capability enables the cluster configuration to keep network resources and application programs running on the network while the failed node is taken offline, repaired, and brought back online.
Each storage system in the cluster is centrally managed by one host system (also called a management station) running EMC ControlCenter™ Navisphere Manager™—a storage management application used to configure Dell | EMC storage systems. Navisphere Manager provides centralized storage management and configuration from a single management console. Using a graphical user interface (GUI), you can select a specific view of your storage arrays, as shown in Table 1-1. Table 1-1.
Storage System NAS Fibre Channel clusters support the Dell | EMC CX200, CX400, and CX600 systems—modular 2-Gb/s Fibre Channel storage systems that provide scalable storage solutions in a direct-attached or SAN-attached environment. These storage systems consist of multiple hardware components that work together as a single system.
NOTICE: The hardware components in the processor enclosures are not compatible with the components in the Dell | EMC FC4500, FC4700, and FC4700-2 DAEs. Exchanging components between these enclosures could damage the enclosures and void your warranty. The supported processor enclosures vary in performance and configuration. Table 1-2 provides the features for each processor enclosure. See Table 1-13 in for more information on the supported processor enclosures. Table 1-2.
additional hard drives to the enclosure as needed. The DAE2 is also available as the DAE2-OS—a DAE2 enclosure with the core software (FLARE) preinstalled on the first five hard drives that provides both the primary storage and the operating system for the CX600 storage system.
The Fibre Channel switch functions as a director, mapping requests and responses between the interconnected devices and provides a dedicated interconnection between the server and storage system. See the Platform Guide for a list of supported Fibre Channel switches. NOTE: The Dell | EMC DS-24M2 (McData Spherion 4500) Fibre Channel switch supports Flexport configurations, which allows you to vary the number of active switch ports.
NAS Fibre Channel clusters support 2-Gb Fibre Channel, which provides high-speed data transfer between the cluster nodes and storage systems and supports the smaller SFP connectors for higher port density. Fibre Channel Switch Fabric A Fibre Channel switch fabric consists of one or more Fibre Channel switches that provide high-speed connections between servers and storage devices.
HBA Connects the cluster node's PCI bus to the storage system. Storage system Provides external storage for the host systems (cluster nodes and stand-alone systems). Supported storage systems include tape libraries, DAEs, and SPEs. SAN storage Provides centralized control of the SAN for easier management.
Access Logix is required when the storage system is connected to two or more clusters, two or more nonclustered systems, or a combination of both clustered and nonclustered systems. MirrorView (optional) Automatically duplicates primary storage system data for a cluster or stand-alone system to a secondary storage system in a campus environment. MirrorView implements synchronous mirroring, which ensures that data is updated on both the primary and secondary storage systems at the same time.
Using EMC PowerPath in a Direct-Attach Cluster EMC PowerPath provides failover capabilities and multiple path detection as well as dynamic load balancing between multiple ports on the same SP. Direct-attached clusters supported by Dell connect to a single port on each SP in the storage system. Because of the single port limitation, PowerPath can only provide failover protection in a direct-attached configuration.
Cluster configurations may include both active and passive cluster nodes. Active nodes are the primary nodes in the cluster. These nodes support the cluster workload by processing application requests and providing client services. Passive nodes are backup nodes that support the active nodes if a hardware or software failure occurs, ensuring that client applications and services are highly available.
Cluster Service. Failover Policies When implementing a failover policy, configure failback if the cluster node lacks the resources (such as memory or processing power) to support one or more cluster node failures.
Table 1-7. Example of an N+I Failover Configuration for an Eight-Node Cluster Cluster Resource Group Primary Node AntiAffinityClassNamesValue A Node 1 AString B Node 2 AString C Node 3 AString D Node 4 AString E Node 5 AString F Node 6 AString Configuring Group Affinity On N + I (active/passive) failover clusters running Windows Storage Server 2003, Enterprise Edition, some resource groups may conflict with other groups if they are running on the same node.
Figure 1-4. Example of a Failover Pair Configuration Table 1-8. Example of a Failover Pair Configuration for a Four-Node Cluster Cluster Resource Group Possible Owners List App1 1,2 App2 3,4 Multiway Failover Multiway failover is an active/active failover policy where running applications from a failed node migrate to multiple nodes in the cluster. This type of failover solution provides automatic failover and load-balancing between cluster nodes.
Table 1-9. Example of a Four-Node Multiway Failover Configuration Cluster Resource Group Failover Order in the Preferred Owners List A Node 1 B Node 2 C Node 3 D Node 4 E Node 5 F Node 6 Failover Ring Failover ring is an active/active failover policy where all running applications migrate from the failed node to the next preassigned cluster node in the Preferred Owners List in Cluster Administrator.
cluster. Failover ring Running applications migrate to the next preassigned cluster node. Easy to scope server capacity for one server failure. The cluster node next in line for failover may not have ample resources available to handle the additional workload of the failed node.
Table 1-12. Cluster Storage Requirements Hardware Components Minimum Requirement Supported storage systems One Dell | EMC CX200 or CX400 with dual SPSs and at least five internal hard drives on the DPE. OR One Dell | EMC CX600 with dual SPSs and one DAE2 enclosure. At least five hard drives are required on the first DAE2 enclosure that is attached to the CX600 SPE.
Ethernet switch (optional) One Ethernet switch for the private network (cluster interconnect). Ethernet cables Standard, crossover, or fiber optic Ethernet cables.
Back to Contents Page Preparing Your Systems for Clustering Dell™ PowerVault™ NAS Fibre Channel Cluster Systems Installation and Troubleshooting Guide Before You Begin Configuring the Hard Drives on the Shared Storage System(s) Installation Overview Updating a Dell | EMC Storage System for Clustering Selecting a Domain Model Installing and Configuring MSCS Configuring Windows Networking Verifying Cluster Functionality Installing the Fibre Channel HBAs Verifying Cluster Resource
See "Selecting a Domain Model" for more information. 3. Reserve static IP addresses for your cluster resources and components. The resources and components include: Public network Private network MSCS software Cluster-aware applications running on the cluster You will use these IP addresses when you install or configure MSCS. See "Assigning Static IP Addresses to Your Cluster Resources and Components" for more information. 4. Install or update the HBA drivers.
The following sections provide detailed information for each step in the "Installation Overview" that is specific to your Windows operating system. Selecting a Domain Model On a cluster running the Windows operating system, the cluster nodes must belong to a common domain or directory model. The following membership configurations are supported: All cluster nodes are member servers in a Windows Server 2003 Active Directory domain.
Configuring IP Addresses for the Private Network Dell recommends using the static IP address assignments for the NICs used for the private network (cluster interconnect). The IP addresses in Table 2-2 are given as examples only. Table 2-2. Examples of IP Address Assignments Usage Cluster Node 1 Cluster Node 2 Public network static IP address (for client and domain controller communications) 192.168.1.101 192.168.1.102 Public network subnet mask 255.255.255.0 255.255.255.
Verifying Cluster Network Communications To ensure proper cluster operations, the cluster nodes must be able to communicate with each other through the private network (cluster interconnect). This communication involves the exchange of heartbeat messages, whereby the two cluster nodes inquire about each other's status, or "health," and acknowledge each inquiry. To verify network communications between the cluster nodes: 1. Open a command prompt on each cluster node. 2. At the prompt, type: ipconfig /all 3.
See your Navisphere documentation for more information. 3. Update the storage system configuration settings using Navisphere Manager. See "Updating the Storage System Configuration Settings" for more information. The following subsections provide an overview of the storage management software and procedures for connecting the host systems to the storage systems. Access Logix Fibre Channel topologies allow multiple clusters and stand-alone systems to share a single storage system.
Access Control Access Control is a feature of Access Logix that connects the host system to the storage system. Enabling Access Control prevents all host systems from accessing any data on the storage system until they are given explicit access to a LUN through a storage group. By installing Access Logix on your storage system(s) and enabling Access Control, you can prevent the host systems from taking ownership of all LUNs on the storage system and prevent unauthorized access to sensitive information.
Name — Name of the LUN Capacity — Amount of allocated storage space on the LUN Used host An additional storage group feature that performs the following tasks: connection Lists all of the paths from the host server to the storage group paths Displays whether the path is enabled or disabled Each path contains the following fields: HBA — Device name of the HBA in the host system HBA Port — Unique ID for the HBA port connected to the storage system SP Port — Unique ID for the storage processor port connected t
verifies that the SPS monitoring system is functioning correctly. Update the storage system configuration settings in order by performing the steps in the following subsections. Updating the Cache Page Size and High/Low Watermark Settings 1. Open a Web browser and connect to the storage management server (usually the IP address of the storage processor on a Dell | EMC storage system). 2. Click the Storage tab. 3. Right-click the storage device you want to manage and select Properties. 4.
Enabling Access Logix and Creating Storage Groups The following subsection provides the required procedures for creating storage groups and connecting your storage systems to the host systems using the Access Logix software. Enabling Access Logix and Creating Storage Groups for Systems Using Navisphere 6.x 1. Ensure that Navisphere Agent is started on all host systems. a. Click the Start button and select Programs→ Administrative Tools, and then select Services. b.
a. Right-click the icon of your storage group and select Properties. b. Click the LUNs tab. c. In the Available LUNs window, click an available LUN. d. Click the right-arrow button to move the selected LUN to the Selected LUNs pane. e. Click Apply. 16. Add new hosts to the Sharable storage group. a. In the Storage Group Properties dialog box, click the Hosts tab. b. In the Available Hosts window pane, click the host system that you want to add to the storage group. c.
Naming and Formatting Drives on the Shared Storage System When the LUNs have completed the binding process, assign drive letters to the LUNs and then format the drives as NTFS drives. Format the drives and assign volume labels from the first cluster node. When completed, the remaining nodes will see the file systems and volume labels. NOTICE: Accessing the hard drives from multiple cluster nodes may corrupt the file system.
10. Right-click the drive icon again and select Format from the submenu. 11. Under Volume Label, enter a descriptive name for the new volume; for example, Disk_Z or Email_Data. 12. In the dialog box, change the file system to NTFS, select Quick Format, and click Start. NOTE: The NTFS file system is required for shared-disk resources under MSCS. 13. Click OK at the warning. 14. Click OK to acknowledge that the format is complete. 15. Click Close to close the dialog box. 16.
1. Shut down all the cluster nodes except node 1. 2. Format the disks, assign the drive letters and volume labels on node 1 by using the Windows Disk Management utility or Array Manager. For example, create volumes labeled "Volume Y" for disk Y and "Volume Z" for disk Z. 3. Shut down node 1 and perform the following steps on the remaining node: a. Turn on the node. b. Open Disk Management. c. Assign the drive letters for the drives. d. Reassign the drive letter, if necessary.
The New Server Cluster Wizard window appears. 4. Click Next to continue. 5. Follow the procedures in the wizard, and then click Finish. 6. Add the additional node to the cluster: a. Turn on the remaining node. b. Click the Start button, select Programs→ Administrative Tools, and then double-click Cluster Administrator. c. From the File menu, select Open Connection. d. In the Action box of the Open Connection to Cluster, select Add nodes to cluster. e.
If Cluster Administrator discovers a problem with the cluster configuration, a warning icon appears in the Checking cluster feasibility window. Click the plus ("+") sign to review any warnings, if needed. 12. Click Next to continue. 13. In the Password field of the Cluster Service Account menu, type the password for the account used to run the Cluster Service, and click Next. The Proposed Cluster Configuration menu appears with a summary with the configuration settings for your cluster. 14.
NOTE: The Majority Node Set Quorum types for Windows Storage Server 2003 are not supported. Preventing Quorum Resource Failure Since the quorum resource plays a crucial role in cluster operation, losing a quorum resource causes the entire cluster to fail. To prevent cluster failure, configure the quorum resource on a RAID volume in the shared storage system. NOTICE: Dell recommends that you use a RAID level other than RAID 0, which is commonly called striping.
1. Start Cluster Administrator on the monitoring node. 2. Click the Start button and select Programs→ Administrative Tools (Common)→ Cluster Administrator. 3. Open a connection to the cluster and observe the running state of each resource group. If a group has failed, one or more of its resources might be offline.
You must use the Dell™ PowerVault™ NAS Manager to manage your shadow copies. Using Cluster Administrator or cluster.exe to manage shadow copies in a cluster is not supported. See the Dell PowerVault 77xN NAS Systems Administrator Guide for more information on managing shadow copies using NAS Manager.
Back to Contents Page Cabling Your Cluster Hardware Dell™ PowerVault™ NAS Fibre Channel Cluster Systems Installation and Troubleshooting Guide Fibre Channel Cable Connectors Cabling Your Public and Private Networks Cabling the Mouse, Keyboard, and Monitor Cabling Your Power Supplies NAS cluster configurations require cabling for the Fibre Channel storage systems, cluster interconnects, client network connections, and power connections.
The duplex SC multimode fiber optic connector (see Figure 3-2) is used to connect a Fibre Channel switch to a Dell™ PowerVault™ 132T or 136T tape library. This type of connection requires fiber optic cables with both duplex LC and SC multimode fiber optic connectors. Figure 3-2. Duplex SC Multimode Fiber Optic Connector Table 3-2 provides the hardware component applications for the duplex multimode fiber optic connectors. Table 3-2.
Table 3-3. Network Connections Network Description Connection Public network All connections to the client LAN. Private network A dedicated connection for sharing cluster health and status information between the cluster nodes. At least one public network must be configured for Mixed mode for private network failover. NICs connected to the LAN can also provide redundancy at the communications level in case the cluster interconnect fails.
The private network connection to the cluster nodes is provided by a second or subsequent NIC that is installed in each node. This network is used for intracluster communications. Table 3-4 lists the required hardware components and connection method for three possible private network configurations. Table 3-4.
Figure 3-4. CX200 Power Cable Configuration Figure 3-5. CX400 Power Cable Configuration Figure 3-6.
Back to Contents Page
Back to Contents Page Installing Your Cluster in a Direct-Attached Environment Dell™ PowerVault™ NAS Fibre Channel Cluster Systems Installation and Troubleshooting Guide Before You Begin Connecting the Storage Systems to Your Cluster Before You Begin Verify that your cluster hardware and storage systems installed and configured properly as explained in the following sections: "Cabling Your Cluster Hardware" "Preparing Your Systems for Clustering" NOTICE: Microsoft® Windows® standby mode and hibernati
Cabling a Cluster to a Dell | EMC Storage System Each cluster node attaches to the storage system using two fiber optic cables with duplex LC multimode connectors that attach to the HBAs in the cluster nodes and the SP ports in the Dell | EMC storage system. These connectors consist of two individual fibre optic connectors with indexed tabs that must be aligned properly into the HBAs and SP ports. NOTICE: Do not remove the connector covers until you are ready to insert the connectors into the system.
Figure 4-3. Cabling the Cluster Nodes to a CX400 Storage System Cabling a Cluster to a CX600 Storage System 1. Connect cluster node 1 to the storage system. a. Install a cable from cluster node 1 HBA0 to SP-A port 0. b. Install a cable from cluster node 1 HBA1 to SP-B port 0. 2. Connect cluster node 2 to the storage system. a. Install a cable from cluster node 2 HBA0 to SP-A port 1. b. Install a cable from cluster node 2 HBA1 to SP-B port 1.
Cabling a Four-Node Cluster to a CX200 and CX400 Storage System The CX200 and CX400 storage systems do not support more than two cluster nodes in a direct-attached cluster configuration. Only the CX600 storage system can support a four-node cluster configuration. Cabling a Four-Node Cluster to a CX600 Storage System 1. Connect cluster node 1 to the storage system. a. Install a cable from cluster node 1 HBA0 to SP-A port 0. b. Install a cable from cluster node 1 HBA1 to SP-B port 0. 2.
The CX200 and CX400 storage systems do not support more than one direct-attached 2-node cluster. The CX600 storage system includes four ports on each storage processor, allowing you to connect two 2-node clusters or a single four-node cluster running Windows Storage Server 2003, Enterprise Edition to the storage system in a direct-attached configuration. NOTE: EMC® Access Logix™ is required if the CX600 storage system is connected to more than one cluster in a direct-attached configuration.
Back to Contents Page Installing Your Cluster in a SAN Environment Dell™ PowerVault™ NAS Fibre Channel Cluster Systems Installation and Troubleshooting Guide SAN Overview Connecting the Storage Systems to Your Cluster Implementing Zoning on a Fibre Channel Switched Fabric SAN-Attached Cluster Configurations SAN Overview A SAN is a high-performance network that is used to move data between server and storage resources, providing scalable storage in a networked environment.
A SAN-attached cluster is a cluster configuration where all cluster nodes are attached to a single storage system or to multiple storage systems through a SAN using a redundant switch fabric. SAN-attached cluster configurations provide more flexibility, expandability, and performance than direct-attached configurations. See "Fibre Channel Switch Fabric" for more information on Fibre Channel switch fabrics. Figure 5-1 shows an example of a SAN-attached cluster.
Cabling a SAN-Attached Cluster to a Dell | EMC Storage System The supported Dell | EMC storage systems are configured as follows: CX200 — One DPE, one DAE2 enclosure, and two SPSs CX400 — One DPE, one or more DAE2 enclosures, and two SPSs CX600 — One SPE, at least one DAE2 (or DAE2-OS) enclosure, and two SPSs The cluster nodes attach to the storage system using a redundant switch fabric and fiber optic cables with duplex LC multimode connectors.
Figure 5-3 and Figure 5-4 illustrate how to cable a SAN-attached cluster to the CX200 and CX400 storage systems, respectively. Figure 5-5 illustrates how to cable a SAN-attached cluster to a CX600 storage system. Table 5-2. Storage System Cabling Description Storage System SP Ports Fiber Optic Cables Required Cabling Description CX200 One port per storage processor 2 Attach one cable from each storage processor port to the Fibre Channel switch.
Cabling a SAN-Attached Cluster to a Dell | EMC CX400 Storage System 1. Connect cluster node 1 to the SAN. a. Connect a cable from HBA0 to Fibre Channel switch 0 (sw0). b. Connect a cable from HBA1 to Fibre Channel switch 1 (sw1). 2. Repeat step 1 for each node. 3. Connect the storage system to the SAN. a. Connect a cable from Fibre Channel switch 0 (sw0) to SP-A port FE 0. b. Connect a cable from Fibre Channel switch 0 (sw0) to SP-B port FE 1. c.
Cabling a SAN-Attached Cluster to the CX600 Storage System 1. Connect cluster node 1 to the SAN. a. Connect a cable from HBA0 to Fibre Channel switch 0 (sw0). b. Connect a cable from HBA1 to Fibre Channel switch 1 (sw1). 2. Repeat step 1 for each node. 3. Connect the storage system to the SAN. a. Connect a cable from Fibre Channel switch 0 (sw0) to SP-A port 0. b. Connect a cable from Fibre Channel switch 0 (sw0) to SP-A port 2. c. Connect a cable from Fibre Channel switch 0 (sw0) to SP-B port 1. d.
Cabling Multiple SAN-Attached Clusters to a Dell | EMC Storage System To cable two clusters to the storage system, connect the cluster nodes to the appropriate Fibre Channel switches and then connect the Fibre Channel switches to the appropriate storage processors on the processor enclosure. See the Platform Guide for rules and guidelines for SAN-attached clusters. NOTE: The following procedures use Figure 5-3, Figure 5-4, and Figure 5-5 as examples for cabling additional clusters.
b. Connect a cable from Fibre Channel switch 0 (sw0) to SP-B port FE A. Cabling Multiple SAN-Attached Clusters to the CX400 Storage System 1. In the first cluster, connect cluster node 1 to the SAN. a. Connect a cable from HBA0 to Fibre Channel switch 0 (sw0). b. Connect a cable from HBA1 to Fibre Channel switch 1 (sw1). 2. In the first cluster, repeat step 1 for each node. 3. For each additional cluster, repeat step 1 and step 2. 4. Connect the storage system to the SAN. a.
CX200 — A single SP port on each SP CX400 and CX600 — One or more SP ports Connecting a Cluster to Multiple Storage Systems You can increase your cluster storage capacity by attaching multiple storage systems to your cluster using a redundant switch fabric. NAS cluster systems can support configurations with multiple storage units attached to clustered servers. In this scenario, the MSCS software can fail over disk drives in any cluster-attached shared storage array between the cluster nodes.
Obtaining More Information See the storage and tape backup documentation for more information on configuring these components. Configuring Your Cluster With SAN Backup You can provide centralized backup for your clusters by sharing your SAN with multiple clusters, storage systems, and a tape library. Figure 5-8 provides an example of cabling the cluster nodes to your storage systems and SAN backup with a tape library. Figure 5-8.
Access Logix allows multiple cluster nodes and servers to share a Dell | EMC storage system, restricts server access to specific volumes on a shared Dell | EMC storage system, and protects your data from unauthorized access. Access Logix is required when the storage system is connected to two or more clusters, two or more nonclustered systems, or a combination of both clustered and nonclustered systems.
NOTICE: Before enabling Access Control, ensure that no hosts are attempting to access the storage. Enabling Access Control prevents all hosts from accessing any data until they are given explicit access to a LUN in the appropriate storage group. You must stop all I/O before enabling Access Control. Dell recommends powering off all hosts connected to the storage system during this procedure or data loss may occur. After you enable Access Control, it cannot be disabled. 8. Click Yes to enable Access Control.
Using Zoning in SAN Configurations Containing Multiple Hosts Using the combination of zoning and Access Logix in SAN configurations containing multiple hosts, you can restrict server access to specific volumes on a shared storage system by preventing the hosts from discovering a storage device that belongs to another host. This configuration allows multiple clustered or nonclustered hosts from sharing a storage system.
Each HBA can be connected to a maximum of four storage systems The integrated bridge on a tape library can be added to any zone. NOTE: If you are sharing a storage system with multiple clusters or a combination of clustered and nonclustered systems (hosts), you must install Access Logix and enable Access Control. Otherwise, you can only have one nonclustered system or one NAS cluster attached to the Dell | EMC storage system.
Back to Contents Page Maintaining Your Cluster Dell™ PowerVault™ NAS Fibre Channel Cluster Systems Installation and Troubleshooting Guide Launching Microsoft Cluster Administrator Recovering From a Corrupt Quorum Resource Adding a NIC to a Cluster Node Replacing a Cluster Node Changing the IP Address of a Cluster Node on the Same IP Subnet Changing the Cluster Service Account Password Removing a Node From a Cluster Reformatting a Cluster Disk Running chkdsk /f on a Quorum Resource
Windows Plug and Play detects the new NIC and installs the appropriate drivers. NOTE: If Plug and Play does not detect the new NIC, the NIC is not supported. a. Update the NIC drivers (if required). b. After the drivers are installed, click the Start button, select Control Panel, and then double-click Network Connections. c. In the Connections box, locate the new NIC that you installed in the system. d. Right-click the new NIC, and select Properties. e.
3. In Cluster Administrator, right-click the icon of the node you want to uninstall and then select Stop Cluster Service. 4. In Cluster Administrator, right-click the icon of the node you want to uninstall and then select Evict Node. If you cannot evict the node or the node is the last node in the cluster, perform the following steps: NOTICE: To avoid problems, you must follow this procedure if the evicted node is the last node in the cluster a. Open a command prompt. b.
chkdsk /f q: where q is the drive letter assigned to your quorum resource. 5. After the chkdsk utility completes, stop MSCS by pressing . 6. Restart the cluster service. To restart MSCS from the Services console: a. Click the Start button and select Programs→ Administrative Tools→ Services. b. In the Services window, right-click Cluster Service. c. In the drop-down menu, click the Start button. To restart MSCS from the command prompt: a.
Changing the Cluster Service Account Password To change the cluster service account password for all nodes in a Windows Storage Server 2003 cluster, open a command prompt and type the following syntax: Cluster /cluster:[cluster_name] /changepass where cluster_name is the name of your cluster.
Back to Contents Page
Back to Contents Page Using MSCS Dell™ PowerVault™ NAS Fibre Channel Cluster Systems Installation and Troubleshooting Guide Cluster Objects Groups Cluster Networks Cluster Resources Network Interfaces File Share Resources Cluster Nodes Failover and Failback This section provides information about Microsoft® Cluster Service (MSCS).
Preventing Network Failure The Cluster Service uses all available private and public-and-private networks for internal communication. Configure multiple networks as private or public-and-private to protect the cluster from a single network failure. If there is only one such network available and it fails, the cluster nodes stop communicating with each other. When two nodes are unable to communicate, they are partitioned and the Cluster Service automatically shuts down on one node.
Table 7-1. Node States and Definitions State Definition Down The node is not actively participating in cluster operations. Joining The node is in the process of becoming an active participant in the cluster operations. Paused The node is actively participating in cluster operations but cannot take ownership of resource groups and cannot bring resources online. Up The node is actively participating in all cluster operations, including hosting cluster groups. Unknown The state cannot be determined.
Cluster Resources A cluster resource is any physical or logical component that has the following characteristics: Can be brought online and taken offline Can be managed in a server cluster Can be hosted (owned) by only one node at a time To manage resources, the Cluster Service communicates to a resource DLL through a Resource Monitor.
Resource Dependencies Groups function properly only if resource dependencies are configured correctly. The Cluster Service uses the dependencies list when bringing resources online and offline. For example, if a group in which a physical disk and a file share are located is brought online, the physical disk containing the file share must be brought online before the file share. Table 7-2 shows resources and their dependencies.
Table 7-3.
To ensure cluster unity, the operating system uses the quorum resource to ensure that only one set of active, communicating nodes is allowed to operate as a cluster. A node can form a cluster only if it can gain control of the quorum resource. A node can join a cluster or remain in an existing cluster only if it can communicate with the node that controls the quorum resource.
NOTE: Do not adjust the Retry Period On Failure settings unless instructed by technical support. Resource Dependencies A dependent resource requires—or depends on—another resource to operate. For example, if a Generic Application resource requires access to clustered physical storage, it would depend on a physical disk resource. The following terms describe resources in a dependency relationship: Dependent resource — A resource that depends on other resources (the dependencies).
8. The New Resource window appears with Available resources and Resource dependencies selections. To add dependencies, under Available resources, click a resource, and then click Add. To remove dependencies, under Resource dependencies, click a resource, and then click Remove. 9. Repeat step 7 for any other resource dependencies, and then click Finish. 10. Set the resource properties. For more information on setting resource properties, see the MSCS online help. Deleting a Resource 1.
10. In the New Resource window, perform the following steps: a. In the Name field, type a name for the new share. b. In the Description field, type a description of the new share (if required). c. In the Resource type drop-down menu, select File Share. d. In the Group drop-down menu, select the appropriate virtual server for your file share. 11. Click Next. The Possible Owners window appears. 12. Select the appropriate cluster node(s) in the Available nodes box on which this resource can be brought online.
3. In the right window pane, right-click the file share you want to remove and select Delete. NOTE: When you delete a resource, Cluster Administrator automatically deletes all the resources that have a a dependency on the deleted resource. DFS File Shares You can use the File Share resource type selection in Cluster Administrator to create a resource that manages a stand-alone DFS root; however, fault-tolerant DFS roots cannot be managed by this resource.
1. Open a command prompt. 2. At the prompt, type: c:\dell\util\cluster 3. In the cluster directory, run the NFSShareEnable.bat file. Failover and Failback This section provides information about the failover and failback capabilities of the Cluster Service. Failover When an individual application or user resource (also known as a cluster resource) fails on a cluster node, the Cluster Service will detect the application failure and try to restart the application on the cluster node.
For example, if an application depends on a Physical Disk resource, the Cluster Service takes the application offline first, allowing the application to write changes to the disk before the disk is taken offline. The resource is taken offline. Cluster Service takes a resource offline by invoking, through the Resource Monitor, the resource DLL that manages the resource. If the resource does not shut down within a specified time limit, the Cluster Service forces the resource to shut down.
You can configure failback to occur immediately, at any given time, or not at all. However, be sure to configure the failback time during your offpeak hours to minimize the effect on user, as they may see a delay in service until the resources come back online.
Back to Contents Page Troubleshooting Dell™ PowerVault™ NAS Fibre Channel Cluster Systems Installation and Troubleshooting Guide This appendix provides troubleshooting information for NAS cluster configurations. Table A-1 describes general cluster problems you may encounter and the probable causes and solutions for each problem. Table A-1.
cluster. shared disks. cluster disks using Windows Disk Administration. If the disks do not appear in Disk Administration, check the The shared disks are enumerated following: by the operating system Check all cable connections. differently on the cluster nodes. Check all zone configurations. Check the Access Control settings on the attached storage systems. Use the "Advanced" with "Minimum" option.
Back to Contents Page Cluster Data Sheets Dell™ PowerVault™ NAS Fibre Channel Cluster Systems Installation and Troubleshooting Guide NAS Cluster Configuration Matrix Form Zoning Configuration Matrix Form The configuration matrix and data sheets on the following pages are provided for the system installer to record pertinent information about Dell™ PowerVault™ NAS cluster configurations.
System Storage Port Connection on Server Node 1, HBA 1 Node 1, HBA 2 Node 2, HBA 1 Node 2, HBA 2 Zoning Configuration Matrix Form Switch Storage Port Storage 1 Storage 2 Storage 3 Storage 4 Back to Contents Page
Back to Contents Page Abbreviations and Acronyms Dell™ PowerVault™ NAS Fibre Channel Cluster Systems Installation and Troubleshooting Guide AC alternating current BIOS basic input/output system CD compact disc DAE disk array enclosure DAE2-OS disk array enclosure 2 (Gb)—operating system DDNS dynamic domain naming system DFS distributed file system DHCP dynamic host configuration protocol DLL dynamic link library DLT digital linear tape
DNS domain naming system DPE disk processor enclosure ESD electrostatic discharge GB gigabyte Gb gigabit GBIC gigabit interface converter Gb/s gigabits per second GMT Greenwich Mean Time GUI graphical user interface HBA host bus adapter Hz hertz ID
identification IIS Internet Information Server I/O input/output IP Internet Protocol ISL interswitched link km kilometers LAN local area network LCC link control card LUN logical unit number m meter MB megabyte(s) MB/sec megabyte(s) per second
MHz megahertz MSCS Microsoft® Cluster Service MS-DOS Microsoft Disk Operating System NetBIOS network basic input/output system NLB network load balancing NIC network interface controller NTFS NT File System OS operating system PDU power distribution unit PERC PowerEdge™ Expandable RAID Controller PCI Peripheral Component Interconnect
PCI-X Peripheral Component Interconnect eXtended POST power-on self-test PSM persistent storage manager RAID redundant array of independent disks RAM random access memory SAN storage area network SCSI small computer serial interface SES SCSI enclosure services SFP small form-factor pluggable SMP symmetric multiprocessing SP storage processor SPE
storage processor enclosure SPS standby power supply SQL Structured Query Language SNMP Simple Network Management Protocol TB terabyte TCP/IP Transmission Control Protocol/Internet Protocol UPS uninterruptible power supply VLAN virtual local area network Back to Contents Page