HP XC System Software Installation Guide Version 3.1

Configured unknown node n14 with 1 CPU and 1 MB of total memory...
After the node has been booted up, re-run the spconfig utility to configure the correct settings.
3. If the system is using a QsNet
II
interconnect, ensure that the number of node entries in the
/opt/hptc/libelanhosts/etc/elanhosts file matches the expected number of operational
nodes in the cluster. If the number does not match, verify the status of the nodes to ensure that they
are all up and running, and re-run the spconfig utility. If the number does not match, it is possible
that the QsNet
II
network card on the missing node is not fully operational.
4. Edit the SLURM configuration file to add required customizations to the SLURM configuration and
run the scontrol reconfig command to update the SLURM daemons with this new information.
Assigning features to nodes is a common customization that is useful if the compute resources of the
cluster are not consistent.
For example, if compute nodes n[1-20] have significantly more memory than the rest of the nodes,
assign a bigmem feature to nodes n[1-20], and assign a lowmem feature to the rest of the nodes. To
do this, add Features to the NodeName entries to look similar to this:
NodeName=n[1-20] Procs=2 RealMemory=4096 Features=bigmem
NodeName=n[21-46] Procs=2 RealMemory=2048 Features=lowmem
Another SLURM customization you might want to make is to create two distinguishable sets of nodes,
such as a prod set of nodes for production use and a test set of nodes for testing use.
In general, when you assign features to SLURM nodes, the nodes can be used in job submissions to
request that the job run within a specific set of nodes. These features are recognized by LSF-HPC with
SLURM and can be used in LSF queue definitions as well as user job submissions. For more information,
see the HP XC System Software Administration Guide and HP XC System Software User's Guide.
A second common customization to the SLURM configuration is to create additional SLURM partitions.
By default, an lsf partition is created with settings that prevent non-root users from requesting node
allocations directly from SLURM. You might want to move a small set of nodes from the lsf partition
into a new, more user-accessible partition.
For more information about configuring SLURM partitions, see slurm.conf(8) and the comments in
the slurm.conf file.
In summary, follow this procedure to make SLURM customizations:
a. Use the text editor of your choice to edit the SLURM configuration file:
/hptc_cluster/slurm/etc/slurm.conf
b. Use the guidelines already described in this step to customize the SLURM configuration according
to your requirements.
c. If you make changes to the slurm.conf file, save your changes.
d. Update the SLURM daemons with this new information.
# scontrol reconfig
3.18.2 Perform LSF Postconfiguration Tasks
Follow this procedure to set up the LSF environment and enable LSF failover (if you assigned the
resource_management role to two or more nodes). Bypass this task if you did not configure LSF.
1. Begin this procedure as the root user on the head node.
2. Set up the LSF environment by sourcing the LSF profile file:
# . /opt/hptc/lsf/top/conf/profile.lsf
3. Verify that the LSF profile file has been sourced by finding an LSF command:
3.18 Task 17: Finalize the Configuration of Compute Resources 87