8 things to double-check, 9 things to note – HP StorageWorks Scalable File Share User Manual
Page 37
![background image](/manuals/398305/37/background.png)
10 Resources configured.
============
Node: n4 (0236b688-3bb7-458a-839b-c19a69d75afa): online
Node: n3 (48610537-c58e-48c5-ae4c-ae44d56527c6): online
Filesystem_1 (heartbeat::ocf:Filesystem): Started n3
Filesystem_2 (heartbeat::ocf:Filesystem): Started n3
Filesystem_3 (heartbeat::ocf:Filesystem): Started n3
Filesystem_4 (heartbeat::ocf:Filesystem): Started n3
Filesystem_5 (heartbeat::ocf:Filesystem): Started n4
Filesystem_6 (heartbeat::ocf:Filesystem): Started n4
Filesystem_7 (heartbeat::ocf:Filesystem): Started n4
Filesystem_8 (heartbeat::ocf:Filesystem): Started n4
Clone Set: clone_9
stonith_9:0 (stonith:external/riloe): Started n4
stonith_9:1 (stonith:external/riloe): Started n3
Clone Set: clone_10
stonith_10:0 (stonith:external/riloe): Started n4
stonith_10:1 (stonith:external/riloe): Started n3
The display updates periodically until you interrupt it and terminate the program.
5.2.7 Moving and Starting Lustre Servers Using Heartbeat
Lustre servers can be moved between nodes in a failover pair, and stopped, or started using the
Heartbeat command crm_resource. The local file systems corresponding to the Lustre servers
appear as file system resources with names of the form Filesystem_n, where n is an integer.
The mapping from file system resource names to Lustre server mount-points is found in cib.xml.
For example, to move Filesystem_7 from its current location to node 11:
# crm_resource -H node11 -M -r Filesystem_7
The destination host name is optional but it is important to note that if it is not specified,
crm_resource
forces the resource to move by creating a rule for the current location with the
value -INFINITY. This prevents the resource from running on that node again until the constraint
is removed with crm_resource -U.
If it is not possible to start a resource on a node, check that node for values of -INFINITY in
/var/lib/heartbeat/crm/cib.xml
. There should be none. For more details, see the
crm_resource
manpage. See also
5.2.8 Things to Double-Check
•
The files .sig and .last should be removed from /var/lib/heartbeat/crm when a
new cib.xml is copied there. Otherwise, Heartbeat ignores the new cib.xml and uses the
last one.
•
The file /var/lib/heartbeat/crm/cib.xml should contain owner=hacluster,
group=haclient
. Heartbeat writes cib.xml to add status information. If cib.xml cannot
be written, Heartbeat will be confused about the state of other nodes in the failover group
and may power-cycle them to put them in a state it understands.
•
The /etc/ha.d/authkeys file must be readable and writable only by root (mode 0600).
•
The host names for each node in /etc/ha.d/ha.cf must be the value that is returned
from executing the hostname or uname -n command on that node.
5.2.9 Things to Note
•
When Heartbeat starts, it waits for a period to give its failover peer time to boot and get
started. This time is specified by the init_dead parameter in the ha.cf file (60 seconds
in the example ha.cf file). Consequently, there may be an unexpected time lag before
Heartbeat starts Lustre the first time. This process is quicker if both nodes start Heartbeat
at about the same time.
5.2 Configuring Heartbeat
37