This chapter describes several system administration procedures that must be performed on the nodes in a cluster to prepare and configure them for IRIS FailSafe. These procedures assume that you have done the planning described in Chapter 2, “Planning IRIS FailSafe Configuration.”
The major sections in this chapter are as follows:
Performing the system administration procedures required to prepare nodes for IRIS FailSafe involves these steps:
Install required software as described in the section “Installing Required Software.”
Check the setting of two important NVRAM variables as described in the section “Setting NVRAM Variables.”
Create the XLV logical volumes and XFS filesystems required by the high-availability applications you plan to run on the cluster. See the section “Creating XLV Logical Volumes and XFS Filesystems.”
Configure the network interfaces on both nodes using the procedure in the section “Configuring Network Interfaces.”
Configure the serial ports used on each node for the serial connection to the other node by following the procedure in the section “Configuring the Serial Ports.”
If the IRIS FailSafe NFS option is used, perform necessary NFS configuration as described in the section “Configuring NFS Filesystems.”
If the IRIS FailSafe Web option is used, configure Netscape servers as described in the section “Configuring a Netscape Server.”
When you are ready configure the nodes so that IRIS FailSafe software starts up when they are rebooted, follow the instructions in the section “Configuring IRIS FailSafe On.”
To complete the configuration of nodes for IRIS FailSafe, you must create and install the IRIS FailSafe configuration file /var/ha/ha.conf as described in Chapter 4, “Creating the IRIS FailSafe Configuration File.”
This section explains software installation requirements and procedures for new IRIS FailSafe clusters. To install a new release of IRIX or IRIS FailSafe software on an IRIS FailSafe cluster already in use, see Chapter 7, “Upgrading an IRIS FailSafe Cluster.”
Several categories of software must be installed on each node in a cluster:
The two nodes in a cluster need not be running the same version of IRIX. However, they must run the same version of IRIS FailSafe software.
Follow this procedure for installing software on the nodes in a new cluster:
On one node, follow normal software installation procedures and use inst to install required subsystems that aren't already installed. The required subsystems are listed in Table 3-1. (See “Related Documentation” for guides about inst).
Table 3-1. Required Software Subsystems
Product | Base System Software Subsystems Required by IRIS FailSafe | IRIS FailSafe Subsystems |
---|---|---|
IRIS FailSafe | eoe.sw.xfs, eoe.sw.xlv, eoe.sw.xlvplex | ha.sw.base |
IRIS FailSafe NFS | nfs.sw.nfs | ha_nfs.sw.base |
IRIS FailSafe Web | ns_admin.sw.server (Netscape
Administration Server, required) and | ha_www.sw.base |
On the same node, install required patches. See the release notes for IRIS FailSafe (called ha_base, ha_nfs, and ha_www) for lists of the required patches.
Install software and patches on the other node by repeating steps 1 and 2 on the other node.
If you are using plexed XLV logical volumes, install a disk plexing license on each node. The license is a FLEXlm license and installed in /var/flexlm/license.dat. For more information see the FLEXlm End User Manual .
If you are using plexed XLV logical volumes, verify that the license has been successfully installed by entering these commands on each node:
# xlv_mgr xlv_mgr> show config ... Plexing license: present ... xlv_mgr> quit |
The line shown about the plexing license indicates that the plexing license has been successfully installed.
During the hardware installation of IRIS FailSafe nodes, two NVRAM variables must be set:
You can check the setting of these variables with these commands:
# nvram AutoLoad Y # nvram scsihostid 0 |
To set these variables, use these commands:
# nvram AutoLoad yes # nvram scsihostid number |
number is the SCSI ID you choose. A node uses its SCSI ID on all buses attached to it. Therefore, you must make sure that no device attached to a node has number as its SCSI unit number. If you change the value of the scsihostid variable, you must reboot the system for the change to take effect.
In Chapter 2 you planned the XLV logical volumes and XFS filesystems to be used by high-availability applications on the cluster. You can create them by following the instructions in the guide IRIX Admin: Disks and Filesystems .
When you create the XLV logical volumes and XFS filesystems you need, remember these important points:
If the shared disks are not in a CHALLENGE RAID storage system, plexed XLV logical volumes should be created.
Each XLV logical volume must be owned by the same node that is the primary node for the high-availability applications that use the logical volume (see “Planning Logical Volumes” in Chapter 2). To simplify the management of the nodenames (owners) of volumes on shared disks, follow these recommendations:
Work with the volumes on a shared disk from only one node in the cluster.
After you create all the volumes on one node, you can selectively change the nodename to the other node using xlv_mgr.
If the XLV logical volumes you create are used as raw volumes (no filesystem) for storing database data, the database system may require that the device names (in /dev/rdsk/xlv and /dev/dsk/xlv on IRIX 6.2 and in /dev/rxlv and /dev/xlv on IRIX 6.4) have specific owners, groups, and modes. If this is the case (see the documentation provided by the database vendor), use the chown and chmod commands (see the chown(1) and chmod(1) reference pages) to set the owner, group, and mode as required.
No filesystem entries are made in /etc/fstab for XFS filesystems on shared disks; IRIS FailSafe software mounts the filesystems on shared disks. However, to simplify system administration, consider adding comments to /etc/fstab that list the XFS filesystems configured for IRIS FailSafe. Thus, a system administrator who sees mounted IRIS FailSafe filesystems in the output of the df command and looks for the filesystems in the /etc/fstab file will learn that they are filesystems managed by IRIS FailSafe.
Be sure to create the mount point directory for each filesystem on both nodes.
The procedure in this section describes how to configure the public and private interfaces on the nodes in an IRIS FailSafe cluster. The example shown in Figure 3-1 is used in the procedure.
If possible, add every IP address, IP name, and IP alias for both nodes to /etc/hosts on one node.
190.0.2.1 xfs-ha1.company.com xfs-ha1 190.0.2.3 stocks 190.0.3.1 priv-xfs-ha1 190.0.2.2 xfs-ha2.company.com xfs-ha2 190.0.2.4 bonds 190.0.3.2 priv-xfs-ha2 |
Add all of the IP addresses from step 1 to /etc/hosts on the other node in the cluster.
If there are IP addresses, IP names, or IP aliases that you did not add to /etc/hosts in steps 1 and 2, verify that NIS is running on both nodes in the cluster by entering this command on each node:
# chkconfig ... yp on |
If the output shows that yp is off, you must start NIS. See the NIS Administration Guide for details.
For IP addresses, IP names, and IP aliases that you did not add to /etc/hosts on the nodes in steps 1 and 2, verify that they are in the NIS database by entering this command for each address:
# ypmatch address hosts 190.0.2.1 xfs-ha1.company.com xfs-ha1 |
address is an IP address, IP name, or IP alias. If ypmatch reports that address doesn't match, it must be added to the NIS database. See the NIS Administration Guide for details.
On one node, add that node's public and private interfaces and their fixed IP addresses to the file /etc/config/netif.options (high availability IP addresses are not added to the netif.options file).
For the example in Figure 3-1, the public interface name and IP address lines are
if1name=ec0 if1addr=$HOSTNAME |
$HOSTNAME is an alias for an IP address that appears in /etc/hosts.
If there are additional public interfaces, their interface names and IP addresses appear on lines like these:
if2name= if2addr= |
In the example, the private interface name and IP address are
if3name=ec3 if3addr=priv-$HOSTNAME |
The private interface IP address in this example, priv-$HOSTNAME, is an alias for an IP address that appears in /etc/hosts.
If there are more than eight interfaces on the node, change the value of if_num to the number of interfaces. For less than eight interfaces (as in the example in Figure 3-1) the line looks like this:
if_num=8 |
Edit the file /etc/config/routed.options on each node and add the -q option so that the routes are not shown over the private network (routing is turned off). An example of the content of /etc/config/routed.options on IRIX 6.2 nodes is
-h -q |
An example of the content of /etc/config/routed.options on IRIX 6.4 nodes is
-h -Prdisc_interval=45 -q |
![]() | Note: The -q option is required for IRIS FailSafe to function correctly. |
Verify that IRIS FailSafe is chkconfig'd off on each node:
# chkconfig ... failsafe off ... |
If failsafe is on on either node, enter this command on that node:
# chkconfig failsafe off |
Configure an e-mail alias on each node that sends the IRIS FailSafe e-mail notifications of cluster transitions to a user outside the IRIS FailSafe cluster and to a user on the other node in the cluster. For example, if the nodes are called xfs-ha1 and xfs-ha2, in /usr/lib/aliases on xfs-ha1, add
fsafe_admin:[email protected],[email protected] |
On xfs-ha2, add this line to /usr/lib/aliases:
fsafe_admin:[email protected],[email protected] |
The alias you choose, fsafe_admin in this case, is the value you will use for the mail-dest-addr parameter in the IRIS FailSafe configuration file ha.conf (see the section “System-Configuration Block” in Chapter 4). In this example, operations is the user outside the cluster and admin_user is a user on each node.
If the nodes use NIS (yp is chkconfig'ed on) or the BIND domain name server (DNS), switching to local name resolution is recommended. Create or modify the file /etc/resolv.conf so that local is listed first for the hostresorder keyword (the order of nis and bind is up to you):
hostresorder local nis bind |
If FDDI is being used, finish configuring and verifying the new FDDI station, as explained in Chapter 2 of the FDDIXpress release notes and Chapter 2 of the FDDIXpress Administration Guide .
Reboot both nodes to put the new network configuration into effect.
The getty process for the tty ports to which the reset serial cables are connected must be turned off. Perform these steps on each node:
Determine which port is used for the reset serial line (see the section “Serial Port Configuration” in Chapter 2).
Open the file /etc/inittab for editing.
Find the line for the port by looking at the comments on the right for the port number from step 1.
Change the third field of this line to off. For example:
t2:23:off:/sbin/getty -N ttyd2 co_9600 # port 2 |
Save the file.
Enter these commands to make the change take effect:
# killall getty # init q |
Follow the procedure below to perform the NFS configuration for the filesystems to be failed over and for the status monitor statd. Note that no entries in /etc/exports are required for these filesystems; IRIS FailSafe software exports NFS filesystems.
Create or identify the filesystems to be failed over. They must follow the guidelines in the section “Planning Filesystems” in Chapter 2.
On each node that exports filesystems, create the statmon directory on any one of the exported filesystems on a shared disk. The name of the directory must be statmon.
On one node, open the file /etc/config/statd.options for editing (it may be new).
Put -h on the first line of the file:
-h |
Close the file.
Set the owner, group, and mode of the file, if you just created it:
# chown root.sys /etc/config/statd.options # chmod 644 /etc/config/statd.options |
![]() | Note: Although IRIS FailSafe supports the Netscape Communications and Commerce Servers, the use of their newer counterparts, the Netscape FastTrack and Enterprise Servers, is recommended. |
To configure a Netscape FastTrack or Enterprise Server, follow these steps:
On one node, verify that the Netscape Administration Server is installed and running by entering this command on that node:
# ps -ef | grep ns-admin root 7457 7461 0 13:45:25 ? 0:00 /usr/ns-home/admserv/ns-admin -d /usr/ns-home/admserv ... |
If ns-admin commands aren't shown in the ps output, enter these commands:
# chkconfig ns_admin on # /etc/init.d/ns_admin start |
ifconfig the interface to the public network up by entering these commands:
# /usr/etc/ifconfig interface alias ip_alias netmask netmask |
interface is the interface to be used access the node, ip_alias is a high availability IP address for the interface, and netmask is the netmask of the IP address.
Start a Web browser on this node. For example:
# netscape |
Open the Netscape server's configuration page:
http://ip_alias:81 |
ip_alias is the high availability IP address used in step 3. A dialog box called “Netscape: Password” appears.
In the dialog box, enter the user ID and password of Netscape server administrator, for example, root.
Click the OK button. A page called “Netscape Server Selector” appears.
Click on the Install a new Netscape FastTrack Server button or the Install a new Netscape Enterprise Server button, as appropriate. A page called “Netscape FastTrack Server Installation” appears.
Fill in the fields on this page. These are the fields that have special requirements for IRIS FailSafe or are used in the remainder of this procedure:
Bind address | This is the high availability IP address (from step 3) in x.x.x.x notation. | |
Server Identifier |
| |
Server Port | The port number you choose is the value that is used for the port-num parameter in the configuration file /var/ha/ha.conf. The default port for the FastTrack Server is 80 and the default port for the Enterprise Server is 443. | |
Document Root |
|
Click the OK button. A page called “Success!” appears.
The configuration of this server is complete. You can click “Return to the server selector” on this page, go to another Web page, or exit the Web browser. Do not click “Start your new server” on this page.
Repeat steps 1 through 11 on the other node in this cluster to configure an identical server on that node.
If this cluster is to have more than one active Netscape FastTrack or Enterprise Server, repeat steps 1 through 12 for each additional Server.
If the log files, access control logs, and document root for the Netscape servers are on local disks, the configuration of the Netscape servers is complete and you can skip the rest of this procedure.
On the node that is to be the primary node for a Netscape server, create the mount point for the filesystem on the shared disk that is to contain the log files, access control lists, and document root if necessary, and mount the filesystem. For example:
# mkdir /shared # mount /dev/dsk/dks0d2s7 /shared |
Move the log files to the shared disk and create a link to their original location. For example:
# mkdir /shared/httpd-server_id # mv /usr/ns-home/httpd-server_id/logs /shared/httpd-server_id # ln -s /shared/httpd-server_id/logs /usr/ns-home/httpd-server_id/logs |
server_id is the server identifier from step 9.
Move the access control lists to the shared disk and create a link to their original location. For example:
# mv /usr/ns-home/httpacl /shared/httpacl # ln -s /shared/httpacl /usr/ns-home/httpacl |
Move the document files to the shared disk and create a link to their original location. For example, if the entire document root is to be placed on the shared disk, enter these commands:
# mv /usr/ns-home/docs /shared # ln -s /shared/docs /usr/ns-home/docs |
Unmount the filesystem on the shared disk:
# umount /shared |
Mount the filesystem on the shared disk on the other node by repeating step 15 on the other node, the backup node.
On the backup node, delete any log files, access control lists, and document files on the node. For example:
# rm -rf /usr/ns-home/httpd-var/logs # rm -rf /usr/ns-home/httpacl # rm -rf /usr/ns-home/docs |
On the backup node, create links from the shared disk to the local disk on the backup node:
# ln -s /shared/httpd-var/logs /usr/ns-home/httpd-var/logs # ln -s /shared/httpacl /usr/ns-home/httpacl # ln -s /shared/docs /usr/ns-home/docs |
If there are other Netscape servers on the cluster, repeat steps 15 through 22 for each additional server.
To configure IRIS FailSafe so that it starts up on a node each time it is rebooted, enter this command on the node:
# chkconfig failsafe on |
![]() | Note: During the testing described in Chapter 5, “Testing IRIS FailSafe Configuration,” IRIS FailSafe should be chkconfig'd off. |