Chapter 3. Configuring Nodes for IRIS FailSafe

This chapter describes several system administration procedures that must be performed on the nodes in a cluster to prepare and configure them for IRIS FailSafe. These procedures assume that you have done the planning described in Chapter 2, “Planning IRIS FailSafe Configuration.”

The major sections in this chapter are as follows:

Overview of Configuring Nodes for IRIS FailSafe

Performing the system administration procedures required to prepare nodes for IRIS FailSafe involves these steps:

  1. Install required software as described in the section “Installing Required Software.”

  2. Check the setting of two important NVRAM variables as described in the section “Setting NVRAM Variables.”

  3. Create the XLV logical volumes and XFS filesystems required by the high-availability applications you plan to run on the cluster. See the section “Creating XLV Logical Volumes and XFS Filesystems.”

  4. Configure the network interfaces on both nodes using the procedure in the section “Configuring Network Interfaces.”

  5. Configure the serial ports used on each node for the serial connection to the other node by following the procedure in the section “Configuring the Serial Ports.”

  6. If the IRIS FailSafe NFS option is used, perform necessary NFS configuration as described in the section “Configuring NFS Filesystems.”

  7. If the IRIS FailSafe Web option is used, configure Netscape servers as described in the section “Configuring a Netscape Server.”

  8. When you are ready configure the nodes so that IRIS FailSafe software starts up when they are rebooted, follow the instructions in the section “Configuring IRIS FailSafe On.”

To complete the configuration of nodes for IRIS FailSafe, you must create and install the IRIS FailSafe configuration file /var/ha/ha.conf as described in Chapter 4, “Creating the IRIS FailSafe Configuration File.”

Installing Required Software

This section explains software installation requirements and procedures for new IRIS FailSafe clusters. To install a new release of IRIX or IRIS FailSafe software on an IRIS FailSafe cluster already in use, see Chapter 7, “Upgrading an IRIS FailSafe Cluster.”

Several categories of software must be installed on each node in a cluster:

Base system software 


IRIX 6.2 or IRIX 6.4 is required. NFS software is required if the IRIS FailSafe NFS option is used. Netscape server software is required if the IRIS FailSafe Web option is used. RAID software is required if RAID is being used. It must be Release 2.3 of the RAID software or later.

Patches 

IRIS FailSafe requires several base system software patches, which are listed in the IRIS FailSafe base software release notes.

IRIS FailSafe software 


The base IRIS FailSafe option product is required. The IRIS FailSafe NFS option is required for failing over filesystems. The IRIS FailSafe Web option is required for failing over Netscape servers.

Disk plexing licenses 


On clusters that use plexed XLV logical volumes on shared disks, a FLEXlm license is required on each node.

Netscape licenses  


On clusters that use licensed Netscape server software, a license is required on each node.

The two nodes in a cluster need not be running the same version of IRIX. However, they must run the same version of IRIS FailSafe software.

Follow this procedure for installing software on the nodes in a new cluster:

  1. On one node, follow normal software installation procedures and use inst to install required subsystems that aren't already installed. The required subsystems are listed in Table 3-1. (See “Related Documentation” for guides about inst).

    Table 3-1. Required Software Subsystems

    Product

    Base System Software Subsystems Required by IRIS FailSafe

    IRIS FailSafe Subsystems

    IRIS FailSafe

    eoe.sw.xfs, eoe.sw.xlv, eoe.sw.xlvplex
    For FDDI (if used): FDDIXPress.sw.FDDIXPress
    For RAID (if used): raid5.sw.cli

    ha.sw.base

    IRIS FailSafe NFS

    nfs.sw.nfs

    ha_nfs.sw.base

    IRIS FailSafe Web

    ns_admin.sw.server (Netscape Administration Server, required) and
    ns_fasttrack.sw.server (Netscape FastTrack server) or
    ns_enterprise.sw.server (Netscape Enterprise server)

    ha_www.sw.base


  2. On the same node, install required patches. See the release notes for IRIS FailSafe (called ha_base, ha_nfs, and ha_www) for lists of the required patches.

  3. Install software and patches on the other node by repeating steps 1 and 2 on the other node.

  4. If you are using plexed XLV logical volumes, install a disk plexing license on each node. The license is a FLEXlm license and installed in /var/flexlm/license.dat. For more information see the FLEXlm End User Manual .

  5. If you are using plexed XLV logical volumes, verify that the license has been successfully installed by entering these commands on each node:

    # xlv_mgr
    xlv_mgr> show config
    ...
    Plexing license: present
    ...
    xlv_mgr> quit
    

    The line shown about the plexing license indicates that the plexing license has been successfully installed.

Setting NVRAM Variables

During the hardware installation of IRIS FailSafe nodes, two NVRAM variables must be set:

  • The boot parameter AutoLoad must be set to yes. The IRIS Failsafe software requires the nodes to be automatically booted when they are reset or when the node is powered on.

  • The SCSI IDs of the nodes in an IRIS FailSafe cluster, specified by the scsihostid variable, must be different.

You can check the setting of these variables with these commands:

# nvram AutoLoad 
Y
# nvram scsihostid 
0

To set these variables, use these commands:

# nvram AutoLoad yes 
# nvram scsihostid number

number is the SCSI ID you choose. A node uses its SCSI ID on all buses attached to it. Therefore, you must make sure that no device attached to a node has number as its SCSI unit number. If you change the value of the scsihostid variable, you must reboot the system for the change to take effect.

Creating XLV Logical Volumes and XFS Filesystems

In Chapter 2 you planned the XLV logical volumes and XFS filesystems to be used by high-availability applications on the cluster. You can create them by following the instructions in the guide IRIX Admin: Disks and Filesystems .

When you create the XLV logical volumes and XFS filesystems you need, remember these important points:

  • If the shared disks are not in a CHALLENGE RAID storage system, plexed XLV logical volumes should be created.

  • Each XLV logical volume must be owned by the same node that is the primary node for the high-availability applications that use the logical volume (see “Planning Logical Volumes” in Chapter 2). To simplify the management of the nodenames (owners) of volumes on shared disks, follow these recommendations:

    • Work with the volumes on a shared disk from only one node in the cluster.

    • After you create all the volumes on one node, you can selectively change the nodename to the other node using xlv_mgr.

  • If the XLV logical volumes you create are used as raw volumes (no filesystem) for storing database data, the database system may require that the device names (in /dev/rdsk/xlv and /dev/dsk/xlv on IRIX 6.2 and in /dev/rxlv and /dev/xlv on IRIX 6.4) have specific owners, groups, and modes. If this is the case (see the documentation provided by the database vendor), use the chown and chmod commands (see the chown(1) and chmod(1) reference pages) to set the owner, group, and mode as required.

  • No filesystem entries are made in /etc/fstab for XFS filesystems on shared disks; IRIS FailSafe software mounts the filesystems on shared disks. However, to simplify system administration, consider adding comments to /etc/fstab that list the XFS filesystems configured for IRIS FailSafe. Thus, a system administrator who sees mounted IRIS FailSafe filesystems in the output of the df command and looks for the filesystems in the /etc/fstab file will learn that they are filesystems managed by IRIS FailSafe.

  • Be sure to create the mount point directory for each filesystem on both nodes.

Configuring Network Interfaces

The procedure in this section describes how to configure the public and private interfaces on the nodes in an IRIS FailSafe cluster. The example shown in Figure 3-1 is used in the procedure.

Figure 3-1. Example Interface Configuration


  1. If possible, add every IP address, IP name, and IP alias for both nodes to /etc/hosts on one node.

    For example:

    190.0.2.1 xfs-ha1.company.com xfs-ha1
    190.0.2.3 stocks
    190.0.3.1 priv-xfs-ha1
    190.0.2.2 xfs-ha2.company.com xfs-ha2
    190.0.2.4 bonds
    190.0.3.2 priv-xfs-ha2
    


    Note: IP aliases that are used exclusively by high-availability services are not added to the file /etc/config/ipaliases.options. Similarly, if all IP aliases are used only by high-availability services, the ipaliases chkconfig flag should be off.


  2. Add all of the IP addresses from step 1 to /etc/hosts on the other node in the cluster.

  3. If there are IP addresses, IP names, or IP aliases that you did not add to /etc/hosts in steps 1 and 2, verify that NIS is running on both nodes in the cluster by entering this command on each node:

    # chkconfig
    ...
            yp                   on
    

    If the output shows that yp is off, you must start NIS. See the NIS Administration Guide for details.

  4. For IP addresses, IP names, and IP aliases that you did not add to /etc/hosts on the nodes in steps 1 and 2, verify that they are in the NIS database by entering this command for each address:

    # ypmatch address hosts
    190.0.2.1 xfs-ha1.company.com xfs-ha1
    

    address is an IP address, IP name, or IP alias. If ypmatch reports that address doesn't match, it must be added to the NIS database. See the NIS Administration Guide for details.

  5. On one node, add that node's public and private interfaces and their fixed IP addresses to the file /etc/config/netif.options (high availability IP addresses are not added to the netif.options file).

    For the example in Figure 3-1, the public interface name and IP address lines are

    if1name=ec0
    if1addr=$HOSTNAME
    

    $HOSTNAME is an alias for an IP address that appears in /etc/hosts.

    If there are additional public interfaces, their interface names and IP addresses appear on lines like these:

    if2name=
    if2addr=
    

    In the example, the private interface name and IP address are

    if3name=ec3
    if3addr=priv-$HOSTNAME
    

    The private interface IP address in this example, priv-$HOSTNAME, is an alias for an IP address that appears in /etc/hosts.

  6. If there are more than eight interfaces on the node, change the value of if_num to the number of interfaces. For less than eight interfaces (as in the example in Figure 3-1) the line looks like this:

    if_num=8
    

  7. Repeat steps 5 and 6 on the other node.

  8. Edit the file /etc/config/routed.options on each node and add the -q option so that the routes are not shown over the private network (routing is turned off). An example of the content of /etc/config/routed.options on IRIX 6.2 nodes is

    -h -q
    

    An example of the content of /etc/config/routed.options on IRIX 6.4 nodes is

    -h -Prdisc_interval=45 -q
    


    Note: The -q option is required for IRIS FailSafe to function correctly.


  9. Verify that IRIS FailSafe is chkconfig'd off on each node:

    # chkconfig 
    ...
            failsafe             off
    ...
    

    If failsafe is on on either node, enter this command on that node:

    # chkconfig failsafe off 
    

  10. Configure an e-mail alias on each node that sends the IRIS FailSafe e-mail notifications of cluster transitions to a user outside the IRIS FailSafe cluster and to a user on the other node in the cluster. For example, if the nodes are called xfs-ha1 and xfs-ha2, in /usr/lib/aliases on xfs-ha1, add

    fsafe_admin:[email protected],[email protected] 
    

    On xfs-ha2, add this line to /usr/lib/aliases:

    fsafe_admin:[email protected],[email protected] 
    

    The alias you choose, fsafe_admin in this case, is the value you will use for the mail-dest-addr parameter in the IRIS FailSafe configuration file ha.conf (see the section “System-Configuration Block” in Chapter 4). In this example, operations is the user outside the cluster and admin_user is a user on each node.

  11. If the nodes use NIS (yp is chkconfig'ed on) or the BIND domain name server (DNS), switching to local name resolution is recommended. Create or modify the file /etc/resolv.conf so that local is listed first for the hostresorder keyword (the order of nis and bind is up to you):

    hostresorder local nis bind 
    

  12. If FDDI is being used, finish configuring and verifying the new FDDI station, as explained in Chapter 2 of the FDDIXpress release notes and Chapter 2 of the FDDIXpress Administration Guide .

  13. Reboot both nodes to put the new network configuration into effect.

Configuring the Serial Ports

The getty process for the tty ports to which the reset serial cables are connected must be turned off. Perform these steps on each node:

  1. Determine which port is used for the reset serial line (see the section “Serial Port Configuration” in Chapter 2).

  2. Open the file /etc/inittab for editing.

  3. Find the line for the port by looking at the comments on the right for the port number from step 1.

  4. Change the third field of this line to off. For example:

    t2:23:off:/sbin/getty -N ttyd2 co_9600          # port 2
    

  5. Save the file.

  6. Enter these commands to make the change take effect:

    # killall getty
    # init q
    

Configuring NFS Filesystems

Follow the procedure below to perform the NFS configuration for the filesystems to be failed over and for the status monitor statd. Note that no entries in /etc/exports are required for these filesystems; IRIS FailSafe software exports NFS filesystems.

  1. Create or identify the filesystems to be failed over. They must follow the guidelines in the section “Planning Filesystems” in Chapter 2.

  2. On each node that exports filesystems, create the statmon directory on any one of the exported filesystems on a shared disk. The name of the directory must be statmon.

  3. On one node, open the file /etc/config/statd.options for editing (it may be new).

  4. Put -h on the first line of the file:

    -h 
    

  5. Close the file.

  6. Set the owner, group, and mode of the file, if you just created it:

    # chown root.sys /etc/config/statd.options
    # chmod 644 /etc/config/statd.options
    

  7. Repeat steps 3 through 6 on the other node.

Configuring a Netscape Server


Note: Although IRIS FailSafe supports the Netscape Communications and Commerce Servers, the use of their newer counterparts, the Netscape FastTrack and Enterprise Servers, is recommended.

To configure a Netscape FastTrack or Enterprise Server, follow these steps:

  1. On one node, verify that the Netscape Administration Server is installed and running by entering this command on that node:

    # ps -ef | grep ns-admin
        root  7457  7461  0 13:45:25 ?       0:00 /usr/ns-home/admserv/ns-admin -d /usr/ns-home/admserv
    ...
    

  2. If ns-admin commands aren't shown in the ps output, enter these commands:

    # chkconfig ns_admin on
    # /etc/init.d/ns_admin start
    

  3. ifconfig the interface to the public network up by entering these commands:

    # /usr/etc/ifconfig interface alias ip_alias netmask netmask
    

    interface is the interface to be used access the node, ip_alias is a high availability IP address for the interface, and netmask is the netmask of the IP address.

  4. Start a Web browser on this node. For example:

    # netscape 
    

  5. Open the Netscape server's configuration page:

    http://ip_alias:81
    

    ip_alias is the high availability IP address used in step 3. A dialog box called “Netscape: Password” appears.

  6. In the dialog box, enter the user ID and password of Netscape server administrator, for example, root.

  7. Click the OK button. A page called “Netscape Server Selector” appears.

  8. Click on the Install a new Netscape FastTrack Server button or the Install a new Netscape Enterprise Server button, as appropriate. A page called “Netscape FastTrack Server Installation” appears.

  9. Fill in the fields on this page. These are the fields that have special requirements for IRIS FailSafe or are used in the remainder of this procedure:

    Bind address 

    This is the high availability IP address (from step 3) in x.x.x.x notation.

    Server Identifier  


    This can be any name. This name is the value of the variable server_id, which is used later in this procedure.

    Server Port  

    The port number you choose is the value that is used for the port-num parameter in the configuration file /var/ha/ha.conf. The default port for the FastTrack Server is 80 and the default port for the Enterprise Server is 443.

    Document Root  


    The remainder of this procedure assumes that you use the default document root, /usr/ns-home/docs. If you change this value, substitute the new document root pathname in the remainder of this procedure.

  10. Click the OK button. A page called “Success!” appears.

  11. The configuration of this server is complete. You can click “Return to the server selector” on this page, go to another Web page, or exit the Web browser. Do not click “Start your new server” on this page.

  12. Repeat steps 1 through 11 on the other node in this cluster to configure an identical server on that node.

  13. If this cluster is to have more than one active Netscape FastTrack or Enterprise Server, repeat steps 1 through 12 for each additional Server.

  14. If the log files, access control logs, and document root for the Netscape servers are on local disks, the configuration of the Netscape servers is complete and you can skip the rest of this procedure.

  15. On the node that is to be the primary node for a Netscape server, create the mount point for the filesystem on the shared disk that is to contain the log files, access control lists, and document root if necessary, and mount the filesystem. For example:

    # mkdir /shared
    # mount /dev/dsk/dks0d2s7 /shared
    

  16. Move the log files to the shared disk and create a link to their original location. For example:

    # mkdir /shared/httpd-server_id 
    # mv /usr/ns-home/httpd-server_id/logs /shared/httpd-server_id
    # ln -s /shared/httpd-server_id/logs /usr/ns-home/httpd-server_id/logs
    

    server_id is the server identifier from step 9.

  17. Move the access control lists to the shared disk and create a link to their original location. For example:

    # mv /usr/ns-home/httpacl /shared/httpacl
    # ln -s /shared/httpacl /usr/ns-home/httpacl
    

  18. Move the document files to the shared disk and create a link to their original location. For example, if the entire document root is to be placed on the shared disk, enter these commands:

    # mv /usr/ns-home/docs /shared
    # ln -s /shared/docs /usr/ns-home/docs
    

  19. Unmount the filesystem on the shared disk:

    # umount /shared
    

  20. Mount the filesystem on the shared disk on the other node by repeating step 15 on the other node, the backup node.

  21. On the backup node, delete any log files, access control lists, and document files on the node. For example:

    # rm -rf /usr/ns-home/httpd-var/logs
    # rm -rf /usr/ns-home/httpacl
    # rm -rf /usr/ns-home/docs
    

  22. On the backup node, create links from the shared disk to the local disk on the backup node:

    # ln -s /shared/httpd-var/logs /usr/ns-home/httpd-var/logs
    # ln -s /shared/httpacl /usr/ns-home/httpacl
    # ln -s /shared/docs /usr/ns-home/docs
    

  23. If there are other Netscape servers on the cluster, repeat steps 15 through 22 for each additional server.

Configuring IRIS FailSafe On

To configure IRIS FailSafe so that it starts up on a node each time it is rebooted, enter this command on the node:

# chkconfig failsafe on 


Note: During the testing described in Chapter 5, “Testing IRIS FailSafe Configuration,” IRIS FailSafe should be chkconfig'd off.